2 * Copyright (C) 2004-2006 Atmel Corporation
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License version 2 as
6 * published by the Free Software Foundation.
9 #include <linux/kernel.h>
11 #include <linux/swap.h>
12 #include <linux/init.h>
13 #include <linux/mmzone.h>
14 #include <linux/bootmem.h>
15 #include <linux/pagemap.h>
16 #include <linux/nodemask.h>
19 #include <asm/mmu_context.h>
23 #include <asm/setup.h>
24 #include <asm/sections.h>
26 DEFINE_PER_CPU(struct mmu_gather
, mmu_gathers
);
28 pgd_t swapper_pg_dir
[PTRS_PER_PGD
];
30 struct page
*empty_zero_page
;
33 * Cache of MMU context last used.
35 unsigned long mmu_context_cache
= NO_CONTEXT
;
37 #define START_PFN (NODE_DATA(0)->bdata->node_boot_start >> PAGE_SHIFT)
38 #define MAX_LOW_PFN (NODE_DATA(0)->bdata->node_low_pfn)
42 int total
= 0, reserved
= 0, cached
= 0;
43 int slab
= 0, free
= 0, shared
= 0;
46 printk("Mem-info:\n");
49 for_each_online_pgdat(pgdat
) {
50 struct page
*page
, *end
;
52 page
= pgdat
->node_mem_map
;
53 end
= page
+ pgdat
->node_spanned_pages
;
57 if (PageReserved(page
))
59 else if (PageSwapCache(page
))
61 else if (PageSlab(page
))
63 else if (!page_count(page
))
66 shared
+= page_count(page
) - 1;
71 printk ("%d pages of RAM\n", total
);
72 printk ("%d free pages\n", free
);
73 printk ("%d reserved pages\n", reserved
);
74 printk ("%d slab pages\n", slab
);
75 printk ("%d pages shared\n", shared
);
76 printk ("%d pages swap cached\n", cached
);
80 * paging_init() sets up the page tables
82 * This routine also unmaps the page at virtual kernel address 0, so
83 * that we can trap those pesky NULL-reference errors in the kernel.
85 void __init
paging_init(void)
87 extern unsigned long _evba
;
92 * Make sure we can handle exceptions before enabling
93 * paging. Not that we should ever _get_ any exceptions this
94 * early, but you never know...
96 printk("Exception vectors start at %p\n", &_evba
);
97 sysreg_write(EVBA
, (unsigned long)&_evba
);
100 * Since we are ready to handle exceptions now, we should let
101 * the CPU generate them...
103 __asm__
__volatile__ ("csrf %0" : : "i"(SR_EM_BIT
));
106 * Allocate the zero page. The allocator will panic if it
107 * can't satisfy the request, so no need to check.
109 zero_page
= alloc_bootmem_low_pages_node(NODE_DATA(0),
116 pg_dir
= swapper_pg_dir
;
117 sysreg_write(PTBR
, (unsigned long)pg_dir
);
119 for (i
= 0; i
< PTRS_PER_PGD
; i
++)
120 pgd_val(pg_dir
[i
]) = 0;
123 printk ("CPU: Paging enabled\n");
126 for_each_online_node(nid
) {
127 pg_data_t
*pgdat
= NODE_DATA(nid
);
128 unsigned long zones_size
[MAX_NR_ZONES
];
129 unsigned long low
, start_pfn
;
131 start_pfn
= pgdat
->bdata
->node_boot_start
;
132 start_pfn
>>= PAGE_SHIFT
;
133 low
= pgdat
->bdata
->node_low_pfn
;
135 memset(zones_size
, 0, sizeof(zones_size
));
136 zones_size
[ZONE_NORMAL
] = low
- start_pfn
;
138 printk("Node %u: start_pfn = 0x%lx, low = 0x%lx\n",
139 nid
, start_pfn
, low
);
141 free_area_init_node(nid
, pgdat
, zones_size
, start_pfn
, NULL
);
143 printk("Node %u: mem_map starts at %p\n",
144 pgdat
->node_id
, pgdat
->node_mem_map
);
147 mem_map
= NODE_DATA(0)->node_mem_map
;
149 memset(zero_page
, 0, PAGE_SIZE
);
150 empty_zero_page
= virt_to_page(zero_page
);
151 flush_dcache_page(empty_zero_page
);
154 void __init
mem_init(void)
156 int codesize
, reservedpages
, datasize
, initsize
;
162 /* this will put all low memory onto the freelists */
163 for_each_online_node(nid
) {
164 pg_data_t
*pgdat
= NODE_DATA(nid
);
165 unsigned long node_pages
= 0;
166 void *node_high_memory
;
168 num_physpages
+= pgdat
->node_present_pages
;
170 if (pgdat
->node_spanned_pages
!= 0)
171 node_pages
= free_all_bootmem_node(pgdat
);
173 totalram_pages
+= node_pages
;
175 for (i
= 0; i
< node_pages
; i
++)
176 if (PageReserved(pgdat
->node_mem_map
+ i
))
179 node_high_memory
= (void *)((pgdat
->node_start_pfn
180 + pgdat
->node_spanned_pages
)
182 if (node_high_memory
> high_memory
)
183 high_memory
= node_high_memory
;
186 max_mapnr
= MAP_NR(high_memory
);
188 codesize
= (unsigned long)_etext
- (unsigned long)_text
;
189 datasize
= (unsigned long)_edata
- (unsigned long)_data
;
190 initsize
= (unsigned long)__init_end
- (unsigned long)__init_begin
;
192 printk ("Memory: %luk/%luk available (%dk kernel code, "
193 "%dk reserved, %dk data, %dk init)\n",
194 (unsigned long)nr_free_pages() << (PAGE_SHIFT
- 10),
195 totalram_pages
<< (PAGE_SHIFT
- 10),
197 reservedpages
<< (PAGE_SHIFT
- 10),
202 static inline void free_area(unsigned long addr
, unsigned long end
, char *s
)
204 unsigned int size
= (end
- addr
) >> 10;
206 for (; addr
< end
; addr
+= PAGE_SIZE
) {
207 struct page
*page
= virt_to_page(addr
);
208 ClearPageReserved(page
);
209 init_page_count(page
);
215 printk(KERN_INFO
"Freeing %s memory: %dK (%lx - %lx)\n",
216 s
, size
, end
- (size
<< 10), end
);
219 void free_initmem(void)
221 free_area((unsigned long)__init_begin
, (unsigned long)__init_end
,
225 #ifdef CONFIG_BLK_DEV_INITRD
227 static int keep_initrd
;
229 void free_initrd_mem(unsigned long start
, unsigned long end
)
232 free_area(start
, end
, "initrd");
235 static int __init
keepinitrd_setup(char *__unused
)
241 __setup("keepinitrd", keepinitrd_setup
);