mirror of
https://github.com/adulau/aha.git
synced 2024-12-30 20:56:23 +00:00
f06a96844a
Fix array initialization in lots of arches The number of zones may now be reduced from 4 to 2 for many arches. Fix the array initialization for the zones array for all architectures so that it is not initializing a fixed number of elements. Signed-off-by: Christoph Lameter <clameter@sgi.com> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
231 lines
6 KiB
C
231 lines
6 KiB
C
/*
|
|
* linux/arch/h8300/mm/init.c
|
|
*
|
|
* Copyright (C) 1998 D. Jeff Dionne <jeff@lineo.ca>,
|
|
* Kenneth Albanowski <kjahds@kjahds.com>,
|
|
* Copyright (C) 2000 Lineo, Inc. (www.lineo.com)
|
|
*
|
|
* Based on:
|
|
*
|
|
* linux/arch/m68knommu/mm/init.c
|
|
* linux/arch/m68k/mm/init.c
|
|
*
|
|
* Copyright (C) 1995 Hamish Macdonald
|
|
*
|
|
* JAN/1999 -- hacked to support ColdFire (gerg@snapgear.com)
|
|
* DEC/2000 -- linux 2.4 support <davidm@snapgear.com>
|
|
*/
|
|
|
|
#include <linux/signal.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/string.h>
|
|
#include <linux/types.h>
|
|
#include <linux/ptrace.h>
|
|
#include <linux/mman.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/swap.h>
|
|
#include <linux/init.h>
|
|
#include <linux/highmem.h>
|
|
#include <linux/pagemap.h>
|
|
#include <linux/bootmem.h>
|
|
#include <linux/slab.h>
|
|
|
|
#include <asm/setup.h>
|
|
#include <asm/segment.h>
|
|
#include <asm/page.h>
|
|
#include <asm/pgtable.h>
|
|
#include <asm/system.h>
|
|
|
|
#undef DEBUG
|
|
|
|
extern void die_if_kernel(char *,struct pt_regs *,long);
|
|
extern void free_initmem(void);
|
|
|
|
/*
|
|
* BAD_PAGE is the page that is used for page faults when linux
|
|
* is out-of-memory. Older versions of linux just did a
|
|
* do_exit(), but using this instead means there is less risk
|
|
* for a process dying in kernel mode, possibly leaving a inode
|
|
* unused etc..
|
|
*
|
|
* BAD_PAGETABLE is the accompanying page-table: it is initialized
|
|
* to point to BAD_PAGE entries.
|
|
*
|
|
* ZERO_PAGE is a special page that is used for zero-initialized
|
|
* data and COW.
|
|
*/
|
|
static unsigned long empty_bad_page_table;
|
|
|
|
static unsigned long empty_bad_page;
|
|
|
|
unsigned long empty_zero_page;
|
|
|
|
extern unsigned long rom_length;
|
|
|
|
void show_mem(void)
|
|
{
|
|
unsigned long i;
|
|
int free = 0, total = 0, reserved = 0, shared = 0;
|
|
int cached = 0;
|
|
|
|
printk("\nMem-info:\n");
|
|
show_free_areas();
|
|
i = max_mapnr;
|
|
while (i-- > 0) {
|
|
total++;
|
|
if (PageReserved(mem_map+i))
|
|
reserved++;
|
|
else if (PageSwapCache(mem_map+i))
|
|
cached++;
|
|
else if (!page_count(mem_map+i))
|
|
free++;
|
|
else
|
|
shared += page_count(mem_map+i) - 1;
|
|
}
|
|
printk("%d pages of RAM\n",total);
|
|
printk("%d free pages\n",free);
|
|
printk("%d reserved pages\n",reserved);
|
|
printk("%d pages shared\n",shared);
|
|
printk("%d pages swap cached\n",cached);
|
|
}
|
|
|
|
extern unsigned long memory_start;
|
|
extern unsigned long memory_end;
|
|
|
|
/*
|
|
* paging_init() continues the virtual memory environment setup which
|
|
* was begun by the code in arch/head.S.
|
|
* The parameters are pointers to where to stick the starting and ending
|
|
* addresses of available kernel virtual memory.
|
|
*/
|
|
void paging_init(void)
|
|
{
|
|
/*
|
|
* Make sure start_mem is page aligned, otherwise bootmem and
|
|
* page_alloc get different views og the world.
|
|
*/
|
|
#ifdef DEBUG
|
|
unsigned long start_mem = PAGE_ALIGN(memory_start);
|
|
#endif
|
|
unsigned long end_mem = memory_end & PAGE_MASK;
|
|
|
|
#ifdef DEBUG
|
|
printk ("start_mem is %#lx\nvirtual_end is %#lx\n",
|
|
start_mem, end_mem);
|
|
#endif
|
|
|
|
/*
|
|
* Initialize the bad page table and bad page to point
|
|
* to a couple of allocated pages.
|
|
*/
|
|
empty_bad_page_table = (unsigned long)alloc_bootmem_pages(PAGE_SIZE);
|
|
empty_bad_page = (unsigned long)alloc_bootmem_pages(PAGE_SIZE);
|
|
empty_zero_page = (unsigned long)alloc_bootmem_pages(PAGE_SIZE);
|
|
memset((void *)empty_zero_page, 0, PAGE_SIZE);
|
|
|
|
/*
|
|
* Set up SFC/DFC registers (user data space).
|
|
*/
|
|
set_fs (USER_DS);
|
|
|
|
#ifdef DEBUG
|
|
printk ("before free_area_init\n");
|
|
|
|
printk ("free_area_init -> start_mem is %#lx\nvirtual_end is %#lx\n",
|
|
start_mem, end_mem);
|
|
#endif
|
|
|
|
{
|
|
unsigned long zones_size[MAX_NR_ZONES] = {0, };
|
|
|
|
zones_size[ZONE_DMA] = 0 >> PAGE_SHIFT;
|
|
zones_size[ZONE_NORMAL] = (end_mem - PAGE_OFFSET) >> PAGE_SHIFT;
|
|
#ifdef CONFIG_HIGHMEM
|
|
zones_size[ZONE_HIGHMEM] = 0;
|
|
#endif
|
|
free_area_init(zones_size);
|
|
}
|
|
}
|
|
|
|
void mem_init(void)
|
|
{
|
|
int codek = 0, datak = 0, initk = 0;
|
|
/* DAVIDM look at setup memory map generically with reserved area */
|
|
unsigned long tmp;
|
|
extern char _etext, _stext, _sdata, _ebss, __init_begin, __init_end;
|
|
extern unsigned long _ramend, _ramstart;
|
|
unsigned long len = &_ramend - &_ramstart;
|
|
unsigned long start_mem = memory_start; /* DAVIDM - these must start at end of kernel */
|
|
unsigned long end_mem = memory_end; /* DAVIDM - this must not include kernel stack at top */
|
|
|
|
#ifdef DEBUG
|
|
printk(KERN_DEBUG "Mem_init: start=%lx, end=%lx\n", start_mem, end_mem);
|
|
#endif
|
|
|
|
end_mem &= PAGE_MASK;
|
|
high_memory = (void *) end_mem;
|
|
|
|
start_mem = PAGE_ALIGN(start_mem);
|
|
max_mapnr = num_physpages = MAP_NR(high_memory);
|
|
|
|
/* this will put all memory onto the freelists */
|
|
totalram_pages = free_all_bootmem();
|
|
|
|
codek = (&_etext - &_stext) >> 10;
|
|
datak = (&_ebss - &_sdata) >> 10;
|
|
initk = (&__init_begin - &__init_end) >> 10;
|
|
|
|
tmp = nr_free_pages() << PAGE_SHIFT;
|
|
printk(KERN_INFO "Memory available: %luk/%luk RAM, %luk/%luk ROM (%dk kernel code, %dk data)\n",
|
|
tmp >> 10,
|
|
len >> 10,
|
|
(rom_length > 0) ? ((rom_length >> 10) - codek) : 0,
|
|
rom_length >> 10,
|
|
codek,
|
|
datak
|
|
);
|
|
}
|
|
|
|
|
|
#ifdef CONFIG_BLK_DEV_INITRD
|
|
void free_initrd_mem(unsigned long start, unsigned long end)
|
|
{
|
|
int pages = 0;
|
|
for (; start < end; start += PAGE_SIZE) {
|
|
ClearPageReserved(virt_to_page(start));
|
|
init_page_count(virt_to_page(start));
|
|
free_page(start);
|
|
totalram_pages++;
|
|
pages++;
|
|
}
|
|
printk ("Freeing initrd memory: %dk freed\n", pages);
|
|
}
|
|
#endif
|
|
|
|
void
|
|
free_initmem()
|
|
{
|
|
#ifdef CONFIG_RAMKERNEL
|
|
unsigned long addr;
|
|
extern char __init_begin, __init_end;
|
|
/*
|
|
* the following code should be cool even if these sections
|
|
* are not page aligned.
|
|
*/
|
|
addr = PAGE_ALIGN((unsigned long)(&__init_begin));
|
|
/* next to check that the page we free is not a partial page */
|
|
for (; addr + PAGE_SIZE < (unsigned long)(&__init_end); addr +=PAGE_SIZE) {
|
|
ClearPageReserved(virt_to_page(addr));
|
|
init_page_count(virt_to_page(addr));
|
|
free_page(addr);
|
|
totalram_pages++;
|
|
}
|
|
printk(KERN_INFO "Freeing unused kernel memory: %ldk freed (0x%x - 0x%x)\n",
|
|
(addr - PAGE_ALIGN((long) &__init_begin)) >> 10,
|
|
(int)(PAGE_ALIGN((unsigned long)(&__init_begin))),
|
|
(int)(addr - PAGE_SIZE));
|
|
#endif
|
|
}
|
|
|