diff options
author | Josh Aas <josha@sgi.com> | 2004-08-23 21:26:54 -0700 |
---|---|---|
committer | Linus Torvalds <torvalds@ppc970.osdl.org> | 2004-08-23 21:26:54 -0700 |
commit | fe92ebf3b40d54ba2a2ceb4806b07c654319133b (patch) | |
tree | 07b2323e45c3cba78f5695ca31601affdb021229 /mm | |
parent | 13c6195264e3f64a1bfbdad3dce78ef5fb431808 (diff) | |
download | history-fe92ebf3b40d54ba2a2ceb4806b07c654319133b.tar.gz |
[PATCH] improve speed of freeing bootmem
Attached is a patch that greatly improves the speed of freeing boot memory.
On ia64 machines with 2GB or more memory (I didn't test with less, but I
can't imagine there being a problem), the speed improvement is about 75%
for the function free_all_bootmem_core. This translates to savings on the
order of 1 minute / TB of memory during boot time. That number comes from
testing on a machine with 512GB, and extrapolating based on profiling of an
unpatched 4TB machine. For 4 and 8 TB machines, the time spent in this
function is about 1 minutes/TB, which is painful especially given that
there is no indication of what is going on put to the console (this issue
to possibly be addressed later).
The basic idea is to free higher order pages instead of going through every
single one. Also, some unnecessary atomic operations are done away with
and replaced with non-atomic equivalents, and prefetching is done where it
helps the most. For a more in-depth discusion of this patch, please see
the linux-ia64 archives (topic is "free bootmem feedback patch").
The patch is originally Tony Luck's, and I added some further optimizations
(non-atomic ops improvements and prefetching).
Signed-off-by: Tony Luck <tony.luck@intel.com>
Signed-off-by: Josh Aas <josha@sgi.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
Diffstat (limited to 'mm')
-rw-r--r-- | mm/bootmem.c | 25 |
1 files changed, 22 insertions, 3 deletions
diff --git a/mm/bootmem.c b/mm/bootmem.c index 966135b90ff28b..092d8517649091 100644 --- a/mm/bootmem.c +++ b/mm/bootmem.c @@ -259,6 +259,7 @@ static unsigned long __init free_all_bootmem_core(pg_data_t *pgdat) unsigned long i, count, total = 0; unsigned long idx; unsigned long *map; + int gofast = 0; BUG_ON(!bdata->node_bootmem_map); @@ -267,14 +268,32 @@ static unsigned long __init free_all_bootmem_core(pg_data_t *pgdat) page = virt_to_page(phys_to_virt(bdata->node_boot_start)); idx = bdata->node_low_pfn - (bdata->node_boot_start >> PAGE_SHIFT); map = bdata->node_bootmem_map; + /* Check physaddr is O(LOG2(BITS_PER_LONG)) page aligned */ + if (bdata->node_boot_start == 0 || + ffs(bdata->node_boot_start) - PAGE_SHIFT > ffs(BITS_PER_LONG)) + gofast = 1; for (i = 0; i < idx; ) { unsigned long v = ~map[i / BITS_PER_LONG]; - if (v) { + if (gofast && v == ~0UL) { + int j; + + count += BITS_PER_LONG; + __ClearPageReserved(page); + set_page_count(page, 1); + for (j = 1; j < BITS_PER_LONG; j++) { + if (j + 16 < BITS_PER_LONG) + prefetchw(page + j + 16); + __ClearPageReserved(page + j); + } + __free_pages(page, ffs(BITS_PER_LONG)-1); + i += BITS_PER_LONG; + page += BITS_PER_LONG; + } else if (v) { unsigned long m; for (m = 1; m && i < idx; m<<=1, page++, i++) { if (v & m) { count++; - ClearPageReserved(page); + __ClearPageReserved(page); set_page_count(page, 1); __free_page(page); } @@ -294,7 +313,7 @@ static unsigned long __init free_all_bootmem_core(pg_data_t *pgdat) count = 0; for (i = 0; i < ((bdata->node_low_pfn-(bdata->node_boot_start >> PAGE_SHIFT))/8 + PAGE_SIZE-1)/PAGE_SIZE; i++,page++) { count++; - ClearPageReserved(page); + __ClearPageReserved(page); set_page_count(page, 1); __free_page(page); } |