X-Git-Url: https://asedeno.scripts.mit.edu/gitweb/?a=blobdiff_plain;f=mm%2Fpage_alloc.c;h=a7a6aac95a6d158690e1ca0981b1b652820248d4;hb=89d790ab31d033d67635f6362d57ea64e47708fa;hp=9f9623d690d69582be8e82885649dd3b60937aa6;hpb=94eae8034002401d71ae950106659e16add36e77;p=linux.git diff --git a/mm/page_alloc.c b/mm/page_alloc.c index 9f9623d690d6..a7a6aac95a6d 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -5925,7 +5925,7 @@ static unsigned long __paginginit calc_memmap_size(unsigned long spanned_pages, * the zone and SPARSEMEM is in use. If there are holes within the * zone, each populated memory region may cost us one or two extra * memmap pages due to alignment because memmap pages for each - * populated regions may not naturally algined on page boundary. + * populated regions may not be naturally aligned on page boundary. * So the (present_pages >> 4) heuristic is a tradeoff for that. */ if (spanned_pages > present_pages + (present_pages >> 4) &&