| #include <linux/kernel.h> |
| #include <linux/types.h> |
| #include <linux/init.h> |
| #include <linux/bitops.h> |
| #include <linux/memblock.h> |
| #include <linux/bootmem.h> |
| #include <linux/mm.h> |
| #include <linux/range.h> |
| |
| /* Check for already reserved areas */ |
| static inline bool __init bad_addr_size(u64 *addrp, u64 *sizep, u64 align) |
| { |
| struct memblock_region *r; |
| u64 addr = *addrp, last; |
| u64 size = *sizep; |
| bool changed = false; |
| |
| again: |
| last = addr + size; |
| for_each_memblock(reserved, r) { |
| if (last > r->base && addr < r->base) { |
| size = r->base - addr; |
| changed = true; |
| goto again; |
| } |
| if (last > (r->base + r->size) && addr < (r->base + r->size)) { |
| addr = round_up(r->base + r->size, align); |
| size = last - addr; |
| changed = true; |
| goto again; |
| } |
| if (last <= (r->base + r->size) && addr >= r->base) { |
| (*sizep)++; |
| return false; |
| } |
| } |
| if (changed) { |
| *addrp = addr; |
| *sizep = size; |
| } |
| return changed; |
| } |
| |
| static u64 __init __memblock_x86_find_in_range_size(u64 ei_start, u64 ei_last, u64 start, |
| u64 *sizep, u64 align) |
| { |
| u64 addr, last; |
| |
| addr = round_up(ei_start, align); |
| if (addr < start) |
| addr = round_up(start, align); |
| if (addr >= ei_last) |
| goto out; |
| *sizep = ei_last - addr; |
| while (bad_addr_size(&addr, sizep, align) && addr + *sizep <= ei_last) |
| ; |
| last = addr + *sizep; |
| if (last > ei_last) |
| goto out; |
| |
| return addr; |
| |
| out: |
| return MEMBLOCK_ERROR; |
| } |
| |
| /* |
| * Find next free range after start, and size is returned in *sizep |
| */ |
| u64 __init memblock_x86_find_in_range_size(u64 start, u64 *sizep, u64 align) |
| { |
| struct memblock_region *r; |
| |
| for_each_memblock(memory, r) { |
| u64 ei_start = r->base; |
| u64 ei_last = ei_start + r->size; |
| u64 addr; |
| |
| addr = __memblock_x86_find_in_range_size(ei_start, ei_last, start, |
| sizep, align); |
| |
| if (addr != MEMBLOCK_ERROR) |
| return addr; |
| } |
| |
| return MEMBLOCK_ERROR; |
| } |
| |
| #ifndef CONFIG_NO_BOOTMEM |
| void __init memblock_x86_to_bootmem(u64 start, u64 end) |
| { |
| int count; |
| u64 final_start, final_end; |
| struct memblock_region *r; |
| |
| /* Take out region array itself */ |
| memblock_free_reserved_regions(); |
| |
| count = memblock.reserved.cnt; |
| pr_info("(%d early reservations) ==> bootmem [%010llx-%010llx]\n", count, start, end - 1); |
| for_each_memblock(reserved, r) { |
| pr_info(" [%010llx-%010llx] ", (u64)r->base, (u64)r->base + r->size - 1); |
| final_start = max(start, r->base); |
| final_end = min(end, r->base + r->size); |
| if (final_start >= final_end) { |
| pr_cont("\n"); |
| continue; |
| } |
| pr_cont(" ==> [%010llx-%010llx]\n", final_start, final_end - 1); |
| reserve_bootmem_generic(final_start, final_end - final_start, BOOTMEM_DEFAULT); |
| } |
| |
| /* Put region array back ? */ |
| memblock_reserve_reserved_regions(); |
| } |
| #endif |
| |
| void __init memblock_x86_reserve_range(u64 start, u64 end, char *name) |
| { |
| if (start == end) |
| return; |
| |
| if (WARN_ONCE(start > end, "memblock_x86_reserve_range: wrong range [%#llx, %#llx]\n", start, end)) |
| return; |
| |
| memblock_reserve(start, end - start); |
| } |
| |
| void __init memblock_x86_free_range(u64 start, u64 end) |
| { |
| if (start == end) |
| return; |
| |
| if (WARN_ONCE(start > end, "memblock_x86_free_range: wrong range [%#llx, %#llx]\n", start, end)) |
| return; |
| |
| memblock_free(start, end - start); |
| } |