Lines Matching refs:req_node
31 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_top_down_numa_simple_check() local
40 ASSERT_LE(SZ_4, req_node->size); in alloc_exact_nid_top_down_numa_simple_check()
41 size = req_node->size / SZ_4; in alloc_exact_nid_top_down_numa_simple_check()
53 ASSERT_EQ(new_rgn->base, region_end(req_node) - size); in alloc_exact_nid_top_down_numa_simple_check()
54 ASSERT_LE(req_node->base, new_rgn->base); in alloc_exact_nid_top_down_numa_simple_check()
83 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_top_down_numa_part_reserved_check() local
93 ASSERT_LE(SZ_8, req_node->size); in alloc_exact_nid_top_down_numa_part_reserved_check()
94 r1.base = req_node->base; in alloc_exact_nid_top_down_numa_part_reserved_check()
95 r1.size = req_node->size / SZ_2; in alloc_exact_nid_top_down_numa_part_reserved_check()
109 ASSERT_EQ(new_rgn->base, region_end(req_node) - size); in alloc_exact_nid_top_down_numa_part_reserved_check()
110 ASSERT_LE(req_node->base, new_rgn->base); in alloc_exact_nid_top_down_numa_part_reserved_check()
144 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_top_down_numa_split_range_low_check() local
154 req_node_end = region_end(req_node); in alloc_exact_nid_top_down_numa_split_range_low_check()
167 ASSERT_LE(req_node->base, new_rgn->base); in alloc_exact_nid_top_down_numa_split_range_low_check()
201 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_top_down_numa_no_overlap_split_check() local
223 ASSERT_EQ(new_rgn->base, region_end(req_node) - size); in alloc_exact_nid_top_down_numa_no_overlap_split_check()
224 ASSERT_LE(req_node->base, new_rgn->base); in alloc_exact_nid_top_down_numa_no_overlap_split_check()
259 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_top_down_numa_no_overlap_low_check() local
281 ASSERT_EQ(new_rgn->base, region_end(req_node) - size); in alloc_exact_nid_top_down_numa_no_overlap_low_check()
300 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_bottom_up_numa_simple_check() local
309 ASSERT_LE(SZ_4, req_node->size); in alloc_exact_nid_bottom_up_numa_simple_check()
310 size = req_node->size / SZ_4; in alloc_exact_nid_bottom_up_numa_simple_check()
322 ASSERT_EQ(new_rgn->base, req_node->base); in alloc_exact_nid_bottom_up_numa_simple_check()
323 ASSERT_LE(region_end(new_rgn), region_end(req_node)); in alloc_exact_nid_bottom_up_numa_simple_check()
352 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_bottom_up_numa_part_reserved_check() local
363 ASSERT_LE(SZ_8, req_node->size); in alloc_exact_nid_bottom_up_numa_part_reserved_check()
364 r1.base = req_node->base; in alloc_exact_nid_bottom_up_numa_part_reserved_check()
365 r1.size = req_node->size / SZ_2; in alloc_exact_nid_bottom_up_numa_part_reserved_check()
380 ASSERT_EQ(new_rgn->base, req_node->base); in alloc_exact_nid_bottom_up_numa_part_reserved_check()
381 ASSERT_LE(region_end(new_rgn), region_end(req_node)); in alloc_exact_nid_bottom_up_numa_part_reserved_check()
415 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_bottom_up_numa_split_range_low_check() local
425 req_node_end = region_end(req_node); in alloc_exact_nid_bottom_up_numa_split_range_low_check()
437 ASSERT_EQ(new_rgn->base, req_node->base); in alloc_exact_nid_bottom_up_numa_split_range_low_check()
472 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_bottom_up_numa_no_overlap_split_check() local
494 ASSERT_EQ(new_rgn->base, req_node->base); in alloc_exact_nid_bottom_up_numa_no_overlap_split_check()
495 ASSERT_LE(region_end(new_rgn), region_end(req_node)); in alloc_exact_nid_bottom_up_numa_no_overlap_split_check()
530 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_bottom_up_numa_no_overlap_low_check() local
552 ASSERT_EQ(new_rgn->base, req_node->base); in alloc_exact_nid_bottom_up_numa_no_overlap_low_check()
553 ASSERT_LE(region_end(new_rgn), region_end(req_node)); in alloc_exact_nid_bottom_up_numa_no_overlap_low_check()
580 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_numa_small_node_generic_check() local
589 size = SZ_2 * req_node->size; in alloc_exact_nid_numa_small_node_generic_check()
621 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_numa_node_reserved_generic_check() local
630 size = req_node->size; in alloc_exact_nid_numa_node_reserved_generic_check()
634 memblock_reserve(req_node->base, req_node->size); in alloc_exact_nid_numa_node_reserved_generic_check()
664 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_numa_part_reserved_fail_generic_check() local
674 ASSERT_LE(SZ_4, req_node->size); in alloc_exact_nid_numa_part_reserved_fail_generic_check()
675 size = req_node->size / SZ_2; in alloc_exact_nid_numa_part_reserved_fail_generic_check()
676 r1.base = req_node->base + (size / SZ_2); in alloc_exact_nid_numa_part_reserved_fail_generic_check()
712 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_numa_split_range_high_generic_check() local
721 min_addr = req_node->base - SZ_256; in alloc_exact_nid_numa_split_range_high_generic_check()
842 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; in alloc_exact_nid_numa_reserved_full_merge_generic_check() local
846 phys_addr_t size = req_node->size; in alloc_exact_nid_numa_reserved_full_merge_generic_check()
877 ASSERT_LE(new_rgn->base, req_node->base); in alloc_exact_nid_numa_reserved_full_merge_generic_check()
878 ASSERT_LE(region_end(req_node), region_end(new_rgn)); in alloc_exact_nid_numa_reserved_full_merge_generic_check()