/linux/drivers/gpu/drm/ttm/ |
A D | ttm_pool.c | 67 static struct ttm_pool_type global_write_combined[MAX_ORDER]; 68 static struct ttm_pool_type global_uncached[MAX_ORDER]; 70 static struct ttm_pool_type global_dma32_write_combined[MAX_ORDER]; 71 static struct ttm_pool_type global_dma32_uncached[MAX_ORDER]; 385 for (order = min_t(unsigned int, MAX_ORDER - 1, __fls(num_pages)); in ttm_pool_alloc() 510 for (j = 0; j < MAX_ORDER; ++j) in ttm_pool_init() 530 for (j = 0; j < MAX_ORDER; ++j) in ttm_pool_fini() 584 for (i = 0; i < MAX_ORDER; ++i) in ttm_pool_debugfs_header() 595 for (i = 0; i < MAX_ORDER; ++i) in ttm_pool_debugfs_orders() 704 for (i = 0; i < MAX_ORDER; ++i) { in ttm_pool_mgr_init() [all …]
|
/linux/include/linux/ |
A D | mmzone.h | 28 #define MAX_ORDER 11 macro 30 #define MAX_ORDER CONFIG_FORCE_MAX_ZONEORDER macro 32 #define MAX_ORDER_NR_PAGES (1 << (MAX_ORDER - 1)) 87 for (order = 0; order < MAX_ORDER; order++) \ 617 struct free_area free_area[MAX_ORDER]; 1286 #if (MAX_ORDER - 1 + PAGE_SHIFT) > SECTION_SIZE_BITS 1287 #error Allocator MAX_ORDER exceeds SECTION_SIZE
|
A D | slab.h | 245 #define KMALLOC_SHIFT_HIGH ((MAX_ORDER + PAGE_SHIFT - 1) <= 25 ? \ 246 (MAX_ORDER + PAGE_SHIFT - 1) : 25) 259 #define KMALLOC_SHIFT_MAX (MAX_ORDER + PAGE_SHIFT - 1) 272 #define KMALLOC_SHIFT_MAX (MAX_ORDER + PAGE_SHIFT - 1)
|
A D | pageblock-flags.h | 48 #define pageblock_order (MAX_ORDER-1)
|
/linux/drivers/staging/media/atomisp/pci/hmm/ |
A D | hmm_reserved_pool.c | 136 order = MAX_ORDER; in hmm_reserved_pool_init() 144 BUG_ON(order > MAX_ORDER); in hmm_reserved_pool_init()
|
/linux/arch/arm64/kvm/hyp/include/nvhe/ |
A D | gfp.h | 19 struct list_head free_area[MAX_ORDER];
|
/linux/mm/ |
A D | page_reporting.c | 14 unsigned int page_reporting_order = MAX_ORDER; 247 for (order = page_reporting_order; order < MAX_ORDER; order++) { in page_reporting_process_zone()
|
A D | shuffle.h | 7 #define SHUFFLE_ORDER (MAX_ORDER-1)
|
A D | page_owner.c | 289 if (freepage_order < MAX_ORDER) in pagetypeinfo_showmixedcount_print() 475 if (freepage_order < MAX_ORDER) in read_page_owner() 564 if (order > 0 && order < MAX_ORDER) in init_pages_in_zone()
|
A D | vmstat.c | 1070 for (order = 0; order < MAX_ORDER; order++) { in fill_contig_page_info() 1103 if (WARN_ON_ONCE(order >= MAX_ORDER)) in __fragmentation_index() 1453 for (order = 0; order < MAX_ORDER; ++order) in frag_show_print() 1482 for (order = 0; order < MAX_ORDER; ++order) { in pagetypeinfo_showfree_print() 1522 for (order = 0; order < MAX_ORDER; ++order) in pagetypeinfo_showfree() 2139 for (order = 0; order < MAX_ORDER; ++order) { in unusable_show_print() 2191 for (order = 0; order < MAX_ORDER; ++order) { in extfrag_show_print()
|
A D | page_alloc.c | 1021 if (order >= MAX_ORDER - 2) in buddy_merge_likely() 1105 if (order < MAX_ORDER - 1) { in __free_one_page() 2249 if (pageblock_order >= MAX_ORDER) { in init_cma_reserved_pageblock() 2254 __free_pages(p, MAX_ORDER - 1); in init_cma_reserved_pageblock() 2947 VM_BUG_ON(current_order == MAX_ORDER); in __rmqueue_fallback() 3879 for (o = order; o < MAX_ORDER; o++) { in __zone_watermark_ok() 5343 if (unlikely(order >= MAX_ORDER)) { in __alloc_pages() 6037 unsigned char types[MAX_ORDER]; in show_free_areas() 7337 order = MAX_ORDER - 1; in set_pageblock_order() 9121 if (++order >= MAX_ORDER) { in alloc_contig_range() [all …]
|
A D | cma.c | 184 max_t(unsigned long, MAX_ORDER - 1, pageblock_order); in cma_init_reserved_mem() 272 max_t(unsigned long, MAX_ORDER - 1, pageblock_order)); in cma_declare_contiguous_nid()
|
A D | page_isolation.c | 91 if (order >= pageblock_order && order < MAX_ORDER - 1) { in unset_migratetype_isolate()
|
/linux/drivers/base/regmap/ |
A D | regmap-debugfs.c | 229 if (count > (PAGE_SIZE << (MAX_ORDER - 1))) in regmap_read_debugfs() 230 count = PAGE_SIZE << (MAX_ORDER - 1); in regmap_read_debugfs() 376 if (count > (PAGE_SIZE << (MAX_ORDER - 1))) in regmap_reg_ranges_read_file() 377 count = PAGE_SIZE << (MAX_ORDER - 1); in regmap_reg_ranges_read_file()
|
/linux/include/drm/ttm/ |
A D | ttm_pool.h | 75 struct ttm_pool_type orders[MAX_ORDER];
|
/linux/drivers/crypto/hisilicon/ |
A D | sgl.c | 74 block_size = 1 << (PAGE_SHIFT + MAX_ORDER <= 32 ? in hisi_acc_create_sgl_pool() 75 PAGE_SHIFT + MAX_ORDER - 1 : 31); in hisi_acc_create_sgl_pool()
|
/linux/arch/powerpc/include/asm/ |
A D | fadump-internal.h | 24 max_t(unsigned long, MAX_ORDER - 1, \
|
/linux/kernel/events/ |
A D | ring_buffer.c | 609 if (order > MAX_ORDER) in rb_alloc_aux_page() 610 order = MAX_ORDER; in rb_alloc_aux_page() 814 if (order_base_2(size) >= PAGE_SHIFT+MAX_ORDER) in rb_alloc()
|
/linux/drivers/gpu/drm/i915/gem/ |
A D | i915_gem_internal.c | 43 max_order = MAX_ORDER; in i915_gem_object_get_pages_internal()
|
/linux/arch/ia64/mm/ |
A D | hugetlbpage.c | 188 size >= (1UL << PAGE_SHIFT << MAX_ORDER)) { in hugetlb_setup_sz()
|
/linux/arch/sparc/mm/ |
A D | tsb.c | 405 if (max_tsb_size > (PAGE_SIZE << MAX_ORDER)) in tsb_grow() 406 max_tsb_size = (PAGE_SIZE << MAX_ORDER); in tsb_grow()
|
/linux/fs/ramfs/ |
A D | file-nommu.c | 73 if (unlikely(order >= MAX_ORDER)) in ramfs_nommu_expand_for_mapping()
|
/linux/arch/arm64/kvm/hyp/nvhe/ |
A D | page_alloc.c | 244 pool->max_order = min(MAX_ORDER, get_order(nr_pages << PAGE_SHIFT)); in hyp_pool_init()
|
/linux/kernel/dma/ |
A D | pool.c | 88 order = min(get_order(pool_size), MAX_ORDER-1); in atomic_pool_expand()
|
A D | contiguous.c | 402 phys_addr_t align = PAGE_SIZE << max(MAX_ORDER - 1, pageblock_order); in rmem_cma_setup()
|