/linux/arch/arm64/kvm/hyp/nvhe/ |
A D | page_alloc.c | 58 if (!buddy || buddy->order != order || buddy->refcount) in __find_buddy_avail() 96 unsigned short order = p->order; in __hyp_attach_page() local 108 for (; (order + 1) < pool->max_order; order++) { in __hyp_attach_page() 120 p->order = order; in __hyp_attach_page() 131 while (p->order > order) { in __hyp_extract_page() 138 p->order--; in __hyp_extract_page() 140 buddy->order = p->order; in __hyp_extract_page() 199 unsigned short order = p->order; in hyp_split_page() local 202 p->order = 0; in hyp_split_page() 206 tail->order = 0; in hyp_split_page() [all …]
|
/linux/include/trace/events/ |
A D | compaction.h | 173 int order, 186 __entry->order = order; 192 __entry->order, 200 int order, 215 __entry->order = order; 222 __entry->order, 229 int order, 238 int order, 262 __entry->order = order; 271 __entry->order, [all …]
|
A D | vmscan.h | 81 __entry->order = order; 86 __entry->order) 105 __entry->order = order; 111 __entry->order, 127 __entry->order = order; 132 __entry->order, 283 int order, 306 __entry->order = order; 322 __entry->order, 454 __entry->order = order; [all …]
|
A D | kmem.h | 164 TP_ARGS(page, order), 168 __field( unsigned int, order ) 173 __entry->order = order; 179 __entry->order) 210 __field( unsigned int, order ) 217 __entry->order = order; 225 __entry->order, 244 __entry->order = order; 251 __entry->order, 253 __entry->order == 0) [all …]
|
A D | oom.h | 35 int order, 42 TP_ARGS(zoneref, order, reclaimable, available, min_wmark, no_progress_loops, wmark_check), 47 __field( int, order) 58 __entry->order = order; 68 __entry->order, 157 TP_PROTO(int order, 164 TP_ARGS(order, priority, result, retries, max_retries, ret), 167 __field( int, order) 176 __entry->order = order; 185 __entry->order,
|
/linux/drivers/gpu/drm/ttm/ |
A D | ttm_pool.c | 79 unsigned int order) in ttm_pool_alloc_page() argument 90 if (order) in ttm_pool_alloc_page() 97 p->private = order; in ttm_pool_alloc_page() 105 if (order) in ttm_pool_alloc_page() 151 if (order) in ttm_pool_free_page() 262 pt->order = order; in ttm_pool_type_init() 367 unsigned int i, order; in ttm_pool_alloc() local 387 order = min_t(unsigned int, order, __fls(num_pages))) { in ttm_pool_alloc() 402 if (order) { in ttm_pool_alloc() 403 --order; in ttm_pool_alloc() [all …]
|
/linux/mm/ |
A D | page_alloc.c | 925 if (!capc || order != capc->cc->order) in compaction_capture() 2824 for (order = 0; order < MAX_ORDER; order++) { in unreserve_highatomic_pageblock() 4253 .order = order, in __alloc_pages_may_oom() 4494 if (!order || order > PAGE_ALLOC_COSTLY_ORDER) in should_compact_retry() 5457 free_the_page(page + (1 << order), order); in __free_pages() 6045 for (order = 0; order < MAX_ORDER; order++) { in show_free_areas() 6050 total += nr[order] << order; in show_free_areas() 6059 for (order = 0; order < MAX_ORDER; order++) { in show_free_areas() 6061 nr[order], K(1UL) << order); in show_free_areas() 9381 for (order = 0; order < MAX_ORDER; order++) { in is_free_buddy_page() [all …]
|
A D | compaction.c | 48 #define block_start_pfn(pfn, order) round_down(pfn, 1UL << (order)) argument 49 #define block_end_pfn(pfn, order) ALIGN((pfn) + 1, 1UL << (order)) argument 100 if (order) in split_map_pages() 1363 order--; in next_search_order() 1365 order = cc->order - 1; in next_search_order() 1388 int order; in fast_isolate_freepages() local 1422 order = next_search_order(cc, order)) { in fast_isolate_freepages() 1744 int order; in fast_find_migrateblock() local 1787 for (order = cc->order - 1; in fast_find_migrateblock() 2091 for (order = cc->order; order < MAX_ORDER; order++) { in __compact_finished() [all …]
|
/linux/drivers/media/pci/cx18/ |
A D | cx18-mailbox.c | 239 mb = &order->mb; in epu_dma_done() 348 order->mb.cmd); in epu_cmd() 355 order->mb.cmd); in epu_cmd() 373 epu_cmd(cx, order); in cx18_in_work_handler() 398 order->rpu, order->mb.cmd); in mb_ack_irq() 407 rpu_str[order->rpu], rpu_str[order->rpu], req); in mb_ack_irq() 422 mb = &order->mb; in epu_dma_done_irq() 488 order->mb.cmd); in epu_cmd_irq() 517 return order; in alloc_in_work_order_irq() 545 order->flags = 0; in cx18_api_epu_cmd_irq() [all …]
|
/linux/drivers/gpu/drm/i915/selftests/ |
A D | i915_buddy.c | 427 for (order = 0; order < max_order; order++) { in igt_buddy_alloc_pessimistic() 431 order); in igt_buddy_alloc_pessimistic() 449 for (order = max_order; order--; ) { in igt_buddy_alloc_pessimistic() 453 order); in igt_buddy_alloc_pessimistic() 465 order = 1; in igt_buddy_alloc_pessimistic() 473 order); in igt_buddy_alloc_pessimistic() 478 order++; in igt_buddy_alloc_pessimistic() 503 int order; in igt_buddy_alloc_optimistic() local 520 for (order = 0; order <= max_order; order++) { in igt_buddy_alloc_optimistic() 579 for (order = top; order--; ) { in igt_buddy_alloc_pathological() [all …]
|
A D | i915_syncmap.c | 274 unsigned int pass, order; in igt_syncmap_join_above() local 296 for (order = 0; order < 64; order += SHIFT) { in igt_syncmap_join_above() 335 unsigned int step, order, idx; in igt_syncmap_join_below() local 345 for (order = 64 - SHIFT; order > 0; order -= SHIFT) { in igt_syncmap_join_below() 362 for (order = SHIFT; order < 64; order += SHIFT) { in igt_syncmap_join_below() 383 for (order = SHIFT; order < 64; order += SHIFT) { in igt_syncmap_join_below() 449 unsigned int idx, order; in igt_syncmap_compact() local 462 for (order = SHIFT; order < 64; order += SHIFT) { in igt_syncmap_compact() 477 context, order, idx, in igt_syncmap_compact() 491 if (sync->height != order) { in igt_syncmap_compact() [all …]
|
A D | i915_random.c | 70 void i915_random_reorder(unsigned int *order, unsigned int count, in i915_random_reorder() argument 73 i915_prandom_shuffle(order, sizeof(*order), count, state); in i915_random_reorder() 78 unsigned int *order, i; in i915_random_order() local 80 order = kmalloc_array(count, sizeof(*order), in i915_random_order() 82 if (!order) in i915_random_order() 83 return order; in i915_random_order() 86 order[i] = i; in i915_random_order() 88 i915_random_reorder(order, count, state); in i915_random_order() 89 return order; in i915_random_order()
|
/linux/Documentation/trace/postprocess/ |
A D | trace-vmscan-postprocess.pl | 317 my $order = $1; 341 my $order = $2; 374 my $order = $3; 557 for (my $order = 0; $order < 20; $order++) { 560 print "direct-$order=$count "; 566 for (my $order = 0; $order < 20; $order++) { 569 print "wakeup-$order=$count "; 614 for (my $order = 0; $order < 20; $order++) { 617 print "wake-$order=$count "; 623 for (my $order = 0; $order < 20; $order++) { [all …]
|
/linux/tools/testing/radix-tree/ |
A D | multiorder.c | 16 unsigned order) in item_insert_order() argument 18 XA_STATE_ORDER(xas, xa, index, order); in item_insert_order() 19 struct item *item = item_create(index, order); in item_insert_order() 58 int height = order[i] / XA_CHUNK_SHIFT; in multiorder_iteration() 60 unsigned long mask = (1UL << order[i]) - 1; in multiorder_iteration() 66 assert(item->order == order[i]); in multiorder_iteration() 112 mask = (1UL << order[k]) - 1; in multiorder_tagged_iteration() 117 assert(item->order == order[k]); in multiorder_tagged_iteration() 139 mask = (1 << order[k]) - 1; in multiorder_tagged_iteration() 144 assert(item->order == order[k]); in multiorder_tagged_iteration() [all …]
|
A D | iteration_check.c | 25 int order; in my_item_insert() local 29 for (order = max_order; order >= 0; order--) { in my_item_insert() 30 xas_set_order(&xas, index, order); in my_item_insert() 31 item->order = order; in my_item_insert() 41 if (order < 0) in my_item_insert() 165 void iteration_test(unsigned order, unsigned test_duration) in iteration_test() argument 170 order > 0 ? "multiorder " : "", test_duration); in iteration_test() 172 max_order = order; in iteration_test()
|
/linux/drivers/gpu/drm/lib/ |
A D | drm_random.c | 15 void drm_random_reorder(unsigned int *order, unsigned int count, in drm_random_reorder() argument 23 swap(order[i], order[j]); in drm_random_reorder() 30 unsigned int *order, i; in drm_random_order() local 32 order = kmalloc_array(count, sizeof(*order), GFP_KERNEL); in drm_random_order() 33 if (!order) in drm_random_order() 34 return order; in drm_random_order() 37 order[i] = i; in drm_random_order() 39 drm_random_reorder(order, count, state); in drm_random_order() 40 return order; in drm_random_order()
|
/linux/lib/ |
A D | test_xarray.c | 207 for (order = 2; order < max_order; order++) { in check_xa_mark_1() 353 for (order = 0; order < max_order; order++) { in check_xa_shrink() 994 for (order = 5; order < order_limit; order++) { in check_multi_find_3() 1158 for (order = 0; order < 20; order++) { in check_find_entry() 1192 for (order = 0; order < order_limit; order++) { in check_pause() 1471 for (order = 0; order < max_order; order++) { in check_create_range() 1564 for (order = 1; order < 2 * XA_CHUNK_SHIFT; order++) { in check_split() 1693 for (order = 1; order < 12; order++) { in check_account() 1704 xa_store_order(xa, 1 << order, order, xa_mk_index(1UL << order), in check_account() 1726 for (order = 0; order < max_order; order++) { in check_get_order() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/therm/ |
A D | gk104.c | 34 const struct gk104_clkgate_engine_info *order = therm->clkgate_order; in gk104_clkgate_enable() local 38 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_enable() 39 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_enable() 42 nvkm_mask(dev, 0x20200 + order[i].offset, 0xff00, 0x4500); in gk104_clkgate_enable() 50 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_enable() 51 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_enable() 54 nvkm_mask(dev, 0x20200 + order[i].offset, 0x00ff, 0x0045); in gk104_clkgate_enable() 63 const struct gk104_clkgate_engine_info *order = therm->clkgate_order; in gk104_clkgate_fini() local 67 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_fini() 68 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_fini() [all …]
|
/linux/include/linux/ |
A D | gfp.h | 570 return __alloc_pages(gfp_mask, order, nid, NULL); in __alloc_pages_node() 579 return __folio_alloc(gfp, order, nid, NULL); in __folio_alloc_node() 588 unsigned int order) in alloc_pages_node() argument 593 return __alloc_pages_node(nid, gfp_mask, order); in alloc_pages_node() 597 struct page *alloc_pages(gfp_t gfp, unsigned int order); 598 struct folio *folio_alloc(gfp_t gfp, unsigned order); 611 return __folio_alloc_node(gfp, order, numa_node_id()); in folio_alloc() 614 alloc_pages(gfp_mask, order) 616 alloc_pages(gfp_mask, order) 632 #define __get_dma_pages(gfp_mask, order) \ argument [all …]
|
A D | compaction.h | 65 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument 80 return 2UL << order; in compact_gap() 92 extern unsigned int extfrag_for_order(struct zone *zone, unsigned int order); 93 extern int fragmentation_index(struct zone *zone, unsigned int order); 95 unsigned int order, unsigned int alloc_flags, 99 extern enum compact_result compaction_suitable(struct zone *zone, int order, 102 extern void compaction_defer_reset(struct zone *zone, int order, 177 bool compaction_zonelist_suitable(struct alloc_context *ac, int order, 182 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx); 189 static inline enum compact_result compaction_suitable(struct zone *zone, int order, in compaction_suitable() argument [all …]
|
A D | page_owner.h | 11 extern void __reset_page_owner(struct page *page, unsigned short order); 13 unsigned short order, gfp_t gfp_mask); 21 static inline void reset_page_owner(struct page *page, unsigned short order) in reset_page_owner() argument 24 __reset_page_owner(page, order); in reset_page_owner() 28 unsigned short order, gfp_t gfp_mask) in set_page_owner() argument 31 __set_page_owner(page, order, gfp_mask); in set_page_owner() 55 static inline void reset_page_owner(struct page *page, unsigned short order) in reset_page_owner() argument 59 unsigned int order, gfp_t gfp_mask) in set_page_owner() argument 63 unsigned short order) in split_page_owner() argument
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
A D | dr_buddy.c | 77 unsigned int *order) in dr_buddy_find_free_seg() argument 101 *order = order_iter; in dr_buddy_find_free_seg() 122 unsigned int order, in mlx5dr_buddy_alloc_mem() argument 139 while (order_iter > order) { in mlx5dr_buddy_alloc_mem() 146 seg <<= order; in mlx5dr_buddy_alloc_mem() 153 unsigned int seg, unsigned int order) in mlx5dr_buddy_free_mem() argument 155 seg >>= order; in mlx5dr_buddy_free_mem() 162 --buddy->num_free[order]; in mlx5dr_buddy_free_mem() 164 ++order; in mlx5dr_buddy_free_mem() 166 bitmap_set(buddy->bitmap[order], seg, 1); in mlx5dr_buddy_free_mem() [all …]
|
/linux/arch/arm/lib/ |
A D | lib1funcs.S | 111 rsb \order, \order, #31 122 addhs \order, \order, #8 126 addhs \order, \order, #4 129 addhi \order, \order, #3 130 addls \order, \order, \divisor, lsr #1 143 sub \order, \order, \spare 148 mov \order, #0 157 addlo \order, \order, #4 165 addlo \order, \order, #1 172 subs \order, \order, #3 @ yes, 3 is intended here [all …]
|
/linux/scripts/atomic/ |
A D | gen-atomic-fallback.sh | 16 local order="$1"; shift 20 local atomicname="arch_${atomic}_${pfx}${name}${sfx}${order}" 42 local order="$1"; shift 44 local tmpl="$(find_fallback_template "${pfx}" "${name}" "${sfx}" "${order}")" 65 local order="$1"; shift 70 printf "#define ${basename}${order} ${basename}${order}\n" 84 local template="$(find_fallback_template "${pfx}" "${name}" "${sfx}" "${order}")" 167 local order="$1"; shift; 174 ___r = arch_cmpxchg${order}((_ptr), ___o, (_new)); \\ 193 for order in "" "_acquire" "_release" "_relaxed"; do [all …]
|
/linux/arch/s390/mm/ |
A D | page-states.c | 71 static inline void set_page_unused(struct page *page, int order) in set_page_unused() argument 75 for (i = 0; i < (1 << order); i++) in set_page_unused() 86 for (i = 0; i < (1 << order); i++) in set_page_stable_dat() 97 for (i = 0; i < (1 << order); i++) in set_page_stable_nodat() 208 void arch_free_page(struct page *page, int order) in arch_free_page() argument 212 set_page_unused(page, order); in arch_free_page() 215 void arch_alloc_page(struct page *page, int order) in arch_alloc_page() argument 220 set_page_stable_dat(page, order); in arch_alloc_page() 222 set_page_stable_nodat(page, order); in arch_alloc_page() 225 void arch_set_page_dat(struct page *page, int order) in arch_set_page_dat() argument [all …]
|