/linux/mm/ |
A D | list_lru.c | 42 list_del(&lru->list); in list_lru_unregister() 63 return &nlru->lru; in list_lru_from_memcg_idx() 108 return &nlru->lru; in list_lru_from_memcg_idx() 117 return &nlru->lru; in list_lru_from_kmem() 409 memcpy(&new->lru, &old->lru, flex_array_size(new, lru, old_size)); in memcg_update_list_lru_node() 584 lru->node = kcalloc(nr_node_ids, sizeof(*lru->node), GFP_KERNEL); in __list_lru_init() 585 if (!lru->node) in __list_lru_init() 592 init_one_lru(&lru->node[i].lru); in __list_lru_init() 613 if (!lru->node) in list_lru_destroy() 621 kfree(lru->node); in list_lru_destroy() [all …]
|
A D | zbud.c | 111 struct list_head lru; member 128 struct list_head lru; member 159 INIT_LIST_HEAD(&zhdr->lru); in init_zbud_page() 232 INIT_LIST_HEAD(&pool->lru); in zbud_create_pool() 323 if (!list_empty(&zhdr->lru)) in zbud_alloc() 324 list_del(&zhdr->lru); in zbud_alloc() 325 list_add(&zhdr->lru, &pool->lru); in zbud_alloc() 368 list_del(&zhdr->lru); in zbud_free() 428 zhdr = list_last_entry(&pool->lru, struct zbud_header, lru); in zbud_reclaim_page() 429 list_del(&zhdr->lru); in zbud_reclaim_page() [all …]
|
A D | vmscan.c | 2921 nr[lru] = scan; in get_scan_count() 2977 if (nr[lru]) { in shrink_lruvec() 3013 lru = LRU_BASE; in shrink_lruvec() 3023 nr[lru] = 0; in shrink_lruvec() 3030 lru = (lru == LRU_FILE) ? LRU_BASE : LRU_FILE; in shrink_lruvec() 3031 nr_scanned = targets[lru] - nr[lru]; in shrink_lruvec() 3032 nr[lru] = targets[lru] * (100 - percentage) / 100; in shrink_lruvec() 3033 nr[lru] -= min(nr[lru], nr_scanned); in shrink_lruvec() 3036 nr_scanned = targets[lru] - nr[lru]; in shrink_lruvec() 3037 nr[lru] = targets[lru] * (100 - percentage) / 100; in shrink_lruvec() [all …]
|
A D | balloon_compaction.c | 48 list_for_each_entry_safe(page, tmp, pages, lru) { in balloon_page_list_enqueue() 49 list_del(&page->lru); in balloon_page_list_enqueue() 84 list_for_each_entry_safe(page, tmp, &b_dev_info->pages, lru) { in balloon_page_list_dequeue() 104 list_add(&page->lru, pages); in balloon_page_list_dequeue() 200 return list_first_entry(&pages, struct page, lru); in balloon_page_dequeue() 213 list_del(&page->lru); in balloon_page_isolate() 226 list_add(&page->lru, &b_dev_info->pages); in balloon_page_putback()
|
A D | page_reporting.c | 151 list_for_each_entry_safe(page, next, list, lru) { in page_reporting_cycle() 186 if (!list_is_first(&page->lru, list)) in page_reporting_cycle() 187 list_rotate_to_front(&page->lru, list); in page_reporting_cycle() 211 next = list_first_entry(list, struct page, lru); in page_reporting_cycle() 219 if (!list_entry_is_head(next, list, lru) && !list_is_first(&next->lru, list)) in page_reporting_cycle() 220 list_rotate_to_front(&next->lru, list); in page_reporting_cycle()
|
A D | pgtable-generic.c | 168 INIT_LIST_HEAD(&pgtable->lru); in pgtable_trans_huge_deposit() 170 list_add(&pgtable->lru, &pmd_huge_pte(mm, pmdp)->lru); in pgtable_trans_huge_deposit() 185 pmd_huge_pte(mm, pmdp) = list_first_entry_or_null(&pgtable->lru, in pgtable_trans_huge_withdraw() 186 struct page, lru); in pgtable_trans_huge_withdraw() 188 list_del(&pgtable->lru); in pgtable_trans_huge_withdraw()
|
A D | z3fold.c | 162 struct list_head lru; member 391 INIT_LIST_HEAD(&page->lru); in init_z3fold_page() 514 list_del_init(&page->lru); in __release_z3fold_page() 999 INIT_LIST_HEAD(&pool->lru); in z3fold_create_pool() 1175 list_del(&page->lru); in z3fold_alloc() 1177 list_add(&page->lru, &pool->lru); in z3fold_alloc() 1216 list_del(&page->lru); in z3fold_free() 1446 list_add(&page->lru, &pool->lru); in z3fold_reclaim_page() 1464 list_add(&page->lru, &pool->lru); in z3fold_reclaim_page() 1654 list_add(&newpage->lru, &pool->lru); in z3fold_page_migrate() [all …]
|
/linux/include/linux/ |
A D | list_lru.h | 37 struct list_lru_one *lru[]; member 44 struct list_lru_one lru; member 61 void list_lru_destroy(struct list_lru *lru); 65 #define list_lru_init(lru) \ argument 66 __list_lru_init((lru), false, NULL, NULL) 67 #define list_lru_init_key(lru, key) \ argument 68 __list_lru_init((lru), false, (key), NULL) 69 #define list_lru_init_memcg(lru, shrinker) \ argument 70 __list_lru_init((lru), true, NULL, shrinker) 123 return list_lru_count_one(lru, sc->nid, sc->memcg); in list_lru_shrink_count() [all …]
|
A D | mm_inline.h | 34 enum lru_list lru, enum zone_type zid, in update_lru_size() argument 41 NR_ZONE_LRU_BASE + lru, nr_pages); in update_lru_size() 79 enum lru_list lru; in folio_lru_list() local 88 lru += LRU_ACTIVE; in folio_lru_list() 90 return lru; in folio_lru_list() 96 enum lru_list lru = folio_lru_list(folio); in lruvec_add_folio() local 98 update_lru_size(lruvec, lru, folio_zonenum(folio), in lruvec_add_folio() 100 list_add(&folio->lru, &lruvec->lists[lru]); in lruvec_add_folio() 112 enum lru_list lru = folio_lru_list(folio); in lruvec_add_folio_tail() local 116 list_add_tail(&folio->lru, &lruvec->lists[lru]); in lruvec_add_folio_tail() [all …]
|
A D | balloon_compaction.h | 105 list_add(&page->lru, &balloon->pages); in balloon_page_insert() 126 list_del(&page->lru); in balloon_page_delete() 149 list_add(&page->lru, &balloon->pages); in balloon_page_insert() 155 list_del(&page->lru); in balloon_page_delete() 190 list_add(&page->lru, pages); in balloon_page_push() 202 struct page *page = list_first_entry_or_null(pages, struct page, lru); in balloon_page_pop() 207 list_del(&page->lru); in balloon_page_pop()
|
A D | mmzone.h | 106 struct page, lru); in get_page_from_free_area() 284 #define for_each_lru(lru) for (lru = 0; lru < NR_LRU_LISTS; lru++) argument 286 #define for_each_evictable_lru(lru) for (lru = 0; lru <= LRU_ACTIVE_FILE; lru++) argument 288 static inline bool is_file_lru(enum lru_list lru) in is_file_lru() argument 290 return (lru == LRU_INACTIVE_FILE || lru == LRU_ACTIVE_FILE); in is_file_lru() 293 static inline bool is_active_lru(enum lru_list lru) in is_active_lru() argument 295 return (lru == LRU_ACTIVE_ANON || lru == LRU_ACTIVE_FILE); in is_active_lru()
|
/linux/kernel/bpf/ |
A D | bpf_lru_list.c | 220 } else if (lru->del_from_htab(lru->del_arg, node)) { in __bpf_lru_list_shrink_inactive() 227 if (++i == lru->nr_scans) in __bpf_lru_list_shrink_inactive() 279 if (lru->del_from_htab(lru->del_arg, node)) { in __bpf_lru_list_shrink() 385 lru->del_from_htab(lru->del_arg, node)) { in __local_list_pop_pending() 496 if (lru->percpu) in bpf_lru_pop_free() 553 if (lru->percpu) in bpf_lru_push_free() 611 if (lru->percpu) in bpf_lru_populate() 653 if (!lru->percpu_lru) in bpf_lru_init() 681 lru->percpu = percpu; in bpf_lru_init() 683 lru->del_arg = del_arg; in bpf_lru_init() [all …]
|
A D | bpf_lru_list.h | 73 int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset, 75 void bpf_lru_populate(struct bpf_lru *lru, void *buf, u32 node_offset, 77 void bpf_lru_destroy(struct bpf_lru *lru); 78 struct bpf_lru_node *bpf_lru_pop_free(struct bpf_lru *lru, u32 hash); 79 void bpf_lru_push_free(struct bpf_lru *lru, struct bpf_lru_node *node); 80 void bpf_lru_promote(struct bpf_lru *lru, struct bpf_lru_node *node);
|
/linux/samples/bpf/ |
A D | test_lru_dist.c | 115 assert(lru->free_nodes); in pfect_lru_init() 118 lru->cur_size = 0; in pfect_lru_init() 119 lru->lru_size = lru_size; in pfect_lru_init() 120 lru->nr_unique = lru->nr_misses = lru->total = 0; in pfect_lru_init() 125 close(lru->map_fd); in pfect_lru_destroy() 126 free(lru->free_nodes); in pfect_lru_destroy() 135 lru->total++; in pfect_lru_lookup_or_insert() 144 if (lru->cur_size < lru->lru_size) { in pfect_lru_lookup_or_insert() 145 node = &lru->free_nodes[lru->cur_size++]; in pfect_lru_lookup_or_insert() 159 lru->nr_misses++; in pfect_lru_lookup_or_insert() [all …]
|
/linux/net/ceph/ |
A D | pagelist.c | 45 lru); in ceph_pagelist_release() 46 list_del(&page->lru); in ceph_pagelist_release() 61 page = list_first_entry(&pl->free_list, struct page, lru); in ceph_pagelist_addpage() 62 list_del(&page->lru); in ceph_pagelist_addpage() 69 list_add_tail(&page->lru, &pl->head); in ceph_pagelist_addpage() 113 list_add_tail(&page->lru, &pl->free_list); in ceph_pagelist_reserve() 125 struct page, lru); in ceph_pagelist_free_reserve() 126 list_del(&page->lru); in ceph_pagelist_free_reserve() 159 page = list_entry(pl->head.prev, struct page, lru); in ceph_pagelist_truncate() 161 list_move_tail(&page->lru, &pl->free_list); in ceph_pagelist_truncate() [all …]
|
/linux/fs/nfs/ |
A D | nfs42xattr.c | 71 struct list_head lru; member 82 struct list_head lru; member 130 struct list_lru *lru; in nfs4_xattr_entry_lru_add() local 135 return list_lru_add(lru, &entry->lru); in nfs4_xattr_entry_lru_add() 141 struct list_lru *lru; in nfs4_xattr_entry_lru_del() local 146 return list_lru_del(lru, &entry->lru); in nfs4_xattr_entry_lru_del() 850 list_lru_isolate(lru, &cache->lru); in cache_lru_isolate() 919 list_lru_isolate(lru, &entry->lru); in entry_lru_isolate() 930 list_lru_isolate(lru, &entry->lru); in entry_lru_isolate() 945 struct list_lru *lru; in nfs4_xattr_entry_scan() local [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
A D | nv50.c | 37 struct list_head lru; member 52 struct list_head lru; member 142 eobj = list_first_entry_or_null(&imem->lru, typeof(*eobj), lru); in nv50_instobj_kmap() 148 list_del_init(&eobj->lru); in nv50_instobj_kmap() 206 list_add_tail(&iobj->lru, &imem->lru); in nv50_instobj_release() 270 list_del_init(&iobj->lru); in nv50_instobj_boot() 271 iobj->lru.next = NULL; in nv50_instobj_boot() 319 if (likely(iobj->lru.next)) in nv50_instobj_dtor() 320 list_del(&iobj->lru); in nv50_instobj_dtor() 366 INIT_LIST_HEAD(&iobj->lru); in nv50_instobj_new() [all …]
|
/linux/arch/mips/mm/ |
A D | cerr-sb1.c | 326 uint8_t lru; in extract_ic() local 348 lru = (taghi >> 14) & 0xff; in extract_ic() 352 (lru & 0x3), in extract_ic() 353 ((lru >> 2) & 0x3), in extract_ic() 354 ((lru >> 4) & 0x3), in extract_ic() 355 ((lru >> 6) & 0x3)); in extract_ic() 481 uint8_t ecc, lru; in extract_dc() local 507 (lru & 0x3), in extract_dc() 508 ((lru >> 2) & 0x3), in extract_dc() 509 ((lru >> 4) & 0x3), in extract_dc() [all …]
|
/linux/drivers/gpu/drm/ttm/ |
A D | ttm_device.c | 157 list_for_each_entry(bo, &man->lru[j], lru) { in ttm_device_swapout() 253 if (list_empty(&man->lru[0])) in ttm_device_fini() 270 bo = list_first_entry(&bdev->pinned, struct ttm_buffer_object, lru); in ttm_device_clear_dma_mappings() 273 list_del_init(&bo->lru); in ttm_device_clear_dma_mappings() 290 while (!list_empty(&man->lru[j])) { in ttm_device_clear_dma_mappings() 291 bo = list_first_entry(&man->lru[j], struct ttm_buffer_object, lru); in ttm_device_clear_dma_mappings() 293 list_del_init(&bo->lru); in ttm_device_clear_dma_mappings()
|
A D | ttm_bo.c | 76 list_move_tail(&bo->lru, &bdev->pinned); in ttm_bo_move_to_pinned() 86 list_del_init(&bo->lru); in ttm_bo_del_from_lru() 119 list_move_tail(&bo->lru, &man->lru[bo->priority]); in ttm_bo_move_to_lru_tail() 153 list_bulk_move_tail(&man->lru[i], &pos->first->lru, in ttm_bo_bulk_move_lru_tail() 154 &pos->last->lru); in ttm_bo_bulk_move_lru_tail() 168 list_bulk_move_tail(&man->lru[i], &pos->first->lru, in ttm_bo_bulk_move_lru_tail() 169 &pos->last->lru); in ttm_bo_bulk_move_lru_tail() 682 list_for_each_entry(bo, &man->lru[i], lru) { in ttm_mem_evict_first() 702 if (&bo->lru != &man->lru[i]) in ttm_mem_evict_first() 972 INIT_LIST_HEAD(&bo->lru); in ttm_bo_init_reserved()
|
/linux/fs/proc/ |
A D | meminfo.c | 40 int lru; in meminfo_proc_show() local 51 for (lru = LRU_BASE; lru < NR_LRU_LISTS; lru++) in meminfo_proc_show() 52 pages[lru] = global_node_page_state(NR_LRU_BASE + lru); in meminfo_proc_show()
|
/linux/drivers/xen/ |
A D | balloon.c | 171 list_add_tail(&page->lru, &ballooned_pages); in balloon_append() 174 list_add(&page->lru, &ballooned_pages); in balloon_append() 188 page = list_entry(ballooned_pages.next, struct page, lru); in balloon_retrieve() 191 list_del(&page->lru); in balloon_retrieve() 204 struct list_head *next = page->lru.next; in balloon_next_page() 207 return list_entry(next, struct page, lru); in balloon_next_page() 412 page = list_first_entry_or_null(&ballooned_pages, struct page, lru); in increase_reservation() 462 list_add(&page->lru, &pages); in decrease_reservation() 479 list_for_each_entry_safe(page, tmp, &pages, lru) { in decrease_reservation() 484 list_del(&page->lru); in decrease_reservation()
|
/linux/drivers/md/ |
A D | dm-writecache.c | 84 struct list_head lru; member 114 struct list_head lru; member 670 list_add(&ins->lru, &wc->lru); in writecache_insert_entry() 676 list_del(&e->lru); in writecache_unlink() 739 list_del(&e->lru); in writecache_pop_from_freelist() 804 e = container_of(wc->lru.next, struct wc_entry, lru); in writecache_flush() 815 if (unlikely(e->lru.next == &wc->lru)) in writecache_flush() 817 e2 = container_of(e->lru.next, struct wc_entry, lru); in writecache_flush() 847 if (unlikely(e->lru.prev == &wc->lru)) in writecache_flush() 849 e = container_of(e->lru.prev, struct wc_entry, lru); in writecache_flush() [all …]
|
/linux/tools/testing/selftests/bpf/progs/ |
A D | timer.c | 47 } lru SEC(".maps"); 79 lru_timer = bpf_map_lookup_elem(&lru, &lru_key); in timer_cb1() 135 bpf_map_update_elem(&lru, &lru_key, &init, 0); in BPF_PROG() 136 lru_timer = bpf_map_lookup_elem(&lru, &lru_key); in BPF_PROG() 139 bpf_timer_init(lru_timer, &lru, CLOCK_MONOTONIC); in BPF_PROG()
|
/linux/kernel/power/ |
A D | wakelock.c | 32 struct list_head lru; member 95 list_add(&wl->lru, &wakelocks_lru_list); in wakelocks_lru_add() 100 list_move(&wl->lru, &wakelocks_lru_list); in wakelocks_lru_most_recent() 111 list_for_each_entry_safe_reverse(wl, aux, &wakelocks_lru_list, lru) { in __wakelocks_gc() 126 list_del(&wl->lru); in __wakelocks_gc()
|