Lines Matching refs:lruvec
693 void __mod_memcg_lruvec_state(struct lruvec *lruvec, enum node_stat_item idx, in __mod_memcg_lruvec_state() argument
699 pn = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in __mod_memcg_lruvec_state()
721 void __mod_lruvec_state(struct lruvec *lruvec, enum node_stat_item idx, in __mod_lruvec_state() argument
725 __mod_node_page_state(lruvec_pgdat(lruvec), idx, val); in __mod_lruvec_state()
729 __mod_memcg_lruvec_state(lruvec, idx, val); in __mod_lruvec_state()
738 struct lruvec *lruvec; in __mod_lruvec_page_state() local
749 lruvec = mem_cgroup_lruvec(memcg, pgdat); in __mod_lruvec_page_state()
750 __mod_lruvec_state(lruvec, idx, val); in __mod_lruvec_page_state()
759 struct lruvec *lruvec; in __mod_lruvec_kmem_state() local
773 lruvec = mem_cgroup_lruvec(memcg, pgdat); in __mod_lruvec_kmem_state()
774 __mod_lruvec_state(lruvec, idx, val); in __mod_lruvec_kmem_state()
1160 void lruvec_memcg_debug(struct lruvec *lruvec, struct folio *folio) in lruvec_memcg_debug() argument
1170 VM_BUG_ON_FOLIO(lruvec_memcg(lruvec) != root_mem_cgroup, folio); in lruvec_memcg_debug()
1172 VM_BUG_ON_FOLIO(lruvec_memcg(lruvec) != memcg, folio); in lruvec_memcg_debug()
1188 struct lruvec *folio_lruvec_lock(struct folio *folio) in folio_lruvec_lock()
1190 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock() local
1192 spin_lock(&lruvec->lru_lock); in folio_lruvec_lock()
1193 lruvec_memcg_debug(lruvec, folio); in folio_lruvec_lock()
1195 return lruvec; in folio_lruvec_lock()
1211 struct lruvec *folio_lruvec_lock_irq(struct folio *folio) in folio_lruvec_lock_irq()
1213 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock_irq() local
1215 spin_lock_irq(&lruvec->lru_lock); in folio_lruvec_lock_irq()
1216 lruvec_memcg_debug(lruvec, folio); in folio_lruvec_lock_irq()
1218 return lruvec; in folio_lruvec_lock_irq()
1235 struct lruvec *folio_lruvec_lock_irqsave(struct folio *folio, in folio_lruvec_lock_irqsave()
1238 struct lruvec *lruvec = folio_lruvec(folio); in folio_lruvec_lock_irqsave() local
1240 spin_lock_irqsave(&lruvec->lru_lock, *flags); in folio_lruvec_lock_irqsave()
1241 lruvec_memcg_debug(lruvec, folio); in folio_lruvec_lock_irqsave()
1243 return lruvec; in folio_lruvec_lock_irqsave()
1257 void mem_cgroup_update_lru_size(struct lruvec *lruvec, enum lru_list lru, in mem_cgroup_update_lru_size() argument
1267 mz = container_of(lruvec, struct mem_cgroup_per_node, lruvec); in mem_cgroup_update_lru_size()
1276 __func__, lruvec, lru, nr_pages, size)) { in mem_cgroup_update_lru_size()
2810 struct lruvec *lruvec; in mod_objcg_mlstate() local
2814 lruvec = mem_cgroup_lruvec(memcg, pgdat); in mod_objcg_mlstate()
2815 mod_memcg_lruvec_state(lruvec, idx, nr); in mod_objcg_mlstate()
3839 struct lruvec *lruvec = mem_cgroup_lruvec(memcg, NODE_DATA(nid)); in mem_cgroup_node_nr_lru_pages() local
3849 nr += lruvec_page_state(lruvec, NR_LRU_BASE + lru); in mem_cgroup_node_nr_lru_pages()
3851 nr += lruvec_page_state_local(lruvec, NR_LRU_BASE + lru); in mem_cgroup_node_nr_lru_pages()
4025 anon_cost += mz->lruvec.anon_cost; in memcg_stat_show()
4026 file_cost += mz->lruvec.file_cost; in memcg_stat_show()
5069 lruvec_init(&pn->lruvec); in alloc_mem_cgroup_per_node_info()
5554 struct lruvec *from_vec, *to_vec; in mem_cgroup_move_account()
6342 static inline unsigned long lruvec_page_state_output(struct lruvec *lruvec, in lruvec_page_state_output() argument
6345 return lruvec_page_state(lruvec, item) * memcg_page_state_unit(item); in lruvec_page_state_output()
6364 struct lruvec *lruvec; in memory_numa_stat_show() local
6366 lruvec = mem_cgroup_lruvec(memcg, NODE_DATA(nid)); in memory_numa_stat_show()
6367 size = lruvec_page_state_output(lruvec, in memory_numa_stat_show()