/linux/tools/perf/util/ |
A D | hist.c | 249 struct rb_node *next = rb_first_cached(&hists->entries); in hists__output_recalc_col_len() 1654 next = rb_first_cached(root); in hists__collapse_resort() 1733 node = rb_first_cached(&hists->entries); in hierarchy_recalc_total_periods() 1795 node = rb_first_cached(root_in); in hists__hierarchy_output_resort() 1918 next = rb_first_cached(root); in output_resort() 1997 node = rb_first_cached(&he->hroot_out); in __rb_hierarchy_next() 2035 node = rb_first_cached(&he->hroot_out); in hist_entry__has_hierarchy_children() 2191 nd = rb_first_cached(&he->hroot_out); in resort_filtered_entry() 2214 nd = rb_first_cached(&hists->entries); in hists__filter_hierarchy() 2258 nd = rb_first_cached(&hists->entries); in hists__filter_hierarchy() [all …]
|
A D | rblist.c | 112 struct rb_node *pos, *next = rb_first_cached(&rblist->entries); in rblist__exit() 133 for (node = rb_first_cached(&rblist->entries); node; in rblist__entry()
|
A D | intlist.h | 48 struct rb_node *rn = rb_first_cached(&ilist->rblist.entries); in intlist__first()
|
A D | strlist.h | 60 struct rb_node *rn = rb_first_cached(&slist->rblist.entries); in strlist__first()
|
A D | symbol_fprintf.c | 69 for (nd = rb_first_cached(&dso->symbol_names); nd; nd = rb_next(nd)) { in dso__fprintf_symbols_by_name()
|
A D | build-id.c | 387 for (nd = rb_first_cached(&session->machines.guests); nd; in perf_session__write_buildid_table() 421 for (nd = rb_first_cached(&session->machines.guests); nd; in dsos__hit_all() 899 for (nd = rb_first_cached(&machines->guests); nd; in machines__for_each_dso() 935 for (nd = rb_first_cached(&session->machines.guests); nd; in perf_session__read_build_ids()
|
A D | machine.c | 202 nd = rb_first_cached(&threads->entries); in machine__delete_threads() 305 for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) { in machines__set_comm_exec() 386 for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) { in machines__process_guests() 399 for (node = rb_first_cached(&machines->guests); node; in machines__set_id_hdr_size() 931 for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) { in machines__fprintf_dsos() 951 for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) { in machines__fprintf_dsos_buildid() 991 for (nd = rb_first_cached(&threads->entries); nd; in machine__fprintf() 1289 struct rb_node *next = rb_first_cached(&machines->guests); in machines__destroy_kernel_maps() 2996 for (nd = rb_first_cached(&threads->entries); nd; in machine__for_each_thread() 3024 for (nd = rb_first_cached(&machines->guests); nd; nd = rb_next(nd)) { in machines__for_each_thread()
|
A D | symbol.c | 194 nd = rb_first_cached(symbols); in symbols__fixup_duplicate() 222 struct rb_node *nd, *prevnd = rb_first_cached(symbols); in symbols__fixup_end() 310 struct rb_node *next = rb_first_cached(symbols); in symbols__delete() 384 struct rb_node *n = rb_first_cached(symbols); in symbols__first() 440 for (nd = rb_first_cached(source); nd; nd = rb_next(nd)) { in symbols__sort_by_name() 750 struct rb_node *next = rb_first_cached(root); in maps__split_kallsyms_for_kcore() 804 struct rb_node *next = rb_first_cached(root); in maps__split_kallsyms()
|
A D | symbol.h | 83 for (nd = rb_first_cached(symbols); \
|
A D | srcline.c | 828 struct rb_node *next = rb_first_cached(tree); in srcline__tree_delete() 912 struct rb_node *next = rb_first_cached(tree); in inlines__tree_delete()
|
A D | stream.c | 115 struct rb_node *next = rb_first_cached(&hists->entries); in init_hot_callchain()
|
/linux/tools/perf/tests/ |
A D | hists_output.c | 107 node = rb_first_cached(root_out); in del_hist_entries() 166 node = rb_first_cached(root); in test1() 266 node = rb_first_cached(root); in test2() 320 node = rb_first_cached(root); in test3() 398 node = rb_first_cached(root); in test4() 501 node = rb_first_cached(root); in test5()
|
A D | hists_common.c | 175 node = rb_first_cached(root); in print_hists_in() 202 node = rb_first_cached(root); in print_hists_out()
|
A D | hists_link.c | 155 node = rb_first_cached(root); in __validate_match() 207 node = rb_first_cached(root); in __validate_link()
|
/linux/kernel/locking/ |
A D | rtmutex_common.h | 106 struct rb_node *leftmost = rb_first_cached(&lock->waiters); in rt_mutex_waiter_is_top_waiter() 113 struct rb_node *leftmost = rb_first_cached(&lock->waiters); in rt_mutex_top_waiter()
|
/linux/include/linux/ |
A D | timerqueue.h | 36 struct rb_node *leftmost = rb_first_cached(&head->rb_root); in timerqueue_getnext()
|
A D | rbtree.h | 106 #define rb_first_cached(root) (root)->rb_leftmost macro
|
/linux/drivers/gpu/drm/i915/ |
A D | i915_scheduler.c | 47 GEM_BUG_ON(rb_first_cached(&sched_engine->queue) != in assert_priolists() 51 for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) { in assert_priolists()
|
/linux/net/sched/ |
A D | sch_etf.c | 113 p = rb_first_cached(&q->head); in etf_peek_timesortedlist() 423 struct rb_node *p = rb_first_cached(&q->head); in timesortedlist_clear()
|
/linux/drivers/gpu/drm/i915/gt/ |
A D | intel_execlists_submission.c | 279 rb = rb_first_cached(&sched_engine->queue); in queue_prio() 288 struct rb_node *rb = rb_first_cached(&el->virtual); in virtual_prio() 1012 struct rb_node *rb = rb_first_cached(&el->virtual); in first_virtual_engine() 1023 rb = rb_first_cached(&el->virtual); in first_virtual_engine() 1472 while ((rb = rb_first_cached(&sched_engine->queue))) { in execlists_dequeue() 3080 while ((rb = rb_first_cached(&sched_engine->queue))) { in execlists_reset_cancel() 3099 while ((rb = rb_first_cached(&execlists->virtual))) { in execlists_reset_cancel() 3707 first = rb_first_cached(&sibling->execlists.virtual) == in virtual_submission_tasklet() 3969 for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) { in intel_execlists_show_requests() 3990 for (rb = rb_first_cached(&execlists->virtual); rb; rb = rb_next(rb)) { in intel_execlists_show_requests()
|
/linux/fs/ |
A D | eventpoll.c | 753 for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) { in ep_free() 769 while ((rbp = rb_first_cached(&ep->rbr)) != NULL) { in ep_free() 864 for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) { in ep_show_fdinfo() 994 for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) { in ep_find_tfd() 1897 for (rbp = rb_first_cached(&ep->rbr); rbp; rbp = rb_next(rbp)) { in ep_loop_check_proc()
|
/linux/tools/include/linux/ |
A D | rbtree.h | 127 #define rb_first_cached(root) (root)->rb_leftmost macro
|
/linux/tools/perf/ui/gtk/ |
A D | hists.c | 358 for (nd = rb_first_cached(&hists->entries); nd; nd = rb_next(nd)) { in perf_gtk__show_hists() 420 for (node = rb_first_cached(root); node; node = rb_next(node)) { in perf_gtk__add_hierarchy_entries()
|
/linux/tools/perf/ |
A D | builtin-diff.c | 549 next = rb_first_cached(root); in hists__baseline_only() 600 struct rb_node *next = rb_first_cached(root); in get_block_pair() 667 struct rb_node *next = rb_first_cached(root); in block_hists_match() 693 next = rb_first_cached(root); in hists__precompute()
|
/linux/drivers/infiniband/hw/hfi1/ |
A D | mmu_rb.c | 95 while ((node = rb_first_cached(&handler->root))) { in hfi1_mmu_rb_unregister()
|