/linux/tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/ |
A D | locks.h | 192 unsigned int prev_count = __sync_fetch_and_sub(&c->count, 1); in wait_for_completion() local 194 assume(prev_count); in wait_for_completion() 199 unsigned int prev_count = __sync_fetch_and_add(&c->count, 1); in complete() local 201 BUG_ON(prev_count == UINT_MAX); in complete()
|
/linux/arch/arm/mm/ |
A D | cache-l2x0-pmu.c | 120 u64 prev_count, new_count, mask; in l2x0_pmu_event_read() local 123 prev_count = local64_read(&hw->prev_count); in l2x0_pmu_event_read() 125 } while (local64_xchg(&hw->prev_count, new_count) != prev_count); in l2x0_pmu_event_read() 128 local64_add((new_count - prev_count) & mask, &event->count); in l2x0_pmu_event_read() 147 local64_set(&hw->prev_count, 0); in l2x0_pmu_event_configure()
|
/linux/drivers/perf/ |
A D | arm_dmc620_pmu.c | 300 u64 delta, prev_count, new_count; in dmc620_pmu_event_update() local 304 prev_count = local64_read(&hwc->prev_count); in dmc620_pmu_event_update() 306 } while (local64_cmpxchg(&hwc->prev_count, in dmc620_pmu_event_update() 307 prev_count, new_count) != prev_count); in dmc620_pmu_event_update() 308 delta = (new_count - prev_count) & DMC620_CNT_MAX_PERIOD; in dmc620_pmu_event_update() 316 local64_set(&event->hw.prev_count, DMC620_CNT_INIT); in dmc620_pmu_event_set_period()
|
A D | arm_dsu_pmu.c | 352 u64 delta, prev_count, new_count; in dsu_pmu_event_update() local 356 prev_count = local64_read(&hwc->prev_count); in dsu_pmu_event_update() 358 } while (local64_cmpxchg(&hwc->prev_count, prev_count, new_count) != in dsu_pmu_event_update() 359 prev_count); in dsu_pmu_event_update() 360 delta = (new_count - prev_count) & DSU_PMU_COUNTER_MASK(hwc->idx); in dsu_pmu_event_update() 386 local64_set(&event->hw.prev_count, val); in dsu_pmu_set_event_period()
|
A D | qcom_l3_pmu.c | 206 local64_set(&event->hw.prev_count, 0); in qcom_l3_cache__64bit_counter_start() 247 prev = local64_read(&event->hw.prev_count); in qcom_l3_cache__64bit_counter_update() 253 } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); in qcom_l3_cache__64bit_counter_update() 285 local64_set(&event->hw.prev_count, 0); in qcom_l3_cache__32bit_counter_start() 323 prev = local64_read(&event->hw.prev_count); in qcom_l3_cache__32bit_counter_update() 325 } while (local64_cmpxchg(&event->hw.prev_count, prev, new) != prev); in qcom_l3_cache__32bit_counter_update()
|
A D | thunderx2_pmu.c | 379 local64_set(&hwc->prev_count, 0); in uncore_start_event_l3c() 409 local64_set(&hwc->prev_count, 0); in uncore_start_event_dmc() 452 local64_set(&event->hw.prev_count, 0ULL); in uncore_start_event_ccpi2() 484 prev = local64_xchg(&hwc->prev_count, new); in tx2_uncore_event_update() 488 prev = local64_xchg(&hwc->prev_count, new); in tx2_uncore_event_update()
|
A D | arm_smmuv3_pmu.c | 212 prev = local64_read(&hwc->prev_count); in smmu_pmu_event_update() 214 } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev); in smmu_pmu_event_update() 249 local64_set(&hwc->prev_count, new); in smmu_pmu_set_period() 447 local64_set(&hwc->prev_count, 0); in smmu_pmu_event_add()
|
A D | qcom_l2_pmu.c | 304 prev = local64_read(&hwc->prev_count); in l2_cache_event_update() 306 } while (local64_cmpxchg(&hwc->prev_count, prev, now) != prev); in l2_cache_event_update() 335 local64_set(&hwc->prev_count, new); in l2_cache_cluster_set_period() 604 local64_set(&hwc->prev_count, 0); in l2_cache_event_add()
|
/linux/arch/powerpc/perf/ |
A D | 8xx-pmu.c | 115 local64_set(&event->hw.prev_count, val); in mpc8xx_pmu_add() 128 prev = local64_read(&event->hw.prev_count); in mpc8xx_pmu_read() 149 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); in mpc8xx_pmu_read()
|
A D | core-fsl-emb.c | 175 prev = local64_read(&event->hw.prev_count); in fsl_emb_pmu_read() 178 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); in fsl_emb_pmu_read() 314 local64_set(&event->hw.prev_count, val); in fsl_emb_pmu_add() 602 prev = local64_read(&event->hw.prev_count); in record_and_restart() 625 local64_set(&event->hw.prev_count, val); in record_and_restart()
|
A D | core-book3s.c | 1127 local64_set(&event->hw.prev_count, val); in power_pmu_read() 1137 prev = local64_read(&event->hw.prev_count); in power_pmu_read() 1143 } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev); in power_pmu_read() 1187 prev = local64_read(&event->hw.prev_count); in freeze_limited_counters() 1206 prev = local64_read(&event->hw.prev_count); in thaw_limited_counters() 1208 local64_set(&event->hw.prev_count, val); in thaw_limited_counters() 1473 val = local64_read(&event->hw.prev_count); in power_pmu_enable() 1481 local64_set(&event->hw.prev_count, val); in power_pmu_enable() 2083 local64_set(&event->hw.prev_count, 0); in power_pmu_event_init() 2158 prev = local64_read(&event->hw.prev_count); in record_and_restart() [all …]
|
A D | hv-gpci.c | 205 prev = local64_xchg(&event->hw.prev_count, now); in h_gpci_event_update() 211 local64_set(&event->hw.prev_count, h_gpci_get_value(event)); in h_gpci_event_start()
|
/linux/arch/x86/events/ |
A D | msr.c | 239 prev = local64_read(&event->hw.prev_count); in msr_event_update() 242 if (local64_cmpxchg(&event->hw.prev_count, prev, now) != prev) in msr_event_update() 262 local64_set(&event->hw.prev_count, now); in msr_event_start()
|
/linux/fs/gfs2/ |
A D | file.c | 773 size_t *prev_count, in should_fault_in_pages() argument 787 if (*prev_count != count || !*window_size) { in should_fault_in_pages() 796 *prev_count = count; in should_fault_in_pages() 807 size_t prev_count = 0, window_size = 0; in gfs2_file_direct_read() local 846 if (should_fault_in_pages(ret, to, &prev_count, &window_size)) { in gfs2_file_direct_read() 873 size_t prev_count = 0, window_size = 0; in gfs2_file_direct_write() local 915 if (should_fault_in_pages(ret, from, &prev_count, &window_size)) { in gfs2_file_direct_write() 941 size_t prev_count = 0, window_size = 0; in gfs2_file_read_iter() local 984 if (should_fault_in_pages(ret, to, &prev_count, &window_size)) { in gfs2_file_read_iter() 1015 size_t prev_count = 0, window_size = 0; in gfs2_file_buffered_write() local [all …]
|
/linux/kernel/ |
A D | softirq.c | 548 int prev_count; in __do_softirq() local 553 prev_count = preempt_count(); in __do_softirq() 560 if (unlikely(prev_count != preempt_count())) { in __do_softirq() 563 prev_count, preempt_count()); in __do_softirq() 564 preempt_count_set(prev_count); in __do_softirq()
|
/linux/drivers/perf/hisilicon/ |
A D | hisi_uncore_pmu.c | 284 local64_set(&hwc->prev_count, val); in hisi_uncore_pmu_set_event_period() 299 prev_raw_count = local64_read(&hwc->prev_count); in hisi_uncore_pmu_event_update() 300 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in hisi_uncore_pmu_event_update() 324 u64 prev_raw_count = local64_read(&hwc->prev_count); in hisi_uncore_pmu_start()
|
/linux/arch/riscv/kernel/ |
A D | perf_event.c | 223 prev_raw_count = local64_read(&hwc->prev_count); in riscv_pmu_read() 226 oldval = local64_cmpxchg(&hwc->prev_count, prev_raw_count, in riscv_pmu_read() 291 local64_set(&hwc->prev_count, read_counter(hwc->idx)); in riscv_pmu_start()
|
/linux/arch/arm/mach-imx/ |
A D | mmdc.c | 312 prev_raw_count = local64_read(&hwc->prev_count); in mmdc_pmu_event_update() 315 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in mmdc_pmu_event_update() 340 local64_set(&hwc->prev_count, 0); in mmdc_pmu_event_start() 375 local64_set(&hwc->prev_count, mmdc_pmu_read_counter(pmu_mmdc, cfg)); in mmdc_pmu_event_add()
|
/linux/arch/xtensa/kernel/ |
A D | perf_event.c | 143 prev_raw_count = local64_read(&hwc->prev_count); in xtensa_perf_event_update() 145 } while (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in xtensa_perf_event_update() 181 local64_set(&hwc->prev_count, -left); in xtensa_perf_event_set_period()
|
/linux/tools/perf/arch/x86/util/ |
A D | iostat.c | 437 struct perf_counts_values *prev_count = in iostat_print_metric() local 440 prev_count_val = prev_count->val; in iostat_print_metric() 441 prev_count->val = count->val; in iostat_print_metric()
|
/linux/arch/arc/kernel/ |
A D | perf_event.c | 121 u64 prev_raw_count = local64_read(&hwc->prev_count); in arc_perf_event_update() 129 local64_set(&hwc->prev_count, new_raw_count); in arc_perf_event_update() 270 local64_set(&hwc->prev_count, value); in arc_pmu_event_set_period() 388 local64_set(&hwc->prev_count, 0); in arc_pmu_add()
|
/linux/arch/sh/kernel/ |
A D | perf_event.c | 178 prev_raw_count = local64_read(&hwc->prev_count); in sh_perf_event_update() 181 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in sh_perf_event_update()
|
/linux/arch/x86/events/intel/ |
A D | uncore.c | 257 u64 prev_count, new_count, delta; in uncore_perf_event_update() local 269 prev_count = local64_read(&event->hw.prev_count); in uncore_perf_event_update() 271 if (local64_xchg(&event->hw.prev_count, new_count) != prev_count) in uncore_perf_event_update() 274 delta = (new_count << shift) - (prev_count << shift); in uncore_perf_event_update() 508 local64_set(&event->hw.prev_count, in uncore_pmu_event_start() 523 local64_set(&event->hw.prev_count, uncore_read_counter(box, event)); in uncore_pmu_event_start()
|
A D | cstate.c | 367 prev_raw_count = local64_read(&hwc->prev_count); in cstate_pmu_event_update() 370 if (local64_cmpxchg(&hwc->prev_count, prev_raw_count, in cstate_pmu_event_update() 379 local64_set(&event->hw.prev_count, cstate_pmu_read_counter(event)); in cstate_pmu_event_start()
|
/linux/arch/x86/events/amd/ |
A D | uncore.c | 93 prev = local64_read(&hwc->prev_count); in amd_uncore_read() 95 local64_set(&hwc->prev_count, new); in amd_uncore_read() 106 wrmsrl(hwc->event_base, (u64)local64_read(&hwc->prev_count)); in amd_uncore_start()
|