| /linux/kernel/rcu/ |
| A D | tree_stall.h | 512 data_race(rcu_state.gp_flags), in rcu_check_gp_kthread_expired_fqs_timer() 514 data_race(READ_ONCE(gpk->__state))); in rcu_check_gp_kthread_expired_fqs_timer() 770 if (data_race(READ_ONCE(rnp->qsmask))) { in rcu_check_boost_fail() 825 data_race(READ_ONCE(rcu_state.gp_state)), in show_rcu_gp_kthreads() 830 data_race(READ_ONCE(rcu_state.gp_max)), in show_rcu_gp_kthreads() 831 data_race(READ_ONCE(rcu_state.gp_flags))); in show_rcu_gp_kthreads() 834 !data_race(READ_ONCE(rnp->qsmask)) && !data_race(READ_ONCE(rnp->boost_tasks)) && in show_rcu_gp_kthreads() 835 !data_race(READ_ONCE(rnp->exp_tasks)) && !data_race(READ_ONCE(rnp->gp_tasks))) in show_rcu_gp_kthreads() 839 (long)data_race(READ_ONCE(rnp->gp_seq)), in show_rcu_gp_kthreads() 841 data_race(READ_ONCE(rnp->qsmask)), in show_rcu_gp_kthreads() [all …]
|
| A D | tasks.h | 142 int i = data_race(rtp->gp_state); // Let KCSAN detect update races in tasks_gp_state_getname() 284 tasks_gp_state_getname(rtp), data_race(rtp->gp_state), in show_rcu_tasks_generic_gp_kthread() 285 jiffies - data_race(rtp->gp_jiffies), in show_rcu_tasks_generic_gp_kthread() 286 data_race(rtp->n_gps), in show_rcu_tasks_generic_gp_kthread() 287 data_race(rtp->n_ipis_fails), data_race(rtp->n_ipis), in show_rcu_tasks_generic_gp_kthread() 288 ".k"[!!data_race(rtp->kthread_ptr)], in show_rcu_tasks_generic_gp_kthread() 289 ".C"[!!data_race(rtp->cbs_head)], in show_rcu_tasks_generic_gp_kthread() 1292 data_race(n_heavy_reader_ofl_updates), in show_rcu_tasks_trace_gp_kthread() 1293 data_race(n_heavy_reader_updates), in show_rcu_tasks_trace_gp_kthread() 1294 data_race(n_heavy_reader_attempts)); in show_rcu_tasks_trace_gp_kthread()
|
| A D | tree_exp.h | 545 data_race(rnp_root->expmask), in synchronize_rcu_expedited_wait() 546 ".T"[!!data_race(rnp_root->exp_tasks)]); in synchronize_rcu_expedited_wait() 556 data_race(rnp->expmask), in synchronize_rcu_expedited_wait() 557 ".T"[!!data_race(rnp->exp_tasks)]); in synchronize_rcu_expedited_wait()
|
| A D | srcutree.c | 1356 u0 = data_race(sdp->srcu_unlock_count[!idx]); in srcu_torture_stats_print() 1357 u1 = data_race(sdp->srcu_unlock_count[idx]); in srcu_torture_stats_print() 1365 l0 = data_race(sdp->srcu_lock_count[!idx]); in srcu_torture_stats_print() 1366 l1 = data_race(sdp->srcu_lock_count[idx]); in srcu_torture_stats_print()
|
| A D | rcutorture.c | 1811 data_race(n_barrier_successes), in rcu_torture_stats_print() 1812 data_race(n_barrier_attempts), in rcu_torture_stats_print() 1813 data_race(n_rcu_torture_barrier_error)); in rcu_torture_stats_print() 1814 pr_cont("read-exits: %ld ", data_race(n_read_exits)); // Statistic. in rcu_torture_stats_print()
|
| /linux/tools/memory-model/Documentation/ |
| A D | access-marking.txt | 19 2. Data-race marking, for example, "data_race(a = b);" 31 WRITE_ONCE(a, b + data_race(c + d) + READ_ONCE(e)); 58 Use of the data_race() Macro 93 data_race(READ_ONCE(a)). 138 instead of data_race(). 226 write is marked with data_race() or WRITE_ONCE(). 231 of these data_race() writes. 275 pr_info("Current value of foo: %d\n", data_race(foo)); 371 if (!data_race(global_flag)) { 535 newold = data_race(foo); /* Checked by cmpxchg(). */ [all …]
|
| /linux/include/linux/ |
| A D | srcutiny.h | 84 idx = ((data_race(READ_ONCE(ssp->srcu_idx)) + 1) & 0x2) >> 1; in srcu_torture_stats_print() 87 data_race(READ_ONCE(ssp->srcu_lock_nesting[!idx])), in srcu_torture_stats_print() 88 data_race(READ_ONCE(ssp->srcu_lock_nesting[idx]))); in srcu_torture_stats_print()
|
| A D | compiler.h | 214 #define data_race(expr) \ macro
|
| /linux/mm/ |
| A D | frontswap.c | 65 data_race(frontswap_loads++); in inc_frontswap_loads() 69 data_race(frontswap_succ_stores++); in inc_frontswap_succ_stores() 73 data_race(frontswap_failed_stores++); in inc_frontswap_failed_stores() 77 data_race(frontswap_invalidates++); in inc_frontswap_invalidates()
|
| A D | page_io.c | 69 if (data_race(!(sis->flags & SWP_BLKDEV))) in swap_slot_free_notify() 293 if (data_race(sis->flags & SWP_FS_OPS)) { in __swap_writepage() 380 if (data_race(sis->flags & SWP_FS_OPS)) { in swap_readpage() 442 if (data_race(sis->flags & SWP_FS_OPS)) { in swap_set_page_dirty()
|
| A D | page_counter.c | 129 data_race(c->failcnt++); in page_counter_try_charge()
|
| A D | swap_state.c | 61 #define INC_CACHE_INFO(x) data_race(swap_cache_info.x++) 62 #define ADD_CACHE_INFO(x, nr) data_race(swap_cache_info.x += (nr))
|
| A D | swap.c | 617 if (data_race(pagevec_count(pvec))) { in lru_add_drain_cpu() 836 data_race(pagevec_count(&per_cpu(lru_rotate.pvec, cpu))) || in __lru_add_drain_all()
|
| A D | vmstat.c | 1079 blocks = data_race(zone->free_area[order].nr_free); in fill_contig_page_info() 1458 seq_printf(m, "%6lu ", data_race(zone->free_area[order].nr_free)); in frag_show_print()
|
| A D | swapfile.c | 954 if (data_race(!si->swap_map[offset])) { in scan_swap_map_slots() 971 if (data_race(!si->swap_map[offset])) { in scan_swap_map_slots() 1134 if (data_race(!(p->flags & SWP_USED))) in __swap_info_get() 1160 if (data_race(!p->swap_map[swp_offset(entry)])) in _swap_info_get()
|
| /linux/security/tomoyo/ |
| A D | util.c | 1069 perm = data_race(container_of(ptr, struct tomoyo_path_acl, head)->perm); in tomoyo_domain_quota_is_ok() 1072 perm = data_race(container_of(ptr, struct tomoyo_path2_acl, head)->perm); in tomoyo_domain_quota_is_ok() 1075 perm = data_race(container_of(ptr, struct tomoyo_path_number_acl, head) in tomoyo_domain_quota_is_ok() 1079 perm = data_race(container_of(ptr, struct tomoyo_mkdev_acl, head)->perm); in tomoyo_domain_quota_is_ok() 1082 perm = data_race(container_of(ptr, struct tomoyo_inet_acl, head)->perm); in tomoyo_domain_quota_is_ok() 1085 perm = data_race(container_of(ptr, struct tomoyo_unix_acl, head)->perm); in tomoyo_domain_quota_is_ok()
|
| /linux/kernel/locking/ |
| A D | osq_lock.c | 161 if (data_race(prev->next) == node && in osq_lock()
|
| A D | locktorture.c | 744 long max = 0, min = statp ? data_race(statp[0].n_lock_acquired) : 0; in __torture_print_stats() 749 if (data_race(statp[i].n_lock_fail)) in __torture_print_stats() 751 cur = data_race(statp[i].n_lock_acquired); in __torture_print_stats()
|
| /linux/kernel/irq/ |
| A D | irqdesc.c | 905 return data_race(desc->tot_count); in kstat_irqs() 908 sum += data_race(*per_cpu_ptr(desc->kstat_irqs, cpu)); in kstat_irqs()
|
| A D | proc.c | 493 any_count |= data_race(*per_cpu_ptr(desc->kstat_irqs, j)); in show_interrupts()
|
| /linux/mm/kfence/ |
| A D | core.c | 975 distance = addr - data_race(meta->addr + meta->size); in kfence_handle_page_fault() 981 if (!to_report || distance > data_race(meta->addr) - addr) in kfence_handle_page_fault()
|
| /linux/fs/jbd2/ |
| A D | transaction.c | 368 if (!data_race(journal->j_running_transaction)) { in start_this_handle() 1499 if (data_race(jh->b_transaction != transaction && in jbd2_journal_dirty_metadata() 1508 if (data_race(jh->b_transaction == transaction && in jbd2_journal_dirty_metadata()
|
| /linux/kernel/ |
| A D | scftorture.c | 164 invoked_count += data_race(per_cpu(scf_invoked_count, cpu)); in scf_torture_stats_print()
|
| /linux/lib/ |
| A D | sbitmap.c | 634 data_race(*per_cpu_ptr(sb->alloc_hint, cpu) = tag); in sbitmap_update_cpu_hint()
|
| /linux/net/ipv4/ |
| A D | af_inet.c | 574 if (data_race(!inet_sk(sk)->inet_num) && inet_autobind(sk)) in inet_dgram_connect() 804 if (data_race(!inet_sk(sk)->inet_num) && !sk->sk_prot->no_autobind && in inet_send_prepare()
|