/linux/kernel/kcsan/ |
A D | kcsan_test.c | 149 } access[2]; member 184 cmp < 0 ? r->access[0].fn : r->access[1].fn, in report_matches() 185 cmp < 0 ? r->access[1].fn : r->access[0].fn); in report_matches() 237 r->access[i].addr, r->access[i].size); in report_matches() 442 .access = { in test_basic() 448 .access = { in test_basic() 473 .access = { in test_concurrent_races() 480 .access = { in test_concurrent_races() 502 .access = { in test_novalue_change() 508 .access = { in test_novalue_change() [all …]
|
/linux/security/ |
A D | device_cgroup.c | 37 short access; member 106 walk->access |= ex->access; in dev_exception_add() 134 walk->access &= ~ex->access; in dev_exception_rm() 135 if (!walk->access) { in dev_exception_rm() 242 if (access & DEVCG_ACC_READ) in set_access() 330 if (access & (~ex->access)) in match_exception() 376 if (!(access & ex->access)) in match_exception_partial() 419 refex->access); in verify_new_ex() 434 refex->access); in verify_new_ex() 808 short access) in devcgroup_legacy_check_permission() argument [all …]
|
/linux/sound/core/ |
A D | control_led.c | 49 unsigned int access; member 165 lctl->access = access; in snd_ctl_led_set_state() 206 if (lctl && (access == 0 || access != lctl->access)) { in snd_ctl_led_remove() 207 ret = lctl->access; in snd_ctl_led_remove() 223 if (access) in snd_ctl_led_notify() 227 access = vd->access & SNDRV_CTL_ELEM_ACCESS_LED_MASK; in snd_ctl_led_notify() 231 if (access) in snd_ctl_led_notify() 236 access = vd->access & SNDRV_CTL_ELEM_ACCESS_LED_MASK; in snd_ctl_led_notify() 237 if (access) in snd_ctl_led_notify() 258 access = vd->access & SNDRV_CTL_ELEM_ACCESS_LED_MASK; in snd_ctl_led_set_id() [all …]
|
/linux/Documentation/admin-guide/LSM/ |
A D | Smack.rst | 131 access 161 access to deny. The access strings may contain only the characters 224 object label, and the third the requested access. The access 239 object label, and the third the requested access. The access 306 Writing a Smack label here sets the access to '-' for all access 351 access to pieces of data. These schemes are called discretionary access 387 Enforcement and avoided by defining access controls in terms of the access 514 of access allowed. The access specification is searched for letters that 515 describe access modes: 710 write access on LABEL1, and access to the subnet @IP2/MASK if it has write [all …]
|
/linux/tools/testing/selftests/landlock/ |
A D | fs_test.c | 401 __u64 access; in TEST_F_FORK() local 415 for (access = 1; access <= ACCESS_LAST; access <<= 1) { in TEST_F_FORK() 452 __u64 access; member 552 .access = ACCESS_RO, in TEST_F_FORK() 575 .access = ACCESS_RO, in TEST_F_FORK() 627 .access = ACCESS_RO, in TEST_F_FORK() 1040 .access = ACCESS_RO, in TEST_F_FORK() 1076 .access = ACCESS_RO, in TEST_F_FORK() 1133 .access = ACCESS_RO, in TEST_F_FORK() 1138 .access = ACCESS_RO, in TEST_F_FORK() [all …]
|
/linux/drivers/platform/chrome/ |
A D | cros_ec_lpc_mec.c | 79 enum cros_ec_lpc_mec_emi_access_mode access, new_access; in cros_ec_lpc_io_bytes_mec() local 91 access = ACCESS_TYPE_BYTE; in cros_ec_lpc_io_bytes_mec() 93 access = ACCESS_TYPE_LONG_AUTO_INCREMENT; in cros_ec_lpc_io_bytes_mec() 98 cros_ec_lpc_mec_emi_write_address(offset, access); in cros_ec_lpc_io_bytes_mec() 126 if (new_access != access || in cros_ec_lpc_io_bytes_mec() 127 access != ACCESS_TYPE_LONG_AUTO_INCREMENT) { in cros_ec_lpc_io_bytes_mec() 128 access = new_access; in cros_ec_lpc_io_bytes_mec() 129 cros_ec_lpc_mec_emi_write_address(offset, access); in cros_ec_lpc_io_bytes_mec()
|
/linux/drivers/infiniband/sw/rxe/ |
A D | rxe_mr.c | 54 u32 rkey = (access & IB_ACCESS_REMOTE) ? lkey : 0; in rxe_mr_init() 156 rxe_mr_init(access, mr); in rxe_mr_init_dma() 159 mr->access = access; in rxe_mr_init_dma() 165 int access, struct rxe_mr *mr) in rxe_mr_init_user() argument 186 rxe_mr_init(access, mr); in rxe_mr_init_user() 229 mr->access = access; in rxe_mr_init_user() 427 int access, in copy_data() argument 475 mr = lookup_mr(pd, access, sge->lkey, in copy_data() 570 mr_pd(mr) != pd || (access && !(access & mr->access)) || in lookup_mr() 632 u32 access = wqe->wr.wr.reg.access; in rxe_reg_fast_mr() local [all …]
|
A D | rxe_mw.c | 49 mw->access = 0; in rxe_do_dealloc_mw() 84 if (unlikely((mw->access & IB_ZERO_BASED))) { in rxe_check_bind_mw() 122 if (unlikely(mr->access & IB_ZERO_BASED)) { in rxe_check_bind_mw() 128 if (unlikely(!(mr->access & IB_ACCESS_MW_BIND))) { in rxe_check_bind_mw() 135 if (unlikely((mw->access & in rxe_check_bind_mw() 137 !(mr->access & IB_ACCESS_LOCAL_WRITE))) { in rxe_check_bind_mw() 144 if (mw->access & IB_ZERO_BASED) { in rxe_check_bind_mw() 169 mw->access = wqe->wr.wr.mw.access; in rxe_do_bind_mw() 274 mw->access = 0; in rxe_do_invalidate_mw() 313 struct rxe_mw *rxe_lookup_mw(struct rxe_qp *qp, int access, u32 rkey) in rxe_lookup_mw() argument [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/engine/dma/ |
A D | user.c | 87 args->v0.version, args->v0.target, args->v0.access, in nvkm_dmaobj_ctor() 90 dmaobj->access = args->v0.access; in nvkm_dmaobj_ctor() 120 switch (dmaobj->access) { in nvkm_dmaobj_ctor() 122 dmaobj->access = NV_MEM_ACCESS_VM; in nvkm_dmaobj_ctor() 125 dmaobj->access = NV_MEM_ACCESS_RO; in nvkm_dmaobj_ctor() 128 dmaobj->access = NV_MEM_ACCESS_WO; in nvkm_dmaobj_ctor() 131 dmaobj->access = NV_MEM_ACCESS_RW; in nvkm_dmaobj_ctor()
|
/linux/Documentation/core-api/ |
A D | unaligned-memory-access.rst | 14 when it comes to memory access. This document presents some details about 19 The definition of an unaligned access 26 access. 47 of memory access. However, we must consider ALL supported architectures; 52 Why unaligned access is bad 66 unaligned access to be corrected. 76 Code that does not cause unaligned access 139 the memory access in a way that does not cause unaligned access. Of course, 145 Code that causes unaligned access 173 (Hint: it'd be an unaligned access.) [all …]
|
/linux/fs/afs/ |
A D | security.c | 131 h += permits->permits[i].access; in afs_hash_permits() 234 new->permits[j].access = caller_access; in afs_cache_permit() 238 new->permits[j].access = permits->permits[i].access; in afs_cache_permit() 245 new->permits[j].access = caller_access; in afs_cache_permit() 328 *_access = permits->permits[i].access; in afs_check_permit_rcu() 368 *_access = permits->permits[i].access; in afs_check_permit() 403 afs_access_t access; in afs_permission() local 443 if (!(access & AFS_ACE_LOOKUP)) in afs_permission() 452 if (!(access & AFS_ACE_LOOKUP)) in afs_permission() 457 if (!(access & AFS_ACE_READ)) in afs_permission() [all …]
|
/linux/include/linux/ |
A D | device_cgroup.h | 16 short access); 19 short type, access = 0; in devcgroup_inode_permission() local 32 access |= DEVCG_ACC_WRITE; in devcgroup_inode_permission() 34 access |= DEVCG_ACC_READ; in devcgroup_inode_permission() 37 access); in devcgroup_inode_permission() 61 short access) in devcgroup_check_permission() argument
|
/linux/arch/arm/include/asm/ |
A D | arch_timer.h | 28 void arch_timer_reg_write_cp15(int access, enum arch_timer_reg reg, u64 val) in arch_timer_reg_write_cp15() argument 30 if (access == ARCH_TIMER_PHYS_ACCESS) { in arch_timer_reg_write_cp15() 42 } else if (access == ARCH_TIMER_VIRT_ACCESS) { in arch_timer_reg_write_cp15() 60 u32 arch_timer_reg_read_cp15(int access, enum arch_timer_reg reg) in arch_timer_reg_read_cp15() argument 64 if (access == ARCH_TIMER_PHYS_ACCESS) { in arch_timer_reg_read_cp15() 72 } else if (access == ARCH_TIMER_VIRT_ACCESS) { in arch_timer_reg_read_cp15()
|
/linux/drivers/net/dsa/sja1105/ |
A D | sja1105_dynamic_config.c | 886 .access = OP_WRITE, 910 .access = (OP_WRITE | OP_DEL), 919 .access = OP_WRITE, 927 .access = OP_WRITE, 935 .access = OP_WRITE, 943 .access = OP_WRITE, 959 .access = OP_WRITE, 1003 .access = OP_WRITE, 1051 .access = OP_WRITE, 1260 if (!(ops->access & OP_READ)) in sja1105_dynamic_config_read() [all …]
|
/linux/Documentation/arm/ |
A D | mem_alignment.rst | 5 Too many problems popped up because of unnoticed misaligned memory access in 14 unaligned memory access in general. If those access are predictable, you 16 alignment trap can fixup misaligned access for the exception cases, but at 20 trap to SIGBUS any code performing unaligned access (good for debugging bad 21 code), or even fixup the access by software like for kernel code. The later 36 0 A user process performing an unaligned memory access 42 performing the unaligned access. This is of course 47 performing the unaligned access. 59 information on unaligned access occurrences plus the current mode of
|
/linux/tools/perf/Documentation/ |
A D | security.txt | 8 Enabling LSM based mandatory access control (MAC) to perf_event_open() syscall 11 LSM hooks for mandatory access control for perf_event_open() syscall can be 13 Targeted policy with perf_event_open() access control capabilities: 127 Opening access to perf_event_open() syscall on Fedora with SELinux 133 and prevent unauthorized access to perf_event_open() syscall. In such case 139 Enforced MAC policy settings (SELinux) can limit access to performance 141 more perf_event access control information and adjusting the policy. 143 access to performance monitoring and observability operations for users 148 >= 0: Disallow raw and ftrace function tracepoint access 149 >= 1: Disallow CPU event access [all …]
|
/linux/arch/x86/kvm/mmu/ |
A D | paging_tmpl.h | 119 *access &= mask; in FNAME() 219 unsigned access; in FNAME() local 229 access ^= (gpte >> PT64_NX_SHIFT); in FNAME() 232 return access; in FNAME() 342 gpa_t addr, u32 access) in FNAME() 547 access); in FNAME() 553 u32 access) in FNAME() 556 addr, access); in FNAME() 1024 u32 access, in FNAME() 1077 .access = 0x7, in FNAME() [all …]
|
/linux/drivers/hwmon/pmbus/ |
A D | zl6100.c | 25 ktime_t access; /* chip access time */ member 129 s64 delta = ktime_us_delta(ktime_get(), data->access); in zl6100_wait() 179 data->access = ktime_get(); in zl6100_read_word_data() 228 data->access = ktime_get(); in zl6100_read_byte_data() 270 data->access = ktime_get(); in zl6100_write_word_data() 286 data->access = ktime_get(); in zl6100_write_byte() 372 data->access = ktime_get(); in zl6100_probe() 407 data->access = ktime_get(); in zl6100_probe() 421 data->access = ktime_get(); in zl6100_probe() 431 data->access = ktime_get(); in zl6100_probe() [all …]
|
A D | max15301.c | 34 ktime_t access; /* Chip access time */ member 62 s64 delta = ktime_us_delta(ktime_get(), data->access); in max15301_wait() 84 data->access = ktime_get(); in max15301_read_word_data() 100 data->access = ktime_get(); in max15301_read_byte_data() 120 data->access = ktime_get(); in max15301_write_word_data() 136 data->access = ktime_get(); in max15301_write_byte()
|
/linux/tools/testing/selftests/bpf/progs/ |
A D | dev_cgroup.c | 17 short access = ctx->access_type >> 16; in bpf_prog1() local 32 if (access & BPF_DEVCG_ACC_READ) in bpf_prog1() 35 if (access & BPF_DEVCG_ACC_WRITE) in bpf_prog1() 38 if (access & BPF_DEVCG_ACC_MKNOD) in bpf_prog1()
|
/linux/drivers/clocksource/ |
A D | arm_arch_timer.c | 128 if (access == ARCH_TIMER_MEM_PHYS_ACCESS) { in arch_timer_reg_write() 144 } else if (access == ARCH_TIMER_MEM_VIRT_ACCESS) { in arch_timer_reg_write() 158 arch_timer_reg_write_cp15(access, reg, val); in arch_timer_reg_write() 163 u32 arch_timer_reg_read(int access, enum arch_timer_reg reg, in arch_timer_reg_read() argument 168 if (access == ARCH_TIMER_MEM_PHYS_ACCESS) { in arch_timer_reg_read() 177 } else if (access == ARCH_TIMER_MEM_VIRT_ACCESS) { in arch_timer_reg_read() 187 val = arch_timer_reg_read_cp15(access, reg); in arch_timer_reg_read() 412 if (access == ARCH_TIMER_PHYS_ACCESS) { in erratum_set_next_event_generic() 688 static __always_inline int timer_shutdown(const int access, in timer_shutdown() argument 730 if (access == ARCH_TIMER_PHYS_ACCESS) in set_next_event() [all …]
|
/linux/arch/arm64/include/asm/ |
A D | arch_timer.h | 106 void arch_timer_reg_write_cp15(int access, enum arch_timer_reg reg, u64 val) in arch_timer_reg_write_cp15() argument 108 if (access == ARCH_TIMER_PHYS_ACCESS) { in arch_timer_reg_write_cp15() 120 } else if (access == ARCH_TIMER_VIRT_ACCESS) { in arch_timer_reg_write_cp15() 138 u64 arch_timer_reg_read_cp15(int access, enum arch_timer_reg reg) in arch_timer_reg_read_cp15() argument 140 if (access == ARCH_TIMER_PHYS_ACCESS) { in arch_timer_reg_read_cp15() 147 } else if (access == ARCH_TIMER_VIRT_ACCESS) { in arch_timer_reg_read_cp15()
|
/linux/Documentation/userspace-api/ |
A D | landlock.rst | 7 Landlock: unprivileged access control 14 filesystem access) for a set of processes. Because Landlock is a stackable 16 in addition to the existing system-wide access-controls. This kind of sandbox 25 file hierarchy, and the related filesystem actions are defined with `access 95 We now have a ruleset with one rule allowing read access to ``/usr`` while 128 Layers of file path access rights 137 One policy layer grants access to a file path if at least one of its rules 138 encountered on the path grants the access. A sandboxed thread can only access 140 the other system access controls (e.g. filesystem DAC, other LSM policies, 147 access rights can be propagated with bind mounts (cf. [all …]
|
/linux/Documentation/vm/damon/ |
A D | design.rst | 10 DAMON provides data access monitoring functionality while making the accuracy 11 and the overhead controllable. The fundamental access monitorings require 25 Also, if some architectures or devices support special optimized access check 32 The low level primitives for the fundamental access monitoring are defined in 79 Accessed-bit for basic access checks. Only one difference is the way of 105 duration. The resolution of the access frequency is controlled by setting 107 access to each page per ``sampling interval`` and aggregates the results. In 133 assumption (pages in a region have the same access frequencies) is kept, only 150 the data access pattern can be dynamically changed. This will result in low 152 adaptively merges and splits each region based on their access frequency. [all …]
|
/linux/arch/powerpc/mm/book3s64/ |
A D | hash_64k.c | 36 int __hash_page_4K(unsigned long ea, unsigned long access, unsigned long vsid, in __hash_page_4K() argument 59 if (unlikely(!check_pte_access(access, old_pte))) in __hash_page_4K() 67 if (access & _PAGE_WRITE) in __hash_page_4K() 211 hash_failure_debug(ea, access, vsid, trap, ssize, in __hash_page_4K() 223 int __hash_page_64K(unsigned long ea, unsigned long access, in __hash_page_64K() argument 245 if (unlikely(!check_pte_access(access, old_pte))) in __hash_page_64K() 259 if (access & _PAGE_WRITE) in __hash_page_64K() 323 hash_failure_debug(ea, access, vsid, trap, ssize, in __hash_page_64K()
|