Home
last modified time | relevance | path

Searched refs:mmap_lock (Results 1 – 25 of 1925) sorted by relevance

12345678910>>...77

/linux/include/linux/
A Dmmap_lock.h12 .mmap_lock = __RWSEM_INITIALIZER((name).mmap_lock),
65 init_rwsem(&mm->mmap_lock); in mmap_init_lock()
71 down_write(&mm->mmap_lock); in mmap_write_lock()
87 ret = down_write_killable(&mm->mmap_lock); in mmap_write_lock_killable()
105 up_write(&mm->mmap_lock); in mmap_write_unlock()
111 downgrade_write(&mm->mmap_lock); in mmap_write_downgrade()
117 down_read(&mm->mmap_lock); in mmap_read_lock()
126 ret = down_read_killable(&mm->mmap_lock); in mmap_read_lock_killable()
144 up_read(&mm->mmap_lock); in mmap_read_unlock()
150 up_read_non_owner(&mm->mmap_lock); in mmap_read_unlock_non_owner()
[all …]
A Dmm_types.h538 struct rw_semaphore mmap_lock; member
/linux/include/trace/events/
A Dmmap_lock.h3 #define TRACE_SYSTEM mmap_lock
16 DECLARE_EVENT_CLASS(mmap_lock,
43 DEFINE_EVENT_FN(mmap_lock, name, \
/linux/drivers/infiniband/hw/cxgb4/
A Diw_cxgb4.h526 spinlock_t mmap_lock; member
549 spin_lock(&ucontext->mmap_lock); in remove_mmap()
555 spin_unlock(&ucontext->mmap_lock); in remove_mmap()
561 spin_unlock(&ucontext->mmap_lock); in remove_mmap()
568 spin_lock(&ucontext->mmap_lock); in insert_mmap()
572 spin_unlock(&ucontext->mmap_lock); in insert_mmap()
A Dprovider.c88 spin_lock_init(&context->mmap_lock); in c4iw_alloc_ucontext()
102 spin_lock(&context->mmap_lock); in c4iw_alloc_ucontext()
105 spin_unlock(&context->mmap_lock); in c4iw_alloc_ucontext()
A Dcq.c1109 spin_lock(&ucontext->mmap_lock); in c4iw_create_cq()
1119 spin_unlock(&ucontext->mmap_lock); in c4iw_create_cq()
A Dqp.c2262 spin_lock(&ucontext->mmap_lock); in c4iw_create_qp()
2279 spin_unlock(&ucontext->mmap_lock); in c4iw_create_qp()
2753 spin_lock(&ucontext->mmap_lock); in c4iw_create_srq()
2758 spin_unlock(&ucontext->mmap_lock); in c4iw_create_srq()
/linux/drivers/media/common/videobuf2/
A Dvideobuf2-core.c785 mutex_lock(&q->mmap_lock); in vb2_core_reqbufs()
797 mutex_unlock(&q->mmap_lock); in vb2_core_reqbufs()
876 mutex_lock(&q->mmap_lock); in vb2_core_reqbufs()
885 mutex_unlock(&q->mmap_lock); in vb2_core_reqbufs()
888 mutex_unlock(&q->mmap_lock); in vb2_core_reqbufs()
980 mutex_lock(&q->mmap_lock); in vb2_core_create_bufs()
989 mutex_unlock(&q->mmap_lock); in vb2_core_create_bufs()
992 mutex_unlock(&q->mmap_lock); in vb2_core_create_bufs()
2292 mutex_lock(&q->mmap_lock); in vb2_mmap()
2405 mutex_init(&q->mmap_lock); in vb2_core_queue_init()
[all …]
/linux/mm/
A D.mmap_lock.o.cmd1mmap_lock.o := /usr/bin/ccache /home/test/workspace/code/optee_3.16/build/../toolchains/aarch64/bi…
3 source_mm/mmap_lock.o := mm/mmap_lock.c
5 deps_mm/mmap_lock.o := \
24 include/trace/events/mmap_lock.h \
954 include/linux/mmap_lock.h \
1239 mm/mmap_lock.o: $(deps_mm/mmap_lock.o)
1241 $(deps_mm/mmap_lock.o):
A Dbuilt-in.a29 mmap_lock.o/
A D.built-in.a.cmd1 …mm/interval_tree.o mm/list_lru.o mm/workingset.o mm/debug.o mm/gup.o mm/mmap_lock.o mm/highmem.o m…
A DMakefile55 debug.o gup.o mmap_lock.o $(mmu-y)
A Dmmu_notifier.c1003 might_lock(&mm->mmap_lock); in mmu_interval_notifier_insert()
A Dmmap.c3510 down_write_nest_lock(&anon_vma->root->rwsem, &mm->mmap_lock); in vm_lock_anon_vma()
3540 down_write_nest_lock(&mapping->i_mmap_rwsem, &mm->mmap_lock); in vm_lock_mapping()
A Dgup.c2852 might_lock_read(&current->mm->mmap_lock); in internal_get_user_pages_fast()
/linux/arch/ia64/mm/
A Dfault.c79 prefetchw(&mm->mmap_lock); in ia64_do_page_fault()
/linux/Documentation/admin-guide/mm/
A Dnuma_memory_policy.rst374 task's mm's mmap_lock for read during the query. The set_mempolicy() and
375 mbind() APIs [see below] always acquire the mmap_lock for write when
381 we hold them mmap_lock for read. Again, because replacing the task or vma
382 policy requires that the mmap_lock be held for write, the policy can't be
386 shared memory policy while another task, with a distinct mmap_lock, is
A Duserfaultfd.rst36 ``userfaultfd`` runtime load never takes the mmap_lock for writing).
/linux/Documentation/vm/
A Dtranshuge.rst96 mmap_lock in read (or write) mode to be sure a huge pmd cannot be
98 takes the mmap_lock in write mode in addition to the anon_vma lock). If
/linux/include/media/
A Dvideobuf2-core.h607 struct mutex mmap_lock; member
/linux/drivers/gpu/drm/etnaviv/
A Detnaviv_gem.c666 might_lock_read(&current->mm->mmap_lock); in etnaviv_gem_userptr_get_pages()
/linux/kernel/bpf/
A Dstackmap.c216 rwsem_release(&current->mm->mmap_lock.dep_map, _RET_IP_); in stack_map_get_build_id_offset()
/linux/arch/x86/mm/
A Dfault.c1502 prefetchw(&current->mm->mmap_lock); in DEFINE_IDTENTRY_RAW_ERRORCODE()
/linux/Documentation/filesystems/
A Dlocking.rst644 ops mmap_lock PageLocked(page)
/linux/fs/
A D.seq_file.o.cmd884 include/linux/mmap_lock.h \

Completed in 69 milliseconds

12345678910>>...77