/linux/drivers/misc/vmw_vmci/ |
A D | vmci_context.c | 107 context = kzalloc(sizeof(*context), GFP_KERNEL); in vmci_ctx_create() 108 if (!context) { in vmci_ctx_create() 173 return context; in vmci_ctx_create() 180 kfree(context); in vmci_ctx_create() 311 if (!context) { in vmci_ctx_enqueue_datagram() 409 context = c; in vmci_ctx_get() 416 return context; in vmci_ctx_get() 435 ctx_fire_notification(context->cid, context->priv_flags); in ctx_free_ctx() 606 if (!context) in vmci_ctx_add_notification() 673 if (!context) in vmci_ctx_remove_notification() [all …]
|
A D | vmci_route.c | 42 if (VMCI_INVALID_ID == dst->context) in vmci_route() 75 if (VMCI_INVALID_ID == src->context && in vmci_route() 77 src->context = vmci_get_context_id(); in vmci_route() 113 if (VMCI_INVALID_ID == src->context) in vmci_route() 130 if (VMCI_INVALID_ID == src->context) { in vmci_route() 139 src->context = VMCI_HOST_CONTEXT_ID; in vmci_route() 153 if (vmci_ctx_exists(dst->context)) { in vmci_route() 165 src->context = VMCI_HOST_CONTEXT_ID; in vmci_route() 167 src->context != dst->context) { in vmci_route() 209 if (VMCI_INVALID_ID == src->context) in vmci_route() [all …]
|
/linux/security/selinux/ss/ |
A D | mls.h | 32 void mls_sid_to_context(struct policydb *p, struct context *context, 41 struct context *context, 48 int mls_range_set(struct context *context, struct mls_range *range); 69 struct context *context, 72 struct context *context, 75 struct context *context, 78 struct context *context, 82 struct context *context, in mls_export_netlbl_lvl() argument 88 struct context *context, in mls_import_netlbl_lvl() argument 94 struct context *context, in mls_export_netlbl_cat() argument [all …]
|
A D | mls.c | 36 int mls_compute_context_len(struct policydb *p, struct context *context) in mls_compute_context_len() argument 90 struct context *context, in mls_sid_to_context() argument 237 struct context *context, in mls_context_to_sid() argument 345 context->range.level[1].sens = context->range.level[0].sens; in mls_context_to_sid() 385 int mls_range_set(struct context *context, in mls_range_set() argument 566 struct context *context, in mls_export_netlbl_lvl() argument 588 struct context *context, in mls_import_netlbl_lvl() argument 595 context->range.level[1].sens = context->range.level[0].sens; in mls_import_netlbl_lvl() 610 struct context *context, in mls_export_netlbl_cat() argument 640 struct context *context, in mls_import_netlbl_cat() argument [all …]
|
/linux/drivers/gpu/drm/etnaviv/ |
A D | etnaviv_mmu.c | 29 unmapped_page = context->global->ops->unmap(context, iova, in etnaviv_context_unmap() 55 ret = context->global->ops->map(context, iova, paddr, pgsize, in etnaviv_context_map() 79 if (!context || !sgt) in etnaviv_iommu_map() 203 m->context = NULL; in etnaviv_iommu_find_iova() 275 context->flush_seq++; in etnaviv_iommu_map_gem() 294 context->flush_seq++; in etnaviv_iommu_unmap_gem() 303 etnaviv_cmdbuf_suballoc_unmap(context, &context->cmdbuf_mapping); in etnaviv_iommu_context_free() 305 context->global->ops->free(context); in etnaviv_iommu_context_free() 351 context->global->ops->restore(gpu, context); in etnaviv_iommu_restore() 425 return context->global->ops->dump_size(context); in etnaviv_iommu_dump_size() [all …]
|
A D | etnaviv_iommu.c | 37 drm_mm_takedown(&context->mm); in etnaviv_iommuv1_free() 140 context = global->v1.shared_context; in etnaviv_iommuv1_context_alloc() 141 etnaviv_iommu_context_get(context); in etnaviv_iommuv1_context_alloc() 143 return context; in etnaviv_iommuv1_context_alloc() 160 context = &v1_context->base; in etnaviv_iommuv1_context_alloc() 161 context->global = global; in etnaviv_iommuv1_context_alloc() 162 kref_init(&context->refcount); in etnaviv_iommuv1_context_alloc() 163 mutex_init(&context->lock); in etnaviv_iommuv1_context_alloc() 164 INIT_LIST_HEAD(&context->mappings); in etnaviv_iommuv1_context_alloc() 166 context->global->v1.shared_context = context; in etnaviv_iommuv1_context_alloc() [all …]
|
A D | etnaviv_iommu_v2.c | 52 drm_mm_takedown(&context->mm); in etnaviv_iommuv2_free() 166 struct etnaviv_iommu_context *context) in etnaviv_iommuv2_restore_nonsec() argument 181 (u32)context->global->bad_page_dma); in etnaviv_iommuv2_restore_nonsec() 190 struct etnaviv_iommu_context *context) in etnaviv_iommuv2_restore_sec() argument 273 struct etnaviv_iommu_context *context; in etnaviv_iommuv2_context_alloc() local 300 context = &v2_context->base; in etnaviv_iommuv2_context_alloc() 301 context->global = global; in etnaviv_iommuv2_context_alloc() 302 kref_init(&context->refcount); in etnaviv_iommuv2_context_alloc() 303 mutex_init(&context->lock); in etnaviv_iommuv2_context_alloc() 304 INIT_LIST_HEAD(&context->mappings); in etnaviv_iommuv2_context_alloc() [all …]
|
/linux/kernel/ |
A D | auditsc.c | 1021 context->context = AUDIT_CTX_UNUSED; in audit_alloc_context() 1648 switch (context->context) { in audit_log_exit() 1654 context->arch, context->major); in audit_log_exit() 1720 context->fds[0], context->fds[1]); in audit_log_exit() 1770 if (context->context == AUDIT_CTX_SYSCALL) in audit_log_exit() 1976 if (context->context != AUDIT_CTX_UNUSED || context->name_count) { in __audit_syscall_entry() 1998 context->context = AUDIT_CTX_SYSCALL; in __audit_syscall_entry() 2018 if (!context || context->dummy || in __audit_syscall_exit() 2188 if (context->context == AUDIT_CTX_UNUSED) in __audit_getname() 2260 if (context->context == AUDIT_CTX_UNUSED) in __audit_inode() [all …]
|
/linux/fs/xfs/ |
A D | xfs_attr_list.c | 88 context->put_listent(context, in xfs_attr_shortform_list() 97 if (context->seen_enough) in xfs_attr_shortform_list() 106 if (context->bufsize == 0) in xfs_attr_shortform_list() 182 context->put_listent(context, in xfs_attr_shortform_list() 410 if (context->resynch) { in xfs_attr3_leaf_list_int() 415 context->dupcnt = 0; in xfs_attr3_leaf_list_int() 418 context->dupcnt++; in xfs_attr3_leaf_list_int() 421 context->dupcnt = 0; in xfs_attr3_leaf_list_int() 433 context->resynch = 0; in xfs_attr3_leaf_list_int() 470 context->put_listent(context, entry->flags, in xfs_attr3_leaf_list_int() [all …]
|
A D | xfs_xattr.c | 108 if (context->count < 0 || context->seen_enough) in __xfs_xattr_put_listent() 111 if (!context->buffer) in __xfs_xattr_put_listent() 117 context->seen_enough = 1; in __xfs_xattr_put_listent() 120 offset = context->buffer + context->count; in __xfs_xattr_put_listent() 201 memset(&context, 0, sizeof(context)); in xfs_vn_listxattr() 202 context.dp = XFS_I(inode); in xfs_vn_listxattr() 203 context.resynch = 1; in xfs_vn_listxattr() 205 context.bufsize = size; in xfs_vn_listxattr() 206 context.firstu = context.bufsize; in xfs_vn_listxattr() 212 if (context.count < 0) in xfs_vn_listxattr() [all …]
|
/linux/drivers/platform/x86/intel/int1092/ |
A D | intel_sar.c | 41 &context->config_data[context->reg_value]; in update_sar_data() 136 out = acpi_evaluate_dsm(context->handle, &context->guid, rev, in sar_get_device_mode() 144 update_sar_data(context); in sar_get_device_mode() 190 context->reg_value = value; in intc_reg_store() 191 update_sar_data(context); in intc_reg_store() 224 out = acpi_evaluate_dsm(context->handle, &context->guid, rev, in sar_get_data() 252 context = kzalloc(sizeof(*context), GFP_KERNEL); in sar_probe() 253 if (!context) in sar_probe() 267 sar_get_data(reg, context); in sar_probe() 292 kfree(context); in sar_probe() [all …]
|
/linux/drivers/net/ethernet/mellanox/mlx4/ |
A D | en_resources.c | 47 memset(context, 0, sizeof(*context)); in mlx4_en_fill_qp_context() 49 context->pd = cpu_to_be32(mdev->priv_pdn); in mlx4_en_fill_qp_context() 50 context->mtu_msgmax = 0xff; in mlx4_en_fill_qp_context() 59 context->sq_size_stride = ilog2(TXBB_SIZE) - 4; in mlx4_en_fill_qp_context() 63 context->local_qpn = cpu_to_be32(qpn); in mlx4_en_fill_qp_context() 64 context->pri_path.ackto = 1 & 0x07; in mlx4_en_fill_qp_context() 68 context->pri_path.sched_queue |= user_prio << 3; in mlx4_en_fill_qp_context() 72 context->cqn_send = cpu_to_be32(cqn); in mlx4_en_fill_qp_context() 73 context->cqn_recv = cpu_to_be32(cqn); in mlx4_en_fill_qp_context() 76 context->pri_path.counter_index != in mlx4_en_fill_qp_context() [all …]
|
/linux/drivers/gpu/drm/tegra/ |
A D | submit.c | 26 #define SUBMIT_ERR(context, fmt, ...) \ argument 121 xa_lock(&context->mappings); in tegra_drm_mapping_get() 127 xa_unlock(&context->mappings); in tegra_drm_mapping_get() 156 struct tegra_drm_context *context, in submit_copy_gather_data() argument 219 SUBMIT_ERR(context, in submit_write_reloc() 379 class = context->client->base.class; in submit_create_job() 399 job->client = &context->client->base; in submit_create_job() 489 struct tegra_drm_context *context; in tegra_drm_ioctl_channel_submit() local 497 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_submit() 498 if (!context) { in tegra_drm_ioctl_channel_submit() [all …]
|
A D | uapi.c | 47 kfree(context); in tegra_drm_channel_context_close() 89 context = kzalloc(sizeof(*context), GFP_KERNEL); in tegra_drm_ioctl_channel_open() 90 if (!context) in tegra_drm_ioctl_channel_open() 109 err = xa_alloc(&fpriv->contexts, &args->context, context, XA_LIMIT(1, U32_MAX), in tegra_drm_ioctl_channel_open() 128 kfree(context); in tegra_drm_ioctl_channel_open() 141 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_close() 142 if (!context) { in tegra_drm_ioctl_channel_close() 169 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_map() 170 if (!context) { in tegra_drm_ioctl_channel_map() 260 context = xa_load(&fpriv->contexts, args->context); in tegra_drm_ioctl_channel_unmap() [all …]
|
/linux/drivers/net/ethernet/qlogic/qed/ |
A D | qed_nvmetcp_fw_funcs.c | 145 struct e5_nvmetcp_task_context *context = task_params->context; in init_default_nvmetcp_task() local 149 memset(context, 0, sizeof(*context)); in init_default_nvmetcp_task() 150 init_nvmetcp_task_params(context, task_params, in init_default_nvmetcp_task() 210 SET_FIELD(context->ustorm_st_context.flags, in set_local_completion_context() 221 struct e5_nvmetcp_task_context *context = task_params->context; in init_rw_nvmetcp_task() local 247 &context->mstorm_st_context.data_desc, in init_rw_nvmetcp_task() 260 &context->ustorm_ag_context, in init_rw_nvmetcp_task() 279 context->ustorm_ag_context.exp_cont_len = 0; in init_rw_nvmetcp_task() 327 struct e5_nvmetcp_task_context *context = task_params->context; in init_common_login_request_task() local 334 &context->ustorm_ag_context, in init_common_login_request_task() [all …]
|
/linux/arch/sparc/mm/ |
A D | tsb.c | 545 spin_lock_init(&mm->context.lock); in init_new_context() 547 mm->context.sparc64_ctx_val = 0UL; in init_new_context() 549 mm->context.tag_store = NULL; in init_new_context() 559 mm->context.hugetlb_pte_count = 0; in init_new_context() 560 mm->context.thp_pte_count = 0; in init_new_context() 570 mm->context.tsb_block[i].tsb = NULL; in init_new_context() 611 if (CTX_VALID(mm->context)) { in destroy_context() 619 if (mm->context.tag_store) { in destroy_context() 624 tag_desc = mm->context.tag_store; in destroy_context() 632 kfree(mm->context.tag_store); in destroy_context() [all …]
|
/linux/arch/s390/include/asm/ |
A D | mmu_context.h | 24 spin_lock_init(&mm->context.lock); in init_new_context() 26 INIT_LIST_HEAD(&mm->context.gmap_list); in init_new_context() 28 atomic_set(&mm->context.flush_count, 0); in init_new_context() 30 mm->context.gmap_asce = 0; in init_new_context() 31 mm->context.flush_mm = 0; in init_new_context() 36 mm->context.has_pgste = 0; in init_new_context() 37 mm->context.uses_skeys = 0; in init_new_context() 38 mm->context.uses_cmm = 0; in init_new_context() 39 mm->context.allow_gmap_hpage_1m = 0; in init_new_context() 41 switch (mm->context.asce_limit) { in init_new_context() [all …]
|
/linux/arch/powerpc/mm/book3s64/ |
A D | mmu_context.c | 98 if (!mm->context.hash_context) in hash__init_new_context() 115 if (mm->context.id == 0) { in hash__init_new_context() 139 kfree(mm->context.hash_context); in hash__init_new_context() 178 mm->context.hash_context = NULL; in radix__init_new_context() 195 mm->context.id = index; in init_new_context() 197 mm->context.pte_frag = NULL; in init_new_context() 198 mm->context.pmd_frag = NULL; in init_new_context() 245 frag = mm->context.pte_frag; in destroy_pagetable_cache() 249 frag = mm->context.pmd_frag; in destroy_pagetable_cache() 277 destroy_contexts(&mm->context); in destroy_context() [all …]
|
/linux/drivers/usb/image/ |
A D | microtek.c | 386 context in mts_int_submit_urb() 404 context->final_callback(context->srb); in mts_transfer_cleanup() 412 context->srb->result |= (unsigned)(*context->scsi_status)<<1; in mts_transfer_done() 438 scsi_set_resid(context->srb, context->data_length - in mts_data_done() 479 context->data, in mts_command_done() 502 context->curr_sg = sg_next(context->curr_sg); in mts_do_sg() 504 context->data_pipe, in mts_do_sg() 531 desc->context.srb = srb; in mts_build_transfer_context() 539 desc->context.data = sg_virt(desc->context.curr_sg); in mts_build_transfer_context() 540 desc->context.data_length = desc->context.curr_sg->length; in mts_build_transfer_context() [all …]
|
/linux/fs/ocfs2/ |
A D | move_extents.c | 118 &context->dealloc); in __ocfs2_move_extent() 136 len, context->meta_ac, in __ocfs2_move_extent() 137 &context->dealloc, 1); in __ocfs2_move_extent() 243 &context->meta_ac, in ocfs2_defrag_extent() 351 if (context->data_ac) { in ocfs2_defrag_extent() 356 if (context->meta_ac) { in ocfs2_defrag_extent() 616 &context->meta_ac, in ocfs2_move_extent() 726 if (context->meta_ac) { in ocfs2_move_extent() 1009 if (!context) { in ocfs2_ioctl_move_extents() 1016 context->file = filp; in ocfs2_ioctl_move_extents() [all …]
|
/linux/drivers/infiniband/hw/hns/ |
A D | hns_roce_cmd.c | 101 complete(&context->done); in hns_roce_cmd_event() 117 context = &cmd->context[cmd->free_head]; in __hns_roce_cmd_mbox_wait() 119 } while (context->busy); in __hns_roce_cmd_mbox_wait() 121 context->busy = 1; in __hns_roce_cmd_mbox_wait() 130 context->token, 1); in __hns_roce_cmd_mbox_wait() 141 context->token, op); in __hns_roce_cmd_mbox_wait() 146 ret = context->result; in __hns_roce_cmd_mbox_wait() 152 context->busy = 0; in __hns_roce_cmd_mbox_wait() 214 hr_cmd->context = in hns_roce_cmd_use_events() 216 if (!hr_cmd->context) { in hns_roce_cmd_use_events() [all …]
|
A D | hns_roce_restrack.c | 13 struct hns_roce_v2_cq_context *context) in hns_roce_fill_cq() argument 16 roce_get_field(context->byte_4_pg_ceqn, in hns_roce_fill_cq() 22 roce_get_field(context->byte_4_pg_ceqn, in hns_roce_fill_cq() 28 roce_get_field(context->byte_8_cqn, in hns_roce_fill_cq() 41 roce_get_field(context->byte_28_cq_pi, in hns_roce_fill_cq() 48 roce_get_field(context->byte_32_cq_ci, in hns_roce_fill_cq() 84 struct hns_roce_v2_cq_context *context; in hns_roce_fill_res_cq_entry() local 92 if (!context) in hns_roce_fill_res_cq_entry() 105 if (hns_roce_fill_cq(msg, context)) { in hns_roce_fill_res_cq_entry() 111 kfree(context); in hns_roce_fill_res_cq_entry() [all …]
|
/linux/drivers/pci/hotplug/ |
A D | acpiphp_glue.c | 63 context = kzalloc(sizeof(*context), GFP_KERNEL); in acpiphp_init_context() 64 if (!context) in acpiphp_init_context() 71 return context; in acpiphp_init_context() 89 return context; in acpiphp_get_context() 107 kfree(context); in acpiphp_put_context() 127 if (!context) in acpiphp_grab_context() 141 return context; in acpiphp_grab_context() 167 context = bridge->context; in free_bridge() 169 if (context) { in free_bridge() 195 if (!context) in acpiphp_post_dock_fixup() [all …]
|
/linux/drivers/gpu/drm/i915/selftests/ |
A D | i915_syncmap.c | 179 context); in check_one() 195 context, seqno); in check_one() 225 err = check_one(&sync, context, in igt_syncmap_one() 264 context, seqno); in check_leaf() 297 u64 context = BIT_ULL(order); in igt_syncmap_join_above() local 300 err = check_leaf(&sync, context, 0); in igt_syncmap_join_above() 346 u64 context = step * BIT_ULL(order); in igt_syncmap_join_below() local 367 context, order, step); in igt_syncmap_join_below() 389 context, order, step); in igt_syncmap_join_below() 434 context, idx, in igt_syncmap_neighbours() [all …]
|
/linux/arch/ia64/include/asm/ |
A D | mmu_context.h | 78 nv_mm_context_t context = mm->context; in get_mmu_context() local 80 if (likely(context)) in get_mmu_context() 85 context = mm->context; in get_mmu_context() 86 if (context == 0) { in get_mmu_context() 96 mm->context = context = ia64_ctx.next++; in get_mmu_context() 107 return context; in get_mmu_context() 118 mm->context = 0; in init_new_context() 157 nv_mm_context_t context; in activate_context() local 160 context = get_mmu_context(mm); in activate_context() 163 reload_context(context); in activate_context() [all …]
|