/linux/include/linux/sunrpc/ |
A D | gss_api.h | 51 struct gss_ctx **ctx_id, 55 struct gss_ctx *ctx_id, 59 struct gss_ctx *ctx_id, 63 struct gss_ctx *ctx_id, 68 struct gss_ctx *ctx_id, 73 struct gss_ctx **ctx_id); 113 struct gss_ctx *ctx_id, 117 struct gss_ctx *ctx_id, 121 struct gss_ctx *ctx_id, 125 struct gss_ctx *ctx_id, [all …]
|
/linux/net/sunrpc/auth_gss/ |
A D | gss_mech_switch.c | 355 struct gss_ctx **ctx_id, in gss_import_sec_context() argument 359 if (!(*ctx_id = kzalloc(sizeof(**ctx_id), gfp_mask))) in gss_import_sec_context() 361 (*ctx_id)->mech_type = gss_mech_get(mech); in gss_import_sec_context() 364 *ctx_id, endtime, gfp_mask); in gss_import_sec_context() 408 gss_wrap(struct gss_ctx *ctx_id, in gss_wrap() argument 413 return ctx_id->mech_type->gm_ops in gss_wrap() 414 ->gss_wrap(ctx_id, offset, buf, inpages); in gss_wrap() 418 gss_unwrap(struct gss_ctx *ctx_id, in gss_unwrap() argument 423 return ctx_id->mech_type->gm_ops in gss_unwrap() 424 ->gss_unwrap(ctx_id, offset, len, buf); in gss_unwrap()
|
/linux/sound/soc/fsl/ |
A D | fsl_easrc.c | 240 unsigned int ctx_id, int mem_type) in fsl_easrc_coeff_mem_ptr_reset() argument 253 if (ctx_id >= EASRC_CTX_MAX_NUM) { in fsl_easrc_coeff_mem_ptr_reset() 258 reg = REG_EASRC_CCE1(ctx_id); in fsl_easrc_coeff_mem_ptr_reset() 470 unsigned int ctx_id) in fsl_easrc_prefilter_config() argument 487 if (ctx_id >= EASRC_CTX_MAX_NUM) { in fsl_easrc_prefilter_config() 494 ctx = easrc->pair[ctx_id]; in fsl_easrc_prefilter_config() 553 REG_EASRC_CCE1(ctx_id), in fsl_easrc_prefilter_config() 863 struct fsl_asrc_pair *ctx = easrc->pair[ctx_id]; in fsl_easrc_config_slot() 926 struct fsl_asrc_pair *ctx = easrc->pair[ctx_id]; in fsl_easrc_release_slot() 976 if (ctx_id >= EASRC_CTX_MAX_NUM) { in fsl_easrc_config_context() [all …]
|
/linux/drivers/gpu/drm/virtio/ |
A D | virtgpu_trace.h | 21 __field(u32, ctx_id) 30 __entry->ctx_id = le32_to_cpu(hdr->ctx_id); 35 __entry->ctx_id)
|
A D | virtgpu_drv.h | 81 uint32_t ctx_id; member 270 uint32_t ctx_id; member 363 uint32_t ctx_id, 366 uint32_t ctx_id, 370 uint32_t ctx_id, 374 uint32_t ctx_id, 382 uint32_t ctx_id,
|
A D | virtgpu_vq.c | 938 uint32_t ctx_id, in virtio_gpu_cmd_context_attach_resource() argument 950 cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id); in virtio_gpu_cmd_context_attach_resource() 956 uint32_t ctx_id, in virtio_gpu_cmd_context_detach_resource() argument 968 cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id); in virtio_gpu_cmd_context_detach_resource() 1007 uint32_t ctx_id, in virtio_gpu_cmd_transfer_to_host_3d() argument 1032 cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id); in virtio_gpu_cmd_transfer_to_host_3d() 1044 uint32_t ctx_id, in virtio_gpu_cmd_transfer_from_host_3d() argument 1062 cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id); in virtio_gpu_cmd_transfer_from_host_3d() 1075 uint32_t ctx_id, in virtio_gpu_cmd_submit() argument 1090 cmd_p->hdr.ctx_id = cpu_to_le32(ctx_id); in virtio_gpu_cmd_submit() [all …]
|
A D | virtgpu_ioctl.c | 78 virtio_gpu_cmd_context_create(vgdev, vfpriv->ctx_id, in virtio_gpu_create_context_locked() 245 vfpriv->ctx_id, buflist, out_fence); in virtio_gpu_execbuffer_ioctl() 433 (vgdev, vfpriv->ctx_id, offset, args->level, args->stride, in virtio_gpu_transfer_from_host_ioctl() 493 vfpriv ? vfpriv->ctx_id : 0, offset, args->level, in virtio_gpu_transfer_to_host_ioctl() 643 params->ctx_id = vfpriv->ctx_id; in verify_blob() 692 vfpriv->ctx_id, NULL, NULL); in virtio_gpu_resource_create_blob_ioctl()
|
A D | virtgpu_kms.c | 327 vfpriv->ctx_id = handle + 1; in virtio_gpu_driver_open() 341 virtio_gpu_cmd_context_destroy(vgdev, vfpriv->ctx_id); in virtio_gpu_driver_postclose() 345 ida_free(&vgdev->ctx_id_ida, vfpriv->ctx_id - 1); in virtio_gpu_driver_postclose()
|
/linux/drivers/media/platform/sti/hva/ |
A D | hva-hw.c | 120 u8 ctx_id = 0; in hva_hw_its_irq_thread() local 130 ctx_id = (hva->sts_reg & 0xFF00) >> 8; in hva_hw_its_irq_thread() 131 if (ctx_id >= HVA_MAX_INSTANCES) { in hva_hw_its_irq_thread() 133 HVA_PREFIX, __func__, ctx_id); in hva_hw_its_irq_thread() 137 ctx = hva->instances[ctx_id]; in hva_hw_its_irq_thread() 223 u8 ctx_id = 0; in hva_hw_err_irq_thread() local 233 ctx_id = (hva->sts_reg & 0xFF00) >> 8; in hva_hw_err_irq_thread() 234 if (ctx_id >= HVA_MAX_INSTANCES) { in hva_hw_err_irq_thread() 236 ctx_id); in hva_hw_err_irq_thread() 240 ctx = hva->instances[ctx_id]; in hva_hw_err_irq_thread()
|
/linux/arch/x86/mm/ |
A D | tlb.c | 205 this_cpu_write(cpu_tlbstate.ctxs[asid].ctx_id, 0); in clear_asid_other() 228 if (this_cpu_read(cpu_tlbstate.ctxs[asid].ctx_id) != in choose_new_asid() 229 next->context.ctx_id) in choose_new_asid() 554 VM_WARN_ON(this_cpu_read(cpu_tlbstate.ctxs[prev_asid].ctx_id) != in switch_mm_irqs_off() 555 next->context.ctx_id); in switch_mm_irqs_off() 625 this_cpu_write(cpu_tlbstate.ctxs[new_asid].ctx_id, next->context.ctx_id); in switch_mm_irqs_off() 708 this_cpu_write(cpu_tlbstate.ctxs[0].ctx_id, mm->context.ctx_id); in initialize_tlbstate_and_flush() 712 this_cpu_write(cpu_tlbstate.ctxs[i].ctx_id, 0); in initialize_tlbstate_and_flush() 756 VM_WARN_ON(this_cpu_read(cpu_tlbstate.ctxs[loaded_mm_asid].ctx_id) != in flush_tlb_func() 757 loaded_mm->context.ctx_id); in flush_tlb_func()
|
/linux/drivers/gpu/drm/imx/dcss/ |
A D | dcss-ctxld.c | 340 void dcss_ctxld_write_irqsafe(struct dcss_ctxld *ctxld, u32 ctx_id, u32 val, in dcss_ctxld_write_irqsafe() argument 349 int item_idx = ctxld->ctx_size[curr_ctx][ctx_id]; in dcss_ctxld_write_irqsafe() 351 if (item_idx + 1 > dcss_ctxld_ctx_size[ctx_id]) { in dcss_ctxld_write_irqsafe() 356 ctx[ctx_id][item_idx].val = val; in dcss_ctxld_write_irqsafe() 357 ctx[ctx_id][item_idx].ofs = reg_ofs; in dcss_ctxld_write_irqsafe() 358 ctxld->ctx_size[curr_ctx][ctx_id] += 1; in dcss_ctxld_write_irqsafe() 361 void dcss_ctxld_write(struct dcss_ctxld *ctxld, u32 ctx_id, in dcss_ctxld_write() argument 365 dcss_ctxld_write_irqsafe(ctxld, ctx_id, val, reg_ofs); in dcss_ctxld_write()
|
A D | dcss-ss.c | 68 u32 ctx_id; member 78 dcss_ctxld_write(ss->ctxld, ss->ctx_id, val, in dcss_ss_write() 102 ss->ctx_id = CTX_SB_HP; in dcss_ss_init()
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
A D | amdgpu_sched.c | 63 unsigned ctx_id, in amdgpu_sched_context_priority_override() argument 80 ctx = amdgpu_ctx_get(fpriv, ctx_id); in amdgpu_sched_context_priority_override() 126 args->in.ctx_id, in amdgpu_sched_ioctl()
|
/linux/drivers/net/ethernet/netronome/nfp/flower/ |
A D | offload.c | 1570 u32 ctx_id; in __nfp_flower_update_merge_stats() local 1573 pkts = priv->stats[ctx_id].pkts; in __nfp_flower_update_merge_stats() 1577 bytes = priv->stats[ctx_id].bytes; in __nfp_flower_update_merge_stats() 1578 used = priv->stats[ctx_id].used; in __nfp_flower_update_merge_stats() 1581 priv->stats[ctx_id].pkts = 0; in __nfp_flower_update_merge_stats() 1582 priv->stats[ctx_id].bytes = 0; in __nfp_flower_update_merge_stats() 1594 priv->stats[ctx_id].used); in __nfp_flower_update_merge_stats() 1628 u32 ctx_id; in nfp_flower_get_stats() local 1651 priv->stats[ctx_id].pkts, 0, priv->stats[ctx_id].used, in nfp_flower_get_stats() 1654 priv->stats[ctx_id].pkts = 0; in nfp_flower_get_stats() [all …]
|
A D | metadata.c | 119 u32 ctx_id; in nfp_flower_rx_flow_stats() local 127 ctx_id = be32_to_cpu(stats->stats_con_id); in nfp_flower_rx_flow_stats() 128 priv->stats[ctx_id].pkts += be32_to_cpu(stats->pkt_count); in nfp_flower_rx_flow_stats() 129 priv->stats[ctx_id].bytes += be64_to_cpu(stats->byte_count); in nfp_flower_rx_flow_stats() 130 priv->stats[ctx_id].used = jiffies; in nfp_flower_rx_flow_stats() 443 nfp_flower_get_fl_payload_from_ctx(struct nfp_app *app, u32 ctx_id) in nfp_flower_get_fl_payload_from_ctx() argument 448 ctx_entry = rhashtable_lookup_fast(&priv->stats_ctx_table, &ctx_id, in nfp_flower_get_fl_payload_from_ctx()
|
A D | conntrack.c | 1535 u32 ctx_id; in nfp_fl_ct_sub_stats() local 1555 priv->stats[ctx_id].bytes, in nfp_fl_ct_sub_stats() 1556 priv->stats[ctx_id].pkts, in nfp_fl_ct_sub_stats() 1557 0, priv->stats[ctx_id].used, in nfp_fl_ct_sub_stats() 1562 priv->stats[ctx_id].bytes, in nfp_fl_ct_sub_stats() 1563 priv->stats[ctx_id].pkts, in nfp_fl_ct_sub_stats() 1564 0, priv->stats[ctx_id].used, in nfp_fl_ct_sub_stats() 1568 priv->stats[ctx_id].bytes, in nfp_fl_ct_sub_stats() 1569 priv->stats[ctx_id].pkts, in nfp_fl_ct_sub_stats() 1574 priv->stats[ctx_id].pkts = 0; in nfp_fl_ct_sub_stats() [all …]
|
/linux/drivers/infiniband/hw/efa/ |
A D | efa_com.c | 246 u16 ctx_id; in efa_com_alloc_ctx_id() local 253 return ctx_id; in efa_com_alloc_ctx_id() 257 u16 ctx_id) in efa_com_dealloc_ctx_id() argument 270 u16 ctx_id = cmd_id & (aq->depth - 1); in efa_com_put_comp_ctx() local 274 efa_com_dealloc_ctx_id(aq, ctx_id); in efa_com_put_comp_ctx() 291 aq->comp_ctx[ctx_id].occupied = 1; in efa_com_get_comp_ctx() 296 return &aq->comp_ctx[ctx_id]; in efa_com_get_comp_ctx() 309 u16 ctx_id; in __efa_com_submit_admin_cmd() local 315 ctx_id = efa_com_alloc_ctx_id(aq); in __efa_com_submit_admin_cmd() 318 cmd_id = ctx_id & queue_size_mask; in __efa_com_submit_admin_cmd() [all …]
|
/linux/fs/ |
A D | aio.c | 1065 if (ctx && ctx->user_id == ctx_id) { in lookup_ioctx() 2037 ctx = lookup_ioctx(ctx_id); in SYSCALL_DEFINE3() 2079 ctx = lookup_ioctx(ctx_id); in COMPAT_SYSCALL_DEFINE3() 2134 ctx = lookup_ioctx(ctx_id); in SYSCALL_DEFINE3() 2163 static long do_io_getevents(aio_context_t ctx_id, in do_io_getevents() argument 2170 struct kioctx *ioctx = lookup_ioctx(ctx_id); in do_io_getevents() 2222 aio_context_t, ctx_id, in SYSCALL_DEFINE6() argument 2257 aio_context_t, ctx_id, in SYSCALL_DEFINE6() argument 2294 SYSCALL_DEFINE5(io_getevents_time32, __u32, ctx_id, in SYSCALL_DEFINE5() argument 2324 compat_aio_context_t, ctx_id, in COMPAT_SYSCALL_DEFINE6() argument [all …]
|
/linux/include/uapi/misc/ |
A D | habanalabs.h | 591 __u32 ctx_id; member 622 __u32 ctx_id; member 833 __u32 ctx_id; member 905 __u32 ctx_id; member 1060 __u32 ctx_id; member 1214 __u32 ctx_id; member
|
/linux/arch/x86/include/asm/ |
A D | mmu.h | 24 u64 ctx_id; member 62 .ctx_id = 1, \
|
/linux/include/uapi/drm/ |
A D | amdgpu_drm.h | 246 __u32 ctx_id; member 253 __u32 ctx_id; member 302 __u32 ctx_id; member 434 __u32 ctx_id; member 448 __u32 ctx_id; member 565 __u32 ctx_id; member 631 __u32 ctx_id; member
|
/linux/drivers/misc/habanalabs/common/ |
A D | command_buffer.c | 182 int ctx_id, bool internal_cb) in hl_cb_alloc() argument 196 if (ctx_id == HL_KERNEL_ASID_ID && !hdev->disabled) in hl_cb_alloc() 215 } else if (ctx_id == HL_KERNEL_ASID_ID) { in hl_cb_alloc() 247 int rc, ctx_id = ctx->asid; in hl_cb_create() local 254 (ctx_id != HL_KERNEL_ASID_ID))) { in hl_cb_create() 273 if (ctx_id == HL_KERNEL_ASID_ID && in hl_cb_create() 291 cb = hl_cb_alloc(hdev, cb_size, ctx_id, internal_cb); in hl_cb_create() 303 if (ctx_id == HL_KERNEL_ASID_ID) { in hl_cb_create()
|
/linux/drivers/net/ethernet/qlogic/qlcnic/ |
A D | qlcnic_83xx_hw.h | 172 u16 ctx_id; member 178 u16 ctx_id; 194 u16 ctx_id; member 196 u16 ctx_id; 238 u16 ctx_id; member 244 u16 ctx_id; member
|
/linux/drivers/gpu/drm/vmwgfx/ |
A D | vmwgfx_cmdbuf.c | 1001 int ctx_id, in vmw_cmdbuf_reserve_cur() argument 1013 ctx_id != cur->cb_header->dxContext))) in vmw_cmdbuf_reserve_cur() 1028 if (ctx_id != SVGA3D_INVALID_ID) { in vmw_cmdbuf_reserve_cur() 1030 cur->cb_header->dxContext = ctx_id; in vmw_cmdbuf_reserve_cur() 1075 int ctx_id, bool interruptible, in vmw_cmdbuf_reserve() argument 1079 return vmw_cmdbuf_reserve_cur(man, size, ctx_id, interruptible); in vmw_cmdbuf_reserve() 1084 if (ctx_id != SVGA3D_INVALID_ID) { in vmw_cmdbuf_reserve() 1086 header->cb_header->dxContext = ctx_id; in vmw_cmdbuf_reserve()
|
/linux/drivers/scsi/cxlflash/ |
A D | sislite.h | 27 u16 ctx_id; /* ctx_hndl_t */ member 304 #define SISL_RHT_CNT_ID(cnt, ctx_id) (((cnt) << 48) | ((ctx_id) << 32)) argument
|