/linux/drivers/staging/media/atomisp/pci/base/circbuf/src/ |
A D | circbuf.c | 80 cb->desc = desc; in ia_css_circbuf_create() 83 cb->desc->end = 0; in ia_css_circbuf_create() 89 cb->elems = elems; in ia_css_circbuf_create() 98 cb->desc = NULL; in ia_css_circbuf_destroy() 100 cb->elems = NULL; in ia_css_circbuf_destroy() 133 max_offset = ia_css_circbuf_get_offset(cb, cb->desc->start, cb->desc->end); in ia_css_circbuf_extract() 178 pos = ia_css_circbuf_get_pos_at_offset(cb, cb->desc->end, offset); in ia_css_circbuf_peek() 277 elem = cb->elems[cb->desc->start]; in ia_css_circbuf_read() 280 ia_css_circbuf_elem_init(&cb->elems[cb->desc->start]); in ia_css_circbuf_read() 283 cb->desc->start = ia_css_circbuf_get_pos_at_offset(cb, cb->desc->start, 1); in ia_css_circbuf_read() [all …]
|
/linux/drivers/staging/media/atomisp/pci/base/circbuf/interface/ |
A D | ia_css_circbuf.h | 153 OP___assert(cb); in ia_css_circbuf_get_pos_at_offset() 186 OP___assert(cb); in ia_css_circbuf_get_offset() 207 OP___assert(cb); in ia_css_circbuf_get_size() 225 OP___assert(cb); in ia_css_circbuf_get_num_elems() 228 num = ia_css_circbuf_get_offset(cb, cb->desc->start, cb->desc->end); in ia_css_circbuf_get_num_elems() 245 OP___assert(cb); in ia_css_circbuf_is_empty() 262 OP___assert(cb); in ia_css_circbuf_is_full() 281 OP___assert(cb); in ia_css_circbuf_write() 287 ia_css_circbuf_elem_cpy(&elem, &cb->elems[cb->desc->end]); in ia_css_circbuf_write() 289 cb->desc->end = ia_css_circbuf_get_pos_at_offset(cb, cb->desc->end, 1); in ia_css_circbuf_write() [all …]
|
/linux/drivers/scsi/ |
A D | myrb.c | 173 cb->prev_cmd_mbox2 = cb->prev_cmd_mbox1; in myrb_qcmd() 194 cb->qcmd(cb, cmd_blk); in myrb_exec_cmd() 706 cb->old_ev_seq, cb->new_ev_seq, in myrb_hba_enquiry() 861 cb->next_cmd_mbox = cb->first_cmd_mbox; in myrb_enable_mmio() 1216 cb->enquiry, cb->enquiry_addr); in myrb_unmap() 1242 cb->disable_intr(cb->io_base); in myrb_cleanup() 1246 free_irq(cb->irq, cb); in myrb_cleanup() 1259 cb->reset(cb->io_base); in myrb_host_reset() 1333 cb->qcmd(cb, cmd_blk); in myrb_pthru_queuecommand() 1600 cb->qcmd(cb, cmd_blk); in myrb_ldev_queuecommand() [all …]
|
/linux/drivers/misc/habanalabs/common/ |
A D | command_buffer.c | 41 bus_addr < cb->bus_address + cb->size; in cb_map_mem() 144 (uintptr_t)cb->kernel_address, cb->size); in cb_fini() 147 cb->kernel_address, cb->bus_address); in cb_fini() 174 cb_unmap_mem(cb->ctx, cb); in cb_release() 197 cb = kzalloc(sizeof(*cb), GFP_ATOMIC); in hl_cb_alloc() 199 if (!cb) in hl_cb_alloc() 200 cb = kzalloc(sizeof(*cb), GFP_KERNEL); in hl_cb_alloc() 202 if (!cb) in hl_cb_alloc() 342 cb_unmap_mem(cb->ctx, cb); in hl_cb_create() 554 cb->bus_address, cb->size); in hl_cb_mmap() [all …]
|
/linux/drivers/mfd/ |
A D | lm3533-ctrlbank.c | 30 return base + cb->id; in lm3533_ctrlbank_get_reg() 38 dev_dbg(cb->dev, "%s - %d\n", __func__, cb->id); in lm3533_ctrlbank_enable() 40 mask = 1 << cb->id; in lm3533_ctrlbank_enable() 44 dev_err(cb->dev, "failed to enable ctrlbank %d\n", cb->id); in lm3533_ctrlbank_enable() 55 dev_dbg(cb->dev, "%s - %d\n", __func__, cb->id); in lm3533_ctrlbank_disable() 57 mask = 1 << cb->id; in lm3533_ctrlbank_disable() 60 dev_err(cb->dev, "failed to disable ctrlbank %d\n", cb->id); in lm3533_ctrlbank_disable() 83 ret = lm3533_write(cb->lm3533, reg, val); in lm3533_ctrlbank_set_max_current() 97 ret = lm3533_write(cb->lm3533, reg, val); in lm3533_ctrlbank_set_brightness() 111 ret = lm3533_read(cb->lm3533, reg, val); in lm3533_ctrlbank_get_brightness() [all …]
|
/linux/block/ |
A D | blk-stat.c | 82 struct blk_stat_callback *cb = from_timer(cb, t, timer); in blk_stat_timer_fn() local 99 cb->timer_fn(cb); in blk_stat_timer_fn() 109 cb = kmalloc(sizeof(*cb), GFP_KERNEL); in blk_stat_alloc_callback() 110 if (!cb) in blk_stat_alloc_callback() 115 if (!cb->stat) { in blk_stat_alloc_callback() 116 kfree(cb); in blk_stat_alloc_callback() 123 kfree(cb); in blk_stat_alloc_callback() 129 cb->data = data; in blk_stat_alloc_callback() 133 return cb; in blk_stat_alloc_callback() 178 kfree(cb); in blk_stat_free_callback_rcu() [all …]
|
/linux/drivers/irqchip/ |
A D | irq-crossbar.c | 47 writel(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]); in crossbar_writel() 52 writew(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]); in crossbar_writew() 57 writeb(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]); in crossbar_writeb() 163 cb->write(d->hwirq, cb->safe_map); in crossbar_domain_free() 202 cb = kzalloc(sizeof(*cb), GFP_KERNEL); in crossbar_of_init() 204 if (!cb) in crossbar_of_init() 226 if (!cb->irq_map) in crossbar_of_init() 229 cb->int_max = max; in crossbar_of_init() 313 cb->write(i, cb->safe_map); in crossbar_of_init() 327 kfree(cb); in crossbar_of_init() [all …]
|
/linux/fs/nfsd/ |
A D | nfs4callback.c | 1061 if (cb->cb_ops && cb->cb_ops->release) in nfsd41_destroy_cb() 1062 cb->cb_ops->release(cb); in nfsd41_destroy_cb() 1128 nfsd4_mark_cb_fault(cb->cb_clp, cb->cb_seq_status); in nfsd4_cb_sequence_done() 1146 nfsd4_mark_cb_fault(cb->cb_clp, cb->cb_seq_status); in nfsd4_cb_sequence_done() 1184 switch (cb->cb_ops->done(cb, task)) { in nfsd4_cb_done() 1331 if (cb->cb_ops && cb->cb_ops->prepare) in nfsd4_run_cb_work() 1332 cb->cb_ops->prepare(cb); in nfsd4_run_cb_work() 1357 cb->cb_ops ? &nfsd4_cb_ops : &nfsd4_cb_probe_ops, cb); in nfsd4_run_cb_work() 1363 cb->cb_clp = clp; in nfsd4_init_cb() 1365 cb->cb_msg.rpc_argp = cb; in nfsd4_init_cb() [all …]
|
/linux/drivers/net/mdio/ |
A D | mdio-mux.c | 89 while (cb) { in mdio_mux_uninit_children() 91 mdiobus_free(cb->mii_bus); in mdio_mux_uninit_children() 92 cb = cb->next; in mdio_mux_uninit_children() 155 cb = devm_kzalloc(dev, sizeof(*cb), GFP_KERNEL); in mdio_mux_init() 156 if (!cb) { in mdio_mux_init() 160 cb->bus_number = v; in mdio_mux_init() 161 cb->parent = pb; in mdio_mux_init() 164 if (!cb->mii_bus) { in mdio_mux_init() 168 cb->mii_bus->priv = cb; in mdio_mux_init() 183 devm_kfree(dev, cb); in mdio_mux_init() [all …]
|
/linux/drivers/misc/mei/ |
A D | interrupt.c | 35 cl = cb->cl; in mei_irq_compl_handler() 113 if (!cb) { in mei_cl_irq_read_msg() 119 if (!cb) in mei_cl_irq_read_msg() 151 if (cb->vtag && cb->vtag != vtag_hdr->vtag) { in mei_cl_irq_read_msg() 180 cb->buf.size, length, cb->buf_idx); in mei_cl_irq_read_msg() 186 mei_dma_ring_read(dev, cb->buf.data + cb->buf_idx, length); in mei_cl_irq_read_msg() 188 mei_read_slots(dev, cb->buf.data + cb->buf_idx, 0); in mei_cl_irq_read_msg() 190 mei_read_slots(dev, cb->buf.data + cb->buf_idx, length); in mei_cl_irq_read_msg() 206 if (cb) in mei_cl_irq_read_msg() 503 cl = cb->cl; in mei_irq_write_handler() [all …]
|
A D | client.c | 325 kfree(cb); in mei_io_cb_free() 394 cb = kzalloc(sizeof(*cb), GFP_KERNEL); in mei_io_cb_init() 395 if (!cb) in mei_io_cb_init() 405 return cb; in mei_io_cb_init() 442 if (cl == cb->cl && (!fp || fp == cb->fp)) in mei_io_tx_list_free_cl() 492 if (!cb) in mei_cl_alloc_cb() 505 return cb; in mei_cl_alloc_cb() 531 if (!cb) in mei_cl_enqueue_ctrl_wr_cb() 1696 if (!cb) in mei_cl_read_start() 1754 if (!cb) in mei_msg_hdr_init() [all …]
|
/linux/drivers/misc/sgi-gru/ |
A D | gru_instructions.h | 22 extern int gru_check_status_proc(void *cb); 23 extern int gru_wait_proc(void *cb); 24 extern void gru_wait_abort_proc(void *cb); 79 unsigned long cb; member 400 struct gru_instruction *ins = (void *)cb; in gru_vstore() 413 struct gru_instruction *ins = (void *)cb; in gru_ivload() 426 struct gru_instruction *ins = (void *)cb; in gru_ivstore() 672 ret = gru_check_status_proc(cb); in gru_check_status() 683 static inline int gru_wait(void *cb) in gru_wait() argument 685 return gru_wait_proc(cb); in gru_wait() [all …]
|
A D | grukservices.c | 367 if (cb) in gru_lock_async_resource() 513 ret = gru_wait_proc(cb); in gru_wait_abort_proc() 592 ret = gru_wait(cb); in send_noop_message() 800 void *cb; in gru_send_message_gpa() local 907 void *cb; in gru_read_gpa() local 916 ret = gru_wait(cb); in gru_read_gpa() 931 void *cb; in gru_copy_gpa() local 940 ret = gru_wait(cb); in gru_copy_gpa() 953 void *cb; in quicktest0() local 1050 void *cb0, *cb; in quicktest2() local [all …]
|
/linux/arch/s390/kernel/ |
A D | runtime_instr.c | 55 cb->rla = 0xfff; in init_runtime_instr_cb() 56 cb->s = 1; in init_runtime_instr_cb() 57 cb->k = 1; in init_runtime_instr_cb() 58 cb->ps = 1; in init_runtime_instr_cb() 59 cb->pc = 1; in init_runtime_instr_cb() 61 cb->v = 1; in init_runtime_instr_cb() 86 cb = kzalloc(sizeof(*cb), GFP_KERNEL); in SYSCALL_DEFINE2() 87 if (!cb) in SYSCALL_DEFINE2() 91 memset(cb, 0, sizeof(*cb)); in SYSCALL_DEFINE2() 94 init_runtime_instr_cb(cb); in SYSCALL_DEFINE2() [all …]
|
/linux/fs/btrfs/ |
A D | compression.c | 276 kfree(cb); in finish_compressed_bio_read() 363 cb->start, cb->len); in end_compressed_writeback() 382 cb->start, cb->start + cb->len - 1, in finish_compressed_bio_write() 401 kfree(cb); in finish_compressed_bio_write() 418 btrfs_record_physical_zoned(cb->inode, cb->start, bio); in end_compressed_bio_write() 523 if (!cb) in btrfs_submit_compressed_write() 529 cb->len = len; in btrfs_submit_compressed_write() 619 wait_var_event(cb, refcount_read(&cb->pending_sectors) == in btrfs_submit_compressed_write() 829 if (!cb) in btrfs_submit_compressed_read() 950 kfree(cb); in btrfs_submit_compressed_read() [all …]
|
/linux/drivers/net/ethernet/netronome/nfp/ |
A D | ccm_mbox.c | 64 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_msg_init() local 75 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_maxlen() local 82 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_done() local 89 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_in_progress() local 97 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_set_busy() local 104 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_is_posted() local 111 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_mark_posted() local 137 cb = (void *)skb->cb; in nfp_ccm_mbox_mark_next_runner() 268 cb = (void *)skb->cb; in nfp_ccm_mbox_copy_out() 333 cb = (void *)skb->cb; in nfp_ccm_mbox_copy_out() [all …]
|
/linux/drivers/dma-buf/ |
A D | st-dma-fence.c | 43 wake_up_process(container_of(cb, struct wait_cb, cb)->task); in mock_wakeup() 51 if (dma_fence_add_callback(f, &cb.cb, mock_wakeup)) in mock_wait() 67 if (!dma_fence_remove_callback(f, &cb.cb)) in mock_wait() 166 if (dma_fence_add_callback(f, &cb.cb, simple_callback)) { in test_add_callback() 195 if (!dma_fence_add_callback(f, &cb.cb, simple_callback)) { in test_late_add_callback() 201 if (cb.seen) { in test_late_add_callback() 222 if (dma_fence_add_callback(f, &cb.cb, simple_callback)) { in test_rm_callback() 227 if (!dma_fence_remove_callback(f, &cb.cb)) { in test_rm_callback() 233 if (cb.seen) { in test_rm_callback() 254 if (dma_fence_add_callback(f, &cb.cb, simple_callback)) { in test_late_rm_callback() [all …]
|
/linux/net/decnet/ |
A D | dn_nsp_in.c | 80 le16_to_cpu(cb->src), in dn_log_martian() 240 cb->info = msg->info; in dn_find_listener() 346 cb->services = *ptr++; in dn_nsp_conn_conf() 347 cb->info = *ptr++; in dn_nsp_conn_conf() 356 scp->info_rem = cb->info; in dn_nsp_conn_conf() 722 cb->nsp_flags = *ptr++; in dn_nsp_rx_packet() 727 if (cb->nsp_flags & 0x83) in dn_nsp_rx_packet() 755 cb->src_port = 0; in dn_nsp_rx_packet() 772 swap(cb->dst_port, cb->src_port); in dn_nsp_rx_packet() 773 swap(cb->dst, cb->src); in dn_nsp_rx_packet() [all …]
|
/linux/tools/testing/selftests/bpf/verifier/ |
A D | ctx_skb.c | 396 offsetof(struct __sk_buff, cb[0])), 404 offsetof(struct __sk_buff, cb[1])), 412 offsetof(struct __sk_buff, cb[2])), 420 offsetof(struct __sk_buff, cb[3])), 428 offsetof(struct __sk_buff, cb[4])), 436 offsetof(struct __sk_buff, cb[0])), 444 offsetof(struct __sk_buff, cb[1])), 452 offsetof(struct __sk_buff, cb[2])), 460 offsetof(struct __sk_buff, cb[3])), 468 offsetof(struct __sk_buff, cb[4])), [all …]
|
/linux/drivers/dma/ |
A D | dmaengine.h | 115 struct dmaengine_desc_callback *cb) in dmaengine_desc_get_callback() argument 117 cb->callback = tx->callback; in dmaengine_desc_get_callback() 118 cb->callback_result = tx->callback_result; in dmaengine_desc_get_callback() 119 cb->callback_param = tx->callback_param; in dmaengine_desc_get_callback() 140 if (cb->callback_result) { in dmaengine_desc_callback_invoke() 143 cb->callback_result(cb->callback_param, result); in dmaengine_desc_callback_invoke() 144 } else if (cb->callback) { in dmaengine_desc_callback_invoke() 145 cb->callback(cb->callback_param); in dmaengine_desc_callback_invoke() 163 struct dmaengine_desc_callback cb; in dmaengine_desc_get_callback_invoke() local 165 dmaengine_desc_get_callback(tx, &cb); in dmaengine_desc_get_callback_invoke() [all …]
|
/linux/net/sctp/ |
A D | diag.c | 294 struct netlink_callback *cb = commp->cb; in sctp_sock_dump() local 303 if (cb->args[4] < cb->args[1]) in sctp_sock_dump() 328 cb->nlh->nlmsg_seq, 0, cb->nlh, in sctp_sock_dump() 334 cb->args[4]++; in sctp_sock_dump() 336 cb->args[1] = 0; in sctp_sock_dump() 337 cb->args[3] = 0; in sctp_sock_dump() 367 struct netlink_callback *cb = commp->cb; in sctp_ep_dump() local 376 if (cb->args[4] < cb->args[1]) in sctp_ep_dump() 403 cb->args[4]++; in sctp_ep_dump() 470 .cb = cb, in sctp_diag_dump() [all …]
|
/linux/drivers/gpu/drm/i915/ |
A D | i915_sw_fence.c | 417 struct i915_sw_dma_fence_cb *cb = container_of(data, typeof(*cb), base); in dma_i915_sw_fence_wake() local 421 kfree(cb); in dma_i915_sw_fence_wake() 426 struct i915_sw_dma_fence_cb_timer *cb = from_timer(cb, t, timer); in timer_i915_sw_fence_wake() local 434 cb->dma->ops->get_driver_name(cb->dma), in timer_i915_sw_fence_wake() 435 cb->dma->ops->get_timeline_name(cb->dma), in timer_i915_sw_fence_wake() 436 cb->dma->seqno, in timer_i915_sw_fence_wake() 467 kfree_rcu(cb, rcu); in irq_i915_sw_fence_work() 491 if (!cb) { in i915_sw_fence_await_dma_fence() 503 cb->fence = fence; in i915_sw_fence_await_dma_fence() 536 struct i915_sw_dma_fence_cb *cb = container_of(data, typeof(*cb), base); in __dma_i915_sw_fence_wake() local [all …]
|
/linux/arch/x86/kernel/cpu/ |
A D | scattered.c | 52 const struct cpuid_bit *cb; in init_scattered_cpuid_features() local 54 for (cb = cpuid_bits; cb->feature; cb++) { in init_scattered_cpuid_features() 57 max_level = cpuid_eax(cb->level & 0xffff0000); in init_scattered_cpuid_features() 58 if (max_level < cb->level || in init_scattered_cpuid_features() 59 max_level > (cb->level | 0xffff)) in init_scattered_cpuid_features() 62 cpuid_count(cb->level, cb->sub_leaf, ®s[CPUID_EAX], in init_scattered_cpuid_features() 66 if (regs[cb->reg] & (1 << cb->bit)) in init_scattered_cpuid_features() 67 set_cpu_cap(c, cb->feature); in init_scattered_cpuid_features()
|
/linux/drivers/net/wireless/marvell/mwifiex/ |
A D | util.h | 55 struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; in MWIFIEX_SKB_RXCB() local 57 BUILD_BUG_ON(sizeof(struct mwifiex_cb) > sizeof(skb->cb)); in MWIFIEX_SKB_RXCB() 58 return &cb->rx_info; in MWIFIEX_SKB_RXCB() 63 struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; in MWIFIEX_SKB_TXCB() local 65 return &cb->tx_info; in MWIFIEX_SKB_TXCB() 71 struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; in mwifiex_store_mapping() local 73 memcpy(&cb->dma_mapping, mapping, sizeof(*mapping)); in mwifiex_store_mapping() 79 struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; in mwifiex_get_mapping() local 81 memcpy(mapping, &cb->dma_mapping, sizeof(*mapping)); in mwifiex_get_mapping()
|
/linux/drivers/s390/net/ |
A D | smsgiucv.c | 55 struct smsg_callback *cb; in smsg_message_pending() local 79 if (strncmp(buffer + 8, cb->prefix, cb->len) == 0) { in smsg_message_pending() 91 struct smsg_callback *cb; in smsg_register_callback() local 94 if (!cb) in smsg_register_callback() 96 cb->prefix = prefix; in smsg_register_callback() 97 cb->len = strlen(prefix); in smsg_register_callback() 98 cb->callback = callback; in smsg_register_callback() 112 cb = NULL; in smsg_unregister_callback() 116 cb = tmp; in smsg_unregister_callback() 117 list_del(&cb->list); in smsg_unregister_callback() [all …]
|