/linux/drivers/gpu/drm/nouveau/nvkm/falcon/ |
A D | cmdq.c | 28 u32 head = nvkm_falcon_rd32(cmdq->qmgr->falcon, cmdq->head_reg); in nvkm_falcon_cmdq_has_room() 29 u32 tail = nvkm_falcon_rd32(cmdq->qmgr->falcon, cmdq->tail_reg); in nvkm_falcon_cmdq_has_room() 35 free = cmdq->offset + cmdq->size - head; in nvkm_falcon_cmdq_has_room() 67 cmdq->position = cmdq->offset; in nvkm_falcon_cmdq_rewind() 84 cmdq->position = nvkm_falcon_rd32(falcon, cmdq->head_reg); in nvkm_falcon_cmdq_open() 95 nvkm_falcon_wr32(cmdq->qmgr->falcon, cmdq->head_reg, cmdq->position); in nvkm_falcon_cmdq_close() 180 cmdq->head_reg = func->cmdq.head + index * func->cmdq.stride; in nvkm_falcon_cmdq_init() 181 cmdq->tail_reg = func->cmdq.tail + index * func->cmdq.stride; in nvkm_falcon_cmdq_init() 187 index, cmdq->offset, cmdq->size); in nvkm_falcon_cmdq_init() 194 if (cmdq) { in nvkm_falcon_cmdq_del() [all …]
|
A D | Kbuild | 3 nvkm-y += nvkm/falcon/cmdq.o
|
/linux/drivers/mailbox/ |
A D | mtk-cmdq-mailbox.c | 65 struct cmdq *cmdq; member 95 struct cmdq *cmdq = container_of(chan->mbox, struct cmdq, mbox); in cmdq_get_shift_pa() argument 126 static void cmdq_init(struct cmdq *cmdq) in cmdq_init() argument 213 struct cmdq *cmdq = task->cmdq; in cmdq_task_handle_error() local 280 struct cmdq *cmdq = dev; in cmdq_irq_handler() local 301 struct cmdq *cmdq = dev_get_drvdata(dev); in cmdq_suspend() local 326 struct cmdq *cmdq = dev_get_drvdata(dev); in cmdq_resume() local 356 task->cmdq = cmdq; in cmdq_mbox_send_data() 527 struct cmdq *cmdq; in cmdq_probe() local 567 dev, cmdq->base, cmdq->irq); in cmdq_probe() [all …]
|
A D | Makefile | 52 obj-$(CONFIG_MTK_CMDQ_MBOX) += mtk-cmdq-mailbox.o
|
/linux/drivers/crypto/cavium/nitrox/ |
A D | nitrox_lib.c | 29 cmdq->qsize = (ndev->qlen * cmdq->instr_size) + align_bytes; in nitrox_cmdq_init() 30 cmdq->unalign_base = dma_alloc_coherent(DEV(ndev), cmdq->qsize, in nitrox_cmdq_init() 36 cmdq->dma = PTR_ALIGN(cmdq->unalign_dma, align_bytes); in nitrox_cmdq_init() 37 cmdq->base = cmdq->unalign_base + (cmdq->dma - cmdq->unalign_dma); in nitrox_cmdq_init() 64 if (!cmdq) in nitrox_cmdq_cleanup() 74 cmdq->unalign_base, cmdq->unalign_dma); in nitrox_cmdq_cleanup() 82 cmdq->dma = 0; in nitrox_cmdq_cleanup() 83 cmdq->qsize = 0; in nitrox_cmdq_cleanup() 106 cmdq = kzalloc_node(sizeof(*cmdq), GFP_KERNEL, ndev->node); in nitrox_alloc_aqm_queues() 107 if (!cmdq) { in nitrox_alloc_aqm_queues() [all …]
|
A D | nitrox_reqmgr.c | 296 idx = cmdq->write_idx; in post_se_instr() 298 ent = cmdq->base + (idx * cmdq->instr_size); in post_se_instr() 302 response_list_add(sr, cmdq); in post_se_instr() 342 post_se_instr(sr, cmdq); in post_backlog_cmds() 351 struct nitrox_cmdq *cmdq = sr->cmdq; in nitrox_enqueue_request() local 355 post_backlog_cmds(cmdq); in nitrox_enqueue_request() 364 backlog_list_add(sr, cmdq); in nitrox_enqueue_request() 367 post_se_instr(sr, cmdq); in nitrox_enqueue_request() 507 struct nitrox_cmdq *cmdq; in backlog_qflush_work() local 510 post_backlog_cmds(cmdq); in backlog_qflush_work() [all …]
|
A D | nitrox_isr.c | 32 struct nitrox_cmdq *cmdq = qvec->cmdq; in nps_pkt_slc_isr() local 34 slc_cnts.value = readq(cmdq->compl_cnt_csr_addr); in nps_pkt_slc_isr() 337 qvec->cmdq = &ndev->pkt_inq[qvec->ring]; in nitrox_register_interrupts()
|
A D | nitrox_hal.c | 125 struct nitrox_cmdq *cmdq = &ndev->pkt_inq[i]; in nitrox_config_pkt_input_rings() local 138 nitrox_write_csr(ndev, offset, cmdq->dma); in nitrox_config_pkt_input_rings() 357 struct nitrox_cmdq *cmdq = ndev->aqmq[ring]; in nitrox_config_aqm_rings() local 380 nitrox_write_csr(ndev, offset, cmdq->dma); in nitrox_config_aqm_rings()
|
A D | nitrox_dev.h | 107 struct nitrox_cmdq *cmdq; member
|
/linux/drivers/net/ethernet/brocade/bna/ |
A D | bfa_msgq.c | 56 cmdq->flags = 0; in cmdq_sm_stopped_entry() 57 cmdq->token = 0; in cmdq_sm_stopped_entry() 58 cmdq->offset = 0; in cmdq_sm_stopped_entry() 195 if (!bfa_nw_ioc_mbox_queue(cmdq->msgq->ioc, &cmdq->dbell_mb, in bfa_msgq_cmdq_dbell() 219 BFA_MSGQ_INDX_ADD(cmdq->producer_index, 1, cmdq->depth); in __cmd_copy() 270 cmdq->token = 0; in bfa_msgq_cmdq_copy_req() 292 cmdq->token++; in bfa_msgq_cmdq_copy_rsp() 296 if (!bfa_nw_ioc_mbox_queue(cmdq->msgq->ioc, &cmdq->copy_mb, in bfa_msgq_cmdq_copy_rsp() 307 cmdq->msgq = msgq; in bfa_msgq_cmdq_attach() 517 bfa_dma_be_addr_set(msgq_cfg->cmdq.addr, msgq->cmdq.addr.pa); in bfa_msgq_init() [all …]
|
A D | bfa_msgq.h | 104 struct bfa_msgq_cmdq cmdq; member
|
/linux/drivers/net/ethernet/huawei/hinic/ |
A D | hinic_hw_cmdq.c | 78 #define cmdq_to_cmdqs(cmdq) container_of((cmdq) - (cmdq)->cmdq_type, \ argument 369 cmdq->wrapped = !cmdq->wrapped; in cmdq_sync_cmd_direct_resp() 448 cmdq->wrapped = !cmdq->wrapped; in cmdq_set_arm_bit() 515 struct hinic_cmdq *cmdq = &cmdqs->cmdq[HINIC_CMDQ_SYNC]; in hinic_set_arm_bit() local 648 struct hinic_cmdq *cmdq = &cmdqs->cmdq[cmdq_type]; in cmdq_ceq_handler() local 748 cmdq->wq = wq; in init_cmdq() 750 cmdq->wrapped = 1; in init_cmdq() 754 cmdq->done = vzalloc(array_size(sizeof(*cmdq->done), wq->q_depth)); in init_cmdq() 755 if (!cmdq->done) in init_cmdq() 758 cmdq->errcode = vzalloc(array_size(sizeof(*cmdq->errcode), in init_cmdq() [all …]
|
A D | hinic_hw_io.c | 534 enum hinic_cmdq_type cmdq, type; in hinic_io_init() local 566 for (cmdq = HINIC_CMDQ_SYNC; cmdq < HINIC_MAX_CMDQ_TYPES; cmdq++) { in hinic_io_init() 574 func_to_io->cmdq_db_area[cmdq] = db_area; in hinic_io_init() 601 for (type = HINIC_CMDQ_SYNC; type < cmdq; type++) in hinic_io_init() 620 enum hinic_cmdq_type cmdq; in hinic_io_free() local 629 for (cmdq = HINIC_CMDQ_SYNC; cmdq < HINIC_MAX_CMDQ_TYPES; cmdq++) in hinic_io_free() 630 return_db_area(func_to_io, func_to_io->cmdq_db_area[cmdq]); in hinic_io_free()
|
/linux/drivers/infiniband/hw/bnxt_re/ |
A D | qplib_rcfw.c | 62 cmdq = &rcfw->cmdq; in __wait_for_resp() 76 cmdq = &rcfw->cmdq; in __block_for_resp() 91 struct bnxt_qplib_cmdq_ctx *cmdq = &rcfw->cmdq; in __send_message() local 182 cmdq->seq_num++; in __send_message() 578 cmdq = &rcfw->cmdq; in bnxt_qplib_alloc_rcfw_channel() 618 if (!cmdq->cmdq_bitmap) in bnxt_qplib_alloc_rcfw_channel() 663 cmdq = &rcfw->cmdq; in bnxt_qplib_disable_rcfw_channel() 670 indx = find_first_bit(cmdq->cmdq_bitmap, cmdq->bmap_size); in bnxt_qplib_disable_rcfw_channel() 785 cmdq = &rcfw->cmdq; in bnxt_qplib_start_rcfw() 811 cmdq = &rcfw->cmdq; in bnxt_qplib_enable_rcfw_channel() [all …]
|
/linux/drivers/iommu/arm/arm-smmu-v3/ |
A D | arm-smmu-v3.c | 340 return &smmu->cmdq; in arm_smmu_get_cmdq() 459 if (atomic_read(&cmdq->lock) == 1) in arm_smmu_cmdq_shared_tryunlock() 462 arm_smmu_cmdq_shared_unlock(cmdq); in arm_smmu_cmdq_shared_tryunlock() 542 ptr = &cmdq->valid_map[swidx]; in __arm_smmu_cmdq_poll_set_valid_map() 596 WRITE_ONCE(cmdq->q.llq.cons, readl_relaxed(cmdq->q.cons_reg)); in arm_smmu_cmdq_poll_until_not_full() 790 arm_smmu_cmdq_shared_lock(cmdq); in arm_smmu_cmdq_issue_cmdlist() 804 &cmdq->q.llq.atomic.prod); in arm_smmu_cmdq_issue_cmdlist() 2923 struct arm_smmu_cmdq *cmdq = &smmu->cmdq; in arm_smmu_cmdq_init() local 2927 atomic_set(&cmdq->owner_prod, 0); in arm_smmu_cmdq_init() 2928 atomic_set(&cmdq->lock, 0); in arm_smmu_cmdq_init() [all …]
|
/linux/drivers/gpu/drm/nouveau/nvkm/engine/sec2/ |
A D | base.c | 61 nvkm_falcon_cmdq_fini(sec2->cmdq); in nvkm_sec2_fini() 73 nvkm_falcon_cmdq_del(&sec2->cmdq); in nvkm_sec2_dtor() 112 (ret = nvkm_falcon_cmdq_new(sec2->qmgr, "cmdq", &sec2->cmdq)) || in nvkm_sec2_new_()
|
A D | gp102.c | 71 return nvkm_falcon_cmdq_send(sec2->cmdq, &cmd.cmd.hdr, in gp102_sec2_acr_bootstrap_falcon() 153 nvkm_falcon_cmdq_init(sec2->cmdq, in gp102_sec2_initmsg() 246 .cmdq = { 0xa00, 0xa04, 8 },
|
A D | tu102.c | 40 .cmdq = { 0xc00, 0xc04, 8 },
|
/linux/drivers/atm/ |
A D | fore200e.c | 557 struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; in fore200e_pca_prom_read() 562 FORE200E_NEXT_ENTRY(cmdq->head, QUEUE_SIZE_CMD); in fore200e_pca_prom_read() 1226 struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; in fore200e_activate_vcin() 1670 struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; in fore200e_getstats() 1716 struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; 1756 struct host_cmdq_entry* entry = &cmdq->host_entry[ cmdq->head ]; in fore200e_set_oc3() 2240 &cmdq->status, in fore200e_init_cmd_queue() 2253 cmdq->host_entry[ i ].status = in fore200e_init_cmd_queue() 2255 cmdq->host_entry[ i ].cp_entry = &cp_entry[ i ]; in fore200e_init_cmd_queue() 2257 *cmdq->host_entry[ i ].status = STATUS_FREE; in fore200e_init_cmd_queue() [all …]
|
/linux/Documentation/devicetree/bindings/iommu/ |
A D | arm,smmu-v3.yaml | 45 - const: cmdq-sync # CMD_SYNC complete 90 interrupt-names = "eventq", "gerror", "priq", "cmdq-sync";
|
/linux/drivers/gpu/drm/nouveau/include/nvkm/engine/ |
A D | sec2.h | 14 struct nvkm_falcon_cmdq *cmdq; member
|
/linux/drivers/soc/mediatek/ |
A D | Makefile | 2 obj-$(CONFIG_MTK_CMDQ) += mtk-cmdq-helper.o
|
/linux/drivers/scsi/aacraid/ |
A D | dpcsup.c | 200 list_add_tail(&fib->fiblink, &q->cmdq); in aac_command_normal() 310 list_add_tail(&fib->fiblink, &q->cmdq); in aac_intr_normal()
|
/linux/drivers/net/ethernet/marvell/prestera/ |
A D | prestera_pci.c | 469 struct prestera_fw_cmdq *cmdq = &fw->cmd_queue[qid]; in prestera_fw_init() local 471 cmdq->len = prestera_fw_read(fw, PRESTERA_CMDQ_LEN_REG(qid)); in prestera_fw_init() 472 cmdq->addr = fw->cmd_mbox + offs; in prestera_fw_init() 473 mutex_init(&cmdq->cmd_mtx); in prestera_fw_init()
|
/linux/drivers/net/ethernet/chelsio/cxgb/ |
A D | sge.c | 1447 struct cmdQ *cmdq = &sge->cmdQ[0]; in update_tx_info() local 1449 cmdq->processed += pr0; in update_tx_info() 1455 clear_bit(CMDQ_STAT_RUNNING, &cmdq->status); in update_tx_info() 1457 if (cmdq->cleaned + cmdq->in_use != cmdq->processed && in update_tx_info() 1458 !test_and_set_bit(CMDQ_STAT_LAST_PKT_DB, &cmdq->status)) { in update_tx_info() 1459 set_bit(CMDQ_STAT_RUNNING, &cmdq->status); in update_tx_info()
|