/linux/drivers/crypto/qat/qat_common/ |
A D | qat_asym_algs.c | 278 if (unlikely(dma_mapping_error(dev, in qat_dh_compute_value() 348 if (!dma_mapping_error(dev, qat_req->phy_out)) in qat_dh_compute_value() 353 if (!dma_mapping_error(dev, qat_req->phy_in)) in qat_dh_compute_value() 362 if (!dma_mapping_error(dev, qat_req->out.dh.r)) in qat_dh_compute_value() 371 if (!dma_mapping_error(dev, qat_req->in.dh.in.b)) in qat_dh_compute_value() 742 if (!dma_mapping_error(dev, qat_req->phy_out)) in qat_rsa_enc() 747 if (!dma_mapping_error(dev, qat_req->phy_in)) in qat_rsa_enc() 764 if (!dma_mapping_error(dev, qat_req->in.rsa.enc.m)) in qat_rsa_enc() 894 if (!dma_mapping_error(dev, qat_req->phy_out)) in qat_rsa_dec() 899 if (!dma_mapping_error(dev, qat_req->phy_in)) in qat_rsa_dec() [all …]
|
A D | qat_algs.c | 747 if (unlikely(dma_mapping_error(dev, bufl->bufers[y].addr))) in qat_alg_sgl_to_bufl() 753 if (unlikely(dma_mapping_error(dev, blp))) in qat_alg_sgl_to_bufl() 783 if (unlikely(dma_mapping_error(dev, bufers[y].addr))) in qat_alg_sgl_to_bufl() 791 if (unlikely(dma_mapping_error(dev, bloutp))) in qat_alg_sgl_to_bufl() 804 if (!dma_mapping_error(dev, bloutp)) in qat_alg_sgl_to_bufl() 809 if (!dma_mapping_error(dev, buflout->bufers[i].addr)) in qat_alg_sgl_to_bufl() 816 if (!dma_mapping_error(dev, blp)) in qat_alg_sgl_to_bufl() 821 if (!dma_mapping_error(dev, bufl->bufers[i].addr)) in qat_alg_sgl_to_bufl()
|
/linux/drivers/crypto/caam/ |
A D | caampkc.c | 406 if (dma_mapping_error(dev, pdb->n_dma)) { in set_rsa_pub_pdb() 412 if (dma_mapping_error(dev, pdb->e_dma)) { in set_rsa_pub_pdb() 451 if (dma_mapping_error(dev, pdb->n_dma)) { in set_rsa_priv_f1_pdb() 457 if (dma_mapping_error(dev, pdb->d_dma)) { in set_rsa_priv_f1_pdb() 500 if (dma_mapping_error(dev, pdb->d_dma)) { in set_rsa_priv_f2_pdb() 506 if (dma_mapping_error(dev, pdb->p_dma)) { in set_rsa_priv_f2_pdb() 512 if (dma_mapping_error(dev, pdb->q_dma)) { in set_rsa_priv_f2_pdb() 577 if (dma_mapping_error(dev, pdb->p_dma)) { in set_rsa_priv_f3_pdb() 583 if (dma_mapping_error(dev, pdb->q_dma)) { in set_rsa_priv_f3_pdb() 589 if (dma_mapping_error(dev, pdb->dp_dma)) { in set_rsa_priv_f3_pdb() [all …]
|
A D | caamalg_qi2.c | 489 if (dma_mapping_error(dev, iv_dma)) { in aead_edesc_alloc() 534 if (dma_mapping_error(dev, qm_sg_dma)) { in aead_edesc_alloc() 1206 if (dma_mapping_error(dev, iv_dma)) { in skcipher_edesc_alloc() 1230 if (dma_mapping_error(dev, edesc->qm_sg_dma)) { in skcipher_edesc_alloc() 1589 if (dma_mapping_error(ctx->dev, dma_addr)) { in caam_cra_init() 3104 if (dma_mapping_error(dev, state->buf_dma)) { in buf_map_to_qm_sg() 3122 if (dma_mapping_error(dev, state->ctx_dma)) { in ctx_map_to_qm_sg() 3233 if (dma_mapping_error(ctx->dev, key_dma)) { in hash_digest_key() 3253 if (dma_mapping_error(ctx->dev, flc_dma)) { in hash_digest_key() 4577 if (dma_mapping_error(ctx->dev, dma_addr)) { in caam_hash_cra_init() [all …]
|
A D | caamhash.c | 148 if (dma_mapping_error(jrdev, state->ctx_dma)) { in map_seq_out_ptr_ctx() 171 if (dma_mapping_error(jrdev, state->buf_dma)) { in buf_map_to_sec4_sg() 189 if (dma_mapping_error(jrdev, state->ctx_dma)) { in ctx_map_to_sec4_sg() 377 if (dma_mapping_error(jrdev, key_dma)) { in hash_digest_key() 736 if (dma_mapping_error(ctx->jrdev, src_dma)) { in ahash_edesc_add_src() 903 if (dma_mapping_error(jrdev, edesc->sec4_sg_dma)) { in ahash_update_ctx() 976 if (dma_mapping_error(jrdev, edesc->sec4_sg_dma)) { in ahash_final_ctx() 1163 if (dma_mapping_error(jrdev, state->buf_dma)) { in ahash_final_no_ctx() 1268 if (dma_mapping_error(jrdev, edesc->sec4_sg_dma)) { in ahash_update_no_ctx() 1816 if (dma_mapping_error(ctx->jrdev, ctx->adata.key_dma)) { in caam_hash_cra_init() [all …]
|
/linux/drivers/net/ethernet/stmicro/stmmac/ |
A D | chain_mode.c | 39 if (dma_mapping_error(priv->device, des2)) in jumbo_frm() 57 if (dma_mapping_error(priv->device, des2)) in jumbo_frm() 70 if (dma_mapping_error(priv->device, des2)) in jumbo_frm()
|
A D | ring_mode.c | 43 if (dma_mapping_error(priv->device, des2)) in jumbo_frm() 64 if (dma_mapping_error(priv->device, des2)) in jumbo_frm() 78 if (dma_mapping_error(priv->device, des2)) in jumbo_frm()
|
/linux/drivers/infiniband/hw/qib/ |
A D | qib_user_pages.c | 64 if (dma_mapping_error(&hwdev->dev, phys)) in qib_map_page() 71 if (dma_mapping_error(&hwdev->dev, phys)) in qib_map_page()
|
/linux/drivers/spi/ |
A D | spi-fsl-cpm.c | 129 if (dma_mapping_error(dev, mspi->tx_dma)) { in fsl_spi_cpm_bufs() 140 if (dma_mapping_error(dev, mspi->rx_dma)) { in fsl_spi_cpm_bufs() 338 if (dma_mapping_error(dev, mspi->dma_dummy_tx)) { in fsl_spi_cpm_init() 345 if (dma_mapping_error(dev, mspi->dma_dummy_rx)) { in fsl_spi_cpm_init()
|
/linux/drivers/crypto/allwinner/sun8i-ce/ |
A D | sun8i-ce-prng.c | 98 if (dma_mapping_error(ce->dev, dma_iv)) { in sun8i_ce_prng_generate() 105 if (dma_mapping_error(ce->dev, dma_dst)) { in sun8i_ce_prng_generate()
|
/linux/drivers/crypto/allwinner/sun8i-ss/ |
A D | sun8i-ss-prng.c | 102 if (dma_mapping_error(ss->dev, dma_iv)) { in sun8i_ss_prng_generate() 109 if (dma_mapping_error(ss->dev, dma_dst)) { in sun8i_ss_prng_generate()
|
/linux/drivers/crypto/ccree/ |
A D | cc_buffer_mgr.c | 270 if (dma_mapping_error(dev, ret)) { in cc_map_sg() 383 if (dma_mapping_error(dev, req_ctx->gen_ctx.iv_dma_addr)) { in cc_map_cipher_request() 559 if (dma_mapping_error(dev, areq_ctx->gen_ctx.iv_dma_addr)) { in cc_aead_chain_iv() 970 if (dma_mapping_error(dev, dma_addr)) { in cc_map_aead_request() 984 if (dma_mapping_error(dev, dma_addr)) { in cc_map_aead_request() 1002 if (dma_mapping_error(dev, dma_addr)) { in cc_map_aead_request() 1012 if (dma_mapping_error(dev, dma_addr)) { in cc_map_aead_request() 1023 if (dma_mapping_error(dev, dma_addr)) { in cc_map_aead_request() 1035 if (dma_mapping_error(dev, dma_addr)) { in cc_map_aead_request()
|
/linux/drivers/net/ethernet/amd/xgbe/ |
A D | xgbe-desc.c | 319 if (dma_mapping_error(pdata->dev, pages_dma)) { in xgbe_alloc_pages() 560 if (dma_mapping_error(pdata->dev, skb_dma)) { in xgbe_map_tx_skb() 584 if (dma_mapping_error(pdata->dev, skb_dma)) { in xgbe_map_tx_skb() 616 if (dma_mapping_error(pdata->dev, skb_dma)) { in xgbe_map_tx_skb()
|
/linux/drivers/net/ethernet/synopsys/ |
A D | dwc-xlgmac-desc.c | 352 if (dma_mapping_error(pdata->dev, pages_dma)) { in xlgmac_alloc_pages() 536 if (dma_mapping_error(pdata->dev, skb_dma)) { in xlgmac_map_tx_skb() 560 if (dma_mapping_error(pdata->dev, skb_dma)) { in xlgmac_map_tx_skb() 592 if (dma_mapping_error(pdata->dev, skb_dma)) { in xlgmac_map_tx_skb()
|
/linux/tools/virtio/linux/ |
A D | dma-mapping.h | 27 #define dma_mapping_error(...) (0) macro
|
/linux/drivers/misc/bcm-vk/ |
A D | bcm_vk_sg.c | 112 if (unlikely(dma_mapping_error(dev, sg_addr))) { in bcm_vk_dma_alloc() 125 if (unlikely(dma_mapping_error(dev, addr))) { in bcm_vk_dma_alloc()
|
/linux/drivers/crypto/ |
A D | mxs-dcp.c | 182 dma_err = dma_mapping_error(sdcp->dev, desc_phys); in mxs_dcp_start_dma() 231 ret = dma_mapping_error(sdcp->dev, key_phys); in mxs_dcp_run_aes() 237 ret = dma_mapping_error(sdcp->dev, src_phys); in mxs_dcp_run_aes() 243 ret = dma_mapping_error(sdcp->dev, dst_phys); in mxs_dcp_run_aes() 559 ret = dma_mapping_error(sdcp->dev, buf_phys); in mxs_dcp_run_sha() 595 ret = dma_mapping_error(sdcp->dev, digest_phys); in mxs_dcp_run_sha()
|
/linux/drivers/crypto/cavium/cpt/ |
A D | cptvf_reqmanager.c | 63 if (unlikely(dma_mapping_error(&pdev->dev, in setup_sgio_components() 191 if (dma_mapping_error(&pdev->dev, info->dptr_baddr)) { in setup_sgio_list() 210 if (dma_mapping_error(&pdev->dev, info->rptr_baddr)) { in setup_sgio_list() 455 if (dma_mapping_error(&pdev->dev, info->comp_baddr)) { in process_request()
|
/linux/drivers/soc/qcom/ |
A D | qcom-geni-se.c | 694 if (dma_mapping_error(wrapper->dev, *iova)) in geni_se_tx_dma_prep() 730 if (dma_mapping_error(wrapper->dev, *iova)) in geni_se_rx_dma_prep() 758 if (!dma_mapping_error(wrapper->dev, iova)) in geni_se_tx_dma_unprep() 775 if (!dma_mapping_error(wrapper->dev, iova)) in geni_se_rx_dma_unprep()
|
/linux/drivers/net/wireless/intel/iwlwifi/pcie/ |
A D | tx-gen2.c | 205 if (dma_mapping_error(trans->dev, phys_addr)) { in iwl_pcie_gen2_enqueue_hcmd() 227 if (dma_mapping_error(trans->dev, phys_addr)) { in iwl_pcie_gen2_enqueue_hcmd()
|
A D | tx.c | 1104 if (dma_mapping_error(trans->dev, phys_addr)) { in iwl_pcie_enqueue_hcmd() 1128 if (dma_mapping_error(trans->dev, phys_addr)) { in iwl_pcie_enqueue_hcmd() 1256 if (unlikely(dma_mapping_error(trans->dev, tb_phys))) in iwl_fill_data_tbs() 1275 if (unlikely(dma_mapping_error(trans->dev, tb_phys))) in iwl_fill_data_tbs() 1383 if (unlikely(dma_mapping_error(trans->dev, hdr_tb_phys))) in iwl_fill_data_tbs_amsdu() 1403 if (unlikely(dma_mapping_error(trans->dev, tb_phys))) in iwl_fill_data_tbs_amsdu() 1554 if (unlikely(dma_mapping_error(trans->dev, tb1_phys))) in iwl_trans_pcie_tx()
|
/linux/drivers/net/wireless/mediatek/mt76/ |
A D | dma.c | 330 if (unlikely(dma_mapping_error(dev->dev, addr))) in mt76_dma_tx_queue_skb_raw() 378 if (unlikely(dma_mapping_error(dev->dev, addr))) in mt76_dma_tx_queue_skb() 392 if (unlikely(dma_mapping_error(dev->dev, addr))) in mt76_dma_tx_queue_skb() 461 if (unlikely(dma_mapping_error(dev->dev, addr))) { in mt76_dma_rx_fill()
|
/linux/drivers/net/ethernet/google/gve/ |
A D | gve_tx.c | 519 if (unlikely(dma_mapping_error(tx->dev, addr))) { in gve_tx_add_skb_no_copy() 520 tx->dma_mapping_error++; in gve_tx_add_skb_no_copy() 552 if (unlikely(dma_mapping_error(tx->dev, addr))) { in gve_tx_add_skb_no_copy() 553 tx->dma_mapping_error++; in gve_tx_add_skb_no_copy()
|
/linux/drivers/misc/ |
A D | pci_endpoint_test.c | 389 if (dma_mapping_error(dev, orig_src_phys_addr)) { in pci_endpoint_test_copy() 421 if (dma_mapping_error(dev, orig_dst_phys_addr)) { in pci_endpoint_test_copy() 524 if (dma_mapping_error(dev, orig_phys_addr)) { in pci_endpoint_test_write() 620 if (dma_mapping_error(dev, orig_phys_addr)) { in pci_endpoint_test_read()
|
/linux/drivers/dma/ioat/ |
A D | init.c | 339 if (dma_mapping_error(dev, dma_src)) { in ioat_dma_self_test() 345 if (dma_mapping_error(dev, dma_dest)) { in ioat_dma_self_test() 852 if (dma_mapping_error(dev, dest_dma)) { in ioat_xor_val_self_test() 860 if (dma_mapping_error(dev, dma_srcs[i])) { in ioat_xor_val_self_test() 929 if (dma_mapping_error(dev, dma_srcs[i])) { in ioat_xor_val_self_test() 982 if (dma_mapping_error(dev, dma_srcs[i])) { in ioat_xor_val_self_test()
|