/linux/drivers/dma/idxd/ |
A D | init.c | 97 GFP_KERNEL, dev_to_node(dev)); in idxd_setup_interrupts() 212 GFP_KERNEL, dev_to_node(dev)); in idxd_setup_wqs() 217 wq = kzalloc_node(sizeof(*wq), GFP_KERNEL, dev_to_node(dev)); in idxd_setup_wqs() 242 wq->wqcfg = kzalloc_node(idxd->wqcfg_size, GFP_KERNEL, dev_to_node(dev)); in idxd_setup_wqs() 270 GFP_KERNEL, dev_to_node(dev)); in idxd_setup_engines() 275 engine = kzalloc_node(sizeof(*engine), GFP_KERNEL, dev_to_node(dev)); in idxd_setup_engines() 317 GFP_KERNEL, dev_to_node(dev)); in idxd_setup_groups() 322 group = kzalloc_node(sizeof(*group), GFP_KERNEL, dev_to_node(dev)); in idxd_setup_groups() 384 dev_to_node(dev)); in idxd_setup_internals() 503 idxd = kzalloc_node(sizeof(*idxd), GFP_KERNEL, dev_to_node(dev)); in idxd_alloc()
|
A D | dma.c | 176 idxd_dma = kzalloc_node(sizeof(*idxd_dma), GFP_KERNEL, dev_to_node(dev)); in idxd_register_dma_device() 227 idxd_chan = kzalloc_node(sizeof(*idxd_chan), GFP_KERNEL, dev_to_node(dev)); in idxd_register_dma_channel()
|
/linux/include/linux/ |
A D | async.h | 90 return async_schedule_node(func, dev, dev_to_node(dev)); in async_schedule_dev() 112 return async_schedule_node_domain(func, dev, dev_to_node(dev), domain); in async_schedule_dev_domain()
|
/linux/drivers/crypto/qat/qat_c62x/ |
A D | adf_drv.c | 91 if (num_possible_nodes() > 1 && dev_to_node(&pdev->dev) < 0) { in adf_probe() 100 dev_to_node(&pdev->dev)); in adf_probe() 119 dev_to_node(&pdev->dev)); in adf_probe()
|
/linux/drivers/crypto/qat/qat_dh895xcc/ |
A D | adf_drv.c | 91 if (num_possible_nodes() > 1 && dev_to_node(&pdev->dev) < 0) { in adf_probe() 100 dev_to_node(&pdev->dev)); in adf_probe() 119 dev_to_node(&pdev->dev)); in adf_probe()
|
/linux/drivers/crypto/qat/qat_c3xxx/ |
A D | adf_drv.c | 91 if (num_possible_nodes() > 1 && dev_to_node(&pdev->dev) < 0) { in adf_probe() 100 dev_to_node(&pdev->dev)); in adf_probe() 119 dev_to_node(&pdev->dev)); in adf_probe()
|
/linux/drivers/usb/host/ |
A D | xhci-mem.c | 50 dev_to_node(dev)); in xhci_segment_alloc() 626 dev_to_node(dev)); in xhci_alloc_stream_info() 636 dev_to_node(dev)); in xhci_alloc_stream_info() 843 dev_to_node(dev)); in xhci_alloc_tt_info() 1658 dev_to_node(dev)); in scratchpad_alloc() 1746 dev_to_node(dev)); in xhci_alloc_command() 2252 flags, dev_to_node(dev)); in xhci_create_rhub_port_array() 2287 flags, dev_to_node(dev)); in xhci_setup_port_arrays() 2298 dev_to_node(dev)); in xhci_setup_port_arrays() 2326 flags, dev_to_node(dev)); in xhci_setup_port_arrays() [all …]
|
/linux/drivers/crypto/virtio/ |
A D | virtio_crypto_mgr.c | 194 if ((node == dev_to_node(&tmp_dev->vdev->dev) || in virtcrypto_get_dev_node() 195 dev_to_node(&tmp_dev->vdev->dev) < 0) && in virtcrypto_get_dev_node()
|
A D | virtio_crypto_algs.c | 373 dev_to_node(&vcrypto->vdev->dev)); in __virtio_crypto_skcipher_do_req() 378 dev_to_node(&vcrypto->vdev->dev)); in __virtio_crypto_skcipher_do_req() 435 dev_to_node(&vcrypto->vdev->dev)); in __virtio_crypto_skcipher_do_req()
|
A D | virtio_crypto_core.c | 311 if (num_possible_nodes() > 1 && dev_to_node(&vdev->dev) < 0) { in virtcrypto_probe() 322 dev_to_node(&vdev->dev)); in virtcrypto_probe()
|
/linux/drivers/crypto/qat/qat_common/ |
A D | qat_crypto.c | 60 if ((node == dev_to_node(&GET_DEV(tmp_dev)) || in qat_crypto_get_instance_node() 61 dev_to_node(&GET_DEV(tmp_dev)) < 0) && in qat_crypto_get_instance_node() 246 dev_to_node(&GET_DEV(accel_dev))); in qat_crypto_create_instances()
|
A D | adf_transport.c | 396 dev_to_node(&GET_DEV(accel_dev))); in adf_init_bank() 419 dev_to_node(&GET_DEV(accel_dev))); in adf_init_bank() 473 dev_to_node(&GET_DEV(accel_dev))); in adf_init_etr_data() 480 dev_to_node(&GET_DEV(accel_dev))); in adf_init_etr_data()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
A D | ptp.c | 217 node = dev_to_node(mlx5_core_dma_dev(mdev)); in mlx5e_ptp_alloc_txqsq() 299 dev_to_node(mlx5_core_dma_dev(c->mdev))); in mlx5e_ptp_open_txqsq() 373 ccp.node = dev_to_node(mlx5_core_dma_dev(c->mdev)); in mlx5e_ptp_open_tx_cqs() 420 ccp.node = dev_to_node(mlx5_core_dma_dev(c->mdev)); in mlx5e_ptp_open_rx_cq() 522 int node = dev_to_node(c->mdev->device); in mlx5e_ptp_open_rq() 675 c = kvzalloc_node(sizeof(*c), GFP_KERNEL, dev_to_node(mlx5_core_dma_dev(mdev))); in mlx5e_ptp_open()
|
/linux/drivers/crypto/qat/qat_c62xvf/ |
A D | adf_drv.c | 93 dev_to_node(&pdev->dev)); in adf_probe() 113 dev_to_node(&pdev->dev)); in adf_probe()
|
/linux/drivers/crypto/qat/qat_c3xxxvf/ |
A D | adf_drv.c | 93 dev_to_node(&pdev->dev)); in adf_probe() 113 dev_to_node(&pdev->dev)); in adf_probe()
|
/linux/drivers/crypto/qat/qat_dh895xccvf/ |
A D | adf_drv.c | 93 dev_to_node(&pdev->dev)); in adf_probe() 113 dev_to_node(&pdev->dev)); in adf_probe()
|
/linux/drivers/base/test/ |
A D | test_async_driver_probe.c | 48 dev_to_node(dev) != numa_node_id()) { in test_probe() 50 dev_to_node(dev), numa_node_id()); in test_probe()
|
/linux/drivers/gpu/drm/ |
A D | drm_managed.c | 145 dev_to_node(dev->dev)); in __drmm_add_action() 192 dr = alloc_dr(NULL, size, gfp, dev_to_node(dev->dev)); in drmm_kmalloc()
|
/linux/arch/riscv/include/asm/ |
A D | pci.h | 42 return dev_to_node(&bus->dev); in pcibus_to_node()
|
/linux/drivers/nvdimm/ |
A D | of_pmem.c | 57 ndr_desc.numa_node = dev_to_node(&pdev->dev); in of_pmem_region_probe()
|
A D | virtio_pmem.c | 35 int nid = dev_to_node(&vdev->dev); in virtio_pmem_probe()
|
/linux/kernel/dma/ |
A D | ops_helpers.c | 69 page = alloc_pages_node(dev_to_node(dev), gfp, get_order(size)); in dma_common_alloc_pages()
|
/linux/kernel/ |
A D | iomem.c | 145 dev_to_node(dev)); in devm_memremap()
|
/linux/arch/powerpc/kernel/ |
A D | dma-iommu.c | 84 dev_to_node(dev)); in dma_iommu_alloc_coherent()
|
/linux/arch/arm64/kernel/ |
A D | pci.c | 61 return dev_to_node(&bus->dev); in pcibus_to_node()
|