Lines Matching refs:pdd
195 struct kfd_process_device *pdd; in svm_range_dma_map() local
199 pdd = kfd_process_device_from_gpuidx(p, gpuidx); in svm_range_dma_map()
200 if (!pdd) { in svm_range_dma_map()
204 adev = (struct amdgpu_device *)pdd->dev->kgd; in svm_range_dma_map()
235 struct kfd_process_device *pdd; in svm_range_free_dma_mappings() local
248 pdd = kfd_process_device_from_gpuidx(p, gpuidx); in svm_range_free_dma_mappings()
249 if (!pdd) { in svm_range_free_dma_mappings()
253 dev = &pdd->dev->pdev->dev; in svm_range_free_dma_mappings()
567 struct kfd_process_device *pdd; in svm_range_get_adev_by_id() local
578 pdd = kfd_process_device_from_gpuidx(p, gpu_idx); in svm_range_get_adev_by_id()
579 if (!pdd) { in svm_range_get_adev_by_id()
584 return (struct amdgpu_device *)pdd->dev->kgd; in svm_range_get_adev_by_id()
1130 struct kfd_process_device *pdd; in svm_range_unmap_from_gpus() local
1143 pdd = kfd_process_device_from_gpuidx(p, gpuidx); in svm_range_unmap_from_gpus()
1144 if (!pdd) { in svm_range_unmap_from_gpus()
1148 adev = (struct amdgpu_device *)pdd->dev->kgd; in svm_range_unmap_from_gpus()
1150 r = svm_range_unmap_from_gpu(adev, drm_priv_to_vm(pdd->drm_priv), in svm_range_unmap_from_gpus()
1258 struct kfd_process_device *pdd; in svm_range_map_to_gpus() local
1274 pdd = kfd_process_device_from_gpuidx(p, gpuidx); in svm_range_map_to_gpus()
1275 if (!pdd) { in svm_range_map_to_gpus()
1279 adev = (struct amdgpu_device *)pdd->dev->kgd; in svm_range_map_to_gpus()
1281 pdd = kfd_bind_process_to_device(pdd->dev, p); in svm_range_map_to_gpus()
1282 if (IS_ERR(pdd)) in svm_range_map_to_gpus()
1291 r = svm_range_map_to_gpu(adev, drm_priv_to_vm(pdd->drm_priv), in svm_range_map_to_gpus()
1324 struct kfd_process_device *pdd; in svm_range_reserve_bos() local
1332 pdd = kfd_process_device_from_gpuidx(ctx->process, gpuidx); in svm_range_reserve_bos()
1333 if (!pdd) { in svm_range_reserve_bos()
1337 adev = (struct amdgpu_device *)pdd->dev->kgd; in svm_range_reserve_bos()
1338 vm = drm_priv_to_vm(pdd->drm_priv); in svm_range_reserve_bos()
1353 pdd = kfd_process_device_from_gpuidx(ctx->process, gpuidx); in svm_range_reserve_bos()
1354 if (!pdd) { in svm_range_reserve_bos()
1359 adev = (struct amdgpu_device *)pdd->dev->kgd; in svm_range_reserve_bos()
1361 r = amdgpu_vm_validate_pt_bos(adev, drm_priv_to_vm(pdd->drm_priv), in svm_range_reserve_bos()
1383 struct kfd_process_device *pdd; in kfd_svm_page_owner() local
1386 pdd = kfd_process_device_from_gpuidx(p, gpuidx); in kfd_svm_page_owner()
1387 adev = (struct amdgpu_device *)pdd->dev->kgd; in kfd_svm_page_owner()
1964 struct kfd_process_device *pdd; in svm_range_drain_retry_fault() local
1978 pdd = p->pdds[i]; in svm_range_drain_retry_fault()
1979 if (!pdd) in svm_range_drain_retry_fault()
1983 adev = (struct amdgpu_device *)pdd->dev->kgd; in svm_range_drain_retry_fault()
2541 struct kfd_process_device *pdd; in svm_range_count_fault() local
2559 pdd = kfd_process_device_from_gpuidx(p, gpuidx); in svm_range_count_fault()
2560 if (pdd) in svm_range_count_fault()
2561 WRITE_ONCE(pdd->faults, pdd->faults + 1); in svm_range_count_fault()
2981 struct kfd_process_device *pdd; in svm_range_best_prefetch_location() local
3006 pdd = kfd_process_device_from_gpuidx(p, gpuidx); in svm_range_best_prefetch_location()
3007 if (!pdd) { in svm_range_best_prefetch_location()
3011 adev = (struct amdgpu_device *)pdd->dev->kgd; in svm_range_best_prefetch_location()