/linux/drivers/gpu/drm/amd/amdkfd/ |
A D | kfd_flat_memory.c | 322 pdd->lds_limit = MAKE_LDS_APP_LIMIT(pdd->lds_base); in kfd_init_apertures_vi() 330 pdd->gpuvm_limit = in kfd_init_apertures_vi() 337 pdd->gpuvm_limit = MAKE_GPUVM_APP_LIMIT(pdd->gpuvm_base, in kfd_init_apertures_vi() 342 pdd->scratch_limit = MAKE_SCRATCH_APP_LIMIT(pdd->scratch_base); in kfd_init_apertures_vi() 348 pdd->lds_limit = MAKE_LDS_APP_LIMIT(pdd->lds_base); in kfd_init_apertures_v9() 357 pdd->gpuvm_limit = in kfd_init_apertures_v9() 361 pdd->scratch_limit = MAKE_SCRATCH_APP_LIMIT(pdd->scratch_base); in kfd_init_apertures_v9() 383 if (!pdd) { in kfd_init_apertures() 393 pdd->lds_base = pdd->lds_limit = 0; in kfd_init_apertures() 394 pdd->gpuvm_base = pdd->gpuvm_limit = 0; in kfd_init_apertures() [all …]
|
A D | kfd_process.c | 115 pdd = workarea->pdd; in kfd_sdma_activity_worker() 116 if (!pdd) in kfd_sdma_activity_worker() 319 sdma_activity_work_handler.pdd = pdd; in kfd_procfs_show() 554 pdd->kobj_stats = kfd_alloc_struct(pdd->kobj_stats); in kfd_procfs_add_sysfs_stats() 969 if (!pdd) in kfd_process_kunmap_signal_bo() 1006 pdd->dev->kgd, pdd->drm_priv); in kfd_process_destroy_pdds() 1010 if (pdd->qpd.cwsr_kaddr && !pdd->qpd.cwsr_base) in kfd_process_destroy_pdds() 1479 pdd = kzalloc(sizeof(*pdd), GFP_KERNEL); in kfd_create_process_device_data() 1480 if (!pdd) in kfd_create_process_device_data() 1512 return pdd; in kfd_create_process_device_data() [all …]
|
A D | kfd_process_queue_manager.c | 69 struct kfd_dev *dev = pdd->dev; in kfd_process_dequeue_from_device() 71 if (pdd->already_dequeued) in kfd_process_dequeue_from_device() 75 pdd->already_dequeued = true; in kfd_process_dequeue_from_device() 83 struct kfd_process_device *pdd; in pqm_set_gws() local 99 if (!pdd) { in pqm_set_gws() 105 if (gws && pdd->qpd.num_gws) in pqm_set_gws() 200 struct kfd_process_device *pdd; in pqm_create_queue() local 211 if (!pdd) { in pqm_create_queue() 289 kq, &pdd->qpd); in pqm_create_queue() 367 if (!pdd) { in pqm_destroy_queue() [all …]
|
A D | kfd_chardev.c | 304 if (IS_ERR(pdd)) { in kfd_ioctl_create_queue() 515 if (IS_ERR(pdd)) { in kfd_ioctl_set_memory_policy() 528 &pdd->qpd, in kfd_ioctl_set_memory_policy() 556 if (IS_ERR(pdd)) { in kfd_ioctl_set_trap_handler() 596 if (IS_ERR(pdd)) { in kfd_ioctl_dbg_register() 1128 if (IS_ERR(pdd)) { in kfd_ioctl_set_scratch_backing_va() 1211 if (!pdd) { in kfd_ioctl_acquire_vm() 1339 WRITE_ONCE(pdd->vram_usage, pdd->vram_usage + args->size); in kfd_ioctl_alloc_memory_of_gpu() 1389 if (!pdd) { in kfd_ioctl_free_memory_of_gpu() 1412 WRITE_ONCE(pdd->vram_usage, pdd->vram_usage - size); in kfd_ioctl_free_memory_of_gpu() [all …]
|
A D | kfd_iommu.c | 112 struct kfd_dev *dev = pdd->dev; in kfd_iommu_bind_process_to_device() 126 pdd->bound = PDD_BOUND; in kfd_iommu_bind_process_to_device() 150 struct kfd_process_device *pdd; in iommu_pasid_shutdown_callback() local 180 if (pdd) in iommu_pasid_shutdown_callback() 220 struct kfd_process_device *pdd; in kfd_bind_processes_to_device() local 231 if (WARN_ON(!pdd) || pdd->bound != PDD_BOUND_SUSPENDED) { in kfd_bind_processes_to_device() 245 pdd->bound = PDD_BOUND; in kfd_bind_processes_to_device() 261 struct kfd_process_device *pdd; in kfd_unbind_processes_from_device() local 271 if (WARN_ON(!pdd)) { in kfd_unbind_processes_from_device() 276 if (pdd->bound == PDD_BOUND) in kfd_unbind_processes_from_device() [all …]
|
A D | kfd_device_queue_manager_v9.c | 43 static uint32_t compute_sh_mem_bases_64bit(struct kfd_process_device *pdd) in compute_sh_mem_bases_64bit() argument 45 uint32_t shared_base = pdd->lds_base >> 48; in compute_sh_mem_bases_64bit() 46 uint32_t private_base = pdd->scratch_base >> 48; in compute_sh_mem_bases_64bit() 55 struct kfd_process_device *pdd; in update_qpd_v9() local 57 pdd = qpd_to_pdd(qpd); in update_qpd_v9() 69 if (!pdd->process->xnack_enabled) in update_qpd_v9() 82 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(pdd); in update_qpd_v9()
|
A D | kfd_doorbell.c | 130 struct kfd_process_device *pdd; in kfd_doorbell_mmap() local 139 pdd = kfd_get_process_device_data(dev, process); in kfd_doorbell_mmap() 140 if (!pdd) in kfd_doorbell_mmap() 144 address = kfd_get_process_doorbells(pdd); in kfd_doorbell_mmap() 230 struct kfd_process_device *pdd, in kfd_get_doorbell_dw_offset_in_bar() argument 240 pdd->doorbell_index in kfd_get_doorbell_dw_offset_in_bar() 255 phys_addr_t kfd_get_process_doorbells(struct kfd_process_device *pdd) in kfd_get_process_doorbells() argument 257 return pdd->dev->doorbell_base + in kfd_get_process_doorbells() 258 pdd->doorbell_index * kfd_doorbell_process_slice(pdd->dev); in kfd_get_process_doorbells()
|
A D | kfd_device_queue_manager_v10.c | 42 static uint32_t compute_sh_mem_bases_64bit(struct kfd_process_device *pdd) in compute_sh_mem_bases_64bit() argument 44 uint32_t shared_base = pdd->lds_base >> 48; in compute_sh_mem_bases_64bit() 45 uint32_t private_base = pdd->scratch_base >> 48; in compute_sh_mem_bases_64bit() 54 struct kfd_process_device *pdd; in update_qpd_v10() local 56 pdd = qpd_to_pdd(qpd); in update_qpd_v10() 77 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(pdd); in update_qpd_v10()
|
A D | kfd_device_queue_manager_cik.c | 119 struct kfd_process_device *pdd; in update_qpd_cik() local 122 pdd = qpd_to_pdd(qpd); in update_qpd_cik() 135 temp = get_sh_mem_bases_32(pdd); in update_qpd_cik() 139 temp = get_sh_mem_bases_nybble_64(pdd); in update_qpd_cik() 153 struct kfd_process_device *pdd; in update_qpd_cik_hawaii() local 156 pdd = qpd_to_pdd(qpd); in update_qpd_cik_hawaii() 171 temp = get_sh_mem_bases_nybble_64(pdd); in update_qpd_cik_hawaii()
|
A D | kfd_device_queue_manager_vi.c | 156 struct kfd_process_device *pdd; in update_qpd_vi() local 159 pdd = qpd_to_pdd(qpd); in update_qpd_vi() 175 temp = get_sh_mem_bases_32(pdd); in update_qpd_vi() 180 temp = get_sh_mem_bases_nybble_64(pdd); in update_qpd_vi() 197 struct kfd_process_device *pdd; in update_qpd_vi_tonga() local 200 pdd = qpd_to_pdd(qpd); in update_qpd_vi_tonga() 219 temp = get_sh_mem_bases_nybble_64(pdd); in update_qpd_vi_tonga()
|
A D | kfd_device_queue_manager.c | 570 if (!pdd) { in update_queue() 668 pdd = qpd_to_pdd(qpd); in evict_process_queues_nocpsch() 670 pdd->process->pasid); in evict_process_queues_nocpsch() 721 pdd = qpd_to_pdd(qpd); in evict_process_queues_cpsch() 723 pdd->process->pasid); in evict_process_queues_cpsch() 758 pdd = qpd_to_pdd(qpd); in restore_process_queues_nocpsch() 771 pdd->process->pasid); in restore_process_queues_nocpsch() 841 pdd = qpd_to_pdd(qpd); in restore_process_queues_cpsch() 854 pdd->process->pasid); in restore_process_queues_cpsch() 893 pdd = qpd_to_pdd(qpd); in register_process() [all …]
|
A D | kfd_dbgdev.c | 275 struct kfd_process_device *pdd; in dbgdev_address_watch_nodiq() local 279 pdd = kfd_get_process_device_data(dbgdev->dev, in dbgdev_address_watch_nodiq() 281 if (!pdd) { in dbgdev_address_watch_nodiq() 303 &cntl, i, pdd->qpd.vmid); in dbgdev_address_watch_nodiq() 324 pdd->dev->kfd2kgd->address_watch_execute( in dbgdev_address_watch_nodiq() 703 struct kfd_process_device *pdd; in dbgdev_wave_control_nodiq() local 710 if (!pdd) { in dbgdev_wave_control_nodiq() 723 reg_sq_cmd.bits.vm_id = pdd->qpd.vmid; in dbgdev_wave_control_nodiq() 767 struct kfd_process_device *pdd; in dbgdev_wave_reset_wavefronts() local 802 pdd = kfd_get_process_device_data(dev, p); in dbgdev_wave_reset_wavefronts() [all …]
|
A D | kfd_svm.c | 200 if (!pdd) { in svm_range_dma_map() 249 if (!pdd) { in svm_range_free_dma_mappings() 579 if (!pdd) { in svm_range_get_adev_by_id() 1144 if (!pdd) { in svm_range_unmap_from_gpus() 1275 if (!pdd) { in svm_range_map_to_gpus() 1281 pdd = kfd_bind_process_to_device(pdd->dev, p); in svm_range_map_to_gpus() 1333 if (!pdd) { in svm_range_reserve_bos() 1354 if (!pdd) { in svm_range_reserve_bos() 1979 if (!pdd) in svm_range_drain_retry_fault() 2560 if (pdd) in svm_range_count_fault() [all …]
|
A D | kfd_device_queue_manager.h | 223 static inline unsigned int get_sh_mem_bases_32(struct kfd_process_device *pdd) in get_sh_mem_bases_32() argument 225 return (pdd->lds_base >> 16) & 0xFF; in get_sh_mem_bases_32() 229 get_sh_mem_bases_nybble_64(struct kfd_process_device *pdd) in get_sh_mem_bases_nybble_64() argument 231 return (pdd->lds_base >> 60) & 0x0E; in get_sh_mem_bases_nybble_64()
|
A D | kfd_iommu.h | 35 int kfd_iommu_bind_process_to_device(struct kfd_process_device *pdd); 58 struct kfd_process_device *pdd) in kfd_iommu_bind_process_to_device() argument
|
A D | kfd_migrate.c | 410 struct kfd_process_device *pdd; in svm_migrate_vma_to_vram() local 471 pdd = svm_range_get_pdd_by_adev(prange, adev); in svm_migrate_vma_to_vram() 472 if (pdd) in svm_migrate_vma_to_vram() 473 WRITE_ONCE(pdd->page_in, pdd->page_in + cpages); in svm_migrate_vma_to_vram() 650 struct kfd_process_device *pdd; in svm_migrate_vma_to_ram() local 711 pdd = svm_range_get_pdd_by_adev(prange, adev); in svm_migrate_vma_to_ram() 712 if (pdd) in svm_migrate_vma_to_ram() 713 WRITE_ONCE(pdd->page_out, pdd->page_out + cpages); in svm_migrate_vma_to_ram()
|
A D | kfd_priv.h | 912 int kfd_process_device_init_vm(struct kfd_process_device *pdd, 927 int kfd_process_device_create_obj_handle(struct kfd_process_device *pdd, 931 void kfd_process_device_remove_obj_handle(struct kfd_process_device *pdd, 955 struct kfd_process_device *pdd, 957 phys_addr_t kfd_get_process_doorbells(struct kfd_process_device *pdd); 1039 void kfd_process_dequeue_from_device(struct kfd_process_device *pdd); 1176 void kfd_flush_tlb(struct kfd_process_device *pdd, enum TLB_FLUSH_TYPE type);
|
/linux/drivers/base/power/ |
A D | domain_governor.c | 124 struct pm_domain_data *pdd; in update_domain_next_wakeup() local 138 list_for_each_entry(pdd, &genpd->dev_list, list_node) { in update_domain_next_wakeup() 139 next_wakeup = to_gpd_data(pdd)->next_wakeup; in update_domain_next_wakeup() 174 struct pm_domain_data *pdd; in __default_power_down_ok() local 209 list_for_each_entry(pdd, &genpd->dev_list, list_node) { in __default_power_down_ok() 218 td = &to_gpd_data(pdd)->td; in __default_power_down_ok()
|
A D | domain.c | 259 struct pm_domain_data *pdd; in _genpd_reeval_performance_state() local 272 pd_data = to_gpd_data(pdd); in _genpd_reeval_performance_state() 616 struct pm_domain_data *pdd; in genpd_power_off() local 650 if (!pm_runtime_suspended(pdd->dev) || in genpd_power_off() 767 struct pm_domain_data *pdd; in genpd_dev_pm_qos_notifier() local 771 pdd = dev->power.subsys_data ? in genpd_dev_pm_qos_notifier() 773 if (pdd) { in genpd_dev_pm_qos_notifier() 1651 struct pm_domain_data *pdd; in genpd_remove_device() local 1656 pdd = dev->power.subsys_data->domain_data; in genpd_remove_device() 1657 gpd_data = to_gpd_data(pdd); in genpd_remove_device() [all …]
|
/linux/drivers/input/touchscreen/ |
A D | wm9705.c | 77 static int pdd = 8; variable 78 module_param(pdd, int, 0); 79 MODULE_PARM_DESC(pdd, "Set pen detect comparator threshold"); 168 dig2 |= (pdd & 0x000f); in wm9705_phy_init() 169 dev_dbg(wm->dev, "setting pdd to Vmid/%d", 1 - (pdd & 0x000f)); in wm9705_phy_init()
|
/linux/drivers/soc/xilinx/ |
A D | zynqmp_pm_domains.c | 104 struct pm_domain_data *pdd, *tmp; in zynqmp_gpd_power_off() local 118 list_for_each_entry_safe(pdd, tmp, &domain->dev_list, list_node) { in zynqmp_gpd_power_off() 120 may_wakeup = zynqmp_gpd_is_active_wakeup_path(pdd->dev, NULL); in zynqmp_gpd_power_off() 122 dev_dbg(pdd->dev, "device is in wakeup path in %s\n", in zynqmp_gpd_power_off()
|
/linux/include/linux/ |
A D | pm_domain.h | 208 static inline struct generic_pm_domain_data *to_gpd_data(struct pm_domain_data *pdd) in to_gpd_data() argument 210 return container_of(pdd, struct generic_pm_domain_data, base); in to_gpd_data()
|
/linux/Documentation/admin-guide/ |
A D | devices.txt | 883 48 = /dev/pdd Fourth parallel port IDE disk
|