/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_device_queue_manager_v11.c | 30 struct qcm_process_device *qpd); 32 struct qcm_process_device *qpd); 52 struct qcm_process_device *qpd) 56 pdd = qpd_to_pdd(qpd); 59 if (qpd->sh_mem_config == 0) { 60 qpd->sh_mem_config = 65 qpd->sh_mem_ape1_limit = 0; 66 qpd->sh_mem_ape1_base = 0; 69 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(pdd); 71 pr_debug("sh_mem_bases 0x%X\n", qpd 51 update_qpd_v11(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 76 init_sdma_vm_v11(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd) argument [all...] |
H A D | kfd_device_queue_manager_v10.c | 31 struct qcm_process_device *qpd); 33 struct qcm_process_device *qpd); 53 struct qcm_process_device *qpd) 57 pdd = qpd_to_pdd(qpd); 60 if (qpd->sh_mem_config == 0) { 61 qpd->sh_mem_config = 65 qpd->sh_mem_ape1_limit = 0; 66 qpd->sh_mem_ape1_base = 0; 69 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(pdd); 71 pr_debug("sh_mem_bases 0x%X\n", qpd 52 update_qpd_v10(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 76 init_sdma_vm_v10(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd) argument [all...] |
H A D | kfd_device_queue_manager_v9.c | 30 struct qcm_process_device *qpd); 32 struct qcm_process_device *qpd); 52 struct qcm_process_device *qpd) 56 pdd = qpd_to_pdd(qpd); 59 if (qpd->sh_mem_config == 0) { 60 qpd->sh_mem_config = SH_MEM_ALIGNMENT_MODE_UNALIGNED << 64 qpd->sh_mem_config |= 1 << SH_MEM_CONFIG__RETRY_DISABLE__SHIFT; 67 qpd->sh_mem_config |= 70 qpd->sh_mem_ape1_limit = 0; 71 qpd 51 update_qpd_v9(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 89 init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd) argument [all...] |
H A D | kfd_device_queue_manager_cik.c | 31 struct qcm_process_device *qpd, 37 struct qcm_process_device *qpd); 40 struct qcm_process_device *qpd); 79 struct qcm_process_device *qpd, 96 qpd->sh_mem_config = (qpd->sh_mem_config & PTR32) 105 struct qcm_process_device *qpd) 110 pdd = qpd_to_pdd(qpd); 113 if (qpd->sh_mem_config == 0) { 114 qpd 78 set_cache_memory_policy_cik(struct device_queue_manager *dqm, struct qcm_process_device *qpd, enum cache_policy default_policy, enum cache_policy alternate_policy, void __user *alternate_aperture_base, uint64_t alternate_aperture_size) argument 104 update_qpd_cik(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 134 init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd) argument [all...] |
H A D | kfd_device_queue_manager_vi.c | 31 struct qcm_process_device *qpd, 37 struct qcm_process_device *qpd); 40 struct qcm_process_device *qpd); 80 struct qcm_process_device *qpd, 97 qpd->sh_mem_config = 107 struct qcm_process_device *qpd) 112 pdd = qpd_to_pdd(qpd); 115 if (qpd->sh_mem_config == 0) { 116 qpd->sh_mem_config = 124 qpd 79 set_cache_memory_policy_vi(struct device_queue_manager *dqm, struct qcm_process_device *qpd, enum cache_policy default_policy, enum cache_policy alternate_policy, void __user *alternate_aperture_base, uint64_t alternate_aperture_size) argument 106 update_qpd_vi(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 140 init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd) argument [all...] |
H A D | kfd_packet_manager_v9.c | 33 uint32_t *buffer, struct qcm_process_device *qpd) 36 uint64_t vm_page_table_base_addr = qpd->page_table_base; 39 container_of(qpd, struct kfd_process_device, qpd); 45 packet->bitfields2.diq_enable = (qpd->is_debug) ? 1 : 0; 47 packet->bitfields2.pasid = qpd->pqm->process->pasid; 48 packet->bitfields14.gds_size = qpd->gds_size & 0x3F; 49 packet->bitfields14.gds_size_hi = (qpd->gds_size >> 6) & 0xF; 50 packet->bitfields14.num_gws = (qpd->mapped_gws_queue) ? qpd 32 pm_map_process_v9(struct packet_manager *pm, uint32_t *buffer, struct qcm_process_device *qpd) argument 86 pm_map_process_aldebaran(struct packet_manager *pm, uint32_t *buffer, struct qcm_process_device *qpd) argument [all...] |
H A D | kfd_doorbell.c | 208 static int init_doorbell_bitmap(struct qcm_process_device *qpd, argument 226 __set_bit(i, qpd->doorbell_bitmap); 228 qpd->doorbell_bitmap); 240 if (!pdd->qpd.proc_doorbells) { 247 pdd->qpd.proc_doorbells, 256 struct qcm_process_device *qpd = &pdd->qpd; local 259 qpd->doorbell_bitmap = bitmap_zalloc(KFD_MAX_NUM_OF_QUEUES_PER_PROCESS, 261 if (!qpd->doorbell_bitmap) { 266 r = init_doorbell_bitmap(&pdd->qpd, kf 295 struct qcm_process_device *qpd = &pdd->qpd; local [all...] |
H A D | kfd_device_queue_manager.h | 43 struct qcm_process_device *qpd; member in struct:device_process_node 134 struct qcm_process_device *qpd, 140 struct qcm_process_device *qpd, 147 struct qcm_process_device *qpd); 150 struct qcm_process_device *qpd); 159 struct qcm_process_device *qpd); 163 struct qcm_process_device *qpd); 166 struct qcm_process_device *qpd, 173 struct qcm_process_device *qpd); 176 struct qcm_process_device *qpd); [all...] |
H A D | kfd_device_queue_manager.c | 141 struct qcm_process_device *qpd) 148 dqm->dev->adev, qpd->vmid, qpd->sh_mem_config, 149 qpd->sh_mem_ape1_base, qpd->sh_mem_ape1_limit, 150 qpd->sh_mem_bases, xcc_id); 189 struct qcm_process_device *qpd) 192 struct kfd_process_device *pdd = qpd_to_pdd(qpd); 201 queue_input.process_id = qpd->pqm->process->pasid; 202 queue_input.page_table_base_addr = qpd 140 program_sh_mem_settings(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 188 add_queue_mes(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd) argument 258 remove_queue_mes(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd) argument 290 struct qcm_process_device *qpd; local 313 increment_queue_count(struct device_queue_manager *dqm, struct qcm_process_device *qpd, struct queue *q) argument 328 decrement_queue_count(struct device_queue_manager *dqm, struct qcm_process_device *qpd, struct queue *q) argument 347 allocate_doorbell(struct qcm_process_device *qpd, struct queue *q, uint32_t const *restore_id) argument 418 deallocate_doorbell(struct qcm_process_device *qpd, struct queue *q) argument 433 program_trap_handler_settings(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 446 allocate_vmid(struct device_queue_manager *dqm, struct qcm_process_device *qpd, struct queue *q) argument 496 flush_texture_cache_nocpsch(struct kfd_node *kdev, struct qcm_process_device *qpd) argument 514 deallocate_vmid(struct device_queue_manager *dqm, struct qcm_process_device *qpd, struct queue *q) argument 535 create_queue_nocpsch(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd, const struct kfd_criu_queue_priv_data *qd, const void *restore_mqd, const void *restore_ctl_stack) argument 766 destroy_queue_nocpsch_locked(struct device_queue_manager *dqm, struct qcm_process_device *qpd, struct queue *q) argument 825 destroy_queue_nocpsch(struct device_queue_manager *dqm, struct qcm_process_device *qpd, struct queue *q) argument 1014 resume_single_queue(struct device_queue_manager *dqm, struct qcm_process_device *qpd, struct queue *q) argument 1046 evict_process_queues_nocpsch(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 1096 evict_process_queues_cpsch(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 1153 restore_process_queues_nocpsch(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 1234 restore_process_queues_cpsch(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 1297 register_process(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 1336 unregister_process(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 1740 create_kernel_queue_cpsch(struct device_queue_manager *dqm, struct kernel_queue *kq, struct qcm_process_device *qpd) argument 1770 destroy_kernel_queue_cpsch(struct device_queue_manager *dqm, struct kernel_queue *kq, struct qcm_process_device *qpd) argument 1790 create_queue_cpsch(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd, const struct kfd_criu_queue_priv_data *qd, const void *restore_mqd, const void *restore_ctl_stack) argument 2078 destroy_queue_cpsch(struct device_queue_manager *dqm, struct qcm_process_device *qpd, struct queue *q) argument 2180 set_cache_memory_policy(struct device_queue_manager *dqm, struct qcm_process_device *qpd, enum cache_policy default_policy, enum cache_policy alternate_policy, void __user *alternate_aperture_base, uint64_t alternate_aperture_size) argument 2242 process_termination_nocpsch(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 2370 process_termination_cpsch(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 2673 reserve_debug_trap_vmid(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 2721 release_debug_trap_vmid(struct device_queue_manager *dqm, struct qcm_process_device *qpd) argument 2820 struct qcm_process_device *qpd = &pdd->qpd; local 2875 struct qcm_process_device *qpd = &pdd->qpd; local 2978 struct qcm_process_device *qpd = &pdd->qpd; local [all...] |
H A D | kfd_packet_manager_vi.c | 43 struct qcm_process_device *qpd) 53 packet->bitfields2.diq_enable = (qpd->is_debug) ? 1 : 0; 55 packet->bitfields2.pasid = qpd->pqm->process->pasid; 56 packet->bitfields3.page_table_base = qpd->page_table_base; 57 packet->bitfields10.gds_size = qpd->gds_size; 58 packet->bitfields10.num_gws = qpd->num_gws; 59 packet->bitfields10.num_oac = qpd->num_oac; 60 packet->bitfields10.num_queues = (qpd->is_debug) ? 0 : qpd->queue_count; 62 packet->sh_mem_config = qpd 42 pm_map_process_vi(struct packet_manager *pm, uint32_t *buffer, struct qcm_process_device *qpd) argument [all...] |
H A D | kfd_process.c | 106 struct qcm_process_device *qpd; local 119 qpd = &pdd->qpd; 120 if (!dqm || !qpd) 132 * 1. Create a temporary list of SDMA queue nodes from the qpd->queues_list, 138 * from the qpd->queues_list. 139 * 3. Do a second pass over qpd->queues_list to check if any nodes got deleted. 151 list_for_each_entry(q, &qpd->queues_list, list) { 170 * qpd->queues_list. Return the past activity count as the total sdma 213 list_for_each_entry(q, &qpd 766 struct qcm_process_device *qpd = &pdd->qpd; local 792 struct qcm_process_device *qpd = &pdd->qpd; local 1293 struct qcm_process_device *qpd = &p->pdds[i]->qpd; local 1329 struct qcm_process_device *qpd = &pdd->qpd; local 1365 struct qcm_process_device *qpd = &pdd->qpd; local 1373 kfd_process_set_trap_handler(struct qcm_process_device *qpd, uint64_t tba_addr, uint64_t tma_addr) argument 1446 kfd_process_set_trap_debug_flag(struct qcm_process_device *qpd, bool enabled) argument 2057 struct qcm_process_device *qpd; local [all...] |
H A D | kfd_process_queue_manager.c | 89 dev->dqm->ops.process_termination(dev->dqm, &pdd->qpd); 123 if (gws && pdd->qpd.num_gws) 126 if (!gws && pdd->qpd.num_gws == 0) 149 pdd->qpd.num_gws = gws ? dev->adev->gds.gws_size : 0; 195 pdd->qpd.num_gws = 0; 311 if ((pdd->qpd.is_debug) || (type == KFD_QUEUE_TYPE_DIQ)) 314 if (pdd->qpd.queue_count >= max_queues) 326 if (list_empty(&pdd->qpd.queues_list) && 327 list_empty(&pdd->qpd.priv_queue_list)) 328 dev->dqm->ops.register_process(dev->dqm, &pdd->qpd); [all...] |
H A D | kfd_packet_manager.c | 130 struct qcm_process_device *qpd; local 150 qpd = cur->qpd; 158 retval = pm->pmf->map_process(pm, &rl_buffer[rl_wptr], qpd); 166 list_for_each_entry(kq, &qpd->priv_queue_list, list) { 171 kq->queue->queue, qpd->is_debug); 176 qpd->is_debug); 185 list_for_each_entry(q, &qpd->queues_list, list) { 190 q->queue, qpd->is_debug); 195 qpd [all...] |
H A D | kfd_flat_memory.c | 337 pdd->qpd.cwsr_base = SVM_CWSR_BASE; 338 pdd->qpd.ib_base = SVM_IB_BASE; 360 pdd->qpd.cwsr_base = AMDGPU_VA_RESERVED_TRAP_START(pdd->dev->adev);
|
H A D | kfd_debug.c | 584 kfd_process_set_trap_debug_flag(&pdd->qpd, false); 597 release_debug_trap_vmid(pdd->dev->dqm, &pdd->qpd)) 675 r = reserve_debug_trap_vmid(pdd->dev->dqm, &pdd->qpd); 714 kfd_process_set_trap_debug_flag(&pdd->qpd, true); 755 if (pdd->qpd.num_gws && (!kfd_dbg_has_gws_support(pdd->dev) || 921 struct qcm_process_device *qpd = &pdd->qpd; local 923 list_for_each_entry(queue, &qpd->queues_list, list) {
|
H A D | kfd_priv.h | 739 struct qcm_process_device qpd; member in struct:kfd_process_device 841 #define qpd_to_pdd(x) container_of(x, struct kfd_process_device, qpd) 1161 void kfd_process_set_trap_handler(struct qcm_process_device *qpd, 1164 void kfd_process_set_trap_debug_flag(struct qcm_process_device *qpd, 1374 struct qcm_process_device *qpd);
|
H A D | kfd_chardev.c | 337 if (!pdd->qpd.proc_doorbells) { 612 &pdd->qpd, 647 kfd_process_set_trap_handler(&pdd->qpd, args->tba_addr, args->tma_addr); 921 pdd->qpd.sh_hidden_private_base = args->va_addr; 926 pdd->qpd.vmid != 0 && dev->kfd2kgd->set_scratch_backing_va) 928 dev->adev, args->va_addr, pdd->qpd.vmid); 2267 if (!pdd->qpd.proc_doorbells) { 2758 if (pdd->qpd.queue_count)
|
/linux-master/drivers/infiniband/hw/cxgb4/ |
H A D | device.c | 246 struct c4iw_debugfs_data *qpd) 253 space = qpd->bufsize - qpd->pos - 1; 267 cc = snprintf(qpd->buf + qpd->pos, space, 288 cc = snprintf(qpd->buf + qpd->pos, space, 304 cc = snprintf(qpd->buf + qpd->pos, space, 310 qpd 245 dump_qp(unsigned long id, struct c4iw_qp *qp, struct c4iw_debugfs_data *qpd) argument 316 struct c4iw_debugfs_data *qpd = file->private_data; local 329 struct c4iw_debugfs_data *qpd; local [all...] |