/linux-4.4.14/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_device_queue_manager_vi.c | 31 struct qcm_process_device *qpd, 37 struct qcm_process_device *qpd); 40 struct qcm_process_device *qpd); 79 struct qcm_process_device *qpd, set_cache_memory_policy_vi() 96 qpd->sh_mem_config = (qpd->sh_mem_config & set_cache_memory_policy_vi() 108 struct qcm_process_device *qpd) register_process_vi() 113 BUG_ON(!dqm || !qpd); register_process_vi() 115 pdd = qpd_to_pdd(qpd); register_process_vi() 118 if (qpd->sh_mem_config == 0) { register_process_vi() 119 qpd->sh_mem_config = register_process_vi() 126 qpd->sh_mem_ape1_limit = 0; register_process_vi() 127 qpd->sh_mem_ape1_base = 0; register_process_vi() 130 if (qpd->pqm->process->is_32bit_user_mode) { register_process_vi() 132 qpd->sh_mem_bases = temp << SH_MEM_BASES__SHARED_BASE__SHIFT; register_process_vi() 133 qpd->sh_mem_config |= SH_MEM_ADDRESS_MODE_HSA32 << register_process_vi() 137 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(temp); register_process_vi() 138 qpd->sh_mem_config |= SH_MEM_ADDRESS_MODE_HSA64 << register_process_vi() 143 qpd->pqm->process->is_32bit_user_mode, temp, qpd->sh_mem_bases); register_process_vi() 149 struct qcm_process_device *qpd) init_sdma_vm() 155 get_sh_mem_bases_32(qpd_to_pdd(qpd)); init_sdma_vm() 157 value |= ((get_sh_mem_bases_nybble_64(qpd_to_pdd(qpd))) << init_sdma_vm() 78 set_cache_memory_policy_vi(struct device_queue_manager *dqm, struct qcm_process_device *qpd, enum cache_policy default_policy, enum cache_policy alternate_policy, void __user *alternate_aperture_base, uint64_t alternate_aperture_size) set_cache_memory_policy_vi() argument 107 register_process_vi(struct device_queue_manager *dqm, struct qcm_process_device *qpd) register_process_vi() argument 148 init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd) init_sdma_vm() argument
|
H A D | kfd_device_queue_manager_cik.c | 29 struct qcm_process_device *qpd, 35 struct qcm_process_device *qpd); 38 struct qcm_process_device *qpd); 76 struct qcm_process_device *qpd, set_cache_memory_policy_cik() 93 qpd->sh_mem_config = (qpd->sh_mem_config & PTR32) set_cache_memory_policy_cik() 102 struct qcm_process_device *qpd) register_process_cik() 107 BUG_ON(!dqm || !qpd); register_process_cik() 109 pdd = qpd_to_pdd(qpd); register_process_cik() 112 if (qpd->sh_mem_config == 0) { register_process_cik() 113 qpd->sh_mem_config = register_process_cik() 117 qpd->sh_mem_ape1_limit = 0; register_process_cik() 118 qpd->sh_mem_ape1_base = 0; register_process_cik() 121 if (qpd->pqm->process->is_32bit_user_mode) { register_process_cik() 123 qpd->sh_mem_bases = SHARED_BASE(temp); register_process_cik() 124 qpd->sh_mem_config |= PTR32; register_process_cik() 127 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(temp); register_process_cik() 131 qpd->pqm->process->is_32bit_user_mode, temp, qpd->sh_mem_bases); register_process_cik() 137 struct qcm_process_device *qpd) init_sdma_vm() 143 get_sh_mem_bases_32(qpd_to_pdd(qpd)); init_sdma_vm() 145 value |= ((get_sh_mem_bases_nybble_64(qpd_to_pdd(qpd))) << init_sdma_vm() 75 set_cache_memory_policy_cik(struct device_queue_manager *dqm, struct qcm_process_device *qpd, enum cache_policy default_policy, enum cache_policy alternate_policy, void __user *alternate_aperture_base, uint64_t alternate_aperture_size) set_cache_memory_policy_cik() argument 101 register_process_cik(struct device_queue_manager *dqm, struct qcm_process_device *qpd) register_process_cik() argument 136 init_sdma_vm(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd) init_sdma_vm() argument
|
H A D | kfd_device_queue_manager.c | 45 struct qcm_process_device *qpd); 53 struct qcm_process_device *qpd); 84 struct qcm_process_device *qpd) program_sh_mem_settings() 87 dqm->dev->kgd, qpd->vmid, program_sh_mem_settings() 88 qpd->sh_mem_config, program_sh_mem_settings() 89 qpd->sh_mem_ape1_base, program_sh_mem_settings() 90 qpd->sh_mem_ape1_limit, program_sh_mem_settings() 91 qpd->sh_mem_bases); program_sh_mem_settings() 95 struct qcm_process_device *qpd, allocate_vmid() 109 qpd->vmid = allocated_vmid; allocate_vmid() 113 program_sh_mem_settings(dqm, qpd); allocate_vmid() 119 struct qcm_process_device *qpd, deallocate_vmid() 122 int bit = qpd->vmid - KFD_VMID_START_OFFSET; deallocate_vmid() 125 set_pasid_vmid_mapping(dqm, 0, qpd->vmid); deallocate_vmid() 128 qpd->vmid = 0; deallocate_vmid() 134 struct qcm_process_device *qpd, create_queue_nocpsch() 139 BUG_ON(!dqm || !q || !qpd || !allocated_vmid); create_queue_nocpsch() 153 if (list_empty(&qpd->queues_list)) { create_queue_nocpsch() 154 retval = allocate_vmid(dqm, qpd, q); create_queue_nocpsch() 160 *allocated_vmid = qpd->vmid; create_queue_nocpsch() 161 q->properties.vmid = qpd->vmid; create_queue_nocpsch() 164 retval = create_compute_queue_nocpsch(dqm, q, qpd); create_queue_nocpsch() 166 retval = create_sdma_queue_nocpsch(dqm, q, qpd); create_queue_nocpsch() 169 if (list_empty(&qpd->queues_list)) { create_queue_nocpsch() 170 deallocate_vmid(dqm, qpd, q); create_queue_nocpsch() 177 list_add(&q->list, &qpd->queues_list); create_queue_nocpsch() 238 struct qcm_process_device *qpd) create_compute_queue_nocpsch() 243 BUG_ON(!dqm || !q || !qpd); create_compute_queue_nocpsch() 276 struct qcm_process_device *qpd, destroy_queue_nocpsch() 282 BUG_ON(!dqm || !q || !q->mqd || !qpd); destroy_queue_nocpsch() 323 if (list_empty(&qpd->queues_list)) destroy_queue_nocpsch() 324 deallocate_vmid(dqm, qpd, q); destroy_queue_nocpsch() 399 struct qcm_process_device *qpd) register_process_nocpsch() 404 BUG_ON(!dqm || !qpd); register_process_nocpsch() 412 n->qpd = qpd; register_process_nocpsch() 417 retval = dqm->ops_asic_specific.register_process(dqm, qpd); register_process_nocpsch() 427 struct qcm_process_device *qpd) unregister_process_nocpsch() 432 BUG_ON(!dqm || !qpd); unregister_process_nocpsch() 436 pr_debug("qpd->queues_list is %s\n", unregister_process_nocpsch() 437 list_empty(&qpd->queues_list) ? "empty" : "not empty"); unregister_process_nocpsch() 443 if (qpd == cur->qpd) { unregister_process_nocpsch() 450 /* qpd not found in dqm list */ unregister_process_nocpsch() 633 struct qcm_process_device *qpd) create_sdma_queue_nocpsch() 653 dqm->ops_asic_specific.init_sdma_vm(dqm, q, qpd); create_sdma_queue_nocpsch() 758 if (node->qpd->pqm->process && dqm->dev) start_cpsch() 760 node->qpd->pqm->process); start_cpsch() 782 pdd = qpd_to_pdd(node->qpd); stop_cpsch() 793 struct qcm_process_device *qpd) create_kernel_queue_cpsch() 795 BUG_ON(!dqm || !kq || !qpd); create_kernel_queue_cpsch() 815 list_add(&kq->list, &qpd->priv_queue_list); create_kernel_queue_cpsch() 817 qpd->is_debug = true; create_kernel_queue_cpsch() 826 struct qcm_process_device *qpd) destroy_kernel_queue_cpsch() 837 qpd->is_debug = false; destroy_kernel_queue_cpsch() 858 struct qcm_process_device *qpd, int *allocate_vmid) create_queue_cpsch() 863 BUG_ON(!dqm || !q || !qpd); create_queue_cpsch() 890 dqm->ops_asic_specific.init_sdma_vm(dqm, q, qpd); create_queue_cpsch() 896 list_add(&q->list, &qpd->queues_list); create_queue_cpsch() 1037 struct qcm_process_device *qpd, destroy_queue_cpsch() 1044 BUG_ON(!dqm || !qpd || !q); destroy_queue_cpsch() 1053 if (qpd->is_debug) { destroy_queue_cpsch() 1109 struct qcm_process_device *qpd, set_cache_memory_policy() 1123 qpd->sh_mem_ape1_base = 1; set_cache_memory_policy() 1124 qpd->sh_mem_ape1_limit = 0; set_cache_memory_policy() 1148 qpd->sh_mem_ape1_base = base >> 16; set_cache_memory_policy() 1149 qpd->sh_mem_ape1_limit = limit >> 16; set_cache_memory_policy() 1154 qpd, set_cache_memory_policy() 1160 if ((sched_policy == KFD_SCHED_POLICY_NO_HWS) && (qpd->vmid != 0)) set_cache_memory_policy() 1161 program_sh_mem_settings(dqm, qpd); set_cache_memory_policy() 1164 qpd->sh_mem_config, qpd->sh_mem_ape1_base, set_cache_memory_policy() 1165 qpd->sh_mem_ape1_limit); set_cache_memory_policy() 83 program_sh_mem_settings(struct device_queue_manager *dqm, struct qcm_process_device *qpd) program_sh_mem_settings() argument 94 allocate_vmid(struct device_queue_manager *dqm, struct qcm_process_device *qpd, struct queue *q) allocate_vmid() argument 118 deallocate_vmid(struct device_queue_manager *dqm, struct qcm_process_device *qpd, struct queue *q) deallocate_vmid() argument 132 create_queue_nocpsch(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd, int *allocated_vmid) create_queue_nocpsch() argument 236 create_compute_queue_nocpsch(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd) create_compute_queue_nocpsch() argument 275 destroy_queue_nocpsch(struct device_queue_manager *dqm, struct qcm_process_device *qpd, struct queue *q) destroy_queue_nocpsch() argument 398 register_process_nocpsch(struct device_queue_manager *dqm, struct qcm_process_device *qpd) register_process_nocpsch() argument 426 unregister_process_nocpsch(struct device_queue_manager *dqm, struct qcm_process_device *qpd) unregister_process_nocpsch() argument 631 create_sdma_queue_nocpsch(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd) create_sdma_queue_nocpsch() argument 791 create_kernel_queue_cpsch(struct device_queue_manager *dqm, struct kernel_queue *kq, struct qcm_process_device *qpd) create_kernel_queue_cpsch() argument 824 destroy_kernel_queue_cpsch(struct device_queue_manager *dqm, struct kernel_queue *kq, struct qcm_process_device *qpd) destroy_kernel_queue_cpsch() argument 857 create_queue_cpsch(struct device_queue_manager *dqm, struct queue *q, struct qcm_process_device *qpd, int *allocate_vmid) create_queue_cpsch() argument 1036 destroy_queue_cpsch(struct device_queue_manager *dqm, struct qcm_process_device *qpd, struct queue *q) destroy_queue_cpsch() argument 1108 set_cache_memory_policy(struct device_queue_manager *dqm, struct qcm_process_device *qpd, enum cache_policy default_policy, enum cache_policy alternate_policy, void __user *alternate_aperture_base, uint64_t alternate_aperture_size) set_cache_memory_policy() argument
|
H A D | kfd_packet_manager.c | 146 struct qcm_process_device *qpd) pm_create_map_process() 152 BUG_ON(!pm || !buffer || !qpd); pm_create_map_process() 162 packet->bitfields2.diq_enable = (qpd->is_debug) ? 1 : 0; pm_create_map_process() 164 packet->bitfields2.pasid = qpd->pqm->process->pasid; pm_create_map_process() 165 packet->bitfields3.page_table_base = qpd->page_table_base; pm_create_map_process() 166 packet->bitfields10.gds_size = qpd->gds_size; pm_create_map_process() 167 packet->bitfields10.num_gws = qpd->num_gws; pm_create_map_process() 168 packet->bitfields10.num_oac = qpd->num_oac; pm_create_map_process() 170 list_for_each_entry(cur, &qpd->queues_list, list) pm_create_map_process() 172 packet->bitfields10.num_queues = (qpd->is_debug) ? 0 : num_queues; pm_create_map_process() 174 packet->sh_mem_config = qpd->sh_mem_config; pm_create_map_process() 175 packet->sh_mem_bases = qpd->sh_mem_bases; pm_create_map_process() 176 packet->sh_mem_ape1_base = qpd->sh_mem_ape1_base; pm_create_map_process() 177 packet->sh_mem_ape1_limit = qpd->sh_mem_ape1_limit; pm_create_map_process() 179 packet->gds_addr_lo = lower_32_bits(qpd->gds_context_area); pm_create_map_process() 180 packet->gds_addr_hi = upper_32_bits(qpd->gds_context_area); pm_create_map_process() 321 struct qcm_process_device *qpd; pm_create_runlist_ib() local 343 qpd = cur->qpd; list_for_each_entry() 351 retval = pm_create_map_process(pm, &rl_buffer[rl_wptr], qpd); list_for_each_entry() 359 list_for_each_entry(kq, &qpd->priv_queue_list, list) { list_for_each_entry() 364 kq->queue->queue, qpd->is_debug); list_for_each_entry() 371 qpd->is_debug); list_for_each_entry() 376 qpd->is_debug); list_for_each_entry() 385 list_for_each_entry(q, &qpd->queues_list, list) { list_for_each_entry() 390 q->queue, qpd->is_debug); list_for_each_entry() 397 qpd->is_debug); list_for_each_entry() 402 qpd->is_debug); list_for_each_entry() 145 pm_create_map_process(struct packet_manager *pm, uint32_t *buffer, struct qcm_process_device *qpd) pm_create_map_process() argument
|
H A D | kfd_device_queue_manager.h | 44 struct qcm_process_device *qpd; member in struct:device_process_node 89 struct qcm_process_device *qpd, 93 struct qcm_process_device *qpd, 104 struct qcm_process_device *qpd); 107 struct qcm_process_device *qpd); 115 struct qcm_process_device *qpd); 119 struct qcm_process_device *qpd); 122 struct qcm_process_device *qpd, 131 struct qcm_process_device *qpd); 134 struct qcm_process_device *qpd, 141 struct qcm_process_device *qpd); 184 struct qcm_process_device *qpd);
|
H A D | kfd_process_queue_manager.c | 183 if ((pdd->qpd.is_debug) || pqm_create_queue() 185 list_for_each_entry(cur, &pdd->qpd.queues_list, list) pqm_create_queue() 196 pdd->qpd.pqm = pqm; pqm_create_queue() 197 dev->dqm->ops.register_process(dev->dqm, &pdd->qpd); pqm_create_queue() 223 retval = dev->dqm->ops.create_queue(dev->dqm, q, &pdd->qpd, pqm_create_queue() 238 kq, &pdd->qpd); pqm_create_queue() 268 dev->dqm->ops.unregister_process(dev->dqm, &pdd->qpd); pqm_create_queue() 309 dqm->ops.destroy_kernel_queue(dqm, pqn->kq, &pdd->qpd); pqm_destroy_queue() 315 retval = dqm->ops.destroy_queue(dqm, &pdd->qpd, pqn->q); pqm_destroy_queue() 327 dqm->ops.unregister_process(dqm, &pdd->qpd); pqm_destroy_queue()
|
H A D | kfd_process.c | 356 INIT_LIST_HEAD(&pdd->qpd.queues_list); kfd_create_process_device_data() 357 INIT_LIST_HEAD(&pdd->qpd.priv_queue_list); kfd_create_process_device_data() 358 pdd->qpd.dqm = dev->dqm; kfd_create_process_device_data()
|
H A D | kfd_dbgdev.c | 318 &cntl, i, pdd->qpd.vmid); dbgdev_address_watch_nodiq() 322 pr_debug("\t\t%20s %08x\n", "vmid is :", pdd->qpd.vmid); dbgdev_address_watch_nodiq() 758 reg_sq_cmd.bits.vm_id = pdd->qpd.vmid; dbgdev_wave_control_nodiq() 767 pr_debug("\t\t vmid is: %u\n", pdd->qpd.vmid); dbgdev_wave_control_nodiq()
|
H A D | kfd_priv.h | 454 struct qcm_process_device qpd; member in struct:kfd_process_device 473 #define qpd_to_pdd(x) container_of(x, struct kfd_process_device, qpd)
|
H A D | kfd_chardev.c | 423 &pdd->qpd, kfd_ioctl_set_memory_policy()
|
/linux-4.4.14/drivers/infiniband/hw/cxgb4/ |
H A D | device.c | 231 struct c4iw_debugfs_data *qpd = data; dump_qp() local 238 space = qpd->bufsize - qpd->pos - 1; dump_qp() 253 cc = snprintf(qpd->buf + qpd->pos, space, dump_qp() 277 cc = snprintf(qpd->buf + qpd->pos, space, dump_qp() 293 cc = snprintf(qpd->buf + qpd->pos, space, dump_qp() 299 qpd->pos += cc; dump_qp() 305 struct c4iw_debugfs_data *qpd = file->private_data; qp_release() local 306 if (!qpd) { qp_release() 307 printk(KERN_INFO "%s null qpd?\n", __func__); qp_release() 310 vfree(qpd->buf); qp_release() 311 kfree(qpd); qp_release() 317 struct c4iw_debugfs_data *qpd; qp_open() local 321 qpd = kmalloc(sizeof *qpd, GFP_KERNEL); qp_open() 322 if (!qpd) { qp_open() 326 qpd->devp = inode->i_private; qp_open() 327 qpd->pos = 0; qp_open() 329 spin_lock_irq(&qpd->devp->lock); qp_open() 330 idr_for_each(&qpd->devp->qpidr, count_idrs, &count); qp_open() 331 spin_unlock_irq(&qpd->devp->lock); qp_open() 333 qpd->bufsize = count * 128; qp_open() 334 qpd->buf = vmalloc(qpd->bufsize); qp_open() 335 if (!qpd->buf) { qp_open() 340 spin_lock_irq(&qpd->devp->lock); qp_open() 341 idr_for_each(&qpd->devp->qpidr, dump_qp, qpd); qp_open() 342 spin_unlock_irq(&qpd->devp->lock); qp_open() 344 qpd->buf[qpd->pos++] = 0; qp_open() 345 file->private_data = qpd; qp_open() 348 kfree(qpd); qp_open() 653 pr_info("%s null qpd?\n", __func__); ep_release()
|