Searched refs:gmc (Results 1 - 25 of 58) sorted by relevance

123

/netbsd-current/sys/external/bsd/drm2/dist/drm/amd/amdgpu/
H A Damdgpu_gfxhub_v1_1.c59 adev->gmc.xgmi.num_physical_nodes = max_region + 1;
60 if (adev->gmc.xgmi.num_physical_nodes > max_num_physical_nodes)
63 adev->gmc.xgmi.physical_node_id =
65 if (adev->gmc.xgmi.physical_node_id > max_physical_node_id)
67 adev->gmc.xgmi.node_segment_size = REG_GET_FIELD(
H A Damdgpu_xgmi.c140 return snprintf(buf, PAGE_SIZE, "%llu\n", adev->gmc.xgmi.node_id);
245 if (!adev->gmc.xgmi.hive_id)
252 if (tmp->hive_id == adev->gmc.xgmi.hive_id) {
273 tmp->hive_id = adev->gmc.xgmi.hive_id;
311 adev->gmc.xgmi.node_id,
312 adev->gmc.xgmi.hive_id, ret);
323 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) {
349 adev->gmc.xgmi.node_id,
350 adev->gmc.xgmi.hive_id, ret);
363 if (top->nodes[i].node_id == peer_adev->gmc
[all...]
H A Damdgpu_xgmi.h58 adev->gmc.xgmi.hive_id &&
59 adev->gmc.xgmi.hive_id == bo_adev->gmc.xgmi.hive_id);
H A Damdgpu_gmc_v7_0.c45 #include "gmc/gmc_7_1_d.h"
46 #include "gmc/gmc_7_1_sh_mask.h"
164 err = request_firmware(&adev->gmc.fw, fw_name, adev->dev);
167 err = amdgpu_ucode_validate(adev->gmc.fw);
172 release_firmware(adev->gmc.fw);
173 adev->gmc.fw = NULL;
194 if (!adev->gmc.fw)
197 hdr = (const struct mc_firmware_header_v1_0 *)adev->gmc.fw->data;
200 adev->gmc.fw_version = le32_to_cpu(hdr->header.ucode_version);
203 (adev->gmc
[all...]
H A Damdgpu_gmc.h226 #define amdgpu_gmc_flush_gpu_tlb(adev, vmid, vmhub, type) ((adev)->gmc.gmc_funcs->flush_gpu_tlb((adev), (vmid), (vmhub), (type)))
228 ((adev)->gmc.gmc_funcs->flush_gpu_tlb_pasid \
230 #define amdgpu_gmc_emit_flush_gpu_tlb(r, vmid, addr) (r)->adev->gmc.gmc_funcs->emit_flush_gpu_tlb((r), (vmid), (addr))
231 #define amdgpu_gmc_emit_pasid_mapping(r, vmid, pasid) (r)->adev->gmc.gmc_funcs->emit_pasid_mapping((r), (vmid), (pasid))
232 #define amdgpu_gmc_map_mtype(adev, flags) (adev)->gmc.gmc_funcs->map_mtype((adev),(flags))
233 #define amdgpu_gmc_get_vm_pde(adev, level, dst, flags) (adev)->gmc.gmc_funcs->get_vm_pde((adev), (level), (dst), (flags))
234 #define amdgpu_gmc_get_vm_pte(adev, mapping, flags) (adev)->gmc.gmc_funcs->get_vm_pte((adev), (mapping), (flags))
244 static inline bool amdgpu_gmc_vram_full_visible(struct amdgpu_gmc *gmc) argument
246 WARN_ON(gmc->real_vram_size < gmc
[all...]
H A Damdgpu_gmc_v9_0.c405 adev->gmc.vm_fault.num_types = 1;
406 adev->gmc.vm_fault.funcs = &gmc_v9_0_irq_funcs;
409 adev->gmc.ecc_irq.num_types = 1;
410 adev->gmc.ecc_irq.funcs = &gmc_v9_0_ecc_funcs;
506 spin_lock(&adev->gmc.invalidate_lock);
553 spin_unlock(&adev->gmc.invalidate_lock);
742 adev->gmc.vram_start;
745 if (!adev->gmc.translate_further)
794 adev->gmc.gmc_funcs = &gmc_v9_0_gmc_funcs;
847 adev->gmc
[all...]
H A Damdgpu_gmc.c154 if (addr + PAGE_SIZE >= adev->gmc.agp_size)
157 return adev->gmc.agp_start + addr;
205 u64 max_mc_address = min(adev->gmc.mc_mask, AMDGPU_GMC_HOLE_START - 1);
295 struct amdgpu_gmc *gmc = &adev->gmc; local
304 if (gmc->fault_ring[gmc->last_fault].timestamp >= stamp)
309 fault = &gmc->fault_ring[gmc->fault_hash[hash].idx];
317 fault = &gmc
[all...]
H A Damdgpu_gmc_v8_0.c40 #include "gmc/gmc_8_1_d.h"
41 #include "gmc/gmc_8_1_sh_mask.h"
281 err = request_firmware(&adev->gmc.fw, fw_name, adev->dev);
284 err = amdgpu_ucode_validate(adev->gmc.fw);
289 release_firmware(adev->gmc.fw);
290 adev->gmc.fw = NULL;
319 if (!adev->gmc.fw)
322 hdr = (const struct mc_firmware_header_v1_0 *)adev->gmc.fw->data;
325 adev->gmc.fw_version = le32_to_cpu(hdr->header.ucode_version);
328 (adev->gmc
[all...]
H A Damdgpu_gmc_v10_0.c206 adev->gmc.vm_fault.num_types = 1;
207 adev->gmc.vm_fault.funcs = &gmc_v10_0_irq_funcs;
276 spin_lock(&adev->gmc.invalidate_lock);
325 spin_unlock(&adev->gmc.invalidate_lock);
574 adev->gmc.vram_start;
577 if (!adev->gmc.translate_further)
624 if (adev->gmc.gmc_funcs == NULL)
625 adev->gmc.gmc_funcs = &gmc_v10_0_gmc_funcs;
635 adev->gmc.shared_aperture_start = 0x2000000000000000ULL;
636 adev->gmc
[all...]
H A Damdgpu_gmc_v6_0.c43 #include "gmc/gmc_6_0_d.h"
44 #include "gmc/gmc_6_0_sh_mask.h"
150 err = request_firmware(&adev->gmc.fw, fw_name, adev->dev);
154 err = amdgpu_ucode_validate(adev->gmc.fw);
161 release_firmware(adev->gmc.fw);
162 adev->gmc.fw = NULL;
175 if (!adev->gmc.fw)
178 hdr = (const struct mc_firmware_header_v1_0 *)adev->gmc.fw->data;
182 adev->gmc.fw_version = le32_to_cpu(hdr->header.ucode_version);
185 (adev->gmc
[all...]
H A Damdgpu_gfxhub_v1_0.c64 (u32)(adev->gmc.gart_start >> 12));
66 (u32)(adev->gmc.gart_start >> 44));
69 (u32)(adev->gmc.gart_end >> 12));
71 (u32)(adev->gmc.gart_end >> 44));
80 WREG32_SOC15_RLC(GC, 0, mmMC_VM_AGP_BOT, adev->gmc.agp_start >> 24);
81 WREG32_SOC15_RLC(GC, 0, mmMC_VM_AGP_TOP, adev->gmc.agp_end >> 24);
86 min(adev->gmc.fb_start, adev->gmc.agp_start) >> 18);
98 max((adev->gmc.fb_end >> 18) + 0x1,
99 adev->gmc
[all...]
H A Damdgpu_test.c52 n = adev->gmc.gart_size - AMDGPU_IB_POOL_SIZE*64*1024;
164 (gart_addr - adev->gmc.gart_start +
167 (vram_addr - adev->gmc.vram_start +
210 (vram_addr - adev->gmc.vram_start +
213 (gart_addr - adev->gmc.gart_start +
223 gart_addr - adev->gmc.gart_start);
H A Damdgpu_object.c141 unsigned visible_pfn = adev->gmc.visible_vram_size >> PAGE_SHIFT;
591 if (!amdgpu_gmc_vram_full_visible(&adev->gmc) &&
593 bo->tbo.mem.start < adev->gmc.visible_vram_size >> PAGE_SHIFT)
1076 arch_io_reserve_memtype_wc(adev->gmc.aper_base,
1077 adev->gmc.aper_size);
1080 adev->gmc.vram_mtrr = arch_phys_wc_add(adev->gmc.aper_base,
1081 adev->gmc.aper_size);
1083 if (adev->gmc.aper_base)
1084 pmap_pv_track(adev->gmc
[all...]
H A Damdgpu_gfxhub_v2_0.c75 (u32)(adev->gmc.gart_start >> 12));
77 (u32)(adev->gmc.gart_start >> 44));
80 (u32)(adev->gmc.gart_end >> 12));
82 (u32)(adev->gmc.gart_end >> 44));
96 adev->gmc.vram_start >> 18);
98 adev->gmc.vram_end >> 18);
101 value = adev->vram_scratch.gpu_addr - adev->gmc.vram_start
163 if (adev->gmc.translate_further) {
271 adev->gmc.vram_start >> 24);
273 adev->gmc
[all...]
H A Damdgpu_mmhub_v2_0.c60 (u32)(adev->gmc.gart_start >> 12));
62 (u32)(adev->gmc.gart_start >> 44));
65 (u32)(adev->gmc.gart_end >> 12));
67 (u32)(adev->gmc.gart_end >> 44));
82 adev->gmc.vram_start >> 18);
84 adev->gmc.vram_end >> 18);
87 value = adev->vram_scratch.gpu_addr - adev->gmc.vram_start +
150 if (adev->gmc.translate_further) {
262 adev->gmc.vram_start >> 24);
264 adev->gmc
[all...]
H A Damdgpu_amdkfd.c77 amdgpu_amdkfd_total_mem_size += adev->gmc.real_vram_size;
394 resource_size_t aper_limit = adev->gmc.aper_base + adev->gmc.aper_size;
397 if (!(adev->gmc.aper_base & address_mask || aper_limit & address_mask)) {
398 mem_info->local_mem_size_public = adev->gmc.visible_vram_size;
399 mem_info->local_mem_size_private = adev->gmc.real_vram_size -
400 adev->gmc.visible_vram_size;
403 mem_info->local_mem_size_private = adev->gmc.real_vram_size;
405 mem_info->vram_width = adev->gmc.vram_width;
408 &adev->gmc
[all...]
H A Damdgpu_vram_mgr.c58 return snprintf(buf, PAGE_SIZE, "%llu\n", adev->gmc.real_vram_size);
75 return snprintf(buf, PAGE_SIZE, "%llu\n", adev->gmc.visible_vram_size);
121 switch (adev->gmc.vram_vendor) {
263 if (start >= adev->gmc.visible_vram_size)
266 return (end > adev->gmc.visible_vram_size ?
267 adev->gmc.visible_vram_size : end) - start;
286 if (amdgpu_gmc_vram_full_visible(&adev->gmc))
289 if (mem->start >= adev->gmc.visible_vram_size >> PAGE_SHIFT)
349 max_bytes = adev->gmc.mc_vram_size;
H A Damdgpu_mmhub_v1_0.c53 adev->gmc.fb_start = base;
54 adev->gmc.fb_end = top;
80 (u32)(adev->gmc.gart_start >> 12));
82 (u32)(adev->gmc.gart_start >> 44));
85 (u32)(adev->gmc.gart_end >> 12));
87 (u32)(adev->gmc.gart_end >> 44));
97 WREG32_SOC15(MMHUB, 0, mmMC_VM_AGP_BOT, adev->gmc.agp_start >> 24);
98 WREG32_SOC15(MMHUB, 0, mmMC_VM_AGP_TOP, adev->gmc.agp_end >> 24);
102 min(adev->gmc.fb_start, adev->gmc
[all...]
H A Damdgpu_fb.c289 tmp = amdgpu_bo_gpu_offset(abo) - adev->gmc.vram_start;
290 info->fix.smem_start = adev->gmc.aper_base + tmp;
299 info->apertures->ranges[0].size = adev->gmc.aper_size;
309 DRM_INFO("vram apper at 0x%lX\n", (unsigned long)adev->gmc.aper_base);
367 if (adev->gmc.real_vram_size <= (32*1024*1024))
H A Damdgpu_ttm.c112 man->gpu_offset = adev->gmc.gart_start;
120 man->gpu_offset = adev->gmc.vram_start;
191 } else if (!amdgpu_gmc_vram_full_visible(&adev->gmc) &&
202 abo->placements[0].fpfn = adev->gmc.visible_vram_size >> PAGE_SHIFT;
618 <= adev->gmc.visible_vram_size;
736 if ((mem->bus.offset + mem->bus.size) > adev->gmc.visible_vram_size)
747 mem->bus.base = adev->gmc.aper_base;
1185 placements.lpfn = adev->gmc.gart_size >> PAGE_SHIFT;
1697 while (len && pos < adev->gmc.mc_vram_size) {
1786 uint64_t vram_size = adev->gmc
[all...]
H A Damdgpu_mmhub_v9_4.c56 adev->gmc.fb_start = base;
57 adev->gmc.fb_end = top;
93 (u32)(adev->gmc.gart_start >> 12));
97 (u32)(adev->gmc.gart_start >> 44));
102 (u32)(adev->gmc.gart_end >> 12));
106 (u32)(adev->gmc.gart_end >> 44));
131 adev->gmc.agp_end >> 24);
134 adev->gmc.agp_start >> 24);
141 min(adev->gmc.fb_start, adev->gmc
[all...]
H A Damdgpu_umc.c60 r = amdgpu_irq_get(adev, &adev->gmc.ecc_irq, 0);
H A Damdgpu_kms.c234 fw_info->ver = adev->gmc.fw_version;
606 vram_gtt.vram_size = adev->gmc.real_vram_size -
610 min(adev->gmc.visible_vram_size -
623 mem.vram.total_heap_size = adev->gmc.real_vram_size;
624 mem.vram.usable_heap_size = adev->gmc.real_vram_size -
632 adev->gmc.visible_vram_size;
634 min(adev->gmc.visible_vram_size -
753 dev_info.vram_type = adev->gmc.vram_type;
754 dev_info.vram_bit_width = adev->gmc.vram_width;
H A Damdgpu_device.c938 u64 space_needed = roundup_pow_of_two(adev->gmc.real_vram_size);
2004 /* need to do gmc hw init early so we can allocate gpu mem */
2075 if (adev->gmc.xgmi.num_physical_nodes > 1)
2269 if (adev->gmc.xgmi.num_physical_nodes > 1) {
2285 if (mgpu_info.num_dgpu == adev->gmc.xgmi.num_physical_nodes) {
2322 if (adev->gmc.xgmi.num_physical_nodes > 1)
2960 adev->gmc.gart_size = 512 * 1024 * 1024;
2967 adev->gmc.gmc_funcs = NULL;
4062 list_for_each_entry(tmp_adev, device_list_handle, gmc.xgmi.head) {
4064 if (tmp_adev->gmc
[all...]
/netbsd-current/crypto/external/bsd/heimdal/dist/lib/gssapi/mech/
H A Dgss_mo.c240 struct gss_mech_compat_desc_struct *gmc; local
248 gmc = m->gm_compat;
250 if (gmc != NULL && gmc->gmc_inquire_saslname_for_mech != NULL) {
251 major = gmc->gmc_inquire_saslname_for_mech(minor,
359 struct gss_mech_compat_desc_struct *gmc; local
374 gmc = m->gm_mech.gm_compat;
375 if (gmc && gmc->gmc_inquire_mech_for_saslname) {
376 major = gmc
468 struct gss_mech_compat_desc_struct *gmc = mi->gm_compat; local
528 struct gss_mech_compat_desc_struct *gmc; local
[all...]

Completed in 160 milliseconds

123