Home
last modified time | relevance | path

Searched refs:gmc (Results 1 - 25 of 144) sorted by relevance

123456

/kernel/linux/linux-6.6/drivers/gpu/drm/amd/amdgpu/
H A Dgfxhub_v2_1.c144 (u32)(adev->gmc.gart_start >> 12)); in gfxhub_v2_1_init_gart_aperture_regs()
146 (u32)(adev->gmc.gart_start >> 44)); in gfxhub_v2_1_init_gart_aperture_regs()
149 (u32)(adev->gmc.gart_end >> 12)); in gfxhub_v2_1_init_gart_aperture_regs()
151 (u32)(adev->gmc.gart_end >> 44)); in gfxhub_v2_1_init_gart_aperture_regs()
160 WREG32_SOC15(GC, 0, mmGCMC_VM_AGP_BOT, adev->gmc.agp_start >> 24); in gfxhub_v2_1_init_system_aperture_regs()
161 WREG32_SOC15(GC, 0, mmGCMC_VM_AGP_TOP, adev->gmc.agp_end >> 24); in gfxhub_v2_1_init_system_aperture_regs()
165 min(adev->gmc.fb_start, adev->gmc.agp_start) >> 18); in gfxhub_v2_1_init_system_aperture_regs()
167 max(adev->gmc.fb_end, adev->gmc in gfxhub_v2_1_init_system_aperture_regs()
[all...]
H A Damdgpu_gmc.c52 u64 vram_size = adev->gmc.xgmi.node_segment_size * adev->gmc.xgmi.num_physical_nodes; in amdgpu_gmc_pdb0_alloc()
53 uint32_t pde0_page_shift = adev->gmc.vmid0_page_table_block_size + 21; in amdgpu_gmc_pdb0_alloc()
66 r = amdgpu_bo_create(adev, &bp, &adev->gmc.pdb0_bo); in amdgpu_gmc_pdb0_alloc()
70 r = amdgpu_bo_reserve(adev->gmc.pdb0_bo, false); in amdgpu_gmc_pdb0_alloc()
74 r = amdgpu_bo_pin(adev->gmc.pdb0_bo, AMDGPU_GEM_DOMAIN_VRAM); in amdgpu_gmc_pdb0_alloc()
77 r = amdgpu_bo_kmap(adev->gmc.pdb0_bo, &adev->gmc.ptr_pdb0); in amdgpu_gmc_pdb0_alloc()
81 amdgpu_bo_unreserve(adev->gmc.pdb0_bo); in amdgpu_gmc_pdb0_alloc()
85 amdgpu_bo_unpin(adev->gmc in amdgpu_gmc_pdb0_alloc()
377 struct amdgpu_gmc *gmc = &adev->gmc; amdgpu_gmc_filter_faults() local
446 struct amdgpu_gmc *gmc = &adev->gmc; amdgpu_gmc_filter_faults_remove() local
650 struct amdgpu_gmc *gmc = &adev->gmc; amdgpu_gmc_noretry_set() local
[all...]
H A Dgmc_v9_0.c734 adev->gmc.vm_fault.num_types = 1; in gmc_v9_0_set_irq_funcs()
735 adev->gmc.vm_fault.funcs = &gmc_v9_0_irq_funcs; in gmc_v9_0_set_irq_funcs()
738 !adev->gmc.xgmi.connected_to_cpu) { in gmc_v9_0_set_irq_funcs()
739 adev->gmc.ecc_irq.num_types = 1; in gmc_v9_0_set_irq_funcs()
740 adev->gmc.ecc_irq.funcs = &gmc_v9_0_ecc_funcs; in gmc_v9_0_set_irq_funcs()
824 if (adev->gmc.xgmi.num_physical_nodes && in gmc_v9_0_flush_gpu_tlb()
859 spin_lock(&adev->gmc.invalidate_lock); in gmc_v9_0_flush_gpu_tlb()
927 spin_unlock(&adev->gmc.invalidate_lock); in gmc_v9_0_flush_gpu_tlb()
969 bool vega20_xgmi_wa = (adev->gmc.xgmi.num_physical_nodes && in gmc_v9_0_flush_gpu_tlb_pasid()
1160 if (!adev->gmc in gmc_v9_0_get_vm_pde()
[all...]
H A Dgmc_v10_0.c190 adev->gmc.vm_fault.num_types = 1; in gmc_v10_0_set_irq_funcs()
191 adev->gmc.vm_fault.funcs = &gmc_v10_0_irq_funcs; in gmc_v10_0_set_irq_funcs()
194 adev->gmc.ecc_irq.num_types = 1; in gmc_v10_0_set_irq_funcs()
195 adev->gmc.ecc_irq.funcs = &gmc_v10_0_ecc_funcs; in gmc_v10_0_set_irq_funcs()
248 spin_lock(&adev->gmc.invalidate_lock); in gmc_v10_0_flush_vm_hub()
306 spin_unlock(&adev->gmc.invalidate_lock); in gmc_v10_0_flush_vm_hub()
597 if (!adev->gmc.translate_further) in gmc_v10_0_get_vm_pde()
677 if (adev->gmc.gmc_funcs == NULL) in gmc_v10_0_set_gmc_funcs()
678 adev->gmc.gmc_funcs = &gmc_v10_0_gmc_funcs; in gmc_v10_0_set_gmc_funcs()
742 adev->gmc in gmc_v10_0_early_init()
[all...]
H A Dgmc_v11_0.c155 adev->gmc.vm_fault.num_types = 1; in gmc_v11_0_set_irq_funcs()
156 adev->gmc.vm_fault.funcs = &gmc_v11_0_irq_funcs; in gmc_v11_0_set_irq_funcs()
159 adev->gmc.ecc_irq.num_types = 1; in gmc_v11_0_set_irq_funcs()
160 adev->gmc.ecc_irq.funcs = &gmc_v11_0_ecc_funcs; in gmc_v11_0_set_irq_funcs()
209 spin_lock(&adev->gmc.invalidate_lock); in gmc_v11_0_flush_vm_hub()
267 spin_unlock(&adev->gmc.invalidate_lock); in gmc_v11_0_flush_vm_hub()
502 adev->gmc.vram_start; in gmc_v11_0_get_vm_pde()
505 if (!adev->gmc.translate_further) in gmc_v11_0_get_vm_pde()
585 adev->gmc.gmc_funcs = &gmc_v11_0_gmc_funcs; in gmc_v11_0_set_gmc_funcs()
648 adev->gmc in gmc_v11_0_early_init()
[all...]
H A Damdgpu_xgmi.c324 return sysfs_emit(buf, "%llu\n", adev->gmc.xgmi.node_id); in amdgpu_xgmi_show_device_id()
485 if (!adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
496 if (hive->hive_id == adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
548 hive->hive_id = adev->gmc.xgmi.hive_id; in amdgpu_get_xgmi_hive()
619 request_adev->gmc.xgmi.node_id, in amdgpu_xgmi_set_pstate()
620 request_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_set_pstate()
651 adev->gmc.xgmi.node_id, in amdgpu_xgmi_update_topology()
652 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_update_topology()
672 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_hops_count()
684 if (top->nodes[i].node_id == peer_adev->gmc in amdgpu_xgmi_get_num_links()
[all...]
H A Dgmc_v7_0.c40 #include "gmc/gmc_7_1_d.h"
41 #include "gmc/gmc_7_1_sh_mask.h"
158 err = amdgpu_ucode_request(adev, &adev->gmc.fw, fw_name); in gmc_v7_0_init_microcode()
161 amdgpu_ucode_release(&adev->gmc.fw); in gmc_v7_0_init_microcode()
182 if (!adev->gmc.fw) in gmc_v7_0_mc_load_microcode()
185 hdr = (const struct mc_firmware_header_v1_0 *)adev->gmc.fw->data; in gmc_v7_0_mc_load_microcode()
188 adev->gmc.fw_version = le32_to_cpu(hdr->header.ucode_version); in gmc_v7_0_mc_load_microcode()
191 (adev->gmc.fw->data + le32_to_cpu(hdr->io_debug_array_offset_bytes)); in gmc_v7_0_mc_load_microcode()
194 (adev->gmc.fw->data + le32_to_cpu(hdr->header.ucode_array_offset_bytes)); in gmc_v7_0_mc_load_microcode()
285 adev->gmc in gmc_v7_0_mc_program()
[all...]
H A Dgfxhub_v1_0.c58 if (adev->gmc.pdb0_bo) in gfxhub_v1_0_init_gart_aperture_regs()
59 pt_base = amdgpu_gmc_pd_addr(adev->gmc.pdb0_bo); in gfxhub_v1_0_init_gart_aperture_regs()
68 if (adev->gmc.pdb0_bo) { in gfxhub_v1_0_init_gart_aperture_regs()
70 (u32)(adev->gmc.fb_start >> 12)); in gfxhub_v1_0_init_gart_aperture_regs()
72 (u32)(adev->gmc.fb_start >> 44)); in gfxhub_v1_0_init_gart_aperture_regs()
75 (u32)(adev->gmc.gart_end >> 12)); in gfxhub_v1_0_init_gart_aperture_regs()
77 (u32)(adev->gmc.gart_end >> 44)); in gfxhub_v1_0_init_gart_aperture_regs()
80 (u32)(adev->gmc.gart_start >> 12)); in gfxhub_v1_0_init_gart_aperture_regs()
82 (u32)(adev->gmc.gart_start >> 44)); in gfxhub_v1_0_init_gart_aperture_regs()
85 (u32)(adev->gmc in gfxhub_v1_0_init_gart_aperture_regs()
[all...]
H A Dgmc_v6_0.c38 #include "gmc/gmc_6_0_d.h"
39 #include "gmc/gmc_6_0_sh_mask.h"
135 err = amdgpu_ucode_request(adev, &adev->gmc.fw, fw_name); in gmc_v6_0_init_microcode()
140 amdgpu_ucode_release(&adev->gmc.fw); in gmc_v6_0_init_microcode()
153 if (!adev->gmc.fw) in gmc_v6_0_mc_load_microcode()
156 hdr = (const struct mc_firmware_header_v1_0 *)adev->gmc.fw->data; in gmc_v6_0_mc_load_microcode()
160 adev->gmc.fw_version = le32_to_cpu(hdr->header.ucode_version); in gmc_v6_0_mc_load_microcode()
163 (adev->gmc.fw->data + le32_to_cpu(hdr->io_debug_array_offset_bytes)); in gmc_v6_0_mc_load_microcode()
166 (adev->gmc.fw->data + le32_to_cpu(hdr->header.ucode_array_offset_bytes)); in gmc_v6_0_mc_load_microcode()
250 adev->gmc in gmc_v6_0_mc_program()
[all...]
H A Dgfxhub_v1_1.c88 if (max_region || adev->gmc.xgmi.connected_to_cpu) { in gfxhub_v1_1_get_xgmi_info()
89 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v1_1_get_xgmi_info()
91 if (adev->gmc.xgmi.num_physical_nodes > max_num_physical_nodes) in gfxhub_v1_1_get_xgmi_info()
95 adev->gmc.xgmi.physical_node_id = in gfxhub_v1_1_get_xgmi_info()
99 adev->gmc.xgmi.physical_node_id = in gfxhub_v1_1_get_xgmi_info()
104 if (adev->gmc.xgmi.physical_node_id > max_physical_node_id) in gfxhub_v1_1_get_xgmi_info()
107 adev->gmc.xgmi.node_segment_size = seg_size; in gfxhub_v1_1_get_xgmi_info()
H A Dgmc_v8_0.c35 #include "gmc/gmc_8_1_d.h"
36 #include "gmc/gmc_8_1_sh_mask.h"
259 err = amdgpu_ucode_request(adev, &adev->gmc.fw, fw_name); in gmc_v8_0_init_microcode()
262 amdgpu_ucode_release(&adev->gmc.fw); in gmc_v8_0_init_microcode()
291 if (!adev->gmc.fw) in gmc_v8_0_tonga_mc_load_microcode()
294 hdr = (const struct mc_firmware_header_v1_0 *)adev->gmc.fw->data; in gmc_v8_0_tonga_mc_load_microcode()
297 adev->gmc.fw_version = le32_to_cpu(hdr->header.ucode_version); in gmc_v8_0_tonga_mc_load_microcode()
300 (adev->gmc.fw->data + le32_to_cpu(hdr->io_debug_array_offset_bytes)); in gmc_v8_0_tonga_mc_load_microcode()
303 (adev->gmc.fw->data + le32_to_cpu(hdr->header.ucode_array_offset_bytes)); in gmc_v8_0_tonga_mc_load_microcode()
360 if (!adev->gmc in gmc_v8_0_polaris_mc_load_microcode()
[all...]
H A Dgfxhub_v1_2.c80 if (adev->gmc.pdb0_bo) in gfxhub_v1_2_xcc_init_gart_aperture_regs()
81 pt_base = amdgpu_gmc_pd_addr(adev->gmc.pdb0_bo); in gfxhub_v1_2_xcc_init_gart_aperture_regs()
91 if (adev->gmc.pdb0_bo) { in gfxhub_v1_2_xcc_init_gart_aperture_regs()
94 (u32)(adev->gmc.fb_start >> 12)); in gfxhub_v1_2_xcc_init_gart_aperture_regs()
97 (u32)(adev->gmc.fb_start >> 44)); in gfxhub_v1_2_xcc_init_gart_aperture_regs()
101 (u32)(adev->gmc.gart_end >> 12)); in gfxhub_v1_2_xcc_init_gart_aperture_regs()
104 (u32)(adev->gmc.gart_end >> 44)); in gfxhub_v1_2_xcc_init_gart_aperture_regs()
108 (u32)(adev->gmc.gart_start >> 12)); in gfxhub_v1_2_xcc_init_gart_aperture_regs()
111 (u32)(adev->gmc.gart_start >> 44)); in gfxhub_v1_2_xcc_init_gart_aperture_regs()
115 (u32)(adev->gmc in gfxhub_v1_2_xcc_init_gart_aperture_regs()
[all...]
H A Damdgpu_gmc.h338 #define amdgpu_gmc_flush_gpu_tlb(adev, vmid, vmhub, type) ((adev)->gmc.gmc_funcs->flush_gpu_tlb((adev), (vmid), (vmhub), (type)))
340 ((adev)->gmc.gmc_funcs->flush_gpu_tlb_pasid \
342 #define amdgpu_gmc_emit_flush_gpu_tlb(r, vmid, addr) (r)->adev->gmc.gmc_funcs->emit_flush_gpu_tlb((r), (vmid), (addr))
343 #define amdgpu_gmc_emit_pasid_mapping(r, vmid, pasid) (r)->adev->gmc.gmc_funcs->emit_pasid_mapping((r), (vmid), (pasid))
344 #define amdgpu_gmc_map_mtype(adev, flags) (adev)->gmc.gmc_funcs->map_mtype((adev),(flags))
345 #define amdgpu_gmc_get_vm_pde(adev, level, dst, flags) (adev)->gmc.gmc_funcs->get_vm_pde((adev), (level), (dst), (flags))
346 #define amdgpu_gmc_get_vm_pte(adev, mapping, flags) (adev)->gmc.gmc_funcs->get_vm_pte((adev), (mapping), (flags))
348 (adev)->gmc.gmc_funcs->override_vm_pte_flags \
350 #define amdgpu_gmc_get_vbios_fb_size(adev) (adev)->gmc.gmc_funcs->get_vbios_fb_size((adev))
360 static inline bool amdgpu_gmc_vram_full_visible(struct amdgpu_gmc *gmc) in amdgpu_gmc_vram_full_visible() argument
[all...]
H A Dmmhub_v1_8.c48 adev->gmc.fb_start = base; in mmhub_v1_8_get_fb_location()
49 adev->gmc.fb_end = top; in mmhub_v1_8_get_fb_location()
82 if (adev->gmc.pdb0_bo) in mmhub_v1_8_init_gart_aperture_regs()
83 pt_base = amdgpu_gmc_pd_addr(adev->gmc.pdb0_bo); in mmhub_v1_8_init_gart_aperture_regs()
94 if (adev->gmc.pdb0_bo) { in mmhub_v1_8_init_gart_aperture_regs()
97 (u32)(adev->gmc.fb_start >> 12)); in mmhub_v1_8_init_gart_aperture_regs()
100 (u32)(adev->gmc.fb_start >> 44)); in mmhub_v1_8_init_gart_aperture_regs()
104 (u32)(adev->gmc.gart_end >> 12)); in mmhub_v1_8_init_gart_aperture_regs()
107 (u32)(adev->gmc.gart_end >> 44)); in mmhub_v1_8_init_gart_aperture_regs()
112 (u32)(adev->gmc in mmhub_v1_8_init_gart_aperture_regs()
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_xgmi.c220 return snprintf(buf, PAGE_SIZE, "%llu\n", adev->gmc.xgmi.node_id); in amdgpu_xgmi_show_device_id()
330 if (!adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
342 if (hive->hive_id == adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive()
367 hive->hive_id = adev->gmc.xgmi.hive_id; in amdgpu_get_xgmi_hive()
433 request_adev->gmc.xgmi.node_id, in amdgpu_xgmi_set_pstate()
434 request_adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_set_pstate()
462 adev->gmc.xgmi.node_id, in amdgpu_xgmi_update_topology()
463 adev->gmc.xgmi.hive_id, ret); in amdgpu_xgmi_update_topology()
483 if (top->nodes[i].node_id == peer_adev->gmc.xgmi.node_id) in amdgpu_xgmi_get_hops_count()
497 if (!adev->gmc in amdgpu_xgmi_add_device()
[all...]
H A Dgmc_v10_0.c150 adev->gmc.vm_fault.num_types = 1; in gmc_v10_0_set_irq_funcs()
151 adev->gmc.vm_fault.funcs = &gmc_v10_0_irq_funcs; in gmc_v10_0_set_irq_funcs()
154 adev->gmc.ecc_irq.num_types = 1; in gmc_v10_0_set_irq_funcs()
155 adev->gmc.ecc_irq.funcs = &gmc_v10_0_ecc_funcs; in gmc_v10_0_set_irq_funcs()
205 spin_lock(&adev->gmc.invalidate_lock); in gmc_v10_0_flush_vm_hub()
257 spin_unlock(&adev->gmc.invalidate_lock); in gmc_v10_0_flush_vm_hub()
539 adev->gmc.vram_start; in gmc_v10_0_get_vm_pde()
542 if (!adev->gmc.translate_further) in gmc_v10_0_get_vm_pde()
612 if (adev->gmc.gmc_funcs == NULL) in gmc_v10_0_set_gmc_funcs()
613 adev->gmc in gmc_v10_0_set_gmc_funcs()
[all...]
H A Dgmc_v7_0.c40 #include "gmc/gmc_7_1_d.h"
41 #include "gmc/gmc_7_1_sh_mask.h"
159 err = request_firmware(&adev->gmc.fw, fw_name, adev->dev); in gmc_v7_0_init_microcode()
162 err = amdgpu_ucode_validate(adev->gmc.fw); in gmc_v7_0_init_microcode()
167 release_firmware(adev->gmc.fw); in gmc_v7_0_init_microcode()
168 adev->gmc.fw = NULL; in gmc_v7_0_init_microcode()
189 if (!adev->gmc.fw) in gmc_v7_0_mc_load_microcode()
192 hdr = (const struct mc_firmware_header_v1_0 *)adev->gmc.fw->data; in gmc_v7_0_mc_load_microcode()
195 adev->gmc.fw_version = le32_to_cpu(hdr->header.ucode_version); in gmc_v7_0_mc_load_microcode()
198 (adev->gmc in gmc_v7_0_mc_load_microcode()
[all...]
H A Dgmc_v9_0.c645 adev->gmc.vm_fault.num_types = 1; in gmc_v9_0_set_irq_funcs()
646 adev->gmc.vm_fault.funcs = &gmc_v9_0_irq_funcs; in gmc_v9_0_set_irq_funcs()
649 adev->gmc.ecc_irq.num_types = 1; in gmc_v9_0_set_irq_funcs()
650 adev->gmc.ecc_irq.funcs = &gmc_v9_0_ecc_funcs; in gmc_v9_0_set_irq_funcs()
729 if (adev->gmc.xgmi.num_physical_nodes && in gmc_v9_0_flush_gpu_tlb()
759 spin_lock(&adev->gmc.invalidate_lock); in gmc_v9_0_flush_gpu_tlb()
817 spin_unlock(&adev->gmc.invalidate_lock); in gmc_v9_0_flush_gpu_tlb()
856 bool vega20_xgmi_wa = (adev->gmc.xgmi.num_physical_nodes && in gmc_v9_0_flush_gpu_tlb_pasid()
1038 adev->gmc.vram_start; in gmc_v9_0_get_vm_pde()
1041 if (!adev->gmc in gmc_v9_0_get_vm_pde()
[all...]
H A Damdgpu_gmc.c131 if (ttm->dma_address[0] + PAGE_SIZE >= adev->gmc.agp_size) in amdgpu_gmc_agp_addr()
134 return adev->gmc.agp_start + ttm->dma_address[0]; in amdgpu_gmc_agp_addr()
182 u64 max_mc_address = min(adev->gmc.mc_mask, AMDGPU_GMC_HOLE_START - 1); in amdgpu_gmc_gart_location()
272 struct amdgpu_gmc *gmc = &adev->gmc; in amdgpu_gmc_filter_faults() local
281 if (gmc->fault_ring[gmc->last_fault].timestamp >= stamp) in amdgpu_gmc_filter_faults()
286 fault = &gmc->fault_ring[gmc->fault_hash[hash].idx]; in amdgpu_gmc_filter_faults()
294 fault = &gmc in amdgpu_gmc_filter_faults()
425 struct amdgpu_gmc *gmc = &adev->gmc; amdgpu_gmc_noretry_set() local
[all...]
H A Dgmc_v6_0.c38 #include "gmc/gmc_6_0_d.h"
39 #include "gmc/gmc_6_0_sh_mask.h"
134 err = request_firmware(&adev->gmc.fw, fw_name, adev->dev); in gmc_v6_0_init_microcode()
138 err = amdgpu_ucode_validate(adev->gmc.fw); in gmc_v6_0_init_microcode()
145 release_firmware(adev->gmc.fw); in gmc_v6_0_init_microcode()
146 adev->gmc.fw = NULL; in gmc_v6_0_init_microcode()
159 if (!adev->gmc.fw) in gmc_v6_0_mc_load_microcode()
162 hdr = (const struct mc_firmware_header_v1_0 *)adev->gmc.fw->data; in gmc_v6_0_mc_load_microcode()
166 adev->gmc.fw_version = le32_to_cpu(hdr->header.ucode_version); in gmc_v6_0_mc_load_microcode()
169 (adev->gmc in gmc_v6_0_mc_load_microcode()
[all...]
H A Damdgpu_gmc.h245 #define amdgpu_gmc_flush_gpu_tlb(adev, vmid, vmhub, type) ((adev)->gmc.gmc_funcs->flush_gpu_tlb((adev), (vmid), (vmhub), (type)))
247 ((adev)->gmc.gmc_funcs->flush_gpu_tlb_pasid \
249 #define amdgpu_gmc_emit_flush_gpu_tlb(r, vmid, addr) (r)->adev->gmc.gmc_funcs->emit_flush_gpu_tlb((r), (vmid), (addr))
250 #define amdgpu_gmc_emit_pasid_mapping(r, vmid, pasid) (r)->adev->gmc.gmc_funcs->emit_pasid_mapping((r), (vmid), (pasid))
251 #define amdgpu_gmc_map_mtype(adev, flags) (adev)->gmc.gmc_funcs->map_mtype((adev),(flags))
252 #define amdgpu_gmc_get_vm_pde(adev, level, dst, flags) (adev)->gmc.gmc_funcs->get_vm_pde((adev), (level), (dst), (flags))
253 #define amdgpu_gmc_get_vm_pte(adev, mapping, flags) (adev)->gmc.gmc_funcs->get_vm_pte((adev), (mapping), (flags))
254 #define amdgpu_gmc_get_vbios_fb_size(adev) (adev)->gmc.gmc_funcs->get_vbios_fb_size((adev))
264 static inline bool amdgpu_gmc_vram_full_visible(struct amdgpu_gmc *gmc) in amdgpu_gmc_vram_full_visible() argument
266 WARN_ON(gmc in amdgpu_gmc_vram_full_visible()
[all...]
H A Dgmc_v8_0.c35 #include "gmc/gmc_8_1_d.h"
36 #include "gmc/gmc_8_1_sh_mask.h"
276 err = request_firmware(&adev->gmc.fw, fw_name, adev->dev); in gmc_v8_0_init_microcode()
279 err = amdgpu_ucode_validate(adev->gmc.fw); in gmc_v8_0_init_microcode()
284 release_firmware(adev->gmc.fw); in gmc_v8_0_init_microcode()
285 adev->gmc.fw = NULL; in gmc_v8_0_init_microcode()
314 if (!adev->gmc.fw) in gmc_v8_0_tonga_mc_load_microcode()
317 hdr = (const struct mc_firmware_header_v1_0 *)adev->gmc.fw->data; in gmc_v8_0_tonga_mc_load_microcode()
320 adev->gmc.fw_version = le32_to_cpu(hdr->header.ucode_version); in gmc_v8_0_tonga_mc_load_microcode()
323 (adev->gmc in gmc_v8_0_tonga_mc_load_microcode()
[all...]
H A Dgfxhub_v1_1.c55 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v1_1_get_xgmi_info()
56 if (adev->gmc.xgmi.num_physical_nodes > max_num_physical_nodes) in gfxhub_v1_1_get_xgmi_info()
59 adev->gmc.xgmi.physical_node_id = in gfxhub_v1_1_get_xgmi_info()
61 if (adev->gmc.xgmi.physical_node_id > max_physical_node_id) in gfxhub_v1_1_get_xgmi_info()
63 adev->gmc.xgmi.node_segment_size = REG_GET_FIELD( in gfxhub_v1_1_get_xgmi_info()
H A Dgfxhub_v1_0.c59 (u32)(adev->gmc.gart_start >> 12)); in gfxhub_v1_0_init_gart_aperture_regs()
61 (u32)(adev->gmc.gart_start >> 44)); in gfxhub_v1_0_init_gart_aperture_regs()
64 (u32)(adev->gmc.gart_end >> 12)); in gfxhub_v1_0_init_gart_aperture_regs()
66 (u32)(adev->gmc.gart_end >> 44)); in gfxhub_v1_0_init_gart_aperture_regs()
75 WREG32_SOC15_RLC(GC, 0, mmMC_VM_AGP_BOT, adev->gmc.agp_start >> 24); in gfxhub_v1_0_init_system_aperture_regs()
76 WREG32_SOC15_RLC(GC, 0, mmMC_VM_AGP_TOP, adev->gmc.agp_end >> 24); in gfxhub_v1_0_init_system_aperture_regs()
81 min(adev->gmc.fb_start, adev->gmc.agp_start) >> 18); in gfxhub_v1_0_init_system_aperture_regs()
93 max((adev->gmc.fb_end >> 18) + 0x1, in gfxhub_v1_0_init_system_aperture_regs()
94 adev->gmc in gfxhub_v1_0_init_system_aperture_regs()
[all...]
H A Dgfxhub_v2_1.c141 (u32)(adev->gmc.gart_start >> 12)); in gfxhub_v2_1_init_gart_aperture_regs()
143 (u32)(adev->gmc.gart_start >> 44)); in gfxhub_v2_1_init_gart_aperture_regs()
146 (u32)(adev->gmc.gart_end >> 12)); in gfxhub_v2_1_init_gart_aperture_regs()
148 (u32)(adev->gmc.gart_end >> 44)); in gfxhub_v2_1_init_gart_aperture_regs()
162 adev->gmc.vram_start >> 18); in gfxhub_v2_1_init_system_aperture_regs()
164 adev->gmc.vram_end >> 18); in gfxhub_v2_1_init_system_aperture_regs()
167 value = adev->vram_scratch.gpu_addr - adev->gmc.vram_start in gfxhub_v2_1_init_system_aperture_regs()
235 if (adev->gmc.translate_further) { in gfxhub_v2_1_init_cache_regs()
322 !adev->gmc.noretry); in gfxhub_v2_1_setup_vmid_config()
360 adev->gmc in gfxhub_v2_1_gart_enable()
[all...]

Completed in 21 milliseconds

123456