Line data Source code
1 : /*
2 : * Copyright 2021 Advanced Micro Devices, Inc.
3 : *
4 : * Permission is hereby granted, free of charge, to any person obtaining a
5 : * copy of this software and associated documentation files (the "Software"),
6 : * to deal in the Software without restriction, including without limitation
7 : * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 : * and/or sell copies of the Software, and to permit persons to whom the
9 : * Software is furnished to do so, subject to the following conditions:
10 : *
11 : * The above copyright notice and this permission notice shall be included in
12 : * all copies or substantial portions of the Software.
13 : *
14 : * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15 : * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16 : * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17 : * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18 : * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19 : * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20 : * OTHER DEALINGS IN THE SOFTWARE.
21 : *
22 : */
23 :
24 : #include "amdgpu.h"
25 : #include "mmhub_v3_0.h"
26 :
27 : #include "mmhub/mmhub_3_0_0_offset.h"
28 : #include "mmhub/mmhub_3_0_0_sh_mask.h"
29 : #include "navi10_enum.h"
30 :
31 : #include "soc15_common.h"
32 :
33 : #define regMMVM_L2_CNTL3_DEFAULT 0x80100007
34 : #define regMMVM_L2_CNTL4_DEFAULT 0x000000c1
35 : #define regMMVM_L2_CNTL5_DEFAULT 0x00003fe0
36 :
37 : static const char *mmhub_client_ids_v3_0_0[][2] = {
38 : [0][0] = "VMC",
39 : [4][0] = "DCEDMC",
40 : [5][0] = "DCEVGA",
41 : [6][0] = "MP0",
42 : [7][0] = "MP1",
43 : [8][0] = "MPIO",
44 : [16][0] = "HDP",
45 : [17][0] = "LSDMA",
46 : [18][0] = "JPEG",
47 : [19][0] = "VCNU0",
48 : [21][0] = "VSCH",
49 : [22][0] = "VCNU1",
50 : [23][0] = "VCN1",
51 : [32+20][0] = "VCN0",
52 : [2][1] = "DBGUNBIO",
53 : [3][1] = "DCEDWB",
54 : [4][1] = "DCEDMC",
55 : [5][1] = "DCEVGA",
56 : [6][1] = "MP0",
57 : [7][1] = "MP1",
58 : [8][1] = "MPIO",
59 : [10][1] = "DBGU0",
60 : [11][1] = "DBGU1",
61 : [12][1] = "DBGU2",
62 : [13][1] = "DBGU3",
63 : [14][1] = "XDP",
64 : [15][1] = "OSSSYS",
65 : [16][1] = "HDP",
66 : [17][1] = "LSDMA",
67 : [18][1] = "JPEG",
68 : [19][1] = "VCNU0",
69 : [20][1] = "VCN0",
70 : [21][1] = "VSCH",
71 : [22][1] = "VCNU1",
72 : [23][1] = "VCN1",
73 : };
74 :
75 0 : static uint32_t mmhub_v3_0_get_invalidate_req(unsigned int vmid,
76 : uint32_t flush_type)
77 : {
78 0 : u32 req = 0;
79 :
80 : /* invalidate using legacy mode on vmid*/
81 0 : req = REG_SET_FIELD(req, MMVM_INVALIDATE_ENG0_REQ,
82 : PER_VMID_INVALIDATE_REQ, 1 << vmid);
83 0 : req = REG_SET_FIELD(req, MMVM_INVALIDATE_ENG0_REQ, FLUSH_TYPE, flush_type);
84 0 : req = REG_SET_FIELD(req, MMVM_INVALIDATE_ENG0_REQ, INVALIDATE_L2_PTES, 1);
85 0 : req = REG_SET_FIELD(req, MMVM_INVALIDATE_ENG0_REQ, INVALIDATE_L2_PDE0, 1);
86 0 : req = REG_SET_FIELD(req, MMVM_INVALIDATE_ENG0_REQ, INVALIDATE_L2_PDE1, 1);
87 0 : req = REG_SET_FIELD(req, MMVM_INVALIDATE_ENG0_REQ, INVALIDATE_L2_PDE2, 1);
88 0 : req = REG_SET_FIELD(req, MMVM_INVALIDATE_ENG0_REQ, INVALIDATE_L1_PTES, 1);
89 0 : req = REG_SET_FIELD(req, MMVM_INVALIDATE_ENG0_REQ,
90 : CLEAR_PROTECTION_FAULT_STATUS_ADDR, 0);
91 :
92 0 : return req;
93 : }
94 :
95 : static void
96 0 : mmhub_v3_0_print_l2_protection_fault_status(struct amdgpu_device *adev,
97 : uint32_t status)
98 : {
99 : uint32_t cid, rw;
100 0 : const char *mmhub_cid = NULL;
101 :
102 0 : cid = REG_GET_FIELD(status,
103 : MMVM_L2_PROTECTION_FAULT_STATUS, CID);
104 0 : rw = REG_GET_FIELD(status,
105 : MMVM_L2_PROTECTION_FAULT_STATUS, RW);
106 :
107 0 : dev_err(adev->dev,
108 : "MMVM_L2_PROTECTION_FAULT_STATUS:0x%08X\n",
109 : status);
110 0 : switch (adev->ip_versions[MMHUB_HWIP][0]) {
111 : case IP_VERSION(3, 0, 0):
112 : case IP_VERSION(3, 0, 1):
113 0 : mmhub_cid = mmhub_client_ids_v3_0_0[cid][rw];
114 0 : break;
115 : default:
116 : mmhub_cid = NULL;
117 : break;
118 : }
119 0 : dev_err(adev->dev, "\t Faulty UTCL2 client ID: %s (0x%x)\n",
120 : mmhub_cid ? mmhub_cid : "unknown", cid);
121 0 : dev_err(adev->dev, "\t MORE_FAULTS: 0x%lx\n",
122 : REG_GET_FIELD(status,
123 : MMVM_L2_PROTECTION_FAULT_STATUS, MORE_FAULTS));
124 0 : dev_err(adev->dev, "\t WALKER_ERROR: 0x%lx\n",
125 : REG_GET_FIELD(status,
126 : MMVM_L2_PROTECTION_FAULT_STATUS, WALKER_ERROR));
127 0 : dev_err(adev->dev, "\t PERMISSION_FAULTS: 0x%lx\n",
128 : REG_GET_FIELD(status,
129 : MMVM_L2_PROTECTION_FAULT_STATUS, PERMISSION_FAULTS));
130 0 : dev_err(adev->dev, "\t MAPPING_ERROR: 0x%lx\n",
131 : REG_GET_FIELD(status,
132 : MMVM_L2_PROTECTION_FAULT_STATUS, MAPPING_ERROR));
133 0 : dev_err(adev->dev, "\t RW: 0x%x\n", rw);
134 0 : }
135 :
136 0 : static void mmhub_v3_0_setup_vm_pt_regs(struct amdgpu_device *adev, uint32_t vmid,
137 : uint64_t page_table_base)
138 : {
139 0 : struct amdgpu_vmhub *hub = &adev->vmhub[AMDGPU_MMHUB_0];
140 :
141 0 : WREG32_SOC15_OFFSET(MMHUB, 0, regMMVM_CONTEXT0_PAGE_TABLE_BASE_ADDR_LO32,
142 : hub->ctx_addr_distance * vmid,
143 : lower_32_bits(page_table_base));
144 :
145 0 : WREG32_SOC15_OFFSET(MMHUB, 0, regMMVM_CONTEXT0_PAGE_TABLE_BASE_ADDR_HI32,
146 : hub->ctx_addr_distance * vmid,
147 : upper_32_bits(page_table_base));
148 0 : }
149 :
150 0 : static void mmhub_v3_0_init_gart_aperture_regs(struct amdgpu_device *adev)
151 : {
152 0 : uint64_t pt_base = amdgpu_gmc_pd_addr(adev->gart.bo);
153 :
154 0 : mmhub_v3_0_setup_vm_pt_regs(adev, 0, pt_base);
155 :
156 0 : WREG32_SOC15(MMHUB, 0, regMMVM_CONTEXT0_PAGE_TABLE_START_ADDR_LO32,
157 : (u32)(adev->gmc.gart_start >> 12));
158 0 : WREG32_SOC15(MMHUB, 0, regMMVM_CONTEXT0_PAGE_TABLE_START_ADDR_HI32,
159 : (u32)(adev->gmc.gart_start >> 44));
160 :
161 0 : WREG32_SOC15(MMHUB, 0, regMMVM_CONTEXT0_PAGE_TABLE_END_ADDR_LO32,
162 : (u32)(adev->gmc.gart_end >> 12));
163 0 : WREG32_SOC15(MMHUB, 0, regMMVM_CONTEXT0_PAGE_TABLE_END_ADDR_HI32,
164 : (u32)(adev->gmc.gart_end >> 44));
165 0 : }
166 :
167 0 : static void mmhub_v3_0_init_system_aperture_regs(struct amdgpu_device *adev)
168 : {
169 : uint64_t value;
170 : uint32_t tmp;
171 :
172 0 : if (!amdgpu_sriov_vf(adev)) {
173 : /*
174 : * the new L1 policy will block SRIOV guest from writing
175 : * these regs, and they will be programed at host.
176 : * so skip programing these regs.
177 : */
178 : /* Disable AGP. */
179 0 : WREG32_SOC15(MMHUB, 0, regMMMC_VM_AGP_BASE, 0);
180 0 : WREG32_SOC15(MMHUB, 0, regMMMC_VM_AGP_TOP, 0);
181 0 : WREG32_SOC15(MMHUB, 0, regMMMC_VM_AGP_BOT, 0x00FFFFFF);
182 :
183 : /* Program the system aperture low logical page number. */
184 0 : WREG32_SOC15(MMHUB, 0, regMMMC_VM_SYSTEM_APERTURE_LOW_ADDR,
185 : adev->gmc.vram_start >> 18);
186 0 : WREG32_SOC15(MMHUB, 0, regMMMC_VM_SYSTEM_APERTURE_HIGH_ADDR,
187 : adev->gmc.vram_end >> 18);
188 : }
189 :
190 : /* Set default page address. */
191 0 : value = adev->vram_scratch.gpu_addr - adev->gmc.vram_start +
192 0 : adev->vm_manager.vram_base_offset;
193 0 : WREG32_SOC15(MMHUB, 0, regMMMC_VM_SYSTEM_APERTURE_DEFAULT_ADDR_LSB,
194 : (u32)(value >> 12));
195 0 : WREG32_SOC15(MMHUB, 0, regMMMC_VM_SYSTEM_APERTURE_DEFAULT_ADDR_MSB,
196 : (u32)(value >> 44));
197 :
198 : /* Program "protection fault". */
199 0 : WREG32_SOC15(MMHUB, 0, regMMVM_L2_PROTECTION_FAULT_DEFAULT_ADDR_LO32,
200 : (u32)(adev->dummy_page_addr >> 12));
201 0 : WREG32_SOC15(MMHUB, 0, regMMVM_L2_PROTECTION_FAULT_DEFAULT_ADDR_HI32,
202 : (u32)((u64)adev->dummy_page_addr >> 44));
203 :
204 0 : tmp = RREG32_SOC15(MMHUB, 0, regMMVM_L2_PROTECTION_FAULT_CNTL2);
205 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_PROTECTION_FAULT_CNTL2,
206 : ACTIVE_PAGE_MIGRATION_PTE_READ_RETRY, 1);
207 0 : WREG32_SOC15(MMHUB, 0, regMMVM_L2_PROTECTION_FAULT_CNTL2, tmp);
208 0 : }
209 :
210 0 : static void mmhub_v3_0_init_tlb_regs(struct amdgpu_device *adev)
211 : {
212 : uint32_t tmp;
213 :
214 : /* Setup TLB control */
215 0 : tmp = RREG32_SOC15(MMHUB, 0, regMMMC_VM_MX_L1_TLB_CNTL);
216 :
217 0 : tmp = REG_SET_FIELD(tmp, MMMC_VM_MX_L1_TLB_CNTL, ENABLE_L1_TLB, 1);
218 0 : tmp = REG_SET_FIELD(tmp, MMMC_VM_MX_L1_TLB_CNTL, SYSTEM_ACCESS_MODE, 3);
219 0 : tmp = REG_SET_FIELD(tmp, MMMC_VM_MX_L1_TLB_CNTL,
220 : ENABLE_ADVANCED_DRIVER_MODEL, 1);
221 0 : tmp = REG_SET_FIELD(tmp, MMMC_VM_MX_L1_TLB_CNTL,
222 : SYSTEM_APERTURE_UNMAPPED_ACCESS, 0);
223 0 : tmp = REG_SET_FIELD(tmp, MMMC_VM_MX_L1_TLB_CNTL, ECO_BITS, 0);
224 0 : tmp = REG_SET_FIELD(tmp, MMMC_VM_MX_L1_TLB_CNTL,
225 : MTYPE, MTYPE_UC); /* UC, uncached */
226 :
227 0 : WREG32_SOC15(MMHUB, 0, regMMMC_VM_MX_L1_TLB_CNTL, tmp);
228 0 : }
229 :
230 0 : static void mmhub_v3_0_init_cache_regs(struct amdgpu_device *adev)
231 : {
232 : uint32_t tmp;
233 :
234 : /* These registers are not accessible to VF-SRIOV.
235 : * The PF will program them instead.
236 : */
237 0 : if (amdgpu_sriov_vf(adev))
238 : return;
239 :
240 : /* Setup L2 cache */
241 0 : tmp = RREG32_SOC15(MMHUB, 0, regMMVM_L2_CNTL);
242 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL, ENABLE_L2_CACHE, 1);
243 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL, ENABLE_L2_FRAGMENT_PROCESSING, 0);
244 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL,
245 : ENABLE_DEFAULT_PAGE_OUT_TO_SYSTEM_MEMORY, 1);
246 : /* XXX for emulation, Refer to closed source code.*/
247 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL, L2_PDE0_CACHE_TAG_GENERATION_MODE,
248 : 0);
249 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL, PDE_FAULT_CLASSIFICATION, 0);
250 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL, CONTEXT1_IDENTITY_ACCESS_MODE, 1);
251 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL, IDENTITY_MODE_FRAGMENT_SIZE, 0);
252 0 : WREG32_SOC15(MMHUB, 0, regMMVM_L2_CNTL, tmp);
253 :
254 0 : tmp = RREG32_SOC15(MMHUB, 0, regMMVM_L2_CNTL2);
255 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL2, INVALIDATE_ALL_L1_TLBS, 1);
256 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL2, INVALIDATE_L2_CACHE, 1);
257 0 : WREG32_SOC15(MMHUB, 0, regMMVM_L2_CNTL2, tmp);
258 :
259 0 : tmp = regMMVM_L2_CNTL3_DEFAULT;
260 0 : if (adev->gmc.translate_further) {
261 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL3, BANK_SELECT, 12);
262 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL3,
263 : L2_CACHE_BIGK_FRAGMENT_SIZE, 9);
264 : } else {
265 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL3, BANK_SELECT, 9);
266 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL3,
267 : L2_CACHE_BIGK_FRAGMENT_SIZE, 6);
268 : }
269 0 : WREG32_SOC15(MMHUB, 0, regMMVM_L2_CNTL3, tmp);
270 :
271 0 : tmp = regMMVM_L2_CNTL4_DEFAULT;
272 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL4, VMC_TAP_PDE_REQUEST_PHYSICAL, 0);
273 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL4, VMC_TAP_PTE_REQUEST_PHYSICAL, 0);
274 0 : WREG32_SOC15(MMHUB, 0, regMMVM_L2_CNTL4, tmp);
275 :
276 0 : tmp = regMMVM_L2_CNTL5_DEFAULT;
277 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL5, L2_CACHE_SMALLK_FRAGMENT_SIZE, 0);
278 0 : WREG32_SOC15(GC, 0, regMMVM_L2_CNTL5, tmp);
279 : }
280 :
281 0 : static void mmhub_v3_0_enable_system_domain(struct amdgpu_device *adev)
282 : {
283 : uint32_t tmp;
284 :
285 0 : tmp = RREG32_SOC15(MMHUB, 0, regMMVM_CONTEXT0_CNTL);
286 0 : tmp = REG_SET_FIELD(tmp, MMVM_CONTEXT0_CNTL, ENABLE_CONTEXT, 1);
287 0 : tmp = REG_SET_FIELD(tmp, MMVM_CONTEXT0_CNTL, PAGE_TABLE_DEPTH, 0);
288 0 : tmp = REG_SET_FIELD(tmp, MMVM_CONTEXT0_CNTL,
289 : RETRY_PERMISSION_OR_INVALID_PAGE_FAULT, 0);
290 0 : WREG32_SOC15(MMHUB, 0, regMMVM_CONTEXT0_CNTL, tmp);
291 0 : }
292 :
293 0 : static void mmhub_v3_0_disable_identity_aperture(struct amdgpu_device *adev)
294 : {
295 : /* These registers are not accessible to VF-SRIOV.
296 : * The PF will program them instead.
297 : */
298 0 : if (amdgpu_sriov_vf(adev))
299 : return;
300 :
301 0 : WREG32_SOC15(MMHUB, 0,
302 : regMMVM_L2_CONTEXT1_IDENTITY_APERTURE_LOW_ADDR_LO32,
303 : 0xFFFFFFFF);
304 0 : WREG32_SOC15(MMHUB, 0,
305 : regMMVM_L2_CONTEXT1_IDENTITY_APERTURE_LOW_ADDR_HI32,
306 : 0x0000000F);
307 :
308 0 : WREG32_SOC15(MMHUB, 0,
309 : regMMVM_L2_CONTEXT1_IDENTITY_APERTURE_HIGH_ADDR_LO32, 0);
310 0 : WREG32_SOC15(MMHUB, 0,
311 : regMMVM_L2_CONTEXT1_IDENTITY_APERTURE_HIGH_ADDR_HI32, 0);
312 :
313 0 : WREG32_SOC15(MMHUB, 0, regMMVM_L2_CONTEXT_IDENTITY_PHYSICAL_OFFSET_LO32,
314 : 0);
315 0 : WREG32_SOC15(MMHUB, 0, regMMVM_L2_CONTEXT_IDENTITY_PHYSICAL_OFFSET_HI32,
316 : 0);
317 : }
318 :
319 0 : static void mmhub_v3_0_setup_vmid_config(struct amdgpu_device *adev)
320 : {
321 0 : struct amdgpu_vmhub *hub = &adev->vmhub[AMDGPU_MMHUB_0];
322 : int i;
323 : uint32_t tmp;
324 :
325 0 : for (i = 0; i <= 14; i++) {
326 0 : tmp = RREG32_SOC15_OFFSET(MMHUB, 0, regMMVM_CONTEXT1_CNTL, i);
327 0 : tmp = REG_SET_FIELD(tmp, MMVM_CONTEXT1_CNTL, ENABLE_CONTEXT, 1);
328 0 : tmp = REG_SET_FIELD(tmp, MMVM_CONTEXT1_CNTL, PAGE_TABLE_DEPTH,
329 : adev->vm_manager.num_level);
330 0 : tmp = REG_SET_FIELD(tmp, MMVM_CONTEXT1_CNTL,
331 : RANGE_PROTECTION_FAULT_ENABLE_DEFAULT, 1);
332 0 : tmp = REG_SET_FIELD(tmp, MMVM_CONTEXT1_CNTL,
333 : DUMMY_PAGE_PROTECTION_FAULT_ENABLE_DEFAULT,
334 : 1);
335 0 : tmp = REG_SET_FIELD(tmp, MMVM_CONTEXT1_CNTL,
336 : PDE0_PROTECTION_FAULT_ENABLE_DEFAULT, 1);
337 0 : tmp = REG_SET_FIELD(tmp, MMVM_CONTEXT1_CNTL,
338 : VALID_PROTECTION_FAULT_ENABLE_DEFAULT, 1);
339 0 : tmp = REG_SET_FIELD(tmp, MMVM_CONTEXT1_CNTL,
340 : READ_PROTECTION_FAULT_ENABLE_DEFAULT, 1);
341 0 : tmp = REG_SET_FIELD(tmp, MMVM_CONTEXT1_CNTL,
342 : WRITE_PROTECTION_FAULT_ENABLE_DEFAULT, 1);
343 0 : tmp = REG_SET_FIELD(tmp, MMVM_CONTEXT1_CNTL,
344 : EXECUTE_PROTECTION_FAULT_ENABLE_DEFAULT, 1);
345 0 : tmp = REG_SET_FIELD(tmp, MMVM_CONTEXT1_CNTL,
346 : PAGE_TABLE_BLOCK_SIZE,
347 : adev->vm_manager.block_size - 9);
348 : /* Send no-retry XNACK on fault to suppress VM fault storm. */
349 0 : tmp = REG_SET_FIELD(tmp, MMVM_CONTEXT1_CNTL,
350 : RETRY_PERMISSION_OR_INVALID_PAGE_FAULT,
351 : !amdgpu_noretry);
352 0 : WREG32_SOC15_OFFSET(MMHUB, 0, regMMVM_CONTEXT1_CNTL,
353 : i * hub->ctx_distance, tmp);
354 0 : WREG32_SOC15_OFFSET(MMHUB, 0, regMMVM_CONTEXT1_PAGE_TABLE_START_ADDR_LO32,
355 : i * hub->ctx_addr_distance, 0);
356 0 : WREG32_SOC15_OFFSET(MMHUB, 0, regMMVM_CONTEXT1_PAGE_TABLE_START_ADDR_HI32,
357 : i * hub->ctx_addr_distance, 0);
358 0 : WREG32_SOC15_OFFSET(MMHUB, 0, regMMVM_CONTEXT1_PAGE_TABLE_END_ADDR_LO32,
359 : i * hub->ctx_addr_distance,
360 : lower_32_bits(adev->vm_manager.max_pfn - 1));
361 0 : WREG32_SOC15_OFFSET(MMHUB, 0, regMMVM_CONTEXT1_PAGE_TABLE_END_ADDR_HI32,
362 : i * hub->ctx_addr_distance,
363 : upper_32_bits(adev->vm_manager.max_pfn - 1));
364 : }
365 :
366 0 : hub->vm_cntx_cntl = tmp;
367 0 : }
368 :
369 0 : static void mmhub_v3_0_program_invalidation(struct amdgpu_device *adev)
370 : {
371 0 : struct amdgpu_vmhub *hub = &adev->vmhub[AMDGPU_MMHUB_0];
372 : unsigned i;
373 :
374 0 : for (i = 0; i < 18; ++i) {
375 0 : WREG32_SOC15_OFFSET(MMHUB, 0, regMMVM_INVALIDATE_ENG0_ADDR_RANGE_LO32,
376 : i * hub->eng_addr_distance, 0xffffffff);
377 0 : WREG32_SOC15_OFFSET(MMHUB, 0, regMMVM_INVALIDATE_ENG0_ADDR_RANGE_HI32,
378 : i * hub->eng_addr_distance, 0x1f);
379 : }
380 0 : }
381 :
382 0 : static int mmhub_v3_0_gart_enable(struct amdgpu_device *adev)
383 : {
384 : /* GART Enable. */
385 0 : mmhub_v3_0_init_gart_aperture_regs(adev);
386 0 : mmhub_v3_0_init_system_aperture_regs(adev);
387 0 : mmhub_v3_0_init_tlb_regs(adev);
388 0 : mmhub_v3_0_init_cache_regs(adev);
389 :
390 0 : mmhub_v3_0_enable_system_domain(adev);
391 0 : mmhub_v3_0_disable_identity_aperture(adev);
392 0 : mmhub_v3_0_setup_vmid_config(adev);
393 0 : mmhub_v3_0_program_invalidation(adev);
394 :
395 0 : return 0;
396 : }
397 :
398 0 : static void mmhub_v3_0_gart_disable(struct amdgpu_device *adev)
399 : {
400 0 : struct amdgpu_vmhub *hub = &adev->vmhub[AMDGPU_MMHUB_0];
401 : u32 tmp;
402 : u32 i;
403 :
404 : /* Disable all tables */
405 0 : for (i = 0; i < 16; i++)
406 0 : WREG32_SOC15_OFFSET(MMHUB, 0, regMMVM_CONTEXT0_CNTL,
407 : i * hub->ctx_distance, 0);
408 :
409 : /* Setup TLB control */
410 0 : tmp = RREG32_SOC15(MMHUB, 0, regMMMC_VM_MX_L1_TLB_CNTL);
411 0 : tmp = REG_SET_FIELD(tmp, MMMC_VM_MX_L1_TLB_CNTL, ENABLE_L1_TLB, 0);
412 0 : tmp = REG_SET_FIELD(tmp, MMMC_VM_MX_L1_TLB_CNTL,
413 : ENABLE_ADVANCED_DRIVER_MODEL, 0);
414 0 : WREG32_SOC15(MMHUB, 0, regMMMC_VM_MX_L1_TLB_CNTL, tmp);
415 :
416 : /* Setup L2 cache */
417 0 : tmp = RREG32_SOC15(MMHUB, 0, regMMVM_L2_CNTL);
418 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_CNTL, ENABLE_L2_CACHE, 0);
419 0 : WREG32_SOC15(MMHUB, 0, regMMVM_L2_CNTL, tmp);
420 0 : WREG32_SOC15(MMHUB, 0, regMMVM_L2_CNTL3, 0);
421 0 : }
422 :
423 : /**
424 : * mmhub_v3_0_set_fault_enable_default - update GART/VM fault handling
425 : *
426 : * @adev: amdgpu_device pointer
427 : * @value: true redirects VM faults to the default page
428 : */
429 0 : static void mmhub_v3_0_set_fault_enable_default(struct amdgpu_device *adev, bool value)
430 : {
431 : u32 tmp;
432 :
433 : /* These registers are not accessible to VF-SRIOV.
434 : * The PF will program them instead.
435 : */
436 0 : if (amdgpu_sriov_vf(adev))
437 : return;
438 :
439 0 : tmp = RREG32_SOC15(MMHUB, 0, regMMVM_L2_PROTECTION_FAULT_CNTL);
440 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_PROTECTION_FAULT_CNTL,
441 : RANGE_PROTECTION_FAULT_ENABLE_DEFAULT, value);
442 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_PROTECTION_FAULT_CNTL,
443 : PDE0_PROTECTION_FAULT_ENABLE_DEFAULT, value);
444 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_PROTECTION_FAULT_CNTL,
445 : PDE1_PROTECTION_FAULT_ENABLE_DEFAULT, value);
446 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_PROTECTION_FAULT_CNTL,
447 : PDE2_PROTECTION_FAULT_ENABLE_DEFAULT, value);
448 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_PROTECTION_FAULT_CNTL,
449 : TRANSLATE_FURTHER_PROTECTION_FAULT_ENABLE_DEFAULT,
450 : value);
451 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_PROTECTION_FAULT_CNTL,
452 : NACK_PROTECTION_FAULT_ENABLE_DEFAULT, value);
453 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_PROTECTION_FAULT_CNTL,
454 : DUMMY_PAGE_PROTECTION_FAULT_ENABLE_DEFAULT, value);
455 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_PROTECTION_FAULT_CNTL,
456 : VALID_PROTECTION_FAULT_ENABLE_DEFAULT, value);
457 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_PROTECTION_FAULT_CNTL,
458 : READ_PROTECTION_FAULT_ENABLE_DEFAULT, value);
459 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_PROTECTION_FAULT_CNTL,
460 : WRITE_PROTECTION_FAULT_ENABLE_DEFAULT, value);
461 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_PROTECTION_FAULT_CNTL,
462 : EXECUTE_PROTECTION_FAULT_ENABLE_DEFAULT, value);
463 0 : if (!value) {
464 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_PROTECTION_FAULT_CNTL,
465 : CRASH_ON_NO_RETRY_FAULT, 1);
466 0 : tmp = REG_SET_FIELD(tmp, MMVM_L2_PROTECTION_FAULT_CNTL,
467 : CRASH_ON_RETRY_FAULT, 1);
468 : }
469 0 : WREG32_SOC15(MMHUB, 0, regMMVM_L2_PROTECTION_FAULT_CNTL, tmp);
470 : }
471 :
472 : static const struct amdgpu_vmhub_funcs mmhub_v3_0_vmhub_funcs = {
473 : .print_l2_protection_fault_status = mmhub_v3_0_print_l2_protection_fault_status,
474 : .get_invalidate_req = mmhub_v3_0_get_invalidate_req,
475 : };
476 :
477 0 : static void mmhub_v3_0_init(struct amdgpu_device *adev)
478 : {
479 0 : struct amdgpu_vmhub *hub = &adev->vmhub[AMDGPU_MMHUB_0];
480 :
481 0 : hub->ctx0_ptb_addr_lo32 =
482 0 : SOC15_REG_OFFSET(MMHUB, 0,
483 : regMMVM_CONTEXT0_PAGE_TABLE_BASE_ADDR_LO32);
484 0 : hub->ctx0_ptb_addr_hi32 =
485 0 : SOC15_REG_OFFSET(MMHUB, 0,
486 : regMMVM_CONTEXT0_PAGE_TABLE_BASE_ADDR_HI32);
487 0 : hub->vm_inv_eng0_sem =
488 0 : SOC15_REG_OFFSET(MMHUB, 0, regMMVM_INVALIDATE_ENG0_SEM);
489 0 : hub->vm_inv_eng0_req =
490 0 : SOC15_REG_OFFSET(MMHUB, 0, regMMVM_INVALIDATE_ENG0_REQ);
491 0 : hub->vm_inv_eng0_ack =
492 0 : SOC15_REG_OFFSET(MMHUB, 0, regMMVM_INVALIDATE_ENG0_ACK);
493 0 : hub->vm_context0_cntl =
494 0 : SOC15_REG_OFFSET(MMHUB, 0, regMMVM_CONTEXT0_CNTL);
495 0 : hub->vm_l2_pro_fault_status =
496 0 : SOC15_REG_OFFSET(MMHUB, 0, regMMVM_L2_PROTECTION_FAULT_STATUS);
497 0 : hub->vm_l2_pro_fault_cntl =
498 0 : SOC15_REG_OFFSET(MMHUB, 0, regMMVM_L2_PROTECTION_FAULT_CNTL);
499 :
500 0 : hub->ctx_distance = regMMVM_CONTEXT1_CNTL - regMMVM_CONTEXT0_CNTL;
501 0 : hub->ctx_addr_distance = regMMVM_CONTEXT1_PAGE_TABLE_BASE_ADDR_LO32 -
502 : regMMVM_CONTEXT0_PAGE_TABLE_BASE_ADDR_LO32;
503 0 : hub->eng_distance = regMMVM_INVALIDATE_ENG1_REQ -
504 : regMMVM_INVALIDATE_ENG0_REQ;
505 0 : hub->eng_addr_distance = regMMVM_INVALIDATE_ENG1_ADDR_RANGE_LO32 -
506 : regMMVM_INVALIDATE_ENG0_ADDR_RANGE_LO32;
507 :
508 0 : hub->vm_cntx_cntl_vm_fault = MMVM_CONTEXT1_CNTL__RANGE_PROTECTION_FAULT_ENABLE_INTERRUPT_MASK |
509 : MMVM_CONTEXT1_CNTL__DUMMY_PAGE_PROTECTION_FAULT_ENABLE_INTERRUPT_MASK |
510 : MMVM_CONTEXT1_CNTL__PDE0_PROTECTION_FAULT_ENABLE_INTERRUPT_MASK |
511 : MMVM_CONTEXT1_CNTL__VALID_PROTECTION_FAULT_ENABLE_INTERRUPT_MASK |
512 : MMVM_CONTEXT1_CNTL__READ_PROTECTION_FAULT_ENABLE_INTERRUPT_MASK |
513 : MMVM_CONTEXT1_CNTL__WRITE_PROTECTION_FAULT_ENABLE_INTERRUPT_MASK |
514 : MMVM_CONTEXT1_CNTL__EXECUTE_PROTECTION_FAULT_ENABLE_INTERRUPT_MASK;
515 :
516 0 : hub->vm_l2_bank_select_reserved_cid2 =
517 0 : SOC15_REG_OFFSET(MMHUB, 0, regMMVM_L2_BANK_SELECT_RESERVED_CID2);
518 :
519 0 : hub->vmhub_funcs = &mmhub_v3_0_vmhub_funcs;
520 0 : }
521 :
522 0 : static u64 mmhub_v3_0_get_fb_location(struct amdgpu_device *adev)
523 : {
524 : u64 base;
525 :
526 0 : base = RREG32_SOC15(MMHUB, 0, regMMMC_VM_FB_LOCATION_BASE);
527 :
528 0 : base &= MMMC_VM_FB_LOCATION_BASE__FB_BASE_MASK;
529 0 : base <<= 24;
530 :
531 0 : return base;
532 : }
533 :
534 0 : static u64 mmhub_v3_0_get_mc_fb_offset(struct amdgpu_device *adev)
535 : {
536 0 : return (u64)RREG32_SOC15(MMHUB, 0, regMMMC_VM_FB_OFFSET) << 24;
537 : }
538 :
539 0 : static void mmhub_v3_0_update_medium_grain_clock_gating(struct amdgpu_device *adev,
540 : bool enable)
541 : {
542 : uint32_t def, data;
543 : #if 0
544 : uint32_t def1, data1, def2 = 0, data2 = 0;
545 : #endif
546 :
547 0 : def = data = RREG32_SOC15(MMHUB, 0, regMM_ATC_L2_MISC_CG);
548 : #if 0
549 : def1 = data1 = RREG32_SOC15(MMHUB, 0, regDAGB0_CNTL_MISC2);
550 : def2 = data2 = RREG32_SOC15(MMHUB, 0, regDAGB1_CNTL_MISC2);
551 : #endif
552 :
553 0 : if (enable) {
554 0 : data |= MM_ATC_L2_MISC_CG__ENABLE_MASK;
555 : #if 0
556 : data1 &= ~(DAGB0_CNTL_MISC2__DISABLE_WRREQ_CG_MASK |
557 : DAGB0_CNTL_MISC2__DISABLE_WRRET_CG_MASK |
558 : DAGB0_CNTL_MISC2__DISABLE_RDREQ_CG_MASK |
559 : DAGB0_CNTL_MISC2__DISABLE_RDRET_CG_MASK |
560 : DAGB0_CNTL_MISC2__DISABLE_TLBWR_CG_MASK |
561 : DAGB0_CNTL_MISC2__DISABLE_TLBRD_CG_MASK);
562 :
563 : data2 &= ~(DAGB1_CNTL_MISC2__DISABLE_WRREQ_CG_MASK |
564 : DAGB1_CNTL_MISC2__DISABLE_WRRET_CG_MASK |
565 : DAGB1_CNTL_MISC2__DISABLE_RDREQ_CG_MASK |
566 : DAGB1_CNTL_MISC2__DISABLE_RDRET_CG_MASK |
567 : DAGB1_CNTL_MISC2__DISABLE_TLBWR_CG_MASK |
568 : DAGB1_CNTL_MISC2__DISABLE_TLBRD_CG_MASK);
569 : #endif
570 : } else {
571 0 : data &= ~MM_ATC_L2_MISC_CG__ENABLE_MASK;
572 : #if 0
573 : data1 |= (DAGB0_CNTL_MISC2__DISABLE_WRREQ_CG_MASK |
574 : DAGB0_CNTL_MISC2__DISABLE_WRRET_CG_MASK |
575 : DAGB0_CNTL_MISC2__DISABLE_RDREQ_CG_MASK |
576 : DAGB0_CNTL_MISC2__DISABLE_RDRET_CG_MASK |
577 : DAGB0_CNTL_MISC2__DISABLE_TLBWR_CG_MASK |
578 : DAGB0_CNTL_MISC2__DISABLE_TLBRD_CG_MASK);
579 :
580 : data2 |= (DAGB1_CNTL_MISC2__DISABLE_WRREQ_CG_MASK |
581 : DAGB1_CNTL_MISC2__DISABLE_WRRET_CG_MASK |
582 : DAGB1_CNTL_MISC2__DISABLE_RDREQ_CG_MASK |
583 : DAGB1_CNTL_MISC2__DISABLE_RDRET_CG_MASK |
584 : DAGB1_CNTL_MISC2__DISABLE_TLBWR_CG_MASK |
585 : DAGB1_CNTL_MISC2__DISABLE_TLBRD_CG_MASK);
586 : #endif
587 : }
588 :
589 0 : if (def != data)
590 0 : WREG32_SOC15(MMHUB, 0, regMM_ATC_L2_MISC_CG, data);
591 : #if 0
592 : if (def1 != data1)
593 : WREG32_SOC15(MMHUB, 0, regDAGB0_CNTL_MISC2, data1);
594 :
595 : if (def2 != data2)
596 : WREG32_SOC15(MMHUB, 0, regDAGB1_CNTL_MISC2, data2);
597 : #endif
598 0 : }
599 :
600 0 : static void mmhub_v3_0_update_medium_grain_light_sleep(struct amdgpu_device *adev,
601 : bool enable)
602 : {
603 : uint32_t def, data;
604 :
605 0 : def = data = RREG32_SOC15(MMHUB, 0, regMM_ATC_L2_MISC_CG);
606 :
607 0 : if (enable)
608 0 : data |= MM_ATC_L2_MISC_CG__MEM_LS_ENABLE_MASK;
609 : else
610 0 : data &= ~MM_ATC_L2_MISC_CG__MEM_LS_ENABLE_MASK;
611 :
612 0 : if (def != data)
613 0 : WREG32_SOC15(MMHUB, 0, regMM_ATC_L2_MISC_CG, data);
614 0 : }
615 :
616 0 : static int mmhub_v3_0_set_clockgating(struct amdgpu_device *adev,
617 : enum amd_clockgating_state state)
618 : {
619 0 : if (amdgpu_sriov_vf(adev))
620 : return 0;
621 :
622 0 : if (adev->cg_flags & AMD_CG_SUPPORT_MC_MGCG)
623 0 : mmhub_v3_0_update_medium_grain_clock_gating(adev,
624 : state == AMD_CG_STATE_GATE);
625 :
626 0 : if (adev->cg_flags & AMD_CG_SUPPORT_MC_LS)
627 0 : mmhub_v3_0_update_medium_grain_light_sleep(adev,
628 : state == AMD_CG_STATE_GATE);
629 :
630 : return 0;
631 : }
632 :
633 0 : static void mmhub_v3_0_get_clockgating(struct amdgpu_device *adev, u64 *flags)
634 : {
635 : int data;
636 :
637 0 : if (amdgpu_sriov_vf(adev))
638 0 : *flags = 0;
639 :
640 0 : data = RREG32_SOC15(MMHUB, 0, regMM_ATC_L2_MISC_CG);
641 :
642 : /* AMD_CG_SUPPORT_MC_MGCG */
643 0 : if (data & MM_ATC_L2_MISC_CG__ENABLE_MASK)
644 0 : *flags |= AMD_CG_SUPPORT_MC_MGCG;
645 :
646 : /* AMD_CG_SUPPORT_MC_LS */
647 0 : if (data & MM_ATC_L2_MISC_CG__MEM_LS_ENABLE_MASK)
648 0 : *flags |= AMD_CG_SUPPORT_MC_LS;
649 0 : }
650 :
651 : const struct amdgpu_mmhub_funcs mmhub_v3_0_funcs = {
652 : .init = mmhub_v3_0_init,
653 : .get_fb_location = mmhub_v3_0_get_fb_location,
654 : .get_mc_fb_offset = mmhub_v3_0_get_mc_fb_offset,
655 : .gart_enable = mmhub_v3_0_gart_enable,
656 : .set_fault_enable_default = mmhub_v3_0_set_fault_enable_default,
657 : .gart_disable = mmhub_v3_0_gart_disable,
658 : .set_clockgating = mmhub_v3_0_set_clockgating,
659 : .get_clockgating = mmhub_v3_0_get_clockgating,
660 : .setup_vm_pt_regs = mmhub_v3_0_setup_vm_pt_regs,
661 : };
|