Linux-libre 5.3-gnu
[librecmc/linux-libre.git] / drivers / gpu / drm / nouveau / nvkm / falcon / v1.c
1 /*
2  * Copyright (c) 2016, NVIDIA CORPORATION. All rights reserved.
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice shall be included in
12  * all copies or substantial portions of the Software.
13  *
14  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
18  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
19  * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
20  * DEALINGS IN THE SOFTWARE.
21  */
22 #include "priv.h"
23
24 #include <core/gpuobj.h>
25 #include <core/memory.h>
26 #include <subdev/timer.h>
27
28 static void
29 nvkm_falcon_v1_load_imem(struct nvkm_falcon *falcon, void *data, u32 start,
30                          u32 size, u16 tag, u8 port, bool secure)
31 {
32         u8 rem = size % 4;
33         u32 reg;
34         int i;
35
36         size -= rem;
37
38         reg = start | BIT(24) | (secure ? BIT(28) : 0);
39         nvkm_falcon_wr32(falcon, 0x180 + (port * 16), reg);
40         for (i = 0; i < size / 4; i++) {
41                 /* write new tag every 256B */
42                 if ((i & 0x3f) == 0)
43                         nvkm_falcon_wr32(falcon, 0x188 + (port * 16), tag++);
44                 nvkm_falcon_wr32(falcon, 0x184 + (port * 16), ((u32 *)data)[i]);
45         }
46
47         /*
48          * If size is not a multiple of 4, mask the last work to ensure garbage
49          * does not get written
50          */
51         if (rem) {
52                 u32 extra = ((u32 *)data)[i];
53
54                 /* write new tag every 256B */
55                 if ((i & 0x3f) == 0)
56                         nvkm_falcon_wr32(falcon, 0x188 + (port * 16), tag++);
57                 nvkm_falcon_wr32(falcon, 0x184 + (port * 16),
58                                  extra & (BIT(rem * 8) - 1));
59                 ++i;
60         }
61
62         /* code must be padded to 0x40 words */
63         for (; i & 0x3f; i++)
64                 nvkm_falcon_wr32(falcon, 0x184 + (port * 16), 0);
65 }
66
67 static void
68 nvkm_falcon_v1_load_emem(struct nvkm_falcon *falcon, void *data, u32 start,
69                          u32 size, u8 port)
70 {
71         u8 rem = size % 4;
72         int i;
73
74         size -= rem;
75
76         nvkm_falcon_wr32(falcon, 0xac0 + (port * 8), start | (0x1 << 24));
77         for (i = 0; i < size / 4; i++)
78                 nvkm_falcon_wr32(falcon, 0xac4 + (port * 8), ((u32 *)data)[i]);
79
80         /*
81          * If size is not a multiple of 4, mask the last word to ensure garbage
82          * does not get written
83          */
84         if (rem) {
85                 u32 extra = ((u32 *)data)[i];
86
87                 nvkm_falcon_wr32(falcon, 0xac4 + (port * 8),
88                                  extra & (BIT(rem * 8) - 1));
89         }
90 }
91
92 static const u32 EMEM_START_ADDR = 0x1000000;
93
94 static void
95 nvkm_falcon_v1_load_dmem(struct nvkm_falcon *falcon, void *data, u32 start,
96                       u32 size, u8 port)
97 {
98         u8 rem = size % 4;
99         int i;
100
101         if (start >= EMEM_START_ADDR && falcon->has_emem)
102                 return nvkm_falcon_v1_load_emem(falcon, data,
103                                                 start - EMEM_START_ADDR, size,
104                                                 port);
105
106         size -= rem;
107
108         nvkm_falcon_wr32(falcon, 0x1c0 + (port * 8), start | (0x1 << 24));
109         for (i = 0; i < size / 4; i++)
110                 nvkm_falcon_wr32(falcon, 0x1c4 + (port * 8), ((u32 *)data)[i]);
111
112         /*
113          * If size is not a multiple of 4, mask the last word to ensure garbage
114          * does not get written
115          */
116         if (rem) {
117                 u32 extra = ((u32 *)data)[i];
118
119                 nvkm_falcon_wr32(falcon, 0x1c4 + (port * 8),
120                                  extra & (BIT(rem * 8) - 1));
121         }
122 }
123
124 static void
125 nvkm_falcon_v1_read_emem(struct nvkm_falcon *falcon, u32 start, u32 size,
126                          u8 port, void *data)
127 {
128         u8 rem = size % 4;
129         int i;
130
131         size -= rem;
132
133         nvkm_falcon_wr32(falcon, 0xac0 + (port * 8), start | (0x1 << 25));
134         for (i = 0; i < size / 4; i++)
135                 ((u32 *)data)[i] = nvkm_falcon_rd32(falcon, 0xac4 + (port * 8));
136
137         /*
138          * If size is not a multiple of 4, mask the last word to ensure garbage
139          * does not get read
140          */
141         if (rem) {
142                 u32 extra = nvkm_falcon_rd32(falcon, 0xac4 + (port * 8));
143
144                 for (i = size; i < size + rem; i++) {
145                         ((u8 *)data)[i] = (u8)(extra & 0xff);
146                         extra >>= 8;
147                 }
148         }
149 }
150
151 static void
152 nvkm_falcon_v1_read_dmem(struct nvkm_falcon *falcon, u32 start, u32 size,
153                          u8 port, void *data)
154 {
155         u8 rem = size % 4;
156         int i;
157
158         if (start >= EMEM_START_ADDR && falcon->has_emem)
159                 return nvkm_falcon_v1_read_emem(falcon, start - EMEM_START_ADDR,
160                                                 size, port, data);
161
162         size -= rem;
163
164         nvkm_falcon_wr32(falcon, 0x1c0 + (port * 8), start | (0x1 << 25));
165         for (i = 0; i < size / 4; i++)
166                 ((u32 *)data)[i] = nvkm_falcon_rd32(falcon, 0x1c4 + (port * 8));
167
168         /*
169          * If size is not a multiple of 4, mask the last word to ensure garbage
170          * does not get read
171          */
172         if (rem) {
173                 u32 extra = nvkm_falcon_rd32(falcon, 0x1c4 + (port * 8));
174
175                 for (i = size; i < size + rem; i++) {
176                         ((u8 *)data)[i] = (u8)(extra & 0xff);
177                         extra >>= 8;
178                 }
179         }
180 }
181
182 static void
183 nvkm_falcon_v1_bind_context(struct nvkm_falcon *falcon, struct nvkm_memory *ctx)
184 {
185         struct nvkm_device *device = falcon->owner->device;
186         u32 inst_loc;
187         u32 fbif;
188
189         /* disable instance block binding */
190         if (ctx == NULL) {
191                 nvkm_falcon_wr32(falcon, 0x10c, 0x0);
192                 return;
193         }
194
195         switch (falcon->owner->index) {
196         case NVKM_ENGINE_NVENC0:
197         case NVKM_ENGINE_NVENC1:
198         case NVKM_ENGINE_NVENC2:
199                 fbif = 0x800;
200                 break;
201         case NVKM_SUBDEV_PMU:
202                 fbif = 0xe00;
203                 break;
204         default:
205                 fbif = 0x600;
206                 break;
207         }
208
209         nvkm_falcon_wr32(falcon, 0x10c, 0x1);
210
211         /* setup apertures - virtual */
212         nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_UCODE, 0x4);
213         nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_VIRT, 0x0);
214         /* setup apertures - physical */
215         nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_VID, 0x4);
216         nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_SYS_COH, 0x5);
217         nvkm_falcon_wr32(falcon, fbif + 4 * FALCON_DMAIDX_PHYS_SYS_NCOH, 0x6);
218
219         /* Set context */
220         switch (nvkm_memory_target(ctx)) {
221         case NVKM_MEM_TARGET_VRAM: inst_loc = 0; break;
222         case NVKM_MEM_TARGET_HOST: inst_loc = 2; break;
223         case NVKM_MEM_TARGET_NCOH: inst_loc = 3; break;
224         default:
225                 WARN_ON(1);
226                 return;
227         }
228
229         /* Enable context */
230         nvkm_falcon_mask(falcon, 0x048, 0x1, 0x1);
231         nvkm_falcon_wr32(falcon, 0x054,
232                          ((nvkm_memory_addr(ctx) >> 12) & 0xfffffff) |
233                          (inst_loc << 28) | (1 << 30));
234
235         nvkm_falcon_mask(falcon, 0x090, 0x10000, 0x10000);
236         nvkm_falcon_mask(falcon, 0x0a4, 0x8, 0x8);
237
238         /* Not sure if this is a WAR for a HW issue, or some additional
239          * programming sequence that's needed to properly complete the
240          * context switch we trigger above.
241          *
242          * Fixes unreliability of booting the SEC2 RTOS on Quadro P620,
243          * particularly when resuming from suspend.
244          *
245          * Also removes the need for an odd workaround where we needed
246          * to program SEC2's FALCON_CPUCTL_ALIAS_STARTCPU twice before
247          * the SEC2 RTOS would begin executing.
248          */
249         switch (falcon->owner->index) {
250         case NVKM_SUBDEV_GSP:
251         case NVKM_ENGINE_SEC2:
252                 nvkm_msec(device, 10,
253                         u32 irqstat = nvkm_falcon_rd32(falcon, 0x008);
254                         u32 flcn0dc = nvkm_falcon_rd32(falcon, 0x0dc);
255                         if ((irqstat & 0x00000008) &&
256                             (flcn0dc & 0x00007000) == 0x00005000)
257                                 break;
258                 );
259
260                 nvkm_falcon_mask(falcon, 0x004, 0x00000008, 0x00000008);
261                 nvkm_falcon_mask(falcon, 0x058, 0x00000002, 0x00000002);
262
263                 nvkm_msec(device, 10,
264                         u32 flcn0dc = nvkm_falcon_rd32(falcon, 0x0dc);
265                         if ((flcn0dc & 0x00007000) == 0x00000000)
266                                 break;
267                 );
268                 break;
269         default:
270                 break;
271         }
272 }
273
274 static void
275 nvkm_falcon_v1_set_start_addr(struct nvkm_falcon *falcon, u32 start_addr)
276 {
277         nvkm_falcon_wr32(falcon, 0x104, start_addr);
278 }
279
280 static void
281 nvkm_falcon_v1_start(struct nvkm_falcon *falcon)
282 {
283         u32 reg = nvkm_falcon_rd32(falcon, 0x100);
284
285         if (reg & BIT(6))
286                 nvkm_falcon_wr32(falcon, 0x130, 0x2);
287         else
288                 nvkm_falcon_wr32(falcon, 0x100, 0x2);
289 }
290
291 static int
292 nvkm_falcon_v1_wait_for_halt(struct nvkm_falcon *falcon, u32 ms)
293 {
294         struct nvkm_device *device = falcon->owner->device;
295         int ret;
296
297         ret = nvkm_wait_msec(device, ms, falcon->addr + 0x100, 0x10, 0x10);
298         if (ret < 0)
299                 return ret;
300
301         return 0;
302 }
303
304 static int
305 nvkm_falcon_v1_clear_interrupt(struct nvkm_falcon *falcon, u32 mask)
306 {
307         struct nvkm_device *device = falcon->owner->device;
308         int ret;
309
310         /* clear interrupt(s) */
311         nvkm_falcon_mask(falcon, 0x004, mask, mask);
312         /* wait until interrupts are cleared */
313         ret = nvkm_wait_msec(device, 10, falcon->addr + 0x008, mask, 0x0);
314         if (ret < 0)
315                 return ret;
316
317         return 0;
318 }
319
320 static int
321 falcon_v1_wait_idle(struct nvkm_falcon *falcon)
322 {
323         struct nvkm_device *device = falcon->owner->device;
324         int ret;
325
326         ret = nvkm_wait_msec(device, 10, falcon->addr + 0x04c, 0xffff, 0x0);
327         if (ret < 0)
328                 return ret;
329
330         return 0;
331 }
332
333 static int
334 nvkm_falcon_v1_enable(struct nvkm_falcon *falcon)
335 {
336         struct nvkm_device *device = falcon->owner->device;
337         int ret;
338
339         ret = nvkm_wait_msec(device, 10, falcon->addr + 0x10c, 0x6, 0x0);
340         if (ret < 0) {
341                 nvkm_error(falcon->user, "Falcon mem scrubbing timeout\n");
342                 return ret;
343         }
344
345         ret = falcon_v1_wait_idle(falcon);
346         if (ret)
347                 return ret;
348
349         /* enable IRQs */
350         nvkm_falcon_wr32(falcon, 0x010, 0xff);
351
352         return 0;
353 }
354
355 static void
356 nvkm_falcon_v1_disable(struct nvkm_falcon *falcon)
357 {
358         /* disable IRQs and wait for any previous code to complete */
359         nvkm_falcon_wr32(falcon, 0x014, 0xff);
360         falcon_v1_wait_idle(falcon);
361 }
362
363 static const struct nvkm_falcon_func
364 nvkm_falcon_v1 = {
365         .load_imem = nvkm_falcon_v1_load_imem,
366         .load_dmem = nvkm_falcon_v1_load_dmem,
367         .read_dmem = nvkm_falcon_v1_read_dmem,
368         .bind_context = nvkm_falcon_v1_bind_context,
369         .start = nvkm_falcon_v1_start,
370         .wait_for_halt = nvkm_falcon_v1_wait_for_halt,
371         .clear_interrupt = nvkm_falcon_v1_clear_interrupt,
372         .enable = nvkm_falcon_v1_enable,
373         .disable = nvkm_falcon_v1_disable,
374         .set_start_addr = nvkm_falcon_v1_set_start_addr,
375 };
376
377 int
378 nvkm_falcon_v1_new(struct nvkm_subdev *owner, const char *name, u32 addr,
379                    struct nvkm_falcon **pfalcon)
380 {
381         struct nvkm_falcon *falcon;
382         if (!(falcon = *pfalcon = kzalloc(sizeof(*falcon), GFP_KERNEL)))
383                 return -ENOMEM;
384         nvkm_falcon_ctor(&nvkm_falcon_v1, owner, name, addr, falcon);
385         return 0;
386 }