| /* |
| * Copyright 2007 Matthieu CASTET <castet.matthieu@free.fr> |
| * All Rights Reserved. |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a |
| * copy of this software and associated documentation files (the "Software"), |
| * to deal in the Software without restriction, including without limitation |
| * the rights to use, copy, modify, merge, publish, distribute, sublicense, |
| * and/or sell copies of the Software, and to permit persons to whom the |
| * Software is furnished to do so, subject to the following conditions: |
| * |
| * The above copyright notice and this permission notice (including the next |
| * paragr) shall be included in all copies or substantial portions of the |
| * Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |
| * PRECISION INSIGHT AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR |
| * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, |
| * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER |
| * DEALINGS IN THE SOFTWARE. |
| */ |
| #include "nv10.h" |
| #include "regs.h" |
| |
| #include <core/client.h> |
| #include <core/gpuobj.h> |
| #include <engine/fifo.h> |
| #include <engine/fifo/chan.h> |
| #include <subdev/fb.h> |
| |
| struct pipe_state { |
| u32 pipe_0x0000[0x040/4]; |
| u32 pipe_0x0040[0x010/4]; |
| u32 pipe_0x0200[0x0c0/4]; |
| u32 pipe_0x4400[0x080/4]; |
| u32 pipe_0x6400[0x3b0/4]; |
| u32 pipe_0x6800[0x2f0/4]; |
| u32 pipe_0x6c00[0x030/4]; |
| u32 pipe_0x7000[0x130/4]; |
| u32 pipe_0x7400[0x0c0/4]; |
| u32 pipe_0x7800[0x0c0/4]; |
| }; |
| |
| static int nv10_gr_ctx_regs[] = { |
| NV10_PGRAPH_CTX_SWITCH(0), |
| NV10_PGRAPH_CTX_SWITCH(1), |
| NV10_PGRAPH_CTX_SWITCH(2), |
| NV10_PGRAPH_CTX_SWITCH(3), |
| NV10_PGRAPH_CTX_SWITCH(4), |
| NV10_PGRAPH_CTX_CACHE(0, 0), |
| NV10_PGRAPH_CTX_CACHE(0, 1), |
| NV10_PGRAPH_CTX_CACHE(0, 2), |
| NV10_PGRAPH_CTX_CACHE(0, 3), |
| NV10_PGRAPH_CTX_CACHE(0, 4), |
| NV10_PGRAPH_CTX_CACHE(1, 0), |
| NV10_PGRAPH_CTX_CACHE(1, 1), |
| NV10_PGRAPH_CTX_CACHE(1, 2), |
| NV10_PGRAPH_CTX_CACHE(1, 3), |
| NV10_PGRAPH_CTX_CACHE(1, 4), |
| NV10_PGRAPH_CTX_CACHE(2, 0), |
| NV10_PGRAPH_CTX_CACHE(2, 1), |
| NV10_PGRAPH_CTX_CACHE(2, 2), |
| NV10_PGRAPH_CTX_CACHE(2, 3), |
| NV10_PGRAPH_CTX_CACHE(2, 4), |
| NV10_PGRAPH_CTX_CACHE(3, 0), |
| NV10_PGRAPH_CTX_CACHE(3, 1), |
| NV10_PGRAPH_CTX_CACHE(3, 2), |
| NV10_PGRAPH_CTX_CACHE(3, 3), |
| NV10_PGRAPH_CTX_CACHE(3, 4), |
| NV10_PGRAPH_CTX_CACHE(4, 0), |
| NV10_PGRAPH_CTX_CACHE(4, 1), |
| NV10_PGRAPH_CTX_CACHE(4, 2), |
| NV10_PGRAPH_CTX_CACHE(4, 3), |
| NV10_PGRAPH_CTX_CACHE(4, 4), |
| NV10_PGRAPH_CTX_CACHE(5, 0), |
| NV10_PGRAPH_CTX_CACHE(5, 1), |
| NV10_PGRAPH_CTX_CACHE(5, 2), |
| NV10_PGRAPH_CTX_CACHE(5, 3), |
| NV10_PGRAPH_CTX_CACHE(5, 4), |
| NV10_PGRAPH_CTX_CACHE(6, 0), |
| NV10_PGRAPH_CTX_CACHE(6, 1), |
| NV10_PGRAPH_CTX_CACHE(6, 2), |
| NV10_PGRAPH_CTX_CACHE(6, 3), |
| NV10_PGRAPH_CTX_CACHE(6, 4), |
| NV10_PGRAPH_CTX_CACHE(7, 0), |
| NV10_PGRAPH_CTX_CACHE(7, 1), |
| NV10_PGRAPH_CTX_CACHE(7, 2), |
| NV10_PGRAPH_CTX_CACHE(7, 3), |
| NV10_PGRAPH_CTX_CACHE(7, 4), |
| NV10_PGRAPH_CTX_USER, |
| NV04_PGRAPH_DMA_START_0, |
| NV04_PGRAPH_DMA_START_1, |
| NV04_PGRAPH_DMA_LENGTH, |
| NV04_PGRAPH_DMA_MISC, |
| NV10_PGRAPH_DMA_PITCH, |
| NV04_PGRAPH_BOFFSET0, |
| NV04_PGRAPH_BBASE0, |
| NV04_PGRAPH_BLIMIT0, |
| NV04_PGRAPH_BOFFSET1, |
| NV04_PGRAPH_BBASE1, |
| NV04_PGRAPH_BLIMIT1, |
| NV04_PGRAPH_BOFFSET2, |
| NV04_PGRAPH_BBASE2, |
| NV04_PGRAPH_BLIMIT2, |
| NV04_PGRAPH_BOFFSET3, |
| NV04_PGRAPH_BBASE3, |
| NV04_PGRAPH_BLIMIT3, |
| NV04_PGRAPH_BOFFSET4, |
| NV04_PGRAPH_BBASE4, |
| NV04_PGRAPH_BLIMIT4, |
| NV04_PGRAPH_BOFFSET5, |
| NV04_PGRAPH_BBASE5, |
| NV04_PGRAPH_BLIMIT5, |
| NV04_PGRAPH_BPITCH0, |
| NV04_PGRAPH_BPITCH1, |
| NV04_PGRAPH_BPITCH2, |
| NV04_PGRAPH_BPITCH3, |
| NV04_PGRAPH_BPITCH4, |
| NV10_PGRAPH_SURFACE, |
| NV10_PGRAPH_STATE, |
| NV04_PGRAPH_BSWIZZLE2, |
| NV04_PGRAPH_BSWIZZLE5, |
| NV04_PGRAPH_BPIXEL, |
| NV10_PGRAPH_NOTIFY, |
| NV04_PGRAPH_PATT_COLOR0, |
| NV04_PGRAPH_PATT_COLOR1, |
| NV04_PGRAPH_PATT_COLORRAM, /* 64 values from 0x400900 to 0x4009fc */ |
| 0x00400904, |
| 0x00400908, |
| 0x0040090c, |
| 0x00400910, |
| 0x00400914, |
| 0x00400918, |
| 0x0040091c, |
| 0x00400920, |
| 0x00400924, |
| 0x00400928, |
| 0x0040092c, |
| 0x00400930, |
| 0x00400934, |
| 0x00400938, |
| 0x0040093c, |
| 0x00400940, |
| 0x00400944, |
| 0x00400948, |
| 0x0040094c, |
| 0x00400950, |
| 0x00400954, |
| 0x00400958, |
| 0x0040095c, |
| 0x00400960, |
| 0x00400964, |
| 0x00400968, |
| 0x0040096c, |
| 0x00400970, |
| 0x00400974, |
| 0x00400978, |
| 0x0040097c, |
| 0x00400980, |
| 0x00400984, |
| 0x00400988, |
| 0x0040098c, |
| 0x00400990, |
| 0x00400994, |
| 0x00400998, |
| 0x0040099c, |
| 0x004009a0, |
| 0x004009a4, |
| 0x004009a8, |
| 0x004009ac, |
| 0x004009b0, |
| 0x004009b4, |
| 0x004009b8, |
| 0x004009bc, |
| 0x004009c0, |
| 0x004009c4, |
| 0x004009c8, |
| 0x004009cc, |
| 0x004009d0, |
| 0x004009d4, |
| 0x004009d8, |
| 0x004009dc, |
| 0x004009e0, |
| 0x004009e4, |
| 0x004009e8, |
| 0x004009ec, |
| 0x004009f0, |
| 0x004009f4, |
| 0x004009f8, |
| 0x004009fc, |
| NV04_PGRAPH_PATTERN, /* 2 values from 0x400808 to 0x40080c */ |
| 0x0040080c, |
| NV04_PGRAPH_PATTERN_SHAPE, |
| NV03_PGRAPH_MONO_COLOR0, |
| NV04_PGRAPH_ROP3, |
| NV04_PGRAPH_CHROMA, |
| NV04_PGRAPH_BETA_AND, |
| NV04_PGRAPH_BETA_PREMULT, |
| 0x00400e70, |
| 0x00400e74, |
| 0x00400e78, |
| 0x00400e7c, |
| 0x00400e80, |
| 0x00400e84, |
| 0x00400e88, |
| 0x00400e8c, |
| 0x00400ea0, |
| 0x00400ea4, |
| 0x00400ea8, |
| 0x00400e90, |
| 0x00400e94, |
| 0x00400e98, |
| 0x00400e9c, |
| NV10_PGRAPH_WINDOWCLIP_HORIZONTAL, /* 8 values from 0x400f00-0x400f1c */ |
| NV10_PGRAPH_WINDOWCLIP_VERTICAL, /* 8 values from 0x400f20-0x400f3c */ |
| 0x00400f04, |
| 0x00400f24, |
| 0x00400f08, |
| 0x00400f28, |
| 0x00400f0c, |
| 0x00400f2c, |
| 0x00400f10, |
| 0x00400f30, |
| 0x00400f14, |
| 0x00400f34, |
| 0x00400f18, |
| 0x00400f38, |
| 0x00400f1c, |
| 0x00400f3c, |
| NV10_PGRAPH_XFMODE0, |
| NV10_PGRAPH_XFMODE1, |
| NV10_PGRAPH_GLOBALSTATE0, |
| NV10_PGRAPH_GLOBALSTATE1, |
| NV04_PGRAPH_STORED_FMT, |
| NV04_PGRAPH_SOURCE_COLOR, |
| NV03_PGRAPH_ABS_X_RAM, /* 32 values from 0x400400 to 0x40047c */ |
| NV03_PGRAPH_ABS_Y_RAM, /* 32 values from 0x400480 to 0x4004fc */ |
| 0x00400404, |
| 0x00400484, |
| 0x00400408, |
| 0x00400488, |
| 0x0040040c, |
| 0x0040048c, |
| 0x00400410, |
| 0x00400490, |
| 0x00400414, |
| 0x00400494, |
| 0x00400418, |
| 0x00400498, |
| 0x0040041c, |
| 0x0040049c, |
| 0x00400420, |
| 0x004004a0, |
| 0x00400424, |
| 0x004004a4, |
| 0x00400428, |
| 0x004004a8, |
| 0x0040042c, |
| 0x004004ac, |
| 0x00400430, |
| 0x004004b0, |
| 0x00400434, |
| 0x004004b4, |
| 0x00400438, |
| 0x004004b8, |
| 0x0040043c, |
| 0x004004bc, |
| 0x00400440, |
| 0x004004c0, |
| 0x00400444, |
| 0x004004c4, |
| 0x00400448, |
| 0x004004c8, |
| 0x0040044c, |
| 0x004004cc, |
| 0x00400450, |
| 0x004004d0, |
| 0x00400454, |
| 0x004004d4, |
| 0x00400458, |
| 0x004004d8, |
| 0x0040045c, |
| 0x004004dc, |
| 0x00400460, |
| 0x004004e0, |
| 0x00400464, |
| 0x004004e4, |
| 0x00400468, |
| 0x004004e8, |
| 0x0040046c, |
| 0x004004ec, |
| 0x00400470, |
| 0x004004f0, |
| 0x00400474, |
| 0x004004f4, |
| 0x00400478, |
| 0x004004f8, |
| 0x0040047c, |
| 0x004004fc, |
| NV03_PGRAPH_ABS_UCLIP_XMIN, |
| NV03_PGRAPH_ABS_UCLIP_XMAX, |
| NV03_PGRAPH_ABS_UCLIP_YMIN, |
| NV03_PGRAPH_ABS_UCLIP_YMAX, |
| 0x00400550, |
| 0x00400558, |
| 0x00400554, |
| 0x0040055c, |
| NV03_PGRAPH_ABS_UCLIPA_XMIN, |
| NV03_PGRAPH_ABS_UCLIPA_XMAX, |
| NV03_PGRAPH_ABS_UCLIPA_YMIN, |
| NV03_PGRAPH_ABS_UCLIPA_YMAX, |
| NV03_PGRAPH_ABS_ICLIP_XMAX, |
| NV03_PGRAPH_ABS_ICLIP_YMAX, |
| NV03_PGRAPH_XY_LOGIC_MISC0, |
| NV03_PGRAPH_XY_LOGIC_MISC1, |
| NV03_PGRAPH_XY_LOGIC_MISC2, |
| NV03_PGRAPH_XY_LOGIC_MISC3, |
| NV03_PGRAPH_CLIPX_0, |
| NV03_PGRAPH_CLIPX_1, |
| NV03_PGRAPH_CLIPY_0, |
| NV03_PGRAPH_CLIPY_1, |
| NV10_PGRAPH_COMBINER0_IN_ALPHA, |
| NV10_PGRAPH_COMBINER1_IN_ALPHA, |
| NV10_PGRAPH_COMBINER0_IN_RGB, |
| NV10_PGRAPH_COMBINER1_IN_RGB, |
| NV10_PGRAPH_COMBINER_COLOR0, |
| NV10_PGRAPH_COMBINER_COLOR1, |
| NV10_PGRAPH_COMBINER0_OUT_ALPHA, |
| NV10_PGRAPH_COMBINER1_OUT_ALPHA, |
| NV10_PGRAPH_COMBINER0_OUT_RGB, |
| NV10_PGRAPH_COMBINER1_OUT_RGB, |
| NV10_PGRAPH_COMBINER_FINAL0, |
| NV10_PGRAPH_COMBINER_FINAL1, |
| 0x00400e00, |
| 0x00400e04, |
| 0x00400e08, |
| 0x00400e0c, |
| 0x00400e10, |
| 0x00400e14, |
| 0x00400e18, |
| 0x00400e1c, |
| 0x00400e20, |
| 0x00400e24, |
| 0x00400e28, |
| 0x00400e2c, |
| 0x00400e30, |
| 0x00400e34, |
| 0x00400e38, |
| 0x00400e3c, |
| NV04_PGRAPH_PASSTHRU_0, |
| NV04_PGRAPH_PASSTHRU_1, |
| NV04_PGRAPH_PASSTHRU_2, |
| NV10_PGRAPH_DIMX_TEXTURE, |
| NV10_PGRAPH_WDIMX_TEXTURE, |
| NV10_PGRAPH_DVD_COLORFMT, |
| NV10_PGRAPH_SCALED_FORMAT, |
| NV04_PGRAPH_MISC24_0, |
| NV04_PGRAPH_MISC24_1, |
| NV04_PGRAPH_MISC24_2, |
| NV03_PGRAPH_X_MISC, |
| NV03_PGRAPH_Y_MISC, |
| NV04_PGRAPH_VALID1, |
| NV04_PGRAPH_VALID2, |
| }; |
| |
| static int nv17_gr_ctx_regs[] = { |
| NV10_PGRAPH_DEBUG_4, |
| 0x004006b0, |
| 0x00400eac, |
| 0x00400eb0, |
| 0x00400eb4, |
| 0x00400eb8, |
| 0x00400ebc, |
| 0x00400ec0, |
| 0x00400ec4, |
| 0x00400ec8, |
| 0x00400ecc, |
| 0x00400ed0, |
| 0x00400ed4, |
| 0x00400ed8, |
| 0x00400edc, |
| 0x00400ee0, |
| 0x00400a00, |
| 0x00400a04, |
| }; |
| |
| #define nv10_gr(p) container_of((p), struct nv10_gr, base) |
| |
| struct nv10_gr { |
| struct nvkm_gr base; |
| struct nv10_gr_chan *chan[32]; |
| spinlock_t lock; |
| }; |
| |
| #define nv10_gr_chan(p) container_of((p), struct nv10_gr_chan, object) |
| |
| struct nv10_gr_chan { |
| struct nvkm_object object; |
| struct nv10_gr *gr; |
| int chid; |
| int nv10[ARRAY_SIZE(nv10_gr_ctx_regs)]; |
| int nv17[ARRAY_SIZE(nv17_gr_ctx_regs)]; |
| struct pipe_state pipe_state; |
| u32 lma_window[4]; |
| }; |
| |
| |
| /******************************************************************************* |
| * Graphics object classes |
| ******************************************************************************/ |
| |
| #define PIPE_SAVE(gr, state, addr) \ |
| do { \ |
| int __i; \ |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, addr); \ |
| for (__i = 0; __i < ARRAY_SIZE(state); __i++) \ |
| state[__i] = nvkm_rd32(device, NV10_PGRAPH_PIPE_DATA); \ |
| } while (0) |
| |
| #define PIPE_RESTORE(gr, state, addr) \ |
| do { \ |
| int __i; \ |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, addr); \ |
| for (__i = 0; __i < ARRAY_SIZE(state); __i++) \ |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, state[__i]); \ |
| } while (0) |
| |
| static void |
| nv17_gr_mthd_lma_window(struct nv10_gr_chan *chan, u32 mthd, u32 data) |
| { |
| struct nvkm_device *device = chan->object.engine->subdev.device; |
| struct nvkm_gr *gr = &chan->gr->base; |
| struct pipe_state *pipe = &chan->pipe_state; |
| u32 pipe_0x0040[1], pipe_0x64c0[8], pipe_0x6a80[3], pipe_0x6ab0[3]; |
| u32 xfmode0, xfmode1; |
| int i; |
| |
| chan->lma_window[(mthd - 0x1638) / 4] = data; |
| |
| if (mthd != 0x1644) |
| return; |
| |
| nv04_gr_idle(gr); |
| |
| PIPE_SAVE(device, pipe_0x0040, 0x0040); |
| PIPE_SAVE(device, pipe->pipe_0x0200, 0x0200); |
| |
| PIPE_RESTORE(device, chan->lma_window, 0x6790); |
| |
| nv04_gr_idle(gr); |
| |
| xfmode0 = nvkm_rd32(device, NV10_PGRAPH_XFMODE0); |
| xfmode1 = nvkm_rd32(device, NV10_PGRAPH_XFMODE1); |
| |
| PIPE_SAVE(device, pipe->pipe_0x4400, 0x4400); |
| PIPE_SAVE(device, pipe_0x64c0, 0x64c0); |
| PIPE_SAVE(device, pipe_0x6ab0, 0x6ab0); |
| PIPE_SAVE(device, pipe_0x6a80, 0x6a80); |
| |
| nv04_gr_idle(gr); |
| |
| nvkm_wr32(device, NV10_PGRAPH_XFMODE0, 0x10000000); |
| nvkm_wr32(device, NV10_PGRAPH_XFMODE1, 0x00000000); |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x000064c0); |
| for (i = 0; i < 4; i++) |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000); |
| for (i = 0; i < 4; i++) |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000); |
| |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006ab0); |
| for (i = 0; i < 3; i++) |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000); |
| |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006a80); |
| for (i = 0; i < 3; i++) |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000); |
| |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00000040); |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000008); |
| |
| PIPE_RESTORE(device, pipe->pipe_0x0200, 0x0200); |
| |
| nv04_gr_idle(gr); |
| |
| PIPE_RESTORE(device, pipe_0x0040, 0x0040); |
| |
| nvkm_wr32(device, NV10_PGRAPH_XFMODE0, xfmode0); |
| nvkm_wr32(device, NV10_PGRAPH_XFMODE1, xfmode1); |
| |
| PIPE_RESTORE(device, pipe_0x64c0, 0x64c0); |
| PIPE_RESTORE(device, pipe_0x6ab0, 0x6ab0); |
| PIPE_RESTORE(device, pipe_0x6a80, 0x6a80); |
| PIPE_RESTORE(device, pipe->pipe_0x4400, 0x4400); |
| |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x000000c0); |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000); |
| |
| nv04_gr_idle(gr); |
| } |
| |
| static void |
| nv17_gr_mthd_lma_enable(struct nv10_gr_chan *chan, u32 mthd, u32 data) |
| { |
| struct nvkm_device *device = chan->object.engine->subdev.device; |
| struct nvkm_gr *gr = &chan->gr->base; |
| |
| nv04_gr_idle(gr); |
| |
| nvkm_mask(device, NV10_PGRAPH_DEBUG_4, 0x00000100, 0x00000100); |
| nvkm_mask(device, 0x4006b0, 0x08000000, 0x08000000); |
| } |
| |
| static bool |
| nv17_gr_mthd_celcius(struct nv10_gr_chan *chan, u32 mthd, u32 data) |
| { |
| void (*func)(struct nv10_gr_chan *, u32, u32); |
| switch (mthd) { |
| case 0x1638 ... 0x1644: |
| func = nv17_gr_mthd_lma_window; break; |
| case 0x1658: func = nv17_gr_mthd_lma_enable; break; |
| default: |
| return false; |
| } |
| func(chan, mthd, data); |
| return true; |
| } |
| |
| static bool |
| nv10_gr_mthd(struct nv10_gr_chan *chan, u8 class, u32 mthd, u32 data) |
| { |
| bool (*func)(struct nv10_gr_chan *, u32, u32); |
| switch (class) { |
| case 0x99: func = nv17_gr_mthd_celcius; break; |
| default: |
| return false; |
| } |
| return func(chan, mthd, data); |
| } |
| |
| /******************************************************************************* |
| * PGRAPH context |
| ******************************************************************************/ |
| |
| static struct nv10_gr_chan * |
| nv10_gr_channel(struct nv10_gr *gr) |
| { |
| struct nvkm_device *device = gr->base.engine.subdev.device; |
| struct nv10_gr_chan *chan = NULL; |
| if (nvkm_rd32(device, 0x400144) & 0x00010000) { |
| int chid = nvkm_rd32(device, 0x400148) >> 24; |
| if (chid < ARRAY_SIZE(gr->chan)) |
| chan = gr->chan[chid]; |
| } |
| return chan; |
| } |
| |
| static void |
| nv10_gr_save_pipe(struct nv10_gr_chan *chan) |
| { |
| struct nv10_gr *gr = chan->gr; |
| struct pipe_state *pipe = &chan->pipe_state; |
| struct nvkm_device *device = gr->base.engine.subdev.device; |
| |
| PIPE_SAVE(gr, pipe->pipe_0x4400, 0x4400); |
| PIPE_SAVE(gr, pipe->pipe_0x0200, 0x0200); |
| PIPE_SAVE(gr, pipe->pipe_0x6400, 0x6400); |
| PIPE_SAVE(gr, pipe->pipe_0x6800, 0x6800); |
| PIPE_SAVE(gr, pipe->pipe_0x6c00, 0x6c00); |
| PIPE_SAVE(gr, pipe->pipe_0x7000, 0x7000); |
| PIPE_SAVE(gr, pipe->pipe_0x7400, 0x7400); |
| PIPE_SAVE(gr, pipe->pipe_0x7800, 0x7800); |
| PIPE_SAVE(gr, pipe->pipe_0x0040, 0x0040); |
| PIPE_SAVE(gr, pipe->pipe_0x0000, 0x0000); |
| } |
| |
| static void |
| nv10_gr_load_pipe(struct nv10_gr_chan *chan) |
| { |
| struct nv10_gr *gr = chan->gr; |
| struct pipe_state *pipe = &chan->pipe_state; |
| struct nvkm_device *device = gr->base.engine.subdev.device; |
| u32 xfmode0, xfmode1; |
| int i; |
| |
| nv04_gr_idle(&gr->base); |
| /* XXX check haiku comments */ |
| xfmode0 = nvkm_rd32(device, NV10_PGRAPH_XFMODE0); |
| xfmode1 = nvkm_rd32(device, NV10_PGRAPH_XFMODE1); |
| nvkm_wr32(device, NV10_PGRAPH_XFMODE0, 0x10000000); |
| nvkm_wr32(device, NV10_PGRAPH_XFMODE1, 0x00000000); |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x000064c0); |
| for (i = 0; i < 4; i++) |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000); |
| for (i = 0; i < 4; i++) |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000); |
| |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006ab0); |
| for (i = 0; i < 3; i++) |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x3f800000); |
| |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00006a80); |
| for (i = 0; i < 3; i++) |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000000); |
| |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_ADDRESS, 0x00000040); |
| nvkm_wr32(device, NV10_PGRAPH_PIPE_DATA, 0x00000008); |
| |
| |
| PIPE_RESTORE(gr, pipe->pipe_0x0200, 0x0200); |
| nv04_gr_idle(&gr->base); |
| |
| /* restore XFMODE */ |
| nvkm_wr32(device, NV10_PGRAPH_XFMODE0, xfmode0); |
| nvkm_wr32(device, NV10_PGRAPH_XFMODE1, xfmode1); |
| PIPE_RESTORE(gr, pipe->pipe_0x6400, 0x6400); |
| PIPE_RESTORE(gr, pipe->pipe_0x6800, 0x6800); |
| PIPE_RESTORE(gr, pipe->pipe_0x6c00, 0x6c00); |
| PIPE_RESTORE(gr, pipe->pipe_0x7000, 0x7000); |
| PIPE_RESTORE(gr, pipe->pipe_0x7400, 0x7400); |
| PIPE_RESTORE(gr, pipe->pipe_0x7800, 0x7800); |
| PIPE_RESTORE(gr, pipe->pipe_0x4400, 0x4400); |
| PIPE_RESTORE(gr, pipe->pipe_0x0000, 0x0000); |
| PIPE_RESTORE(gr, pipe->pipe_0x0040, 0x0040); |
| nv04_gr_idle(&gr->base); |
| } |
| |
| static void |
| nv10_gr_create_pipe(struct nv10_gr_chan *chan) |
| { |
| struct nv10_gr *gr = chan->gr; |
| struct nvkm_subdev *subdev = &gr->base.engine.subdev; |
| struct pipe_state *pipe_state = &chan->pipe_state; |
| u32 *pipe_state_addr; |
| int i; |
| #define PIPE_INIT(addr) \ |
| do { \ |
| pipe_state_addr = pipe_state->pipe_##addr; \ |
| } while (0) |
| #define PIPE_INIT_END(addr) \ |
| do { \ |
| u32 *__end_addr = pipe_state->pipe_##addr + \ |
| ARRAY_SIZE(pipe_state->pipe_##addr); \ |
| if (pipe_state_addr != __end_addr) \ |
| nvkm_error(subdev, "incomplete pipe init for 0x%x : %p/%p\n", \ |
| addr, pipe_state_addr, __end_addr); \ |
| } while (0) |
| #define NV_WRITE_PIPE_INIT(value) *(pipe_state_addr++) = value |
| |
| PIPE_INIT(0x0200); |
| for (i = 0; i < 48; i++) |
| NV_WRITE_PIPE_INIT(0x00000000); |
| PIPE_INIT_END(0x0200); |
| |
| PIPE_INIT(0x6400); |
| for (i = 0; i < 211; i++) |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x3f800000); |
| NV_WRITE_PIPE_INIT(0x40000000); |
| NV_WRITE_PIPE_INIT(0x40000000); |
| NV_WRITE_PIPE_INIT(0x40000000); |
| NV_WRITE_PIPE_INIT(0x40000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x3f800000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x3f000000); |
| NV_WRITE_PIPE_INIT(0x3f000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x3f800000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x3f800000); |
| NV_WRITE_PIPE_INIT(0x3f800000); |
| NV_WRITE_PIPE_INIT(0x3f800000); |
| NV_WRITE_PIPE_INIT(0x3f800000); |
| PIPE_INIT_END(0x6400); |
| |
| PIPE_INIT(0x6800); |
| for (i = 0; i < 162; i++) |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x3f800000); |
| for (i = 0; i < 25; i++) |
| NV_WRITE_PIPE_INIT(0x00000000); |
| PIPE_INIT_END(0x6800); |
| |
| PIPE_INIT(0x6c00); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0xbf800000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| PIPE_INIT_END(0x6c00); |
| |
| PIPE_INIT(0x7000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x7149f2ca); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x7149f2ca); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x7149f2ca); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x7149f2ca); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x7149f2ca); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x7149f2ca); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x7149f2ca); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x00000000); |
| NV_WRITE_PIPE_INIT(0x7149f2ca); |
| for (i = 0; i < 35; i++) |
| NV_WRITE_PIPE_INIT(0x00000000); |
| PIPE_INIT_END(0x7000); |
| |
| PIPE_INIT(0x7400); |
| for (i = 0; i < 48; i++) |
| NV_WRITE_PIPE_INIT(0x00000000); |
| PIPE_INIT_END(0x7400); |
| |
| PIPE_INIT(0x7800); |
| for (i = 0; i < 48; i++) |
| NV_WRITE_PIPE_INIT(0x00000000); |
| PIPE_INIT_END(0x7800); |
| |
| PIPE_INIT(0x4400); |
| for (i = 0; i < 32; i++) |
| NV_WRITE_PIPE_INIT(0x00000000); |
| PIPE_INIT_END(0x4400); |
| |
| PIPE_INIT(0x0000); |
| for (i = 0; i < 16; i++) |
| NV_WRITE_PIPE_INIT(0x00000000); |
| PIPE_INIT_END(0x0000); |
| |
| PIPE_INIT(0x0040); |
| for (i = 0; i < 4; i++) |
| NV_WRITE_PIPE_INIT(0x00000000); |
| PIPE_INIT_END(0x0040); |
| |
| #undef PIPE_INIT |
| #undef PIPE_INIT_END |
| #undef NV_WRITE_PIPE_INIT |
| } |
| |
| static int |
| nv10_gr_ctx_regs_find_offset(struct nv10_gr *gr, int reg) |
| { |
| struct nvkm_subdev *subdev = &gr->base.engine.subdev; |
| int i; |
| for (i = 0; i < ARRAY_SIZE(nv10_gr_ctx_regs); i++) { |
| if (nv10_gr_ctx_regs[i] == reg) |
| return i; |
| } |
| nvkm_error(subdev, "unknown offset nv10_ctx_regs %d\n", reg); |
| return -1; |
| } |
| |
| static int |
| nv17_gr_ctx_regs_find_offset(struct nv10_gr *gr, int reg) |
| { |
| struct nvkm_subdev *subdev = &gr->base.engine.subdev; |
| int i; |
| for (i = 0; i < ARRAY_SIZE(nv17_gr_ctx_regs); i++) { |
| if (nv17_gr_ctx_regs[i] == reg) |
| return i; |
| } |
| nvkm_error(subdev, "unknown offset nv17_ctx_regs %d\n", reg); |
| return -1; |
| } |
| |
| static void |
| nv10_gr_load_dma_vtxbuf(struct nv10_gr_chan *chan, int chid, u32 inst) |
| { |
| struct nv10_gr *gr = chan->gr; |
| struct nvkm_device *device = gr->base.engine.subdev.device; |
| u32 st2, st2_dl, st2_dh, fifo_ptr, fifo[0x60/4]; |
| u32 ctx_user, ctx_switch[5]; |
| int i, subchan = -1; |
| |
| /* NV10TCL_DMA_VTXBUF (method 0x18c) modifies hidden state |
| * that cannot be restored via MMIO. Do it through the FIFO |
| * instead. |
| */ |
| |
| /* Look for a celsius object */ |
| for (i = 0; i < 8; i++) { |
| int class = nvkm_rd32(device, NV10_PGRAPH_CTX_CACHE(i, 0)) & 0xfff; |
| |
| if (class == 0x56 || class == 0x96 || class == 0x99) { |
| subchan = i; |
| break; |
| } |
| } |
| |
| if (subchan < 0 || !inst) |
| return; |
| |
| /* Save the current ctx object */ |
| ctx_user = nvkm_rd32(device, NV10_PGRAPH_CTX_USER); |
| for (i = 0; i < 5; i++) |
| ctx_switch[i] = nvkm_rd32(device, NV10_PGRAPH_CTX_SWITCH(i)); |
| |
| /* Save the FIFO state */ |
| st2 = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_ST2); |
| st2_dl = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_ST2_DL); |
| st2_dh = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_ST2_DH); |
| fifo_ptr = nvkm_rd32(device, NV10_PGRAPH_FFINTFC_FIFO_PTR); |
| |
| for (i = 0; i < ARRAY_SIZE(fifo); i++) |
| fifo[i] = nvkm_rd32(device, 0x4007a0 + 4 * i); |
| |
| /* Switch to the celsius subchannel */ |
| for (i = 0; i < 5; i++) |
| nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(i), |
| nvkm_rd32(device, NV10_PGRAPH_CTX_CACHE(subchan, i))); |
| nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xe000, subchan << 13); |
| |
| /* Inject NV10TCL_DMA_VTXBUF */ |
| nvkm_wr32(device, NV10_PGRAPH_FFINTFC_FIFO_PTR, 0); |
| nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2, |
| 0x2c000000 | chid << 20 | subchan << 16 | 0x18c); |
| nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2_DL, inst); |
| nvkm_mask(device, NV10_PGRAPH_CTX_CONTROL, 0, 0x10000); |
| nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000001); |
| nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000000); |
| |
| /* Restore the FIFO state */ |
| for (i = 0; i < ARRAY_SIZE(fifo); i++) |
| nvkm_wr32(device, 0x4007a0 + 4 * i, fifo[i]); |
| |
| nvkm_wr32(device, NV10_PGRAPH_FFINTFC_FIFO_PTR, fifo_ptr); |
| nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2, st2); |
| nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2_DL, st2_dl); |
| nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2_DH, st2_dh); |
| |
| /* Restore the current ctx object */ |
| for (i = 0; i < 5; i++) |
| nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(i), ctx_switch[i]); |
| nvkm_wr32(device, NV10_PGRAPH_CTX_USER, ctx_user); |
| } |
| |
| static int |
| nv10_gr_load_context(struct nv10_gr_chan *chan, int chid) |
| { |
| struct nv10_gr *gr = chan->gr; |
| struct nvkm_device *device = gr->base.engine.subdev.device; |
| u32 inst; |
| int i; |
| |
| for (i = 0; i < ARRAY_SIZE(nv10_gr_ctx_regs); i++) |
| nvkm_wr32(device, nv10_gr_ctx_regs[i], chan->nv10[i]); |
| |
| if (device->card_type >= NV_11 && device->chipset >= 0x17) { |
| for (i = 0; i < ARRAY_SIZE(nv17_gr_ctx_regs); i++) |
| nvkm_wr32(device, nv17_gr_ctx_regs[i], chan->nv17[i]); |
| } |
| |
| nv10_gr_load_pipe(chan); |
| |
| inst = nvkm_rd32(device, NV10_PGRAPH_GLOBALSTATE1) & 0xffff; |
| nv10_gr_load_dma_vtxbuf(chan, chid, inst); |
| |
| nvkm_wr32(device, NV10_PGRAPH_CTX_CONTROL, 0x10010100); |
| nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xff000000, chid << 24); |
| nvkm_mask(device, NV10_PGRAPH_FFINTFC_ST2, 0x30000000, 0x00000000); |
| return 0; |
| } |
| |
| static int |
| nv10_gr_unload_context(struct nv10_gr_chan *chan) |
| { |
| struct nv10_gr *gr = chan->gr; |
| struct nvkm_device *device = gr->base.engine.subdev.device; |
| int i; |
| |
| for (i = 0; i < ARRAY_SIZE(nv10_gr_ctx_regs); i++) |
| chan->nv10[i] = nvkm_rd32(device, nv10_gr_ctx_regs[i]); |
| |
| if (device->card_type >= NV_11 && device->chipset >= 0x17) { |
| for (i = 0; i < ARRAY_SIZE(nv17_gr_ctx_regs); i++) |
| chan->nv17[i] = nvkm_rd32(device, nv17_gr_ctx_regs[i]); |
| } |
| |
| nv10_gr_save_pipe(chan); |
| |
| nvkm_wr32(device, NV10_PGRAPH_CTX_CONTROL, 0x10000000); |
| nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xff000000, 0x1f000000); |
| return 0; |
| } |
| |
| static void |
| nv10_gr_context_switch(struct nv10_gr *gr) |
| { |
| struct nvkm_device *device = gr->base.engine.subdev.device; |
| struct nv10_gr_chan *prev = NULL; |
| struct nv10_gr_chan *next = NULL; |
| int chid; |
| |
| nv04_gr_idle(&gr->base); |
| |
| /* If previous context is valid, we need to save it */ |
| prev = nv10_gr_channel(gr); |
| if (prev) |
| nv10_gr_unload_context(prev); |
| |
| /* load context for next channel */ |
| chid = (nvkm_rd32(device, NV04_PGRAPH_TRAPPED_ADDR) >> 20) & 0x1f; |
| next = gr->chan[chid]; |
| if (next) |
| nv10_gr_load_context(next, chid); |
| } |
| |
| static int |
| nv10_gr_chan_fini(struct nvkm_object *object, bool suspend) |
| { |
| struct nv10_gr_chan *chan = nv10_gr_chan(object); |
| struct nv10_gr *gr = chan->gr; |
| struct nvkm_device *device = gr->base.engine.subdev.device; |
| unsigned long flags; |
| |
| spin_lock_irqsave(&gr->lock, flags); |
| nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000000); |
| if (nv10_gr_channel(gr) == chan) |
| nv10_gr_unload_context(chan); |
| nvkm_mask(device, NV04_PGRAPH_FIFO, 0x00000001, 0x00000001); |
| spin_unlock_irqrestore(&gr->lock, flags); |
| return 0; |
| } |
| |
| static void * |
| nv10_gr_chan_dtor(struct nvkm_object *object) |
| { |
| struct nv10_gr_chan *chan = nv10_gr_chan(object); |
| struct nv10_gr *gr = chan->gr; |
| unsigned long flags; |
| |
| spin_lock_irqsave(&gr->lock, flags); |
| gr->chan[chan->chid] = NULL; |
| spin_unlock_irqrestore(&gr->lock, flags); |
| return chan; |
| } |
| |
| static const struct nvkm_object_func |
| nv10_gr_chan = { |
| .dtor = nv10_gr_chan_dtor, |
| .fini = nv10_gr_chan_fini, |
| }; |
| |
| #define NV_WRITE_CTX(reg, val) do { \ |
| int offset = nv10_gr_ctx_regs_find_offset(gr, reg); \ |
| if (offset > 0) \ |
| chan->nv10[offset] = val; \ |
| } while (0) |
| |
| #define NV17_WRITE_CTX(reg, val) do { \ |
| int offset = nv17_gr_ctx_regs_find_offset(gr, reg); \ |
| if (offset > 0) \ |
| chan->nv17[offset] = val; \ |
| } while (0) |
| |
| int |
| nv10_gr_chan_new(struct nvkm_gr *base, struct nvkm_fifo_chan *fifoch, |
| const struct nvkm_oclass *oclass, struct nvkm_object **pobject) |
| { |
| struct nv10_gr *gr = nv10_gr(base); |
| struct nv10_gr_chan *chan; |
| struct nvkm_device *device = gr->base.engine.subdev.device; |
| unsigned long flags; |
| |
| if (!(chan = kzalloc(sizeof(*chan), GFP_KERNEL))) |
| return -ENOMEM; |
| nvkm_object_ctor(&nv10_gr_chan, oclass, &chan->object); |
| chan->gr = gr; |
| chan->chid = fifoch->chid; |
| *pobject = &chan->object; |
| |
| NV_WRITE_CTX(0x00400e88, 0x08000000); |
| NV_WRITE_CTX(0x00400e9c, 0x4b7fffff); |
| NV_WRITE_CTX(NV03_PGRAPH_XY_LOGIC_MISC0, 0x0001ffff); |
| NV_WRITE_CTX(0x00400e10, 0x00001000); |
| NV_WRITE_CTX(0x00400e14, 0x00001000); |
| NV_WRITE_CTX(0x00400e30, 0x00080008); |
| NV_WRITE_CTX(0x00400e34, 0x00080008); |
| if (device->card_type >= NV_11 && device->chipset >= 0x17) { |
| /* is it really needed ??? */ |
| NV17_WRITE_CTX(NV10_PGRAPH_DEBUG_4, |
| nvkm_rd32(device, NV10_PGRAPH_DEBUG_4)); |
| NV17_WRITE_CTX(0x004006b0, nvkm_rd32(device, 0x004006b0)); |
| NV17_WRITE_CTX(0x00400eac, 0x0fff0000); |
| NV17_WRITE_CTX(0x00400eb0, 0x0fff0000); |
| NV17_WRITE_CTX(0x00400ec0, 0x00000080); |
| NV17_WRITE_CTX(0x00400ed0, 0x00000080); |
| } |
| NV_WRITE_CTX(NV10_PGRAPH_CTX_USER, chan->chid << 24); |
| |
| nv10_gr_create_pipe(chan); |
| |
| spin_lock_irqsave(&gr->lock, flags); |
| gr->chan[chan->chid] = chan; |
| spin_unlock_irqrestore(&gr->lock, flags); |
| return 0; |
| } |
| |
| /******************************************************************************* |
| * PGRAPH engine/subdev functions |
| ******************************************************************************/ |
| |
| void |
| nv10_gr_tile(struct nvkm_gr *base, int i, struct nvkm_fb_tile *tile) |
| { |
| struct nv10_gr *gr = nv10_gr(base); |
| struct nvkm_device *device = gr->base.engine.subdev.device; |
| struct nvkm_fifo *fifo = device->fifo; |
| unsigned long flags; |
| |
| nvkm_fifo_pause(fifo, &flags); |
| nv04_gr_idle(&gr->base); |
| |
| nvkm_wr32(device, NV10_PGRAPH_TLIMIT(i), tile->limit); |
| nvkm_wr32(device, NV10_PGRAPH_TSIZE(i), tile->pitch); |
| nvkm_wr32(device, NV10_PGRAPH_TILE(i), tile->addr); |
| |
| nvkm_fifo_start(fifo, &flags); |
| } |
| |
| const struct nvkm_bitfield nv10_gr_intr_name[] = { |
| { NV_PGRAPH_INTR_NOTIFY, "NOTIFY" }, |
| { NV_PGRAPH_INTR_ERROR, "ERROR" }, |
| {} |
| }; |
| |
| const struct nvkm_bitfield nv10_gr_nstatus[] = { |
| { NV10_PGRAPH_NSTATUS_STATE_IN_USE, "STATE_IN_USE" }, |
| { NV10_PGRAPH_NSTATUS_INVALID_STATE, "INVALID_STATE" }, |
| { NV10_PGRAPH_NSTATUS_BAD_ARGUMENT, "BAD_ARGUMENT" }, |
| { NV10_PGRAPH_NSTATUS_PROTECTION_FAULT, "PROTECTION_FAULT" }, |
| {} |
| }; |
| |
| void |
| nv10_gr_intr(struct nvkm_gr *base) |
| { |
| struct nv10_gr *gr = nv10_gr(base); |
| struct nvkm_subdev *subdev = &gr->base.engine.subdev; |
| struct nvkm_device *device = subdev->device; |
| u32 stat = nvkm_rd32(device, NV03_PGRAPH_INTR); |
| u32 nsource = nvkm_rd32(device, NV03_PGRAPH_NSOURCE); |
| u32 nstatus = nvkm_rd32(device, NV03_PGRAPH_NSTATUS); |
| u32 addr = nvkm_rd32(device, NV04_PGRAPH_TRAPPED_ADDR); |
| u32 chid = (addr & 0x01f00000) >> 20; |
| u32 subc = (addr & 0x00070000) >> 16; |
| u32 mthd = (addr & 0x00001ffc); |
| u32 data = nvkm_rd32(device, NV04_PGRAPH_TRAPPED_DATA); |
| u32 class = nvkm_rd32(device, 0x400160 + subc * 4) & 0xfff; |
| u32 show = stat; |
| char msg[128], src[128], sta[128]; |
| struct nv10_gr_chan *chan; |
| unsigned long flags; |
| |
| spin_lock_irqsave(&gr->lock, flags); |
| chan = gr->chan[chid]; |
| |
| if (stat & NV_PGRAPH_INTR_ERROR) { |
| if (chan && (nsource & NV03_PGRAPH_NSOURCE_ILLEGAL_MTHD)) { |
| if (!nv10_gr_mthd(chan, class, mthd, data)) |
| show &= ~NV_PGRAPH_INTR_ERROR; |
| } |
| } |
| |
| if (stat & NV_PGRAPH_INTR_CONTEXT_SWITCH) { |
| nvkm_wr32(device, NV03_PGRAPH_INTR, NV_PGRAPH_INTR_CONTEXT_SWITCH); |
| stat &= ~NV_PGRAPH_INTR_CONTEXT_SWITCH; |
| show &= ~NV_PGRAPH_INTR_CONTEXT_SWITCH; |
| nv10_gr_context_switch(gr); |
| } |
| |
| nvkm_wr32(device, NV03_PGRAPH_INTR, stat); |
| nvkm_wr32(device, NV04_PGRAPH_FIFO, 0x00000001); |
| |
| if (show) { |
| nvkm_snprintbf(msg, sizeof(msg), nv10_gr_intr_name, show); |
| nvkm_snprintbf(src, sizeof(src), nv04_gr_nsource, nsource); |
| nvkm_snprintbf(sta, sizeof(sta), nv10_gr_nstatus, nstatus); |
| nvkm_error(subdev, "intr %08x [%s] nsource %08x [%s] " |
| "nstatus %08x [%s] ch %d [%s] subc %d " |
| "class %04x mthd %04x data %08x\n", |
| show, msg, nsource, src, nstatus, sta, chid, |
| chan ? chan->object.client->name : "unknown", |
| subc, class, mthd, data); |
| } |
| |
| spin_unlock_irqrestore(&gr->lock, flags); |
| } |
| |
| int |
| nv10_gr_init(struct nvkm_gr *base) |
| { |
| struct nv10_gr *gr = nv10_gr(base); |
| struct nvkm_device *device = gr->base.engine.subdev.device; |
| |
| nvkm_wr32(device, NV03_PGRAPH_INTR , 0xFFFFFFFF); |
| nvkm_wr32(device, NV03_PGRAPH_INTR_EN, 0xFFFFFFFF); |
| |
| nvkm_wr32(device, NV04_PGRAPH_DEBUG_0, 0xFFFFFFFF); |
| nvkm_wr32(device, NV04_PGRAPH_DEBUG_0, 0x00000000); |
| nvkm_wr32(device, NV04_PGRAPH_DEBUG_1, 0x00118700); |
| /* nvkm_wr32(device, NV04_PGRAPH_DEBUG_2, 0x24E00810); */ /* 0x25f92ad9 */ |
| nvkm_wr32(device, NV04_PGRAPH_DEBUG_2, 0x25f92ad9); |
| nvkm_wr32(device, NV04_PGRAPH_DEBUG_3, 0x55DE0830 | (1 << 29) | (1 << 31)); |
| |
| if (device->card_type >= NV_11 && device->chipset >= 0x17) { |
| nvkm_wr32(device, NV10_PGRAPH_DEBUG_4, 0x1f000000); |
| nvkm_wr32(device, 0x400a10, 0x03ff3fb6); |
| nvkm_wr32(device, 0x400838, 0x002f8684); |
| nvkm_wr32(device, 0x40083c, 0x00115f3f); |
| nvkm_wr32(device, 0x4006b0, 0x40000020); |
| } else { |
| nvkm_wr32(device, NV10_PGRAPH_DEBUG_4, 0x00000000); |
| } |
| |
| nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(0), 0x00000000); |
| nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(1), 0x00000000); |
| nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(2), 0x00000000); |
| nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(3), 0x00000000); |
| nvkm_wr32(device, NV10_PGRAPH_CTX_SWITCH(4), 0x00000000); |
| nvkm_wr32(device, NV10_PGRAPH_STATE, 0xFFFFFFFF); |
| |
| nvkm_mask(device, NV10_PGRAPH_CTX_USER, 0xff000000, 0x1f000000); |
| nvkm_wr32(device, NV10_PGRAPH_CTX_CONTROL, 0x10000100); |
| nvkm_wr32(device, NV10_PGRAPH_FFINTFC_ST2, 0x08000000); |
| return 0; |
| } |
| |
| int |
| nv10_gr_new_(const struct nvkm_gr_func *func, struct nvkm_device *device, |
| enum nvkm_subdev_type type, int inst, struct nvkm_gr **pgr) |
| { |
| struct nv10_gr *gr; |
| |
| if (!(gr = kzalloc(sizeof(*gr), GFP_KERNEL))) |
| return -ENOMEM; |
| spin_lock_init(&gr->lock); |
| *pgr = &gr->base; |
| |
| return nvkm_gr_ctor(func, device, type, inst, true, &gr->base); |
| } |
| |
| static const struct nvkm_gr_func |
| nv10_gr = { |
| .init = nv10_gr_init, |
| .intr = nv10_gr_intr, |
| .tile = nv10_gr_tile, |
| .chan_new = nv10_gr_chan_new, |
| .sclass = { |
| { -1, -1, 0x0012, &nv04_gr_object }, /* beta1 */ |
| { -1, -1, 0x0019, &nv04_gr_object }, /* clip */ |
| { -1, -1, 0x0030, &nv04_gr_object }, /* null */ |
| { -1, -1, 0x0039, &nv04_gr_object }, /* m2mf */ |
| { -1, -1, 0x0043, &nv04_gr_object }, /* rop */ |
| { -1, -1, 0x0044, &nv04_gr_object }, /* pattern */ |
| { -1, -1, 0x004a, &nv04_gr_object }, /* gdi */ |
| { -1, -1, 0x0052, &nv04_gr_object }, /* swzsurf */ |
| { -1, -1, 0x005f, &nv04_gr_object }, /* blit */ |
| { -1, -1, 0x0062, &nv04_gr_object }, /* surf2d */ |
| { -1, -1, 0x0072, &nv04_gr_object }, /* beta4 */ |
| { -1, -1, 0x0089, &nv04_gr_object }, /* sifm */ |
| { -1, -1, 0x008a, &nv04_gr_object }, /* ifc */ |
| { -1, -1, 0x009f, &nv04_gr_object }, /* blit */ |
| { -1, -1, 0x0093, &nv04_gr_object }, /* surf3d */ |
| { -1, -1, 0x0094, &nv04_gr_object }, /* ttri */ |
| { -1, -1, 0x0095, &nv04_gr_object }, /* mtri */ |
| { -1, -1, 0x0056, &nv04_gr_object }, /* celcius */ |
| {} |
| } |
| }; |
| |
| int |
| nv10_gr_new(struct nvkm_device *device, enum nvkm_subdev_type type, int inst, struct nvkm_gr **pgr) |
| { |
| return nv10_gr_new_(&nv10_gr, device, type, inst, pgr); |
| } |