| /* |
| * Copyright 2012 Red Hat Inc. |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a |
| * copy of this software and associated documentation files (the "Software"), |
| * to deal in the Software without restriction, including without limitation |
| * the rights to use, copy, modify, merge, publish, distribute, sublicense, |
| * and/or sell copies of the Software, and to permit persons to whom the |
| * Software is furnished to do so, subject to the following conditions: |
| * |
| * The above copyright notice and this permission notice shall be included in |
| * all copies or substantial portions of the Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |
| * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR |
| * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, |
| * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR |
| * OTHER DEALINGS IN THE SOFTWARE. |
| * |
| * Authors: Ben Skeggs |
| */ |
| #include "priv.h" |
| |
| #include <core/option.h> |
| #include <subdev/top.h> |
| |
| void |
| nvkm_mc_unk260(struct nvkm_device *device, u32 data) |
| { |
| struct nvkm_mc *mc = device->mc; |
| if (likely(mc) && mc->func->unk260) |
| mc->func->unk260(mc, data); |
| } |
| |
| void |
| nvkm_mc_intr_mask(struct nvkm_device *device, enum nvkm_devidx devidx, bool en) |
| { |
| struct nvkm_mc *mc = device->mc; |
| const struct nvkm_mc_map *map; |
| if (likely(mc) && mc->func->intr_mask) { |
| u32 mask = nvkm_top_intr_mask(device, devidx); |
| for (map = mc->func->intr; !mask && map->stat; map++) { |
| if (map->unit == devidx) |
| mask = map->stat; |
| } |
| mc->func->intr_mask(mc, mask, en ? mask : 0); |
| } |
| } |
| |
| void |
| nvkm_mc_intr_unarm(struct nvkm_device *device) |
| { |
| struct nvkm_mc *mc = device->mc; |
| if (likely(mc)) |
| mc->func->intr_unarm(mc); |
| } |
| |
| void |
| nvkm_mc_intr_rearm(struct nvkm_device *device) |
| { |
| struct nvkm_mc *mc = device->mc; |
| if (likely(mc)) |
| mc->func->intr_rearm(mc); |
| } |
| |
| static u32 |
| nvkm_mc_intr_stat(struct nvkm_mc *mc) |
| { |
| u32 intr = mc->func->intr_stat(mc); |
| if (WARN_ON_ONCE(intr == 0xffffffff)) |
| intr = 0; /* likely fallen off the bus */ |
| return intr; |
| } |
| |
| void |
| nvkm_mc_intr(struct nvkm_device *device, bool *handled) |
| { |
| struct nvkm_mc *mc = device->mc; |
| struct nvkm_subdev *subdev; |
| const struct nvkm_mc_map *map; |
| u32 stat, intr; |
| u64 subdevs; |
| |
| if (unlikely(!mc)) |
| return; |
| |
| intr = nvkm_mc_intr_stat(mc); |
| stat = nvkm_top_intr(device, intr, &subdevs); |
| while (subdevs) { |
| enum nvkm_devidx subidx = __ffs64(subdevs); |
| subdev = nvkm_device_subdev(device, subidx); |
| if (subdev) |
| nvkm_subdev_intr(subdev); |
| subdevs &= ~BIT_ULL(subidx); |
| } |
| |
| for (map = mc->func->intr; map->stat; map++) { |
| if (intr & map->stat) { |
| subdev = nvkm_device_subdev(device, map->unit); |
| if (subdev) |
| nvkm_subdev_intr(subdev); |
| stat &= ~map->stat; |
| } |
| } |
| |
| if (stat) |
| nvkm_error(&mc->subdev, "intr %08x\n", stat); |
| *handled = intr != 0; |
| } |
| |
| static u32 |
| nvkm_mc_reset_mask(struct nvkm_device *device, bool isauto, |
| enum nvkm_devidx devidx) |
| { |
| struct nvkm_mc *mc = device->mc; |
| const struct nvkm_mc_map *map; |
| u64 pmc_enable = 0; |
| if (likely(mc)) { |
| if (!(pmc_enable = nvkm_top_reset(device, devidx))) { |
| for (map = mc->func->reset; map && map->stat; map++) { |
| if (!isauto || !map->noauto) { |
| if (map->unit == devidx) { |
| pmc_enable = map->stat; |
| break; |
| } |
| } |
| } |
| } |
| } |
| return pmc_enable; |
| } |
| |
| void |
| nvkm_mc_reset(struct nvkm_device *device, enum nvkm_devidx devidx) |
| { |
| u64 pmc_enable = nvkm_mc_reset_mask(device, true, devidx); |
| if (pmc_enable) { |
| nvkm_mask(device, 0x000200, pmc_enable, 0x00000000); |
| nvkm_mask(device, 0x000200, pmc_enable, pmc_enable); |
| nvkm_rd32(device, 0x000200); |
| } |
| } |
| |
| void |
| nvkm_mc_disable(struct nvkm_device *device, enum nvkm_devidx devidx) |
| { |
| u64 pmc_enable = nvkm_mc_reset_mask(device, false, devidx); |
| if (pmc_enable) |
| nvkm_mask(device, 0x000200, pmc_enable, 0x00000000); |
| } |
| |
| void |
| nvkm_mc_enable(struct nvkm_device *device, enum nvkm_devidx devidx) |
| { |
| u64 pmc_enable = nvkm_mc_reset_mask(device, false, devidx); |
| if (pmc_enable) { |
| nvkm_mask(device, 0x000200, pmc_enable, pmc_enable); |
| nvkm_rd32(device, 0x000200); |
| } |
| } |
| |
| bool |
| nvkm_mc_enabled(struct nvkm_device *device, enum nvkm_devidx devidx) |
| { |
| u64 pmc_enable = nvkm_mc_reset_mask(device, false, devidx); |
| |
| return (pmc_enable != 0) && |
| ((nvkm_rd32(device, 0x000200) & pmc_enable) == pmc_enable); |
| } |
| |
| |
| static int |
| nvkm_mc_fini(struct nvkm_subdev *subdev, bool suspend) |
| { |
| nvkm_mc_intr_unarm(subdev->device); |
| return 0; |
| } |
| |
| static int |
| nvkm_mc_init(struct nvkm_subdev *subdev) |
| { |
| struct nvkm_mc *mc = nvkm_mc(subdev); |
| if (mc->func->init) |
| mc->func->init(mc); |
| nvkm_mc_intr_rearm(subdev->device); |
| return 0; |
| } |
| |
| static void * |
| nvkm_mc_dtor(struct nvkm_subdev *subdev) |
| { |
| return nvkm_mc(subdev); |
| } |
| |
| static const struct nvkm_subdev_func |
| nvkm_mc = { |
| .dtor = nvkm_mc_dtor, |
| .init = nvkm_mc_init, |
| .fini = nvkm_mc_fini, |
| }; |
| |
| void |
| nvkm_mc_ctor(const struct nvkm_mc_func *func, struct nvkm_device *device, |
| int index, struct nvkm_mc *mc) |
| { |
| nvkm_subdev_ctor(&nvkm_mc, device, index, &mc->subdev); |
| mc->func = func; |
| } |
| |
| int |
| nvkm_mc_new_(const struct nvkm_mc_func *func, struct nvkm_device *device, |
| int index, struct nvkm_mc **pmc) |
| { |
| struct nvkm_mc *mc; |
| if (!(mc = *pmc = kzalloc(sizeof(*mc), GFP_KERNEL))) |
| return -ENOMEM; |
| nvkm_mc_ctor(func, device, index, *pmc); |
| return 0; |
| } |