/* * Copyright 2018 Red Hat Inc. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR * OTHER DEALINGS IN THE SOFTWARE. * * Authors: Lyude Paul */ #include #include "priv.h" #include "gk104.h" void gk104_clkgate_enable(struct nvkm_therm *base) { struct gk104_therm *therm = gk104_therm(base); struct nvkm_device *dev = therm->base.subdev.device; const struct gk104_clkgate_engine_info *order = therm->clkgate_order; int i; /* Program ENG_MANT, ENG_FILTER */ for (i = 0; order[i].engine != NVKM_SUBDEV_NR; i++) { if (!nvkm_device_subdev(dev, order[i].engine)) continue; nvkm_mask(dev, 0x20200 + order[i].offset, 0xff00, 0x4500); } /* magic */ nvkm_wr32(dev, 0x020288, therm->idle_filter->fecs); nvkm_wr32(dev, 0x02028c, therm->idle_filter->hubmmu); /* Enable clockgating (ENG_CLK = RUN->AUTO) */ for (i = 0; order[i].engine != NVKM_SUBDEV_NR; i++) { if (!nvkm_device_subdev(dev, order[i].engine)) continue; nvkm_mask(dev, 0x20200 + order[i].offset, 0x00ff, 0x0045); } } void gk104_clkgate_fini(struct nvkm_therm *base, bool suspend) { struct gk104_therm *therm = gk104_therm(base); struct nvkm_device *dev = therm->base.subdev.device; const struct gk104_clkgate_engine_info *order = therm->clkgate_order; int i; /* ENG_CLK = AUTO->RUN, ENG_PWR = RUN->AUTO */ for (i = 0; order[i].engine != NVKM_SUBDEV_NR; i++) { if (!nvkm_device_subdev(dev, order[i].engine)) continue; nvkm_mask(dev, 0x20200 + order[i].offset, 0xff, 0x54); } } const struct gk104_clkgate_engine_info gk104_clkgate_engine_info[] = { { NVKM_ENGINE_GR, 0x00 }, { NVKM_ENGINE_MSPDEC, 0x04 }, { NVKM_ENGINE_MSPPP, 0x08 }, { NVKM_ENGINE_MSVLD, 0x0c }, { NVKM_ENGINE_CE0, 0x10 }, { NVKM_ENGINE_CE1, 0x14 }, { NVKM_ENGINE_MSENC, 0x18 }, { NVKM_ENGINE_CE2, 0x1c }, { NVKM_SUBDEV_NR, 0 }, }; const struct gf100_idle_filter gk104_idle_filter = { .fecs = 0x00001000, .hubmmu = 0x00001000, }; static const struct nvkm_therm_func gk104_therm_func = { .init = gf119_therm_init, .fini = g84_therm_fini, .pwm_ctrl = gf119_fan_pwm_ctrl, .pwm_get = gf119_fan_pwm_get, .pwm_set = gf119_fan_pwm_set, .pwm_clock = gf119_fan_pwm_clock, .temp_get = g84_temp_get, .fan_sense = gt215_therm_fan_sense, .program_alarms = nvkm_therm_program_alarms_polling, .clkgate_init = gf100_clkgate_init, .clkgate_enable = gk104_clkgate_enable, .clkgate_fini = gk104_clkgate_fini, }; static int gk104_therm_new_(const struct nvkm_therm_func *func, struct nvkm_device *device, int index, const struct gk104_clkgate_engine_info *clkgate_order, const struct gf100_idle_filter *idle_filter, struct nvkm_therm **ptherm) { struct gk104_therm *therm = kzalloc(sizeof(*therm), GFP_KERNEL); if (!therm) return -ENOMEM; nvkm_therm_ctor(&therm->base, device, index, func); *ptherm = &therm->base; therm->clkgate_order = clkgate_order; therm->idle_filter = idle_filter; return 0; } int gk104_therm_new(struct nvkm_device *device, int index, struct nvkm_therm **ptherm) { return gk104_therm_new_(&gk104_therm_func, device, index, gk104_clkgate_engine_info, &gk104_idle_filter, ptherm); }