|  | /* | 
|  | * Copyright 2012 Red Hat Inc. | 
|  | * | 
|  | * Permission is hereby granted, free of charge, to any person obtaining a | 
|  | * copy of this software and associated documentation files (the "Software"), | 
|  | * to deal in the Software without restriction, including without limitation | 
|  | * the rights to use, copy, modify, merge, publish, distribute, sublicense, | 
|  | * and/or sell copies of the Software, and to permit persons to whom the | 
|  | * Software is furnished to do so, subject to the following conditions: | 
|  | * | 
|  | * The above copyright notice and this permission notice shall be included in | 
|  | * all copies or substantial portions of the Software. | 
|  | * | 
|  | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | 
|  | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | 
|  | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL | 
|  | * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR | 
|  | * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, | 
|  | * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR | 
|  | * OTHER DEALINGS IN THE SOFTWARE. | 
|  | * | 
|  | * Authors: Ben Skeggs | 
|  | */ | 
|  |  | 
|  | #include <linux/delay.h> | 
|  | #include <linux/module.h> | 
|  | #include <linux/pci.h> | 
|  | #include <linux/pm_runtime.h> | 
|  | #include <linux/vga_switcheroo.h> | 
|  | #include <linux/mmu_notifier.h> | 
|  | #include <linux/dynamic_debug.h> | 
|  |  | 
|  | #include <drm/drm_aperture.h> | 
|  | #include <drm/drm_client_setup.h> | 
|  | #include <drm/drm_drv.h> | 
|  | #include <drm/drm_fbdev_ttm.h> | 
|  | #include <drm/drm_gem_ttm_helper.h> | 
|  | #include <drm/drm_ioctl.h> | 
|  | #include <drm/drm_vblank.h> | 
|  |  | 
|  | #include <core/gpuobj.h> | 
|  | #include <core/option.h> | 
|  | #include <core/pci.h> | 
|  | #include <core/tegra.h> | 
|  |  | 
|  | #include <nvif/driver.h> | 
|  | #include <nvif/fifo.h> | 
|  | #include <nvif/push006c.h> | 
|  | #include <nvif/user.h> | 
|  |  | 
|  | #include <nvif/class.h> | 
|  | #include <nvif/cl0002.h> | 
|  |  | 
|  | #include "nouveau_drv.h" | 
|  | #include "nouveau_dma.h" | 
|  | #include "nouveau_ttm.h" | 
|  | #include "nouveau_gem.h" | 
|  | #include "nouveau_vga.h" | 
|  | #include "nouveau_led.h" | 
|  | #include "nouveau_hwmon.h" | 
|  | #include "nouveau_acpi.h" | 
|  | #include "nouveau_bios.h" | 
|  | #include "nouveau_ioctl.h" | 
|  | #include "nouveau_abi16.h" | 
|  | #include "nouveau_fence.h" | 
|  | #include "nouveau_debugfs.h" | 
|  | #include "nouveau_connector.h" | 
|  | #include "nouveau_platform.h" | 
|  | #include "nouveau_svm.h" | 
|  | #include "nouveau_dmem.h" | 
|  | #include "nouveau_exec.h" | 
|  | #include "nouveau_uvmm.h" | 
|  | #include "nouveau_sched.h" | 
|  |  | 
|  | DECLARE_DYNDBG_CLASSMAP(drm_debug_classes, DD_CLASS_TYPE_DISJOINT_BITS, 0, | 
|  | "DRM_UT_CORE", | 
|  | "DRM_UT_DRIVER", | 
|  | "DRM_UT_KMS", | 
|  | "DRM_UT_PRIME", | 
|  | "DRM_UT_ATOMIC", | 
|  | "DRM_UT_VBL", | 
|  | "DRM_UT_STATE", | 
|  | "DRM_UT_LEASE", | 
|  | "DRM_UT_DP", | 
|  | "DRM_UT_DRMRES"); | 
|  |  | 
|  | MODULE_PARM_DESC(config, "option string to pass to driver core"); | 
|  | static char *nouveau_config; | 
|  | module_param_named(config, nouveau_config, charp, 0400); | 
|  |  | 
|  | MODULE_PARM_DESC(debug, "debug string to pass to driver core"); | 
|  | static char *nouveau_debug; | 
|  | module_param_named(debug, nouveau_debug, charp, 0400); | 
|  |  | 
|  | MODULE_PARM_DESC(noaccel, "disable kernel/abi16 acceleration"); | 
|  | static int nouveau_noaccel = 0; | 
|  | module_param_named(noaccel, nouveau_noaccel, int, 0400); | 
|  |  | 
|  | MODULE_PARM_DESC(modeset, "enable driver (default: auto, " | 
|  | "0 = disabled, 1 = enabled, 2 = headless)"); | 
|  | int nouveau_modeset = -1; | 
|  | module_param_named(modeset, nouveau_modeset, int, 0400); | 
|  |  | 
|  | MODULE_PARM_DESC(atomic, "Expose atomic ioctl (default: disabled)"); | 
|  | static int nouveau_atomic = 0; | 
|  | module_param_named(atomic, nouveau_atomic, int, 0400); | 
|  |  | 
|  | MODULE_PARM_DESC(runpm, "disable (0), force enable (1), optimus only default (-1)"); | 
|  | static int nouveau_runtime_pm = -1; | 
|  | module_param_named(runpm, nouveau_runtime_pm, int, 0400); | 
|  |  | 
|  | static struct drm_driver driver_stub; | 
|  | static struct drm_driver driver_pci; | 
|  | static struct drm_driver driver_platform; | 
|  |  | 
|  | static u64 | 
|  | nouveau_pci_name(struct pci_dev *pdev) | 
|  | { | 
|  | u64 name = (u64)pci_domain_nr(pdev->bus) << 32; | 
|  | name |= pdev->bus->number << 16; | 
|  | name |= PCI_SLOT(pdev->devfn) << 8; | 
|  | return name | PCI_FUNC(pdev->devfn); | 
|  | } | 
|  |  | 
|  | static u64 | 
|  | nouveau_platform_name(struct platform_device *platformdev) | 
|  | { | 
|  | return platformdev->id; | 
|  | } | 
|  |  | 
|  | static u64 | 
|  | nouveau_name(struct drm_device *dev) | 
|  | { | 
|  | if (dev_is_pci(dev->dev)) | 
|  | return nouveau_pci_name(to_pci_dev(dev->dev)); | 
|  | else | 
|  | return nouveau_platform_name(to_platform_device(dev->dev)); | 
|  | } | 
|  |  | 
|  | static inline bool | 
|  | nouveau_cli_work_ready(struct dma_fence *fence) | 
|  | { | 
|  | bool ret = true; | 
|  |  | 
|  | spin_lock_irq(fence->lock); | 
|  | if (!dma_fence_is_signaled_locked(fence)) | 
|  | ret = false; | 
|  | spin_unlock_irq(fence->lock); | 
|  |  | 
|  | if (ret == true) | 
|  | dma_fence_put(fence); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void | 
|  | nouveau_cli_work(struct work_struct *w) | 
|  | { | 
|  | struct nouveau_cli *cli = container_of(w, typeof(*cli), work); | 
|  | struct nouveau_cli_work *work, *wtmp; | 
|  | mutex_lock(&cli->lock); | 
|  | list_for_each_entry_safe(work, wtmp, &cli->worker, head) { | 
|  | if (!work->fence || nouveau_cli_work_ready(work->fence)) { | 
|  | list_del(&work->head); | 
|  | work->func(work); | 
|  | } | 
|  | } | 
|  | mutex_unlock(&cli->lock); | 
|  | } | 
|  |  | 
|  | static void | 
|  | nouveau_cli_work_fence(struct dma_fence *fence, struct dma_fence_cb *cb) | 
|  | { | 
|  | struct nouveau_cli_work *work = container_of(cb, typeof(*work), cb); | 
|  | schedule_work(&work->cli->work); | 
|  | } | 
|  |  | 
|  | void | 
|  | nouveau_cli_work_queue(struct nouveau_cli *cli, struct dma_fence *fence, | 
|  | struct nouveau_cli_work *work) | 
|  | { | 
|  | work->fence = dma_fence_get(fence); | 
|  | work->cli = cli; | 
|  | mutex_lock(&cli->lock); | 
|  | list_add_tail(&work->head, &cli->worker); | 
|  | if (dma_fence_add_callback(fence, &work->cb, nouveau_cli_work_fence)) | 
|  | nouveau_cli_work_fence(fence, &work->cb); | 
|  | mutex_unlock(&cli->lock); | 
|  | } | 
|  |  | 
|  | static void | 
|  | nouveau_cli_fini(struct nouveau_cli *cli) | 
|  | { | 
|  | struct nouveau_uvmm *uvmm = nouveau_cli_uvmm_locked(cli); | 
|  |  | 
|  | /* All our channels are dead now, which means all the fences they | 
|  | * own are signalled, and all callback functions have been called. | 
|  | * | 
|  | * So, after flushing the workqueue, there should be nothing left. | 
|  | */ | 
|  | flush_work(&cli->work); | 
|  | WARN_ON(!list_empty(&cli->worker)); | 
|  |  | 
|  | if (cli->sched) | 
|  | nouveau_sched_destroy(&cli->sched); | 
|  | if (uvmm) | 
|  | nouveau_uvmm_fini(uvmm); | 
|  | nouveau_vmm_fini(&cli->svm); | 
|  | nouveau_vmm_fini(&cli->vmm); | 
|  | nvif_mmu_dtor(&cli->mmu); | 
|  | cli->device.object.map.ptr = NULL; | 
|  | nvif_device_dtor(&cli->device); | 
|  | mutex_lock(&cli->drm->client_mutex); | 
|  | nvif_client_dtor(&cli->base); | 
|  | mutex_unlock(&cli->drm->client_mutex); | 
|  | } | 
|  |  | 
|  | static int | 
|  | nouveau_cli_init(struct nouveau_drm *drm, const char *sname, | 
|  | struct nouveau_cli *cli) | 
|  | { | 
|  | static const struct nvif_mclass | 
|  | mems[] = { | 
|  | { NVIF_CLASS_MEM_GF100, -1 }, | 
|  | { NVIF_CLASS_MEM_NV50 , -1 }, | 
|  | { NVIF_CLASS_MEM_NV04 , -1 }, | 
|  | {} | 
|  | }; | 
|  | static const struct nvif_mclass | 
|  | vmms[] = { | 
|  | { NVIF_CLASS_VMM_GP100, -1 }, | 
|  | { NVIF_CLASS_VMM_GM200, -1 }, | 
|  | { NVIF_CLASS_VMM_GF100, -1 }, | 
|  | { NVIF_CLASS_VMM_NV50 , -1 }, | 
|  | { NVIF_CLASS_VMM_NV04 , -1 }, | 
|  | {} | 
|  | }; | 
|  | int ret; | 
|  |  | 
|  | snprintf(cli->name, sizeof(cli->name), "%s", sname); | 
|  | cli->drm = drm; | 
|  | mutex_init(&cli->mutex); | 
|  |  | 
|  | INIT_WORK(&cli->work, nouveau_cli_work); | 
|  | INIT_LIST_HEAD(&cli->worker); | 
|  | mutex_init(&cli->lock); | 
|  |  | 
|  | mutex_lock(&drm->client_mutex); | 
|  | ret = nvif_client_ctor(&drm->_client, cli->name, &cli->base); | 
|  | mutex_unlock(&drm->client_mutex); | 
|  | if (ret) { | 
|  | NV_PRINTK(err, cli, "Client allocation failed: %d\n", ret); | 
|  | goto done; | 
|  | } | 
|  |  | 
|  | ret = nvif_device_ctor(&cli->base, "drmDevice", &cli->device); | 
|  | if (ret) { | 
|  | NV_PRINTK(err, cli, "Device allocation failed: %d\n", ret); | 
|  | goto done; | 
|  | } | 
|  |  | 
|  | cli->device.object.map.ptr = drm->device.object.map.ptr; | 
|  |  | 
|  | ret = nvif_mmu_ctor(&cli->device.object, "drmMmu", drm->mmu.object.oclass, | 
|  | &cli->mmu); | 
|  | if (ret) { | 
|  | NV_PRINTK(err, cli, "MMU allocation failed: %d\n", ret); | 
|  | goto done; | 
|  | } | 
|  |  | 
|  | ret = nvif_mclass(&cli->mmu.object, vmms); | 
|  | if (ret < 0) { | 
|  | NV_PRINTK(err, cli, "No supported VMM class\n"); | 
|  | goto done; | 
|  | } | 
|  |  | 
|  | ret = nouveau_vmm_init(cli, vmms[ret].oclass, &cli->vmm); | 
|  | if (ret) { | 
|  | NV_PRINTK(err, cli, "VMM allocation failed: %d\n", ret); | 
|  | goto done; | 
|  | } | 
|  |  | 
|  | ret = nvif_mclass(&cli->mmu.object, mems); | 
|  | if (ret < 0) { | 
|  | NV_PRINTK(err, cli, "No supported MEM class\n"); | 
|  | goto done; | 
|  | } | 
|  |  | 
|  | cli->mem = &mems[ret]; | 
|  |  | 
|  | /* Don't pass in the (shared) sched_wq in order to let | 
|  | * nouveau_sched_create() create a dedicated one for VM_BIND jobs. | 
|  | * | 
|  | * This is required to ensure that for VM_BIND jobs free_job() work and | 
|  | * run_job() work can always run concurrently and hence, free_job() work | 
|  | * can never stall run_job() work. For EXEC jobs we don't have this | 
|  | * requirement, since EXEC job's free_job() does not require to take any | 
|  | * locks which indirectly or directly are held for allocations | 
|  | * elsewhere. | 
|  | */ | 
|  | ret = nouveau_sched_create(&cli->sched, drm, NULL, 1); | 
|  | if (ret) | 
|  | goto done; | 
|  |  | 
|  | return 0; | 
|  | done: | 
|  | if (ret) | 
|  | nouveau_cli_fini(cli); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void | 
|  | nouveau_accel_ce_fini(struct nouveau_drm *drm) | 
|  | { | 
|  | nouveau_channel_idle(drm->cechan); | 
|  | nvif_object_dtor(&drm->ttm.copy); | 
|  | nouveau_channel_del(&drm->cechan); | 
|  | } | 
|  |  | 
|  | static void | 
|  | nouveau_accel_ce_init(struct nouveau_drm *drm) | 
|  | { | 
|  | struct nvif_device *device = &drm->client.device; | 
|  | u64 runm; | 
|  | int ret = 0; | 
|  |  | 
|  | /* Allocate channel that has access to a (preferably async) copy | 
|  | * engine, to use for TTM buffer moves. | 
|  | */ | 
|  | runm = nvif_fifo_runlist_ce(device); | 
|  | if (!runm) { | 
|  | NV_DEBUG(drm, "no ce runlist\n"); | 
|  | return; | 
|  | } | 
|  |  | 
|  | ret = nouveau_channel_new(&drm->client, true, runm, NvDmaFB, NvDmaTT, &drm->cechan); | 
|  | if (ret) | 
|  | NV_ERROR(drm, "failed to create ce channel, %d\n", ret); | 
|  | } | 
|  |  | 
|  | static void | 
|  | nouveau_accel_gr_fini(struct nouveau_drm *drm) | 
|  | { | 
|  | nouveau_channel_idle(drm->channel); | 
|  | nvif_object_dtor(&drm->ntfy); | 
|  | nvkm_gpuobj_del(&drm->notify); | 
|  | nouveau_channel_del(&drm->channel); | 
|  | } | 
|  |  | 
|  | static void | 
|  | nouveau_accel_gr_init(struct nouveau_drm *drm) | 
|  | { | 
|  | struct nvif_device *device = &drm->client.device; | 
|  | u64 runm; | 
|  | int ret; | 
|  |  | 
|  | /* Allocate channel that has access to the graphics engine. */ | 
|  | runm = nvif_fifo_runlist(device, NV_DEVICE_HOST_RUNLIST_ENGINES_GR); | 
|  | if (!runm) { | 
|  | NV_DEBUG(drm, "no gr runlist\n"); | 
|  | return; | 
|  | } | 
|  |  | 
|  | ret = nouveau_channel_new(&drm->client, false, runm, NvDmaFB, NvDmaTT, &drm->channel); | 
|  | if (ret) { | 
|  | NV_ERROR(drm, "failed to create kernel channel, %d\n", ret); | 
|  | nouveau_accel_gr_fini(drm); | 
|  | return; | 
|  | } | 
|  |  | 
|  | /* A SW class is used on pre-NV50 HW to assist with handling the | 
|  | * synchronisation of page flips, as well as to implement fences | 
|  | * on TNT/TNT2 HW that lacks any kind of support in host. | 
|  | */ | 
|  | if (!drm->channel->nvsw.client && device->info.family < NV_DEVICE_INFO_V0_TESLA) { | 
|  | ret = nvif_object_ctor(&drm->channel->user, "drmNvsw", | 
|  | NVDRM_NVSW, nouveau_abi16_swclass(drm), | 
|  | NULL, 0, &drm->channel->nvsw); | 
|  |  | 
|  | if (ret == 0 && device->info.chipset >= 0x11) { | 
|  | ret = nvif_object_ctor(&drm->channel->user, "drmBlit", | 
|  | 0x005f, 0x009f, | 
|  | NULL, 0, &drm->channel->blit); | 
|  | } | 
|  |  | 
|  | if (ret == 0) { | 
|  | struct nvif_push *push = &drm->channel->chan.push; | 
|  |  | 
|  | ret = PUSH_WAIT(push, 8); | 
|  | if (ret == 0) { | 
|  | if (device->info.chipset >= 0x11) { | 
|  | PUSH_NVSQ(push, NV05F, 0x0000, drm->channel->blit.handle); | 
|  | PUSH_NVSQ(push, NV09F, 0x0120, 0, | 
|  | 0x0124, 1, | 
|  | 0x0128, 2); | 
|  | } | 
|  | PUSH_NVSQ(push, NV_SW, 0x0000, drm->channel->nvsw.handle); | 
|  | } | 
|  | } | 
|  |  | 
|  | if (ret) { | 
|  | NV_ERROR(drm, "failed to allocate sw or blit class, %d\n", ret); | 
|  | nouveau_accel_gr_fini(drm); | 
|  | return; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* NvMemoryToMemoryFormat requires a notifier ctxdma for some reason, | 
|  | * even if notification is never requested, so, allocate a ctxdma on | 
|  | * any GPU where it's possible we'll end up using M2MF for BO moves. | 
|  | */ | 
|  | if (device->info.family < NV_DEVICE_INFO_V0_FERMI) { | 
|  | ret = nvkm_gpuobj_new(nvxx_device(drm), 32, 0, false, NULL, &drm->notify); | 
|  | if (ret) { | 
|  | NV_ERROR(drm, "failed to allocate notifier, %d\n", ret); | 
|  | nouveau_accel_gr_fini(drm); | 
|  | return; | 
|  | } | 
|  |  | 
|  | ret = nvif_object_ctor(&drm->channel->user, "drmM2mfNtfy", | 
|  | NvNotify0, NV_DMA_IN_MEMORY, | 
|  | &(struct nv_dma_v0) { | 
|  | .target = NV_DMA_V0_TARGET_VRAM, | 
|  | .access = NV_DMA_V0_ACCESS_RDWR, | 
|  | .start = drm->notify->addr, | 
|  | .limit = drm->notify->addr + 31 | 
|  | }, sizeof(struct nv_dma_v0), | 
|  | &drm->ntfy); | 
|  | if (ret) { | 
|  | nouveau_accel_gr_fini(drm); | 
|  | return; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | static void | 
|  | nouveau_accel_fini(struct nouveau_drm *drm) | 
|  | { | 
|  | nouveau_accel_ce_fini(drm); | 
|  | nouveau_accel_gr_fini(drm); | 
|  | if (drm->fence) | 
|  | nouveau_fence(drm)->dtor(drm); | 
|  | nouveau_channels_fini(drm); | 
|  | } | 
|  |  | 
|  | static void | 
|  | nouveau_accel_init(struct nouveau_drm *drm) | 
|  | { | 
|  | struct nvif_device *device = &drm->client.device; | 
|  | struct nvif_sclass *sclass; | 
|  | int ret, i, n; | 
|  |  | 
|  | if (nouveau_noaccel) | 
|  | return; | 
|  |  | 
|  | /* Initialise global support for channels, and synchronisation. */ | 
|  | ret = nouveau_channels_init(drm); | 
|  | if (ret) | 
|  | return; | 
|  |  | 
|  | /*XXX: this is crap, but the fence/channel stuff is a little | 
|  | *     backwards in some places.  this will be fixed. | 
|  | */ | 
|  | ret = n = nvif_object_sclass_get(&device->object, &sclass); | 
|  | if (ret < 0) | 
|  | return; | 
|  |  | 
|  | for (ret = -ENOSYS, i = 0; i < n; i++) { | 
|  | switch (sclass[i].oclass) { | 
|  | case NV03_CHANNEL_DMA: | 
|  | ret = nv04_fence_create(drm); | 
|  | break; | 
|  | case NV10_CHANNEL_DMA: | 
|  | ret = nv10_fence_create(drm); | 
|  | break; | 
|  | case NV17_CHANNEL_DMA: | 
|  | case NV40_CHANNEL_DMA: | 
|  | ret = nv17_fence_create(drm); | 
|  | break; | 
|  | case NV50_CHANNEL_GPFIFO: | 
|  | ret = nv50_fence_create(drm); | 
|  | break; | 
|  | case G82_CHANNEL_GPFIFO: | 
|  | ret = nv84_fence_create(drm); | 
|  | break; | 
|  | case FERMI_CHANNEL_GPFIFO: | 
|  | case KEPLER_CHANNEL_GPFIFO_A: | 
|  | case KEPLER_CHANNEL_GPFIFO_B: | 
|  | case MAXWELL_CHANNEL_GPFIFO_A: | 
|  | case PASCAL_CHANNEL_GPFIFO_A: | 
|  | case VOLTA_CHANNEL_GPFIFO_A: | 
|  | case TURING_CHANNEL_GPFIFO_A: | 
|  | case AMPERE_CHANNEL_GPFIFO_A: | 
|  | case AMPERE_CHANNEL_GPFIFO_B: | 
|  | ret = nvc0_fence_create(drm); | 
|  | break; | 
|  | default: | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | nvif_object_sclass_put(&sclass); | 
|  | if (ret) { | 
|  | NV_ERROR(drm, "failed to initialise sync subsystem, %d\n", ret); | 
|  | nouveau_accel_fini(drm); | 
|  | return; | 
|  | } | 
|  |  | 
|  | /* Volta requires access to a doorbell register for kickoff. */ | 
|  | if (drm->client.device.info.family >= NV_DEVICE_INFO_V0_VOLTA) { | 
|  | ret = nvif_user_ctor(device, "drmUsermode"); | 
|  | if (ret) | 
|  | return; | 
|  | } | 
|  |  | 
|  | /* Allocate channels we need to support various functions. */ | 
|  | nouveau_accel_gr_init(drm); | 
|  | nouveau_accel_ce_init(drm); | 
|  |  | 
|  | /* Initialise accelerated TTM buffer moves. */ | 
|  | nouveau_bo_move_init(drm); | 
|  | } | 
|  |  | 
|  | static void __printf(2, 3) | 
|  | nouveau_drm_errorf(struct nvif_object *object, const char *fmt, ...) | 
|  | { | 
|  | struct nouveau_drm *drm = container_of(object->parent, typeof(*drm), parent); | 
|  | struct va_format vaf; | 
|  | va_list va; | 
|  |  | 
|  | va_start(va, fmt); | 
|  | vaf.fmt = fmt; | 
|  | vaf.va = &va; | 
|  | NV_ERROR(drm, "%pV", &vaf); | 
|  | va_end(va); | 
|  | } | 
|  |  | 
|  | static void __printf(2, 3) | 
|  | nouveau_drm_debugf(struct nvif_object *object, const char *fmt, ...) | 
|  | { | 
|  | struct nouveau_drm *drm = container_of(object->parent, typeof(*drm), parent); | 
|  | struct va_format vaf; | 
|  | va_list va; | 
|  |  | 
|  | va_start(va, fmt); | 
|  | vaf.fmt = fmt; | 
|  | vaf.va = &va; | 
|  | NV_DEBUG(drm, "%pV", &vaf); | 
|  | va_end(va); | 
|  | } | 
|  |  | 
|  | static const struct nvif_parent_func | 
|  | nouveau_parent = { | 
|  | .debugf = nouveau_drm_debugf, | 
|  | .errorf = nouveau_drm_errorf, | 
|  | }; | 
|  |  | 
|  | static void | 
|  | nouveau_drm_device_fini(struct nouveau_drm *drm) | 
|  | { | 
|  | struct drm_device *dev = drm->dev; | 
|  | struct nouveau_cli *cli, *temp_cli; | 
|  |  | 
|  | if (nouveau_pmops_runtime()) { | 
|  | pm_runtime_get_sync(dev->dev); | 
|  | pm_runtime_forbid(dev->dev); | 
|  | } | 
|  |  | 
|  | nouveau_led_fini(dev); | 
|  | nouveau_dmem_fini(drm); | 
|  | nouveau_svm_fini(drm); | 
|  | nouveau_hwmon_fini(dev); | 
|  | nouveau_debugfs_fini(drm); | 
|  |  | 
|  | if (dev->mode_config.num_crtc) | 
|  | nouveau_display_fini(dev, false, false); | 
|  | nouveau_display_destroy(dev); | 
|  |  | 
|  | nouveau_accel_fini(drm); | 
|  | nouveau_bios_takedown(dev); | 
|  |  | 
|  | nouveau_ttm_fini(drm); | 
|  | nouveau_vga_fini(drm); | 
|  |  | 
|  | /* | 
|  | * There may be existing clients from as-yet unclosed files. For now, | 
|  | * clean them up here rather than deferring until the file is closed, | 
|  | * but this likely not correct if we want to support hot-unplugging | 
|  | * properly. | 
|  | */ | 
|  | mutex_lock(&drm->clients_lock); | 
|  | list_for_each_entry_safe(cli, temp_cli, &drm->clients, head) { | 
|  | list_del(&cli->head); | 
|  | mutex_lock(&cli->mutex); | 
|  | if (cli->abi16) | 
|  | nouveau_abi16_fini(cli->abi16); | 
|  | mutex_unlock(&cli->mutex); | 
|  | nouveau_cli_fini(cli); | 
|  | kfree(cli); | 
|  | } | 
|  | mutex_unlock(&drm->clients_lock); | 
|  |  | 
|  | nouveau_cli_fini(&drm->client); | 
|  | destroy_workqueue(drm->sched_wq); | 
|  | mutex_destroy(&drm->clients_lock); | 
|  | } | 
|  |  | 
|  | static int | 
|  | nouveau_drm_device_init(struct nouveau_drm *drm) | 
|  | { | 
|  | struct drm_device *dev = drm->dev; | 
|  | int ret; | 
|  |  | 
|  | drm->sched_wq = alloc_workqueue("nouveau_sched_wq_shared", 0, | 
|  | WQ_MAX_ACTIVE); | 
|  | if (!drm->sched_wq) | 
|  | return -ENOMEM; | 
|  |  | 
|  | ret = nouveau_cli_init(drm, "DRM", &drm->client); | 
|  | if (ret) | 
|  | goto fail_wq; | 
|  |  | 
|  | INIT_LIST_HEAD(&drm->clients); | 
|  | mutex_init(&drm->clients_lock); | 
|  | spin_lock_init(&drm->tile.lock); | 
|  |  | 
|  | /* workaround an odd issue on nvc1 by disabling the device's | 
|  | * nosnoop capability.  hopefully won't cause issues until a | 
|  | * better fix is found - assuming there is one... | 
|  | */ | 
|  | if (drm->client.device.info.chipset == 0xc1) | 
|  | nvif_mask(&drm->client.device.object, 0x00088080, 0x00000800, 0x00000000); | 
|  |  | 
|  | nouveau_vga_init(drm); | 
|  |  | 
|  | ret = nouveau_ttm_init(drm); | 
|  | if (ret) | 
|  | goto fail_ttm; | 
|  |  | 
|  | ret = nouveau_bios_init(dev); | 
|  | if (ret) | 
|  | goto fail_bios; | 
|  |  | 
|  | nouveau_accel_init(drm); | 
|  |  | 
|  | ret = nouveau_display_create(dev); | 
|  | if (ret) | 
|  | goto fail_dispctor; | 
|  |  | 
|  | if (dev->mode_config.num_crtc) { | 
|  | ret = nouveau_display_init(dev, false, false); | 
|  | if (ret) | 
|  | goto fail_dispinit; | 
|  | } | 
|  |  | 
|  | nouveau_debugfs_init(drm); | 
|  | nouveau_hwmon_init(dev); | 
|  | nouveau_svm_init(drm); | 
|  | nouveau_dmem_init(drm); | 
|  | nouveau_led_init(dev); | 
|  |  | 
|  | if (nouveau_pmops_runtime()) { | 
|  | pm_runtime_use_autosuspend(dev->dev); | 
|  | pm_runtime_set_autosuspend_delay(dev->dev, 5000); | 
|  | pm_runtime_set_active(dev->dev); | 
|  | pm_runtime_allow(dev->dev); | 
|  | pm_runtime_mark_last_busy(dev->dev); | 
|  | pm_runtime_put(dev->dev); | 
|  | } | 
|  |  | 
|  | ret = drm_dev_register(drm->dev, 0); | 
|  | if (ret) { | 
|  | nouveau_drm_device_fini(drm); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | fail_dispinit: | 
|  | nouveau_display_destroy(dev); | 
|  | fail_dispctor: | 
|  | nouveau_accel_fini(drm); | 
|  | nouveau_bios_takedown(dev); | 
|  | fail_bios: | 
|  | nouveau_ttm_fini(drm); | 
|  | fail_ttm: | 
|  | nouveau_vga_fini(drm); | 
|  | nouveau_cli_fini(&drm->client); | 
|  | fail_wq: | 
|  | destroy_workqueue(drm->sched_wq); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void | 
|  | nouveau_drm_device_del(struct nouveau_drm *drm) | 
|  | { | 
|  | if (drm->dev) | 
|  | drm_dev_put(drm->dev); | 
|  |  | 
|  | nvif_mmu_dtor(&drm->mmu); | 
|  | nvif_device_dtor(&drm->device); | 
|  | nvif_client_dtor(&drm->_client); | 
|  | nvif_parent_dtor(&drm->parent); | 
|  |  | 
|  | mutex_destroy(&drm->client_mutex); | 
|  | kfree(drm); | 
|  | } | 
|  |  | 
|  | static struct nouveau_drm * | 
|  | nouveau_drm_device_new(const struct drm_driver *drm_driver, struct device *parent, | 
|  | struct nvkm_device *device) | 
|  | { | 
|  | static const struct nvif_mclass | 
|  | mmus[] = { | 
|  | { NVIF_CLASS_MMU_GF100, -1 }, | 
|  | { NVIF_CLASS_MMU_NV50 , -1 }, | 
|  | { NVIF_CLASS_MMU_NV04 , -1 }, | 
|  | {} | 
|  | }; | 
|  | struct nouveau_drm *drm; | 
|  | int ret; | 
|  |  | 
|  | drm = kzalloc(sizeof(*drm), GFP_KERNEL); | 
|  | if (!drm) | 
|  | return ERR_PTR(-ENOMEM); | 
|  |  | 
|  | drm->nvkm = device; | 
|  |  | 
|  | drm->dev = drm_dev_alloc(drm_driver, parent); | 
|  | if (IS_ERR(drm->dev)) { | 
|  | ret = PTR_ERR(drm->dev); | 
|  | goto done; | 
|  | } | 
|  |  | 
|  | drm->dev->dev_private = drm; | 
|  | dev_set_drvdata(parent, drm); | 
|  |  | 
|  | nvif_parent_ctor(&nouveau_parent, &drm->parent); | 
|  | mutex_init(&drm->client_mutex); | 
|  | drm->_client.object.parent = &drm->parent; | 
|  |  | 
|  | ret = nvif_driver_init(NULL, nouveau_config, nouveau_debug, "drm", | 
|  | nouveau_name(drm->dev), &drm->_client); | 
|  | if (ret) | 
|  | goto done; | 
|  |  | 
|  | ret = nvif_device_ctor(&drm->_client, "drmDevice", &drm->device); | 
|  | if (ret) { | 
|  | NV_ERROR(drm, "Device allocation failed: %d\n", ret); | 
|  | goto done; | 
|  | } | 
|  |  | 
|  | ret = nvif_device_map(&drm->device); | 
|  | if (ret) { | 
|  | NV_ERROR(drm, "Failed to map PRI: %d\n", ret); | 
|  | goto done; | 
|  | } | 
|  |  | 
|  | ret = nvif_mclass(&drm->device.object, mmus); | 
|  | if (ret < 0) { | 
|  | NV_ERROR(drm, "No supported MMU class\n"); | 
|  | goto done; | 
|  | } | 
|  |  | 
|  | ret = nvif_mmu_ctor(&drm->device.object, "drmMmu", mmus[ret].oclass, &drm->mmu); | 
|  | if (ret) { | 
|  | NV_ERROR(drm, "MMU allocation failed: %d\n", ret); | 
|  | goto done; | 
|  | } | 
|  |  | 
|  | done: | 
|  | if (ret) { | 
|  | nouveau_drm_device_del(drm); | 
|  | drm = NULL; | 
|  | } | 
|  |  | 
|  | return ret ? ERR_PTR(ret) : drm; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * On some Intel PCIe bridge controllers doing a | 
|  | * D0 -> D3hot -> D3cold -> D0 sequence causes Nvidia GPUs to not reappear. | 
|  | * Skipping the intermediate D3hot step seems to make it work again. This is | 
|  | * probably caused by not meeting the expectation the involved AML code has | 
|  | * when the GPU is put into D3hot state before invoking it. | 
|  | * | 
|  | * This leads to various manifestations of this issue: | 
|  | *  - AML code execution to power on the GPU hits an infinite loop (as the | 
|  | *    code waits on device memory to change). | 
|  | *  - kernel crashes, as all PCI reads return -1, which most code isn't able | 
|  | *    to handle well enough. | 
|  | * | 
|  | * In all cases dmesg will contain at least one line like this: | 
|  | * 'nouveau 0000:01:00.0: Refused to change power state, currently in D3' | 
|  | * followed by a lot of nouveau timeouts. | 
|  | * | 
|  | * In the \_SB.PCI0.PEG0.PG00._OFF code deeper down writes bit 0x80 to the not | 
|  | * documented PCI config space register 0x248 of the Intel PCIe bridge | 
|  | * controller (0x1901) in order to change the state of the PCIe link between | 
|  | * the PCIe port and the GPU. There are alternative code paths using other | 
|  | * registers, which seem to work fine (executed pre Windows 8): | 
|  | *  - 0xbc bit 0x20 (publicly available documentation claims 'reserved') | 
|  | *  - 0xb0 bit 0x10 (link disable) | 
|  | * Changing the conditions inside the firmware by poking into the relevant | 
|  | * addresses does resolve the issue, but it seemed to be ACPI private memory | 
|  | * and not any device accessible memory at all, so there is no portable way of | 
|  | * changing the conditions. | 
|  | * On a XPS 9560 that means bits [0,3] on \CPEX need to be cleared. | 
|  | * | 
|  | * The only systems where this behavior can be seen are hybrid graphics laptops | 
|  | * with a secondary Nvidia Maxwell, Pascal or Turing GPU. It's unclear whether | 
|  | * this issue only occurs in combination with listed Intel PCIe bridge | 
|  | * controllers and the mentioned GPUs or other devices as well. | 
|  | * | 
|  | * documentation on the PCIe bridge controller can be found in the | 
|  | * "7th Generation Intel® Processor Families for H Platforms Datasheet Volume 2" | 
|  | * Section "12 PCI Express* Controller (x16) Registers" | 
|  | */ | 
|  |  | 
|  | static void quirk_broken_nv_runpm(struct pci_dev *pdev) | 
|  | { | 
|  | struct nouveau_drm *drm = pci_get_drvdata(pdev); | 
|  | struct pci_dev *bridge = pci_upstream_bridge(pdev); | 
|  |  | 
|  | if (!bridge || bridge->vendor != PCI_VENDOR_ID_INTEL) | 
|  | return; | 
|  |  | 
|  | switch (bridge->device) { | 
|  | case 0x1901: | 
|  | drm->old_pm_cap = pdev->pm_cap; | 
|  | pdev->pm_cap = 0; | 
|  | NV_INFO(drm, "Disabling PCI power management to avoid bug\n"); | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | static int nouveau_drm_probe(struct pci_dev *pdev, | 
|  | const struct pci_device_id *pent) | 
|  | { | 
|  | struct nvkm_device *device; | 
|  | struct nouveau_drm *drm; | 
|  | const struct drm_format_info *format; | 
|  | int ret; | 
|  |  | 
|  | if (vga_switcheroo_client_probe_defer(pdev)) | 
|  | return -EPROBE_DEFER; | 
|  |  | 
|  | /* We need to check that the chipset is supported before booting | 
|  | * fbdev off the hardware, as there's no way to put it back. | 
|  | */ | 
|  | ret = nvkm_device_pci_new(pdev, nouveau_config, nouveau_debug, &device); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | /* Remove conflicting drivers (vesafb, efifb etc). */ | 
|  | ret = drm_aperture_remove_conflicting_pci_framebuffers(pdev, &driver_pci); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | pci_set_master(pdev); | 
|  |  | 
|  | if (nouveau_atomic) | 
|  | driver_pci.driver_features |= DRIVER_ATOMIC; | 
|  |  | 
|  | drm = nouveau_drm_device_new(&driver_pci, &pdev->dev, device); | 
|  | if (IS_ERR(drm)) { | 
|  | ret = PTR_ERR(drm); | 
|  | goto fail_nvkm; | 
|  | } | 
|  |  | 
|  | ret = pci_enable_device(pdev); | 
|  | if (ret) | 
|  | goto fail_drm; | 
|  |  | 
|  | ret = nouveau_drm_device_init(drm); | 
|  | if (ret) | 
|  | goto fail_pci; | 
|  |  | 
|  | if (drm->client.device.info.ram_size <= 32 * 1024 * 1024) | 
|  | format = drm_format_info(DRM_FORMAT_C8); | 
|  | else | 
|  | format = NULL; | 
|  |  | 
|  | drm_client_setup(drm->dev, format); | 
|  |  | 
|  | quirk_broken_nv_runpm(pdev); | 
|  | return 0; | 
|  |  | 
|  | fail_pci: | 
|  | pci_disable_device(pdev); | 
|  | fail_drm: | 
|  | nouveau_drm_device_del(drm); | 
|  | fail_nvkm: | 
|  | nvkm_device_del(&device); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | void | 
|  | nouveau_drm_device_remove(struct nouveau_drm *drm) | 
|  | { | 
|  | struct nvkm_device *device = drm->nvkm; | 
|  |  | 
|  | drm_dev_unplug(drm->dev); | 
|  |  | 
|  | nouveau_drm_device_fini(drm); | 
|  | nouveau_drm_device_del(drm); | 
|  | nvkm_device_del(&device); | 
|  | } | 
|  |  | 
|  | static void | 
|  | nouveau_drm_remove(struct pci_dev *pdev) | 
|  | { | 
|  | struct nouveau_drm *drm = pci_get_drvdata(pdev); | 
|  |  | 
|  | /* revert our workaround */ | 
|  | if (drm->old_pm_cap) | 
|  | pdev->pm_cap = drm->old_pm_cap; | 
|  | nouveau_drm_device_remove(drm); | 
|  | pci_disable_device(pdev); | 
|  | } | 
|  |  | 
|  | static int | 
|  | nouveau_do_suspend(struct nouveau_drm *drm, bool runtime) | 
|  | { | 
|  | struct drm_device *dev = drm->dev; | 
|  | struct ttm_resource_manager *man; | 
|  | int ret; | 
|  |  | 
|  | nouveau_svm_suspend(drm); | 
|  | nouveau_dmem_suspend(drm); | 
|  | nouveau_led_suspend(dev); | 
|  |  | 
|  | if (dev->mode_config.num_crtc) { | 
|  | NV_DEBUG(drm, "suspending display...\n"); | 
|  | ret = nouveau_display_suspend(dev, runtime); | 
|  | if (ret) | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | NV_DEBUG(drm, "evicting buffers...\n"); | 
|  |  | 
|  | man = ttm_manager_type(&drm->ttm.bdev, TTM_PL_VRAM); | 
|  | ttm_resource_manager_evict_all(&drm->ttm.bdev, man); | 
|  |  | 
|  | NV_DEBUG(drm, "waiting for kernel channels to go idle...\n"); | 
|  | if (drm->cechan) { | 
|  | ret = nouveau_channel_idle(drm->cechan); | 
|  | if (ret) | 
|  | goto fail_display; | 
|  | } | 
|  |  | 
|  | if (drm->channel) { | 
|  | ret = nouveau_channel_idle(drm->channel); | 
|  | if (ret) | 
|  | goto fail_display; | 
|  | } | 
|  |  | 
|  | NV_DEBUG(drm, "suspending fence...\n"); | 
|  | if (drm->fence && nouveau_fence(drm)->suspend) { | 
|  | if (!nouveau_fence(drm)->suspend(drm)) { | 
|  | ret = -ENOMEM; | 
|  | goto fail_display; | 
|  | } | 
|  | } | 
|  |  | 
|  | NV_DEBUG(drm, "suspending object tree...\n"); | 
|  | ret = nvif_client_suspend(&drm->_client); | 
|  | if (ret) | 
|  | goto fail_client; | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | fail_client: | 
|  | if (drm->fence && nouveau_fence(drm)->resume) | 
|  | nouveau_fence(drm)->resume(drm); | 
|  |  | 
|  | fail_display: | 
|  | if (dev->mode_config.num_crtc) { | 
|  | NV_DEBUG(drm, "resuming display...\n"); | 
|  | nouveau_display_resume(dev, runtime); | 
|  | } | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int | 
|  | nouveau_do_resume(struct nouveau_drm *drm, bool runtime) | 
|  | { | 
|  | struct drm_device *dev = drm->dev; | 
|  | int ret = 0; | 
|  |  | 
|  | NV_DEBUG(drm, "resuming object tree...\n"); | 
|  | ret = nvif_client_resume(&drm->_client); | 
|  | if (ret) { | 
|  | NV_ERROR(drm, "Client resume failed with error: %d\n", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | NV_DEBUG(drm, "resuming fence...\n"); | 
|  | if (drm->fence && nouveau_fence(drm)->resume) | 
|  | nouveau_fence(drm)->resume(drm); | 
|  |  | 
|  | nouveau_run_vbios_init(dev); | 
|  |  | 
|  | if (dev->mode_config.num_crtc) { | 
|  | NV_DEBUG(drm, "resuming display...\n"); | 
|  | nouveau_display_resume(dev, runtime); | 
|  | } | 
|  |  | 
|  | nouveau_led_resume(dev); | 
|  | nouveau_dmem_resume(drm); | 
|  | nouveau_svm_resume(drm); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | int | 
|  | nouveau_pmops_suspend(struct device *dev) | 
|  | { | 
|  | struct pci_dev *pdev = to_pci_dev(dev); | 
|  | struct nouveau_drm *drm = pci_get_drvdata(pdev); | 
|  | int ret; | 
|  |  | 
|  | if (drm->dev->switch_power_state == DRM_SWITCH_POWER_OFF || | 
|  | drm->dev->switch_power_state == DRM_SWITCH_POWER_DYNAMIC_OFF) | 
|  | return 0; | 
|  |  | 
|  | ret = nouveau_do_suspend(drm, false); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | pci_save_state(pdev); | 
|  | pci_disable_device(pdev); | 
|  | pci_set_power_state(pdev, PCI_D3hot); | 
|  | udelay(200); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | int | 
|  | nouveau_pmops_resume(struct device *dev) | 
|  | { | 
|  | struct pci_dev *pdev = to_pci_dev(dev); | 
|  | struct nouveau_drm *drm = pci_get_drvdata(pdev); | 
|  | int ret; | 
|  |  | 
|  | if (drm->dev->switch_power_state == DRM_SWITCH_POWER_OFF || | 
|  | drm->dev->switch_power_state == DRM_SWITCH_POWER_DYNAMIC_OFF) | 
|  | return 0; | 
|  |  | 
|  | pci_set_power_state(pdev, PCI_D0); | 
|  | pci_restore_state(pdev); | 
|  | ret = pci_enable_device(pdev); | 
|  | if (ret) | 
|  | return ret; | 
|  | pci_set_master(pdev); | 
|  |  | 
|  | ret = nouveau_do_resume(drm, false); | 
|  |  | 
|  | /* Monitors may have been connected / disconnected during suspend */ | 
|  | nouveau_display_hpd_resume(drm); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int | 
|  | nouveau_pmops_freeze(struct device *dev) | 
|  | { | 
|  | struct nouveau_drm *drm = dev_get_drvdata(dev); | 
|  |  | 
|  | return nouveau_do_suspend(drm, false); | 
|  | } | 
|  |  | 
|  | static int | 
|  | nouveau_pmops_thaw(struct device *dev) | 
|  | { | 
|  | struct nouveau_drm *drm = dev_get_drvdata(dev); | 
|  |  | 
|  | return nouveau_do_resume(drm, false); | 
|  | } | 
|  |  | 
|  | bool | 
|  | nouveau_pmops_runtime(void) | 
|  | { | 
|  | if (nouveau_runtime_pm == -1) | 
|  | return nouveau_is_optimus() || nouveau_is_v1_dsm(); | 
|  | return nouveau_runtime_pm == 1; | 
|  | } | 
|  |  | 
|  | static int | 
|  | nouveau_pmops_runtime_suspend(struct device *dev) | 
|  | { | 
|  | struct pci_dev *pdev = to_pci_dev(dev); | 
|  | struct nouveau_drm *drm = pci_get_drvdata(pdev); | 
|  | int ret; | 
|  |  | 
|  | if (!nouveau_pmops_runtime()) { | 
|  | pm_runtime_forbid(dev); | 
|  | return -EBUSY; | 
|  | } | 
|  |  | 
|  | nouveau_switcheroo_optimus_dsm(); | 
|  | ret = nouveau_do_suspend(drm, true); | 
|  | pci_save_state(pdev); | 
|  | pci_disable_device(pdev); | 
|  | pci_ignore_hotplug(pdev); | 
|  | pci_set_power_state(pdev, PCI_D3cold); | 
|  | drm->dev->switch_power_state = DRM_SWITCH_POWER_DYNAMIC_OFF; | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int | 
|  | nouveau_pmops_runtime_resume(struct device *dev) | 
|  | { | 
|  | struct pci_dev *pdev = to_pci_dev(dev); | 
|  | struct nouveau_drm *drm = pci_get_drvdata(pdev); | 
|  | struct nvif_device *device = &drm->client.device; | 
|  | int ret; | 
|  |  | 
|  | if (!nouveau_pmops_runtime()) { | 
|  | pm_runtime_forbid(dev); | 
|  | return -EBUSY; | 
|  | } | 
|  |  | 
|  | pci_set_power_state(pdev, PCI_D0); | 
|  | pci_restore_state(pdev); | 
|  | ret = pci_enable_device(pdev); | 
|  | if (ret) | 
|  | return ret; | 
|  | pci_set_master(pdev); | 
|  |  | 
|  | ret = nouveau_do_resume(drm, true); | 
|  | if (ret) { | 
|  | NV_ERROR(drm, "resume failed with: %d\n", ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* do magic */ | 
|  | nvif_mask(&device->object, 0x088488, (1 << 25), (1 << 25)); | 
|  | drm->dev->switch_power_state = DRM_SWITCH_POWER_ON; | 
|  |  | 
|  | /* Monitors may have been connected / disconnected during suspend */ | 
|  | nouveau_display_hpd_resume(drm); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int | 
|  | nouveau_pmops_runtime_idle(struct device *dev) | 
|  | { | 
|  | if (!nouveau_pmops_runtime()) { | 
|  | pm_runtime_forbid(dev); | 
|  | return -EBUSY; | 
|  | } | 
|  |  | 
|  | pm_runtime_mark_last_busy(dev); | 
|  | pm_runtime_autosuspend(dev); | 
|  | /* we don't want the main rpm_idle to call suspend - we want to autosuspend */ | 
|  | return 1; | 
|  | } | 
|  |  | 
|  | static int | 
|  | nouveau_drm_open(struct drm_device *dev, struct drm_file *fpriv) | 
|  | { | 
|  | struct nouveau_drm *drm = nouveau_drm(dev); | 
|  | struct nouveau_cli *cli; | 
|  | char name[32], tmpname[TASK_COMM_LEN]; | 
|  | int ret; | 
|  |  | 
|  | /* need to bring up power immediately if opening device */ | 
|  | ret = pm_runtime_get_sync(dev->dev); | 
|  | if (ret < 0 && ret != -EACCES) { | 
|  | pm_runtime_put_autosuspend(dev->dev); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | get_task_comm(tmpname, current); | 
|  | rcu_read_lock(); | 
|  | snprintf(name, sizeof(name), "%s[%d]", | 
|  | tmpname, pid_nr(rcu_dereference(fpriv->pid))); | 
|  | rcu_read_unlock(); | 
|  |  | 
|  | if (!(cli = kzalloc(sizeof(*cli), GFP_KERNEL))) { | 
|  | ret = -ENOMEM; | 
|  | goto done; | 
|  | } | 
|  |  | 
|  | ret = nouveau_cli_init(drm, name, cli); | 
|  | if (ret) | 
|  | goto done; | 
|  |  | 
|  | fpriv->driver_priv = cli; | 
|  |  | 
|  | mutex_lock(&drm->clients_lock); | 
|  | list_add(&cli->head, &drm->clients); | 
|  | mutex_unlock(&drm->clients_lock); | 
|  |  | 
|  | done: | 
|  | if (ret && cli) { | 
|  | nouveau_cli_fini(cli); | 
|  | kfree(cli); | 
|  | } | 
|  |  | 
|  | pm_runtime_mark_last_busy(dev->dev); | 
|  | pm_runtime_put_autosuspend(dev->dev); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void | 
|  | nouveau_drm_postclose(struct drm_device *dev, struct drm_file *fpriv) | 
|  | { | 
|  | struct nouveau_cli *cli = nouveau_cli(fpriv); | 
|  | struct nouveau_drm *drm = nouveau_drm(dev); | 
|  | int dev_index; | 
|  |  | 
|  | /* | 
|  | * The device is gone, and as it currently stands all clients are | 
|  | * cleaned up in the removal codepath. In the future this may change | 
|  | * so that we can support hot-unplugging, but for now we immediately | 
|  | * return to avoid a double-free situation. | 
|  | */ | 
|  | if (!drm_dev_enter(dev, &dev_index)) | 
|  | return; | 
|  |  | 
|  | pm_runtime_get_sync(dev->dev); | 
|  |  | 
|  | mutex_lock(&cli->mutex); | 
|  | if (cli->abi16) | 
|  | nouveau_abi16_fini(cli->abi16); | 
|  | mutex_unlock(&cli->mutex); | 
|  |  | 
|  | mutex_lock(&drm->clients_lock); | 
|  | list_del(&cli->head); | 
|  | mutex_unlock(&drm->clients_lock); | 
|  |  | 
|  | nouveau_cli_fini(cli); | 
|  | kfree(cli); | 
|  | pm_runtime_mark_last_busy(dev->dev); | 
|  | pm_runtime_put_autosuspend(dev->dev); | 
|  | drm_dev_exit(dev_index); | 
|  | } | 
|  |  | 
|  | static const struct drm_ioctl_desc | 
|  | nouveau_ioctls[] = { | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_GETPARAM, nouveau_abi16_ioctl_getparam, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_SETPARAM, drm_invalid_op, DRM_AUTH|DRM_MASTER|DRM_ROOT_ONLY), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_CHANNEL_ALLOC, nouveau_abi16_ioctl_channel_alloc, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_CHANNEL_FREE, nouveau_abi16_ioctl_channel_free, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_GROBJ_ALLOC, nouveau_abi16_ioctl_grobj_alloc, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_NOTIFIEROBJ_ALLOC, nouveau_abi16_ioctl_notifierobj_alloc, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_GPUOBJ_FREE, nouveau_abi16_ioctl_gpuobj_free, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_SVM_INIT, nouveau_svmm_init, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_SVM_BIND, nouveau_svmm_bind, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_GEM_NEW, nouveau_gem_ioctl_new, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_GEM_PUSHBUF, nouveau_gem_ioctl_pushbuf, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_GEM_CPU_PREP, nouveau_gem_ioctl_cpu_prep, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_GEM_CPU_FINI, nouveau_gem_ioctl_cpu_fini, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_GEM_INFO, nouveau_gem_ioctl_info, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_VM_INIT, nouveau_uvmm_ioctl_vm_init, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_VM_BIND, nouveau_uvmm_ioctl_vm_bind, DRM_RENDER_ALLOW), | 
|  | DRM_IOCTL_DEF_DRV(NOUVEAU_EXEC, nouveau_exec_ioctl_exec, DRM_RENDER_ALLOW), | 
|  | }; | 
|  |  | 
|  | long | 
|  | nouveau_drm_ioctl(struct file *file, unsigned int cmd, unsigned long arg) | 
|  | { | 
|  | struct drm_file *filp = file->private_data; | 
|  | struct drm_device *dev = filp->minor->dev; | 
|  | long ret; | 
|  |  | 
|  | ret = pm_runtime_get_sync(dev->dev); | 
|  | if (ret < 0 && ret != -EACCES) { | 
|  | pm_runtime_put_autosuspend(dev->dev); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | switch (_IOC_NR(cmd) - DRM_COMMAND_BASE) { | 
|  | case DRM_NOUVEAU_NVIF: | 
|  | ret = nouveau_abi16_ioctl(filp, (void __user *)arg, _IOC_SIZE(cmd)); | 
|  | break; | 
|  | default: | 
|  | ret = drm_ioctl(file, cmd, arg); | 
|  | break; | 
|  | } | 
|  |  | 
|  | pm_runtime_mark_last_busy(dev->dev); | 
|  | pm_runtime_put_autosuspend(dev->dev); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static const struct file_operations | 
|  | nouveau_driver_fops = { | 
|  | .owner = THIS_MODULE, | 
|  | .open = drm_open, | 
|  | .release = drm_release, | 
|  | .unlocked_ioctl = nouveau_drm_ioctl, | 
|  | .mmap = drm_gem_mmap, | 
|  | .poll = drm_poll, | 
|  | .read = drm_read, | 
|  | #if defined(CONFIG_COMPAT) | 
|  | .compat_ioctl = nouveau_compat_ioctl, | 
|  | #endif | 
|  | .llseek = noop_llseek, | 
|  | .fop_flags = FOP_UNSIGNED_OFFSET, | 
|  | }; | 
|  |  | 
|  | static struct drm_driver | 
|  | driver_stub = { | 
|  | .driver_features = DRIVER_GEM | | 
|  | DRIVER_SYNCOBJ | DRIVER_SYNCOBJ_TIMELINE | | 
|  | DRIVER_GEM_GPUVA | | 
|  | DRIVER_MODESET | | 
|  | DRIVER_RENDER, | 
|  | .open = nouveau_drm_open, | 
|  | .postclose = nouveau_drm_postclose, | 
|  |  | 
|  | #if defined(CONFIG_DEBUG_FS) | 
|  | .debugfs_init = nouveau_drm_debugfs_init, | 
|  | #endif | 
|  |  | 
|  | .ioctls = nouveau_ioctls, | 
|  | .num_ioctls = ARRAY_SIZE(nouveau_ioctls), | 
|  | .fops = &nouveau_driver_fops, | 
|  |  | 
|  | .gem_prime_import_sg_table = nouveau_gem_prime_import_sg_table, | 
|  |  | 
|  | .dumb_create = nouveau_display_dumb_create, | 
|  | .dumb_map_offset = drm_gem_ttm_dumb_map_offset, | 
|  |  | 
|  | DRM_FBDEV_TTM_DRIVER_OPS, | 
|  |  | 
|  | .name = DRIVER_NAME, | 
|  | .desc = DRIVER_DESC, | 
|  | #ifdef GIT_REVISION | 
|  | .date = GIT_REVISION, | 
|  | #else | 
|  | .date = DRIVER_DATE, | 
|  | #endif | 
|  | .major = DRIVER_MAJOR, | 
|  | .minor = DRIVER_MINOR, | 
|  | .patchlevel = DRIVER_PATCHLEVEL, | 
|  | }; | 
|  |  | 
|  | static struct pci_device_id | 
|  | nouveau_drm_pci_table[] = { | 
|  | { | 
|  | PCI_DEVICE(PCI_VENDOR_ID_NVIDIA, PCI_ANY_ID), | 
|  | .class = PCI_BASE_CLASS_DISPLAY << 16, | 
|  | .class_mask  = 0xff << 16, | 
|  | }, | 
|  | { | 
|  | PCI_DEVICE(PCI_VENDOR_ID_NVIDIA_SGS, PCI_ANY_ID), | 
|  | .class = PCI_BASE_CLASS_DISPLAY << 16, | 
|  | .class_mask  = 0xff << 16, | 
|  | }, | 
|  | {} | 
|  | }; | 
|  |  | 
|  | static void nouveau_display_options(void) | 
|  | { | 
|  | DRM_DEBUG_DRIVER("Loading Nouveau with parameters:\n"); | 
|  |  | 
|  | DRM_DEBUG_DRIVER("... tv_disable   : %d\n", nouveau_tv_disable); | 
|  | DRM_DEBUG_DRIVER("... ignorelid    : %d\n", nouveau_ignorelid); | 
|  | DRM_DEBUG_DRIVER("... duallink     : %d\n", nouveau_duallink); | 
|  | DRM_DEBUG_DRIVER("... config       : %s\n", nouveau_config); | 
|  | DRM_DEBUG_DRIVER("... debug        : %s\n", nouveau_debug); | 
|  | DRM_DEBUG_DRIVER("... noaccel      : %d\n", nouveau_noaccel); | 
|  | DRM_DEBUG_DRIVER("... modeset      : %d\n", nouveau_modeset); | 
|  | DRM_DEBUG_DRIVER("... runpm        : %d\n", nouveau_runtime_pm); | 
|  | DRM_DEBUG_DRIVER("... vram_pushbuf : %d\n", nouveau_vram_pushbuf); | 
|  | DRM_DEBUG_DRIVER("... hdmimhz      : %d\n", nouveau_hdmimhz); | 
|  | } | 
|  |  | 
|  | static const struct dev_pm_ops nouveau_pm_ops = { | 
|  | .suspend = nouveau_pmops_suspend, | 
|  | .resume = nouveau_pmops_resume, | 
|  | .freeze = nouveau_pmops_freeze, | 
|  | .thaw = nouveau_pmops_thaw, | 
|  | .poweroff = nouveau_pmops_freeze, | 
|  | .restore = nouveau_pmops_resume, | 
|  | .runtime_suspend = nouveau_pmops_runtime_suspend, | 
|  | .runtime_resume = nouveau_pmops_runtime_resume, | 
|  | .runtime_idle = nouveau_pmops_runtime_idle, | 
|  | }; | 
|  |  | 
|  | static struct pci_driver | 
|  | nouveau_drm_pci_driver = { | 
|  | .name = "nouveau", | 
|  | .id_table = nouveau_drm_pci_table, | 
|  | .probe = nouveau_drm_probe, | 
|  | .remove = nouveau_drm_remove, | 
|  | .driver.pm = &nouveau_pm_ops, | 
|  | }; | 
|  |  | 
|  | struct drm_device * | 
|  | nouveau_platform_device_create(const struct nvkm_device_tegra_func *func, | 
|  | struct platform_device *pdev, | 
|  | struct nvkm_device **pdevice) | 
|  | { | 
|  | struct nouveau_drm *drm; | 
|  | int err; | 
|  |  | 
|  | err = nvkm_device_tegra_new(func, pdev, nouveau_config, nouveau_debug, pdevice); | 
|  | if (err) | 
|  | goto err_free; | 
|  |  | 
|  | drm = nouveau_drm_device_new(&driver_platform, &pdev->dev, *pdevice); | 
|  | if (IS_ERR(drm)) { | 
|  | err = PTR_ERR(drm); | 
|  | goto err_free; | 
|  | } | 
|  |  | 
|  | err = nouveau_drm_device_init(drm); | 
|  | if (err) | 
|  | goto err_put; | 
|  |  | 
|  | return drm->dev; | 
|  |  | 
|  | err_put: | 
|  | nouveau_drm_device_del(drm); | 
|  | err_free: | 
|  | nvkm_device_del(pdevice); | 
|  |  | 
|  | return ERR_PTR(err); | 
|  | } | 
|  |  | 
|  | static int __init | 
|  | nouveau_drm_init(void) | 
|  | { | 
|  | driver_pci = driver_stub; | 
|  | driver_platform = driver_stub; | 
|  |  | 
|  | nouveau_display_options(); | 
|  |  | 
|  | if (nouveau_modeset == -1) { | 
|  | if (drm_firmware_drivers_only()) | 
|  | nouveau_modeset = 0; | 
|  | } | 
|  |  | 
|  | if (!nouveau_modeset) | 
|  | return 0; | 
|  |  | 
|  | #ifdef CONFIG_NOUVEAU_PLATFORM_DRIVER | 
|  | platform_driver_register(&nouveau_platform_driver); | 
|  | #endif | 
|  |  | 
|  | nouveau_register_dsm_handler(); | 
|  | nouveau_backlight_ctor(); | 
|  |  | 
|  | #ifdef CONFIG_PCI | 
|  | return pci_register_driver(&nouveau_drm_pci_driver); | 
|  | #else | 
|  | return 0; | 
|  | #endif | 
|  | } | 
|  |  | 
|  | static void __exit | 
|  | nouveau_drm_exit(void) | 
|  | { | 
|  | if (!nouveau_modeset) | 
|  | return; | 
|  |  | 
|  | #ifdef CONFIG_PCI | 
|  | pci_unregister_driver(&nouveau_drm_pci_driver); | 
|  | #endif | 
|  | nouveau_backlight_dtor(); | 
|  | nouveau_unregister_dsm_handler(); | 
|  |  | 
|  | #ifdef CONFIG_NOUVEAU_PLATFORM_DRIVER | 
|  | platform_driver_unregister(&nouveau_platform_driver); | 
|  | #endif | 
|  | if (IS_ENABLED(CONFIG_DRM_NOUVEAU_SVM)) | 
|  | mmu_notifier_synchronize(); | 
|  | } | 
|  |  | 
|  | module_init(nouveau_drm_init); | 
|  | module_exit(nouveau_drm_exit); | 
|  |  | 
|  | MODULE_DEVICE_TABLE(pci, nouveau_drm_pci_table); | 
|  | MODULE_AUTHOR(DRIVER_AUTHOR); | 
|  | MODULE_DESCRIPTION(DRIVER_DESC); | 
|  | MODULE_LICENSE("GPL and additional rights"); |