| // SPDX-License-Identifier: MIT |
| /* |
| * Copyright © 2022 Intel Corporation |
| */ |
| |
| #include "xe_debugfs.h" |
| |
| #include <linux/debugfs.h> |
| #include <linux/fault-inject.h> |
| #include <linux/string_helpers.h> |
| |
| #include <drm/drm_debugfs.h> |
| |
| #include "regs/xe_pmt.h" |
| #include "xe_bo.h" |
| #include "xe_device.h" |
| #include "xe_force_wake.h" |
| #include "xe_gt_debugfs.h" |
| #include "xe_gt_printk.h" |
| #include "xe_guc_ads.h" |
| #include "xe_mmio.h" |
| #include "xe_pm.h" |
| #include "xe_psmi.h" |
| #include "xe_pxp_debugfs.h" |
| #include "xe_sriov.h" |
| #include "xe_sriov_pf.h" |
| #include "xe_sriov_vf.h" |
| #include "xe_step.h" |
| #include "xe_tile_debugfs.h" |
| #include "xe_wa.h" |
| #include "xe_vsec.h" |
| |
| #ifdef CONFIG_DRM_XE_DEBUG |
| #include "xe_bo_evict.h" |
| #include "xe_migrate.h" |
| #include "xe_vm.h" |
| #endif |
| |
| DECLARE_FAULT_ATTR(gt_reset_failure); |
| DECLARE_FAULT_ATTR(inject_csc_hw_error); |
| |
| static void read_residency_counter(struct xe_device *xe, struct xe_mmio *mmio, |
| u32 offset, const char *name, struct drm_printer *p) |
| { |
| u64 residency = 0; |
| int ret; |
| |
| ret = xe_pmt_telem_read(to_pci_dev(xe->drm.dev), |
| xe_mmio_read32(mmio, PUNIT_TELEMETRY_GUID), |
| &residency, offset, sizeof(residency)); |
| if (ret != sizeof(residency)) { |
| drm_warn(&xe->drm, "%s counter failed to read, ret %d\n", name, ret); |
| return; |
| } |
| |
| drm_printf(p, "%s : %llu\n", name, residency); |
| } |
| |
| static struct xe_device *node_to_xe(struct drm_info_node *node) |
| { |
| return to_xe_device(node->minor->dev); |
| } |
| |
| static int info(struct seq_file *m, void *data) |
| { |
| struct xe_device *xe = node_to_xe(m->private); |
| struct drm_printer p = drm_seq_file_printer(m); |
| struct xe_gt *gt; |
| u8 id; |
| |
| xe_pm_runtime_get(xe); |
| |
| drm_printf(&p, "graphics_verx100 %d\n", xe->info.graphics_verx100); |
| drm_printf(&p, "media_verx100 %d\n", xe->info.media_verx100); |
| drm_printf(&p, "stepping G:%s M:%s B:%s\n", |
| xe_step_name(xe->info.step.graphics), |
| xe_step_name(xe->info.step.media), |
| xe_step_name(xe->info.step.basedie)); |
| drm_printf(&p, "is_dgfx %s\n", str_yes_no(xe->info.is_dgfx)); |
| drm_printf(&p, "platform %d\n", xe->info.platform); |
| drm_printf(&p, "subplatform %d\n", |
| xe->info.subplatform > XE_SUBPLATFORM_NONE ? xe->info.subplatform : 0); |
| drm_printf(&p, "devid 0x%x\n", xe->info.devid); |
| drm_printf(&p, "revid %d\n", xe->info.revid); |
| drm_printf(&p, "tile_count %d\n", xe->info.tile_count); |
| drm_printf(&p, "vm_max_level %d\n", xe->info.vm_max_level); |
| drm_printf(&p, "force_execlist %s\n", str_yes_no(xe->info.force_execlist)); |
| drm_printf(&p, "has_flat_ccs %s\n", str_yes_no(xe->info.has_flat_ccs)); |
| drm_printf(&p, "has_usm %s\n", str_yes_no(xe->info.has_usm)); |
| drm_printf(&p, "skip_guc_pc %s\n", str_yes_no(xe->info.skip_guc_pc)); |
| for_each_gt(gt, xe, id) { |
| drm_printf(&p, "gt%d force wake %d\n", id, |
| xe_force_wake_ref(gt_to_fw(gt), XE_FW_GT)); |
| drm_printf(&p, "gt%d engine_mask 0x%llx\n", id, |
| gt->info.engine_mask); |
| } |
| |
| xe_pm_runtime_put(xe); |
| return 0; |
| } |
| |
| static int sriov_info(struct seq_file *m, void *data) |
| { |
| struct xe_device *xe = node_to_xe(m->private); |
| struct drm_printer p = drm_seq_file_printer(m); |
| |
| xe_sriov_print_info(xe, &p); |
| return 0; |
| } |
| |
| static int workarounds(struct xe_device *xe, struct drm_printer *p) |
| { |
| xe_pm_runtime_get(xe); |
| xe_wa_device_dump(xe, p); |
| xe_pm_runtime_put(xe); |
| |
| return 0; |
| } |
| |
| static int workaround_info(struct seq_file *m, void *data) |
| { |
| struct xe_device *xe = node_to_xe(m->private); |
| struct drm_printer p = drm_seq_file_printer(m); |
| |
| workarounds(xe, &p); |
| return 0; |
| } |
| |
| static int dgfx_pkg_residencies_show(struct seq_file *m, void *data) |
| { |
| struct xe_device *xe; |
| struct xe_mmio *mmio; |
| struct drm_printer p; |
| |
| xe = node_to_xe(m->private); |
| p = drm_seq_file_printer(m); |
| xe_pm_runtime_get(xe); |
| mmio = xe_root_tile_mmio(xe); |
| static const struct { |
| u32 offset; |
| const char *name; |
| } residencies[] = { |
| {BMG_G2_RESIDENCY_OFFSET, "Package G2"}, |
| {BMG_G6_RESIDENCY_OFFSET, "Package G6"}, |
| {BMG_G8_RESIDENCY_OFFSET, "Package G8"}, |
| {BMG_G10_RESIDENCY_OFFSET, "Package G10"}, |
| {BMG_MODS_RESIDENCY_OFFSET, "Package ModS"} |
| }; |
| |
| for (int i = 0; i < ARRAY_SIZE(residencies); i++) |
| read_residency_counter(xe, mmio, residencies[i].offset, residencies[i].name, &p); |
| |
| xe_pm_runtime_put(xe); |
| return 0; |
| } |
| |
| static int dgfx_pcie_link_residencies_show(struct seq_file *m, void *data) |
| { |
| struct xe_device *xe; |
| struct xe_mmio *mmio; |
| struct drm_printer p; |
| |
| xe = node_to_xe(m->private); |
| p = drm_seq_file_printer(m); |
| xe_pm_runtime_get(xe); |
| mmio = xe_root_tile_mmio(xe); |
| |
| static const struct { |
| u32 offset; |
| const char *name; |
| } residencies[] = { |
| {BMG_PCIE_LINK_L0_RESIDENCY_OFFSET, "PCIE LINK L0 RESIDENCY"}, |
| {BMG_PCIE_LINK_L1_RESIDENCY_OFFSET, "PCIE LINK L1 RESIDENCY"}, |
| {BMG_PCIE_LINK_L1_2_RESIDENCY_OFFSET, "PCIE LINK L1.2 RESIDENCY"} |
| }; |
| |
| for (int i = 0; i < ARRAY_SIZE(residencies); i++) |
| read_residency_counter(xe, mmio, residencies[i].offset, residencies[i].name, &p); |
| |
| xe_pm_runtime_put(xe); |
| return 0; |
| } |
| |
| static const struct drm_info_list debugfs_list[] = { |
| {"info", info, 0}, |
| { .name = "sriov_info", .show = sriov_info, }, |
| { .name = "workarounds", .show = workaround_info, }, |
| }; |
| |
| static const struct drm_info_list debugfs_residencies[] = { |
| { .name = "dgfx_pkg_residencies", .show = dgfx_pkg_residencies_show, }, |
| { .name = "dgfx_pcie_link_residencies", .show = dgfx_pcie_link_residencies_show, }, |
| }; |
| |
| static int forcewake_open(struct inode *inode, struct file *file) |
| { |
| struct xe_device *xe = inode->i_private; |
| struct xe_gt *gt; |
| u8 id, last_gt; |
| unsigned int fw_ref; |
| |
| xe_pm_runtime_get(xe); |
| for_each_gt(gt, xe, id) { |
| last_gt = id; |
| |
| fw_ref = xe_force_wake_get(gt_to_fw(gt), XE_FORCEWAKE_ALL); |
| if (!xe_force_wake_ref_has_domain(fw_ref, XE_FORCEWAKE_ALL)) |
| goto err_fw_get; |
| } |
| |
| return 0; |
| |
| err_fw_get: |
| for_each_gt(gt, xe, id) { |
| if (id < last_gt) |
| xe_force_wake_put(gt_to_fw(gt), XE_FORCEWAKE_ALL); |
| else if (id == last_gt) |
| xe_force_wake_put(gt_to_fw(gt), fw_ref); |
| else |
| break; |
| } |
| |
| xe_pm_runtime_put(xe); |
| return -ETIMEDOUT; |
| } |
| |
| static int forcewake_release(struct inode *inode, struct file *file) |
| { |
| struct xe_device *xe = inode->i_private; |
| struct xe_gt *gt; |
| u8 id; |
| |
| for_each_gt(gt, xe, id) |
| xe_force_wake_put(gt_to_fw(gt), XE_FORCEWAKE_ALL); |
| xe_pm_runtime_put(xe); |
| |
| return 0; |
| } |
| |
| static const struct file_operations forcewake_all_fops = { |
| .owner = THIS_MODULE, |
| .open = forcewake_open, |
| .release = forcewake_release, |
| }; |
| |
| static ssize_t wedged_mode_show(struct file *f, char __user *ubuf, |
| size_t size, loff_t *pos) |
| { |
| struct xe_device *xe = file_inode(f)->i_private; |
| char buf[32]; |
| int len = 0; |
| |
| len = scnprintf(buf, sizeof(buf), "%d\n", xe->wedged.mode); |
| |
| return simple_read_from_buffer(ubuf, size, pos, buf, len); |
| } |
| |
| static ssize_t wedged_mode_set(struct file *f, const char __user *ubuf, |
| size_t size, loff_t *pos) |
| { |
| struct xe_device *xe = file_inode(f)->i_private; |
| struct xe_gt *gt; |
| u32 wedged_mode; |
| ssize_t ret; |
| u8 id; |
| |
| ret = kstrtouint_from_user(ubuf, size, 0, &wedged_mode); |
| if (ret) |
| return ret; |
| |
| if (wedged_mode > 2) |
| return -EINVAL; |
| |
| if (xe->wedged.mode == wedged_mode) |
| return size; |
| |
| xe->wedged.mode = wedged_mode; |
| |
| xe_pm_runtime_get(xe); |
| for_each_gt(gt, xe, id) { |
| ret = xe_guc_ads_scheduler_policy_toggle_reset(>->uc.guc.ads); |
| if (ret) { |
| xe_gt_err(gt, "Failed to update GuC ADS scheduler policy. GuC may still cause engine reset even with wedged_mode=2\n"); |
| xe_pm_runtime_put(xe); |
| return -EIO; |
| } |
| } |
| xe_pm_runtime_put(xe); |
| |
| return size; |
| } |
| |
| static const struct file_operations wedged_mode_fops = { |
| .owner = THIS_MODULE, |
| .read = wedged_mode_show, |
| .write = wedged_mode_set, |
| }; |
| |
| static ssize_t atomic_svm_timeslice_ms_show(struct file *f, char __user *ubuf, |
| size_t size, loff_t *pos) |
| { |
| struct xe_device *xe = file_inode(f)->i_private; |
| char buf[32]; |
| int len = 0; |
| |
| len = scnprintf(buf, sizeof(buf), "%d\n", xe->atomic_svm_timeslice_ms); |
| |
| return simple_read_from_buffer(ubuf, size, pos, buf, len); |
| } |
| |
| static ssize_t atomic_svm_timeslice_ms_set(struct file *f, |
| const char __user *ubuf, |
| size_t size, loff_t *pos) |
| { |
| struct xe_device *xe = file_inode(f)->i_private; |
| u32 atomic_svm_timeslice_ms; |
| ssize_t ret; |
| |
| ret = kstrtouint_from_user(ubuf, size, 0, &atomic_svm_timeslice_ms); |
| if (ret) |
| return ret; |
| |
| xe->atomic_svm_timeslice_ms = atomic_svm_timeslice_ms; |
| |
| return size; |
| } |
| |
| static const struct file_operations atomic_svm_timeslice_ms_fops = { |
| .owner = THIS_MODULE, |
| .read = atomic_svm_timeslice_ms_show, |
| .write = atomic_svm_timeslice_ms_set, |
| }; |
| |
| static ssize_t disable_late_binding_show(struct file *f, char __user *ubuf, |
| size_t size, loff_t *pos) |
| { |
| struct xe_device *xe = file_inode(f)->i_private; |
| struct xe_late_bind *late_bind = &xe->late_bind; |
| char buf[32]; |
| int len; |
| |
| len = scnprintf(buf, sizeof(buf), "%d\n", late_bind->disable); |
| |
| return simple_read_from_buffer(ubuf, size, pos, buf, len); |
| } |
| |
| static ssize_t disable_late_binding_set(struct file *f, const char __user *ubuf, |
| size_t size, loff_t *pos) |
| { |
| struct xe_device *xe = file_inode(f)->i_private; |
| struct xe_late_bind *late_bind = &xe->late_bind; |
| u32 uval; |
| ssize_t ret; |
| |
| ret = kstrtouint_from_user(ubuf, size, sizeof(uval), &uval); |
| if (ret) |
| return ret; |
| |
| if (uval > 1) |
| return -EINVAL; |
| |
| late_bind->disable = !!uval; |
| return size; |
| } |
| |
| static const struct file_operations disable_late_binding_fops = { |
| .owner = THIS_MODULE, |
| .read = disable_late_binding_show, |
| .write = disable_late_binding_set, |
| }; |
| |
| void xe_debugfs_register(struct xe_device *xe) |
| { |
| struct ttm_device *bdev = &xe->ttm; |
| struct drm_minor *minor = xe->drm.primary; |
| struct dentry *root = minor->debugfs_root; |
| struct ttm_resource_manager *man; |
| struct xe_tile *tile; |
| struct xe_gt *gt; |
| u32 mem_type; |
| u8 tile_id; |
| u8 id; |
| |
| drm_debugfs_create_files(debugfs_list, |
| ARRAY_SIZE(debugfs_list), |
| root, minor); |
| |
| if (xe->info.platform == XE_BATTLEMAGE && !IS_SRIOV_VF(xe)) { |
| drm_debugfs_create_files(debugfs_residencies, |
| ARRAY_SIZE(debugfs_residencies), |
| root, minor); |
| fault_create_debugfs_attr("inject_csc_hw_error", root, |
| &inject_csc_hw_error); |
| } |
| |
| debugfs_create_file("forcewake_all", 0400, root, xe, |
| &forcewake_all_fops); |
| |
| debugfs_create_file("wedged_mode", 0600, root, xe, |
| &wedged_mode_fops); |
| |
| debugfs_create_file("atomic_svm_timeslice_ms", 0600, root, xe, |
| &atomic_svm_timeslice_ms_fops); |
| |
| debugfs_create_file("disable_late_binding", 0600, root, xe, |
| &disable_late_binding_fops); |
| |
| for (mem_type = XE_PL_VRAM0; mem_type <= XE_PL_VRAM1; ++mem_type) { |
| man = ttm_manager_type(bdev, mem_type); |
| |
| if (man) { |
| char name[16]; |
| |
| snprintf(name, sizeof(name), "vram%d_mm", mem_type - XE_PL_VRAM0); |
| ttm_resource_manager_create_debugfs(man, root, name); |
| } |
| } |
| |
| man = ttm_manager_type(bdev, XE_PL_TT); |
| ttm_resource_manager_create_debugfs(man, root, "gtt_mm"); |
| |
| man = ttm_manager_type(bdev, XE_PL_STOLEN); |
| if (man) |
| ttm_resource_manager_create_debugfs(man, root, "stolen_mm"); |
| |
| for_each_tile(tile, xe, tile_id) |
| xe_tile_debugfs_register(tile); |
| |
| for_each_gt(gt, xe, id) |
| xe_gt_debugfs_register(gt); |
| |
| xe_pxp_debugfs_register(xe->pxp); |
| |
| xe_psmi_debugfs_register(xe); |
| |
| fault_create_debugfs_attr("fail_gt_reset", root, >_reset_failure); |
| |
| if (IS_SRIOV_PF(xe)) |
| xe_sriov_pf_debugfs_register(xe, root); |
| else if (IS_SRIOV_VF(xe)) |
| xe_sriov_vf_debugfs_register(xe, root); |
| } |