/* * Copyright © 2016 Intel Corporation * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. * */ #include "gt/intel_reset.h" #include "intel_uc.h" #include "intel_guc.h" #include "intel_guc_ads.h" #include "intel_guc_submission.h" #include "i915_drv.h" static void guc_free_load_err_log(struct intel_guc *guc); /* Reset GuC providing us with fresh state for both GuC and HuC. */ static int __intel_uc_reset_hw(struct drm_i915_private *dev_priv) { int ret; u32 guc_status; ret = intel_reset_guc(dev_priv); if (ret) { DRM_ERROR("Failed to reset GuC, ret = %d\n", ret); return ret; } guc_status = I915_READ(GUC_STATUS); WARN(!(guc_status & GS_MIA_IN_RESET), "GuC status: 0x%x, MIA core expected to be in reset\n", guc_status); return ret; } static int __get_platform_enable_guc(struct drm_i915_private *i915) { struct intel_uc_fw *guc_fw = &i915->guc.fw; struct intel_uc_fw *huc_fw = &i915->huc.fw; int enable_guc = 0; /* Default is to use HuC if we know GuC and HuC firmwares */ if (intel_uc_fw_is_selected(guc_fw) && intel_uc_fw_is_selected(huc_fw)) enable_guc |= ENABLE_GUC_LOAD_HUC; /* Any platform specific fine-tuning can be done here */ return enable_guc; } static int __get_default_guc_log_level(struct drm_i915_private *i915) { int guc_log_level; if (!HAS_GUC(i915) || !intel_uc_is_using_guc(i915)) guc_log_level = GUC_LOG_LEVEL_DISABLED; else if (IS_ENABLED(CONFIG_DRM_I915_DEBUG) || IS_ENABLED(CONFIG_DRM_I915_DEBUG_GEM)) guc_log_level = GUC_LOG_LEVEL_MAX; else guc_log_level = GUC_LOG_LEVEL_NON_VERBOSE; /* Any platform specific fine-tuning can be done here */ return guc_log_level; } /** * sanitize_options_early - sanitize uC related modparam options * @i915: device private * * In case of "enable_guc" option this function will attempt to modify * it only if it was initially set to "auto(-1)". Default value for this * modparam varies between platforms and it is hardcoded in driver code. * Any other modparam value is only monitored against availability of the * related hardware or firmware definitions. * * In case of "guc_log_level" option this function will attempt to modify * it only if it was initially set to "auto(-1)" or if initial value was * "enable(1..4)" on platforms without the GuC. Default value for this * modparam varies between platforms and is usually set to "disable(0)" * unless GuC is enabled on given platform and the driver is compiled with * debug config when this modparam will default to "enable(1..4)". */ static void sanitize_options_early(struct drm_i915_private *i915) { struct intel_uc_fw *guc_fw = &i915->guc.fw; struct intel_uc_fw *huc_fw = &i915->huc.fw; /* A negative value means "use platform default" */ if (i915_modparams.enable_guc < 0) i915_modparams.enable_guc = __get_platform_enable_guc(i915); DRM_DEBUG_DRIVER("enable_guc=%d (submission:%s huc:%s)\n", i915_modparams.enable_guc, yesno(intel_uc_is_using_guc_submission(i915)), yesno(intel_uc_is_using_huc(i915))); /* Verify GuC firmware availability */ if (intel_uc_is_using_guc(i915) && !intel_uc_fw_is_selected(guc_fw)) { DRM_WARN("Incompatible option detected: %s=%d, %s!\n", "enable_guc", i915_modparams.enable_guc, !HAS_GUC(i915) ? "no GuC hardware" : "no GuC firmware"); } /* Verify HuC firmware availability */ if (intel_uc_is_using_huc(i915) && !intel_uc_fw_is_selected(huc_fw)) { DRM_WARN("Incompatible option detected: %s=%d, %s!\n", "enable_guc", i915_modparams.enable_guc, !HAS_HUC(i915) ? "no HuC hardware" : "no HuC firmware"); } /* XXX: GuC submission is unavailable for now */ if (intel_uc_is_using_guc_submission(i915)) { DRM_INFO("Incompatible option detected: %s=%d, %s!\n", "enable_guc", i915_modparams.enable_guc, "GuC submission not supported"); DRM_INFO("Switching to non-GuC submission mode!\n"); i915_modparams.enable_guc &= ~ENABLE_GUC_SUBMISSION; } /* A negative value means "use platform/config default" */ if (i915_modparams.guc_log_level < 0) i915_modparams.guc_log_level = __get_default_guc_log_level(i915); if (i915_modparams.guc_log_level > 0 && !intel_uc_is_using_guc(i915)) { DRM_WARN("Incompatible option detected: %s=%d, %s!\n", "guc_log_level", i915_modparams.guc_log_level, !HAS_GUC(i915) ? "no GuC hardware" : "GuC not enabled"); i915_modparams.guc_log_level = 0; } if (i915_modparams.guc_log_level > GUC_LOG_LEVEL_MAX) { DRM_WARN("Incompatible option detected: %s=%d, %s!\n", "guc_log_level", i915_modparams.guc_log_level, "verbosity too high"); i915_modparams.guc_log_level = GUC_LOG_LEVEL_MAX; } DRM_DEBUG_DRIVER("guc_log_level=%d (enabled:%s, verbose:%s, verbosity:%d)\n", i915_modparams.guc_log_level, yesno(i915_modparams.guc_log_level), yesno(GUC_LOG_LEVEL_IS_VERBOSE(i915_modparams.guc_log_level)), GUC_LOG_LEVEL_TO_VERBOSITY(i915_modparams.guc_log_level)); /* Make sure that sanitization was done */ GEM_BUG_ON(i915_modparams.enable_guc < 0); GEM_BUG_ON(i915_modparams.guc_log_level < 0); } void intel_uc_init_early(struct drm_i915_private *i915) { struct intel_guc *guc = &i915->guc; struct intel_huc *huc = &i915->huc; intel_guc_init_early(guc); intel_huc_init_early(huc); sanitize_options_early(i915); } void intel_uc_cleanup_early(struct drm_i915_private *i915) { struct intel_guc *guc = &i915->guc; guc_free_load_err_log(guc); } /** * intel_uc_init_mmio - setup uC MMIO access * @i915: device private * * Setup minimal state necessary for MMIO accesses later in the * initialization sequence. */ void intel_uc_init_mmio(struct drm_i915_private *i915) { intel_guc_init_send_regs(&i915->guc); } static void guc_capture_load_err_log(struct intel_guc *guc) { if (!guc->log.vma || !intel_guc_log_get_level(&guc->log)) return; if (!guc->load_err_log) guc->load_err_log = i915_gem_object_get(guc->log.vma->obj); return; } static void guc_free_load_err_log(struct intel_guc *guc) { if (guc->load_err_log) i915_gem_object_put(guc->load_err_log); } /* * Events triggered while CT buffers are disabled are logged in the SCRATCH_15 * register using the same bits used in the CT message payload. Since our * communication channel with guc is turned off at this point, we can save the * message and handle it after we turn it back on. */ static void guc_clear_mmio_msg(struct intel_guc *guc) { intel_uncore_write(&guc_to_i915(guc)->uncore, SOFT_SCRATCH(15), 0); } static void guc_get_mmio_msg(struct intel_guc *guc) { u32 val; spin_lock_irq(&guc->irq_lock); val = intel_uncore_read(&guc_to_i915(guc)->uncore, SOFT_SCRATCH(15)); guc->mmio_msg |= val & guc->msg_enabled_mask; /* * clear all events, including the ones we're not currently servicing, * to make sure we don't try to process a stale message if we enable * handling of more events later. */ guc_clear_mmio_msg(guc); spin_unlock_irq(&guc->irq_lock); } static void guc_handle_mmio_msg(struct intel_guc *guc) { struct drm_i915_private *i915 = guc_to_i915(guc); /* we need communication to be enabled to reply to GuC */ GEM_BUG_ON(guc->handler == intel_guc_to_host_event_handler_nop); if (!guc->mmio_msg) return; spin_lock_irq(&i915->irq_lock); intel_guc_to_host_process_recv_msg(guc, &guc->mmio_msg, 1); spin_unlock_irq(&i915->irq_lock); guc->mmio_msg = 0; } static void guc_reset_interrupts(struct intel_guc *guc) { guc->interrupts.reset(guc_to_i915(guc)); } static void guc_enable_interrupts(struct intel_guc *guc) { guc->interrupts.enable(guc_to_i915(guc)); } static void guc_disable_interrupts(struct intel_guc *guc) { guc->interrupts.disable(guc_to_i915(guc)); } static int guc_enable_communication(struct intel_guc *guc) { struct drm_i915_private *i915 = guc_to_i915(guc); int ret; ret = intel_guc_ct_enable(&guc->ct); if (ret) return ret; guc->send = intel_guc_send_ct; guc->handler = intel_guc_to_host_event_handler_ct; /* check for mmio messages received before/during the CT enable */ guc_get_mmio_msg(guc); guc_handle_mmio_msg(guc); guc_enable_interrupts(guc); /* check for CT messages received before we enabled interrupts */ spin_lock_irq(&i915->irq_lock); intel_guc_to_host_event_handler_ct(guc); spin_unlock_irq(&i915->irq_lock); DRM_INFO("GuC communication enabled\n"); return 0; } static void guc_stop_communication(struct intel_guc *guc) { intel_guc_ct_stop(&guc->ct); guc->send = intel_guc_send_nop; guc->handler = intel_guc_to_host_event_handler_nop; guc_clear_mmio_msg(guc); } static void guc_disable_communication(struct intel_guc *guc) { /* * Events generated during or after CT disable are logged by guc in * via mmio. Make sure the register is clear before disabling CT since * all events we cared about have already been processed via CT. */ guc_clear_mmio_msg(guc); guc_disable_interrupts(guc); guc->send = intel_guc_send_nop; guc->handler = intel_guc_to_host_event_handler_nop; intel_guc_ct_disable(&guc->ct); /* * Check for messages received during/after the CT disable. We do not * expect any messages to have arrived via CT between the interrupt * disable and the CT disable because GuC should've been idle until we * triggered the CT disable protocol. */ guc_get_mmio_msg(guc); DRM_INFO("GuC communication disabled\n"); } int intel_uc_init_misc(struct drm_i915_private *i915) { struct intel_guc *guc = &i915->guc; struct intel_huc *huc = &i915->huc; int ret; if (!USES_GUC(i915)) return 0; ret = intel_guc_init_misc(guc); if (ret) return ret; if (USES_HUC(i915)) { ret = intel_huc_init_misc(huc); if (ret) goto err_guc; } return 0; err_guc: intel_guc_fini_misc(guc); return ret; } void intel_uc_fini_misc(struct drm_i915_private *i915) { struct intel_guc *guc = &i915->guc; struct intel_huc *huc = &i915->huc; if (!USES_GUC(i915)) return; if (USES_HUC(i915)) intel_huc_fini_misc(huc); intel_guc_fini_misc(guc); } int intel_uc_init(struct drm_i915_private *i915) { struct intel_guc *guc = &i915->guc; struct intel_huc *huc = &i915->huc; int ret; if (!USES_GUC(i915)) return 0; if (!HAS_GUC(i915)) return -ENODEV; /* XXX: GuC submission is unavailable for now */ GEM_BUG_ON(USES_GUC_SUBMISSION(i915)); ret = intel_guc_init(guc); if (ret) return ret; if (USES_HUC(i915)) { ret = intel_huc_init(huc); if (ret) goto err_guc; } if (USES_GUC_SUBMISSION(i915)) { /* * This is stuff we need to have available at fw load time * if we are planning to enable submission later */ ret = intel_guc_submission_init(guc); if (ret) goto err_huc; } return 0; err_huc: if (USES_HUC(i915)) intel_huc_fini(huc); err_guc: intel_guc_fini(guc); return ret; } void intel_uc_fini(struct drm_i915_private *i915) { struct intel_guc *guc = &i915->guc; if (!USES_GUC(i915)) return; GEM_BUG_ON(!HAS_GUC(i915)); if (USES_GUC_SUBMISSION(i915)) intel_guc_submission_fini(guc); if (USES_HUC(i915)) intel_huc_fini(&i915->huc); intel_guc_fini(guc); } static void __uc_sanitize(struct drm_i915_private *i915) { struct intel_guc *guc = &i915->guc; struct intel_huc *huc = &i915->huc; GEM_BUG_ON(!HAS_GUC(i915)); intel_huc_sanitize(huc); intel_guc_sanitize(guc); __intel_uc_reset_hw(i915); } void intel_uc_sanitize(struct drm_i915_private *i915) { if (!USES_GUC(i915)) return; __uc_sanitize(i915); } int intel_uc_init_hw(struct drm_i915_private *i915) { struct intel_guc *guc = &i915->guc; struct intel_huc *huc = &i915->huc; int ret, attempts; if (!USES_GUC(i915)) return 0; GEM_BUG_ON(!HAS_GUC(i915)); guc_reset_interrupts(guc); /* WaEnableuKernelHeaderValidFix:skl */ /* WaEnableGuCBootHashCheckNotSet:skl,bxt,kbl */ if (IS_GEN(i915, 9)) attempts = 3; else attempts = 1; while (attempts--) { /* * Always reset the GuC just before (re)loading, so * that the state and timing are fairly predictable */ ret = __intel_uc_reset_hw(i915); if (ret) goto err_out; if (USES_HUC(i915)) { ret = intel_huc_fw_upload(huc); if (ret) goto err_out; } intel_guc_ads_reset(guc); intel_guc_init_params(guc); ret = intel_guc_fw_upload(guc); if (ret == 0) break; DRM_DEBUG_DRIVER("GuC fw load failed: %d; will reset and " "retry %d more time(s)\n", ret, attempts); } /* Did we succeded or run out of retries? */ if (ret) goto err_log_capture; ret = guc_enable_communication(guc); if (ret) goto err_log_capture; if (USES_HUC(i915)) { ret = intel_huc_auth(huc); if (ret) goto err_communication; } ret = intel_guc_sample_forcewake(guc); if (ret) goto err_communication; if (USES_GUC_SUBMISSION(i915)) { ret = intel_guc_submission_enable(guc); if (ret) goto err_communication; } dev_info(i915->drm.dev, "GuC firmware version %u.%u\n", guc->fw.major_ver_found, guc->fw.minor_ver_found); dev_info(i915->drm.dev, "GuC submission %s\n", enableddisabled(USES_GUC_SUBMISSION(i915))); dev_info(i915->drm.dev, "HuC %s\n", enableddisabled(USES_HUC(i915))); return 0; /* * We've failed to load the firmware :( */ err_communication: guc_disable_communication(guc); err_log_capture: guc_capture_load_err_log(guc); err_out: __uc_sanitize(i915); /* * Note that there is no fallback as either user explicitly asked for * the GuC or driver default option was to run with the GuC enabled. */ if (GEM_WARN_ON(ret == -EIO)) ret = -EINVAL; dev_err(i915->drm.dev, "GuC initialization failed %d\n", ret); return ret; } void intel_uc_fini_hw(struct drm_i915_private *i915) { struct intel_guc *guc = &i915->guc; if (!intel_guc_is_loaded(guc)) return; GEM_BUG_ON(!HAS_GUC(i915)); if (USES_GUC_SUBMISSION(i915)) intel_guc_submission_disable(guc); guc_disable_communication(guc); __uc_sanitize(i915); } /** * intel_uc_reset_prepare - Prepare for reset * @i915: device private * * Preparing for full gpu reset. */ void intel_uc_reset_prepare(struct drm_i915_private *i915) { struct intel_guc *guc = &i915->guc; if (!intel_guc_is_loaded(guc)) return; guc_stop_communication(guc); __uc_sanitize(i915); } void intel_uc_runtime_suspend(struct drm_i915_private *i915) { struct intel_guc *guc = &i915->guc; int err; if (!intel_guc_is_loaded(guc)) return; err = intel_guc_suspend(guc); if (err) DRM_DEBUG_DRIVER("Failed to suspend GuC, err=%d", err); guc_disable_communication(guc); } void intel_uc_suspend(struct drm_i915_private *i915) { struct intel_guc *guc = &i915->guc; intel_wakeref_t wakeref; if (!intel_guc_is_loaded(guc)) return; with_intel_runtime_pm(&i915->runtime_pm, wakeref) intel_uc_runtime_suspend(i915); } int intel_uc_resume(struct drm_i915_private *i915) { struct intel_guc *guc = &i915->guc; int err; if (!intel_guc_is_loaded(guc)) return 0; guc_enable_communication(guc); err = intel_guc_resume(guc); if (err) { DRM_DEBUG_DRIVER("Failed to resume GuC, err=%d", err); return err; } return 0; }