drm/amdgpu: Add reset on init handler for XGMI
In some cases, device needs to be reset before first use. Add handlers for doing device reset during driver init sequence. Signed-off-by: Lijo Lazar <lijo.lazar@amd.com> Reviewed-by: Feifei Xu <feifxu@amd.com> Acked-by: Rajneesh Bhardwaj <rajneesh.bhardwaj@amd.com> Tested-by: Rajneesh Bhardwaj <rajneesh.bhardwaj@amd.com> Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
This commit is contained in:
parent
e82b973d14
commit
1e4acf4d93
3 changed files with 155 additions and 0 deletions
|
@ -566,6 +566,7 @@ enum amd_reset_method {
|
||||||
AMD_RESET_METHOD_MODE2,
|
AMD_RESET_METHOD_MODE2,
|
||||||
AMD_RESET_METHOD_BACO,
|
AMD_RESET_METHOD_BACO,
|
||||||
AMD_RESET_METHOD_PCI,
|
AMD_RESET_METHOD_PCI,
|
||||||
|
AMD_RESET_METHOD_ON_INIT,
|
||||||
};
|
};
|
||||||
|
|
||||||
struct amdgpu_video_codec_info {
|
struct amdgpu_video_codec_info {
|
||||||
|
|
|
@ -26,6 +26,155 @@
|
||||||
#include "sienna_cichlid.h"
|
#include "sienna_cichlid.h"
|
||||||
#include "smu_v13_0_10.h"
|
#include "smu_v13_0_10.h"
|
||||||
|
|
||||||
|
static int amdgpu_reset_xgmi_reset_on_init_suspend(struct amdgpu_device *adev)
|
||||||
|
{
|
||||||
|
int i, r;
|
||||||
|
|
||||||
|
for (i = adev->num_ip_blocks - 1; i >= 0; i--) {
|
||||||
|
if (!adev->ip_blocks[i].status.valid)
|
||||||
|
continue;
|
||||||
|
if (!adev->ip_blocks[i].status.hw)
|
||||||
|
continue;
|
||||||
|
/* displays are handled in phase1 */
|
||||||
|
if (adev->ip_blocks[i].version->type == AMD_IP_BLOCK_TYPE_DCE)
|
||||||
|
continue;
|
||||||
|
|
||||||
|
/* XXX handle errors */
|
||||||
|
r = adev->ip_blocks[i].version->funcs->suspend(adev);
|
||||||
|
/* XXX handle errors */
|
||||||
|
if (r) {
|
||||||
|
dev_err(adev->dev, "suspend of IP block <%s> failed %d",
|
||||||
|
adev->ip_blocks[i].version->funcs->name, r);
|
||||||
|
}
|
||||||
|
adev->ip_blocks[i].status.hw = false;
|
||||||
|
}
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int amdgpu_reset_xgmi_reset_on_init_prep_hwctxt(
|
||||||
|
struct amdgpu_reset_control *reset_ctl,
|
||||||
|
struct amdgpu_reset_context *reset_context)
|
||||||
|
{
|
||||||
|
struct list_head *reset_device_list = reset_context->reset_device_list;
|
||||||
|
struct amdgpu_device *tmp_adev;
|
||||||
|
int r;
|
||||||
|
|
||||||
|
list_for_each_entry(tmp_adev, reset_device_list, reset_list) {
|
||||||
|
amdgpu_unregister_gpu_instance(tmp_adev);
|
||||||
|
r = amdgpu_reset_xgmi_reset_on_init_suspend(tmp_adev);
|
||||||
|
if (r) {
|
||||||
|
dev_err(tmp_adev->dev,
|
||||||
|
"xgmi reset on init: prepare for reset failed");
|
||||||
|
return r;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return r;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int amdgpu_reset_xgmi_reset_on_init_restore_hwctxt(
|
||||||
|
struct amdgpu_reset_control *reset_ctl,
|
||||||
|
struct amdgpu_reset_context *reset_context)
|
||||||
|
{
|
||||||
|
struct list_head *reset_device_list = reset_context->reset_device_list;
|
||||||
|
struct amdgpu_device *tmp_adev = NULL;
|
||||||
|
int r;
|
||||||
|
|
||||||
|
r = amdgpu_device_reinit_after_reset(reset_context);
|
||||||
|
if (r)
|
||||||
|
return r;
|
||||||
|
list_for_each_entry(tmp_adev, reset_device_list, reset_list) {
|
||||||
|
if (!tmp_adev->kfd.init_complete) {
|
||||||
|
kgd2kfd_init_zone_device(tmp_adev);
|
||||||
|
amdgpu_amdkfd_device_init(tmp_adev);
|
||||||
|
amdgpu_amdkfd_drm_client_create(tmp_adev);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return r;
|
||||||
|
}
|
||||||
|
|
||||||
|
static int amdgpu_reset_xgmi_reset_on_init_perform_reset(
|
||||||
|
struct amdgpu_reset_control *reset_ctl,
|
||||||
|
struct amdgpu_reset_context *reset_context)
|
||||||
|
{
|
||||||
|
struct amdgpu_device *adev = (struct amdgpu_device *)reset_ctl->handle;
|
||||||
|
struct list_head *reset_device_list = reset_context->reset_device_list;
|
||||||
|
struct amdgpu_device *tmp_adev = NULL;
|
||||||
|
int r;
|
||||||
|
|
||||||
|
dev_dbg(adev->dev, "xgmi roi - hw reset\n");
|
||||||
|
|
||||||
|
list_for_each_entry(tmp_adev, reset_device_list, reset_list) {
|
||||||
|
mutex_lock(&tmp_adev->reset_cntl->reset_lock);
|
||||||
|
tmp_adev->reset_cntl->active_reset =
|
||||||
|
amdgpu_asic_reset_method(adev);
|
||||||
|
}
|
||||||
|
r = 0;
|
||||||
|
/* Mode1 reset needs to be triggered on all devices together */
|
||||||
|
list_for_each_entry(tmp_adev, reset_device_list, reset_list) {
|
||||||
|
/* For XGMI run all resets in parallel to speed up the process */
|
||||||
|
if (!queue_work(system_unbound_wq, &tmp_adev->xgmi_reset_work))
|
||||||
|
r = -EALREADY;
|
||||||
|
if (r) {
|
||||||
|
dev_err(tmp_adev->dev,
|
||||||
|
"xgmi reset on init: reset failed with error, %d",
|
||||||
|
r);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/* For XGMI wait for all resets to complete before proceed */
|
||||||
|
if (!r) {
|
||||||
|
list_for_each_entry(tmp_adev, reset_device_list, reset_list) {
|
||||||
|
flush_work(&tmp_adev->xgmi_reset_work);
|
||||||
|
r = tmp_adev->asic_reset_res;
|
||||||
|
if (r)
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
list_for_each_entry(tmp_adev, reset_device_list, reset_list) {
|
||||||
|
mutex_unlock(&tmp_adev->reset_cntl->reset_lock);
|
||||||
|
tmp_adev->reset_cntl->active_reset = AMD_RESET_METHOD_NONE;
|
||||||
|
}
|
||||||
|
|
||||||
|
return r;
|
||||||
|
}
|
||||||
|
|
||||||
|
int amdgpu_reset_do_xgmi_reset_on_init(
|
||||||
|
struct amdgpu_reset_context *reset_context)
|
||||||
|
{
|
||||||
|
struct list_head *reset_device_list = reset_context->reset_device_list;
|
||||||
|
struct amdgpu_device *adev;
|
||||||
|
int r;
|
||||||
|
|
||||||
|
if (!reset_device_list || list_empty(reset_device_list) ||
|
||||||
|
list_is_singular(reset_device_list))
|
||||||
|
return -EINVAL;
|
||||||
|
|
||||||
|
adev = list_first_entry(reset_device_list, struct amdgpu_device,
|
||||||
|
reset_list);
|
||||||
|
r = amdgpu_reset_prepare_hwcontext(adev, reset_context);
|
||||||
|
if (r)
|
||||||
|
return r;
|
||||||
|
|
||||||
|
r = amdgpu_reset_perform_reset(adev, reset_context);
|
||||||
|
|
||||||
|
return r;
|
||||||
|
}
|
||||||
|
|
||||||
|
struct amdgpu_reset_handler xgmi_reset_on_init_handler = {
|
||||||
|
.reset_method = AMD_RESET_METHOD_ON_INIT,
|
||||||
|
.prepare_env = NULL,
|
||||||
|
.prepare_hwcontext = amdgpu_reset_xgmi_reset_on_init_prep_hwctxt,
|
||||||
|
.perform_reset = amdgpu_reset_xgmi_reset_on_init_perform_reset,
|
||||||
|
.restore_hwcontext = amdgpu_reset_xgmi_reset_on_init_restore_hwctxt,
|
||||||
|
.restore_env = NULL,
|
||||||
|
.do_reset = NULL,
|
||||||
|
};
|
||||||
|
|
||||||
int amdgpu_reset_init(struct amdgpu_device *adev)
|
int amdgpu_reset_init(struct amdgpu_device *adev)
|
||||||
{
|
{
|
||||||
int ret = 0;
|
int ret = 0;
|
||||||
|
|
|
@ -153,4 +153,9 @@ void amdgpu_reset_get_desc(struct amdgpu_reset_context *rst_ctxt, char *buf,
|
||||||
for (i = 0; (i < AMDGPU_RESET_MAX_HANDLERS) && \
|
for (i = 0; (i < AMDGPU_RESET_MAX_HANDLERS) && \
|
||||||
(handler = (*reset_ctl->reset_handlers)[i]); \
|
(handler = (*reset_ctl->reset_handlers)[i]); \
|
||||||
++i)
|
++i)
|
||||||
|
|
||||||
|
extern struct amdgpu_reset_handler xgmi_reset_on_init_handler;
|
||||||
|
int amdgpu_reset_do_xgmi_reset_on_init(
|
||||||
|
struct amdgpu_reset_context *reset_context);
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
|
Loading…
Add table
Reference in a new issue