mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2026-05-01 03:44:27 -04:00
drm/amdgpu/sdma4.4.2: implement ring reset callback for sdma4.4.2
Implement sdma queue reset callback via SMU interface.
v2: Leverage inst_stop/start functions in reset sequence.
Use GET_INST for physical SDMA instance.
Disable apu for sdma reset.
v3: Rephrase error prints.
v4: Remove redundant prints. Remove setting PREEMPT registers as
soft reset handles it.
Signed-off-by: Jiadong Zhu <Jiadong.Zhu@amd.com>
Reviewed-by: Alex Deucher <alexander.deucher@amd.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
This commit is contained in:
committed by
Alex Deucher
parent
610696505c
commit
52b10d55c1
@@ -667,11 +667,12 @@ static uint32_t sdma_v4_4_2_rb_cntl(struct amdgpu_ring *ring, uint32_t rb_cntl)
|
||||
*
|
||||
* @adev: amdgpu_device pointer
|
||||
* @i: instance to resume
|
||||
* @restore: used to restore wptr when restart
|
||||
*
|
||||
* Set up the gfx DMA ring buffers and enable them.
|
||||
* Returns 0 for success, error for failure.
|
||||
*/
|
||||
static void sdma_v4_4_2_gfx_resume(struct amdgpu_device *adev, unsigned int i)
|
||||
static void sdma_v4_4_2_gfx_resume(struct amdgpu_device *adev, unsigned int i, bool restore)
|
||||
{
|
||||
struct amdgpu_ring *ring = &adev->sdma.instance[i].ring;
|
||||
u32 rb_cntl, ib_cntl, wptr_poll_cntl;
|
||||
@@ -698,16 +699,24 @@ static void sdma_v4_4_2_gfx_resume(struct amdgpu_device *adev, unsigned int i)
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_BASE, ring->gpu_addr >> 8);
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_BASE_HI, ring->gpu_addr >> 40);
|
||||
|
||||
ring->wptr = 0;
|
||||
if (!restore)
|
||||
ring->wptr = 0;
|
||||
|
||||
/* before programing wptr to a less value, need set minor_ptr_update first */
|
||||
WREG32_SDMA(i, regSDMA_GFX_MINOR_PTR_UPDATE, 1);
|
||||
|
||||
/* Initialize the ring buffer's read and write pointers */
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_RPTR, 0);
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_RPTR_HI, 0);
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_WPTR, 0);
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_WPTR_HI, 0);
|
||||
if (restore) {
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_RPTR, lower_32_bits(ring->wptr << 2));
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_RPTR_HI, upper_32_bits(ring->wptr << 2));
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_WPTR, lower_32_bits(ring->wptr << 2));
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_WPTR_HI, upper_32_bits(ring->wptr << 2));
|
||||
} else {
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_RPTR, 0);
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_RPTR_HI, 0);
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_WPTR, 0);
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_WPTR_HI, 0);
|
||||
}
|
||||
|
||||
doorbell = RREG32_SDMA(i, regSDMA_GFX_DOORBELL);
|
||||
doorbell_offset = RREG32_SDMA(i, regSDMA_GFX_DOORBELL_OFFSET);
|
||||
@@ -759,7 +768,7 @@ static void sdma_v4_4_2_gfx_resume(struct amdgpu_device *adev, unsigned int i)
|
||||
* Set up the page DMA ring buffers and enable them.
|
||||
* Returns 0 for success, error for failure.
|
||||
*/
|
||||
static void sdma_v4_4_2_page_resume(struct amdgpu_device *adev, unsigned int i)
|
||||
static void sdma_v4_4_2_page_resume(struct amdgpu_device *adev, unsigned int i, bool restore)
|
||||
{
|
||||
struct amdgpu_ring *ring = &adev->sdma.instance[i].page;
|
||||
u32 rb_cntl, ib_cntl, wptr_poll_cntl;
|
||||
@@ -775,10 +784,17 @@ static void sdma_v4_4_2_page_resume(struct amdgpu_device *adev, unsigned int i)
|
||||
WREG32_SDMA(i, regSDMA_PAGE_RB_CNTL, rb_cntl);
|
||||
|
||||
/* Initialize the ring buffer's read and write pointers */
|
||||
WREG32_SDMA(i, regSDMA_PAGE_RB_RPTR, 0);
|
||||
WREG32_SDMA(i, regSDMA_PAGE_RB_RPTR_HI, 0);
|
||||
WREG32_SDMA(i, regSDMA_PAGE_RB_WPTR, 0);
|
||||
WREG32_SDMA(i, regSDMA_PAGE_RB_WPTR_HI, 0);
|
||||
if (restore) {
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_RPTR, lower_32_bits(ring->wptr << 2));
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_RPTR_HI, upper_32_bits(ring->wptr << 2));
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_WPTR, lower_32_bits(ring->wptr << 2));
|
||||
WREG32_SDMA(i, regSDMA_GFX_RB_WPTR_HI, upper_32_bits(ring->wptr << 2));
|
||||
} else {
|
||||
WREG32_SDMA(i, regSDMA_PAGE_RB_RPTR, 0);
|
||||
WREG32_SDMA(i, regSDMA_PAGE_RB_RPTR_HI, 0);
|
||||
WREG32_SDMA(i, regSDMA_PAGE_RB_WPTR, 0);
|
||||
WREG32_SDMA(i, regSDMA_PAGE_RB_WPTR_HI, 0);
|
||||
}
|
||||
|
||||
/* set the wb address whether it's enabled or not */
|
||||
WREG32_SDMA(i, regSDMA_PAGE_RB_RPTR_ADDR_HI,
|
||||
@@ -792,7 +808,8 @@ static void sdma_v4_4_2_page_resume(struct amdgpu_device *adev, unsigned int i)
|
||||
WREG32_SDMA(i, regSDMA_PAGE_RB_BASE, ring->gpu_addr >> 8);
|
||||
WREG32_SDMA(i, regSDMA_PAGE_RB_BASE_HI, ring->gpu_addr >> 40);
|
||||
|
||||
ring->wptr = 0;
|
||||
if (!restore)
|
||||
ring->wptr = 0;
|
||||
|
||||
/* before programing wptr to a less value, need set minor_ptr_update first */
|
||||
WREG32_SDMA(i, regSDMA_PAGE_MINOR_PTR_UPDATE, 1);
|
||||
@@ -916,7 +933,7 @@ static int sdma_v4_4_2_inst_load_microcode(struct amdgpu_device *adev,
|
||||
* Returns 0 for success, error for failure.
|
||||
*/
|
||||
static int sdma_v4_4_2_inst_start(struct amdgpu_device *adev,
|
||||
uint32_t inst_mask)
|
||||
uint32_t inst_mask, bool restore)
|
||||
{
|
||||
struct amdgpu_ring *ring;
|
||||
uint32_t tmp_mask;
|
||||
@@ -927,7 +944,7 @@ static int sdma_v4_4_2_inst_start(struct amdgpu_device *adev,
|
||||
sdma_v4_4_2_inst_enable(adev, false, inst_mask);
|
||||
} else {
|
||||
/* bypass sdma microcode loading on Gopher */
|
||||
if (adev->firmware.load_type != AMDGPU_FW_LOAD_PSP &&
|
||||
if (!restore && adev->firmware.load_type != AMDGPU_FW_LOAD_PSP &&
|
||||
adev->sdma.instance[0].fw) {
|
||||
r = sdma_v4_4_2_inst_load_microcode(adev, inst_mask);
|
||||
if (r)
|
||||
@@ -946,9 +963,9 @@ static int sdma_v4_4_2_inst_start(struct amdgpu_device *adev,
|
||||
uint32_t temp;
|
||||
|
||||
WREG32_SDMA(i, regSDMA_SEM_WAIT_FAIL_TIMER_CNTL, 0);
|
||||
sdma_v4_4_2_gfx_resume(adev, i);
|
||||
sdma_v4_4_2_gfx_resume(adev, i, restore);
|
||||
if (adev->sdma.has_page_queue)
|
||||
sdma_v4_4_2_page_resume(adev, i);
|
||||
sdma_v4_4_2_page_resume(adev, i, restore);
|
||||
|
||||
/* set utc l1 enable flag always to 1 */
|
||||
temp = RREG32_SDMA(i, regSDMA_CNTL);
|
||||
@@ -1486,7 +1503,7 @@ static int sdma_v4_4_2_hw_init(struct amdgpu_ip_block *ip_block)
|
||||
if (!amdgpu_sriov_vf(adev))
|
||||
sdma_v4_4_2_inst_init_golden_registers(adev, inst_mask);
|
||||
|
||||
r = sdma_v4_4_2_inst_start(adev, inst_mask);
|
||||
r = sdma_v4_4_2_inst_start(adev, inst_mask, false);
|
||||
|
||||
return r;
|
||||
}
|
||||
@@ -1573,6 +1590,42 @@ static int sdma_v4_4_2_soft_reset(struct amdgpu_ip_block *ip_block)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int sdma_v4_4_2_reset_queue(struct amdgpu_ring *ring, unsigned int vmid)
|
||||
{
|
||||
struct amdgpu_device *adev = ring->adev;
|
||||
int i, r;
|
||||
u32 inst_mask;
|
||||
|
||||
if ((adev->flags & AMD_IS_APU) || amdgpu_sriov_vf(adev))
|
||||
return -EINVAL;
|
||||
|
||||
/* stop queue */
|
||||
inst_mask = 1 << ring->me;
|
||||
sdma_v4_4_2_inst_gfx_stop(adev, inst_mask);
|
||||
if (adev->sdma.has_page_queue)
|
||||
sdma_v4_4_2_inst_page_stop(adev, inst_mask);
|
||||
|
||||
r = amdgpu_dpm_reset_sdma(adev, 1 << GET_INST(SDMA0, ring->me));
|
||||
if (r)
|
||||
return r;
|
||||
|
||||
udelay(50);
|
||||
|
||||
for (i = 0; i < adev->usec_timeout; i++) {
|
||||
if (!REG_GET_FIELD(RREG32_SDMA(ring->me, regSDMA_F32_CNTL), SDMA_F32_CNTL, HALT))
|
||||
break;
|
||||
udelay(1);
|
||||
}
|
||||
|
||||
if (i == adev->usec_timeout) {
|
||||
dev_err(adev->dev, "timed out waiting for SDMA%d unhalt after reset\n",
|
||||
ring->me);
|
||||
return -ETIMEDOUT;
|
||||
}
|
||||
|
||||
return sdma_v4_4_2_inst_start(adev, inst_mask, true);
|
||||
}
|
||||
|
||||
static int sdma_v4_4_2_set_trap_irq_state(struct amdgpu_device *adev,
|
||||
struct amdgpu_irq_src *source,
|
||||
unsigned type,
|
||||
@@ -1955,6 +2008,7 @@ static const struct amdgpu_ring_funcs sdma_v4_4_2_ring_funcs = {
|
||||
.emit_wreg = sdma_v4_4_2_ring_emit_wreg,
|
||||
.emit_reg_wait = sdma_v4_4_2_ring_emit_reg_wait,
|
||||
.emit_reg_write_reg_wait = amdgpu_ring_emit_reg_write_reg_wait_helper,
|
||||
.reset = sdma_v4_4_2_reset_queue,
|
||||
};
|
||||
|
||||
static const struct amdgpu_ring_funcs sdma_v4_4_2_page_ring_funcs = {
|
||||
@@ -2167,7 +2221,7 @@ static int sdma_v4_4_2_xcp_resume(void *handle, uint32_t inst_mask)
|
||||
if (!amdgpu_sriov_vf(adev))
|
||||
sdma_v4_4_2_inst_init_golden_registers(adev, inst_mask);
|
||||
|
||||
r = sdma_v4_4_2_inst_start(adev, inst_mask);
|
||||
r = sdma_v4_4_2_inst_start(adev, inst_mask, false);
|
||||
|
||||
return r;
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user