AMD-GFX Archive mirror
 help / color / mirror / Atom feed
From: "Khatri, Sunil" <sukhatri@amd.com>
To: "Deucher, Alexander" <Alexander.Deucher@amd.com>,
	"Khatri, Sunil" <Sunil.Khatri@amd.com>,
	"Koenig, Christian" <Christian.Koenig@amd.com>
Cc: "amd-gfx@lists.freedesktop.org" <amd-gfx@lists.freedesktop.org>
Subject: Re: [PATCH v3 3/4] drm/amdgpu: add support to dump gfx10 queue registers
Date: Thu, 16 May 2024 09:10:28 +0530	[thread overview]
Message-ID: <62b9661e-9d70-4152-af4b-2fa9ddac49c4@amd.com> (raw)
In-Reply-To: <BL1PR12MB5144866E6DD7FD0EC9226EB9F7EC2@BL1PR12MB5144.namprd12.prod.outlook.com>


On 5/16/2024 1:42 AM, Deucher, Alexander wrote:
> [Public]
>
>> -----Original Message-----
>> From: Sunil Khatri <sunil.khatri@amd.com>
>> Sent: Wednesday, May 15, 2024 8:18 AM
>> To: Deucher, Alexander <Alexander.Deucher@amd.com>; Koenig, Christian
>> <Christian.Koenig@amd.com>
>> Cc: amd-gfx@lists.freedesktop.org; Khatri, Sunil <Sunil.Khatri@amd.com>
>> Subject: [PATCH v3 3/4] drm/amdgpu: add support to dump gfx10 queue
>> registers
>>
>> Add gfx queue register for all instances in ip dump for gfx10.
>>
>> Signed-off-by: Sunil Khatri <sunil.khatri@amd.com>
>> ---
>>   drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h |  1 +
>> drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c  | 86
>> +++++++++++++++++++++++++
>>   2 files changed, 87 insertions(+)
>>
>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
>> index d96873c154ed..54232066cd3b 100644
>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
>> @@ -437,6 +437,7 @@ struct amdgpu_gfx {
>>        /* IP reg dump */
>>        uint32_t                        *ipdump_core;
>>        uint32_t                        *ipdump_cp;
>> +     uint32_t                        *ipdump_gfx_queue;
> I'd call this ip_dump_gfx or ip_dump_gfx_queues to better align with that it stores.
>
>>   };
>>
>>   struct amdgpu_gfx_ras_reg_entry {
>> diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
>> b/drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
>> index daf9a3571183..5b8132ecc039 100644
>> --- a/drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
>> +++ b/drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
>> @@ -424,6 +424,33 @@ static const struct amdgpu_hwip_reg_entry
>> gc_cp_reg_list_10[] = {
>>        SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_DEQUEUE_STATUS)  };
>>
>> +static const struct amdgpu_hwip_reg_entry gc_gfx_queue_reg_list_10[] = {
>> +     /* gfx queue registers */
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_ACTIVE),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_QUEUE_PRIORITY),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_BASE),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_BASE_HI),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_OFFSET),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_CSMD_RPTR),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_WPTR),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_WPTR_HI),
>> +     SOC15_REG_ENTRY_STR(GC, 0,
>> mmCP_GFX_HQD_DEQUEUE_REQUEST),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_MAPPED),
>> +     SOC15_REG_ENTRY_STR(GC, 0,
>> mmCP_GFX_HQD_QUE_MGR_CONTROL),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_HQ_CONTROL0),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_HQ_STATUS0),
>> +     SOC15_REG_ENTRY_STR(GC, 0,
>> mmCP_GFX_HQD_CE_WPTR_POLL_ADDR_LO),
>> +     SOC15_REG_ENTRY_STR(GC, 0,
>> mmCP_GFX_HQD_CE_WPTR_POLL_ADDR_HI),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_CE_OFFSET),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_CE_CSMD_RPTR),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_CE_WPTR),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_CE_WPTR_HI),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_MQD_BASE_ADDR),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_MQD_BASE_ADDR_HI),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_RB_WPTR_POLL_ADDR_LO),
>> +     SOC15_REG_ENTRY_STR(GC, 0, mmCP_RB_WPTR_POLL_ADDR_HI) };
>> +
>>   static const struct soc15_reg_golden golden_settings_gc_10_1[] = {
>>        SOC15_REG_GOLDEN_VALUE(GC, 0, mmCB_HW_CONTROL_4,
>> 0xffffffff, 0x00400014),
>>        SOC15_REG_GOLDEN_VALUE(GC, 0, mmCGTT_CPF_CLK_CTRL,
>> 0xfcff8fff, 0xf8000100), @@ -4664,6 +4691,19 @@ static void
>> gfx_v10_0_alloc_ip_dump(struct amdgpu_device *adev)
>>        } else {
>>                adev->gfx.ipdump_cp = ptr;
>>        }
>> +
>> +     /* Allocate memory for gfx cp queue registers for all the instances */
>> +     reg_count = ARRAY_SIZE(gc_gfx_queue_reg_list_10);
>> +     inst = adev->gfx.me.num_me * adev->gfx.me.num_pipe_per_me *
>> +             adev->gfx.me.num_queue_per_pipe;
>> +
>> +     ptr = kcalloc(reg_count * inst, sizeof(uint32_t), GFP_KERNEL);
>> +     if (ptr == NULL) {
>> +             DRM_ERROR("Failed to allocate memory for GFX CP IP
>> Dump\n");
>> +             adev->gfx.ipdump_gfx_queue = NULL;
>> +     } else {
>> +             adev->gfx.ipdump_gfx_queue = ptr;
>> +     }
>>   }
>>
>>   static int gfx_v10_0_sw_init(void *handle) @@ -4874,6 +4914,7 @@ static
>> int gfx_v10_0_sw_fini(void *handle)
>>
>>        kfree(adev->gfx.ipdump_core);
>>        kfree(adev->gfx.ipdump_cp);
>> +     kfree(adev->gfx.ipdump_gfx_queue);
>>
>>        return 0;
>>   }
>> @@ -9368,6 +9409,26 @@ static void gfx_v10_ip_print(void *handle, struct
>> drm_printer *p)
>>                        }
>>                }
>>        }
>> +
>> +     /* print gfx queue registers for all instances */
>> +     if (!adev->gfx.ipdump_gfx_queue)
>> +             return;
>> +
>> +     reg_count = ARRAY_SIZE(gc_gfx_queue_reg_list_10);
>> +
>> +     for (i = 0; i < adev->gfx.me.num_me; i++) {
>> +             for (j = 0; j < adev->gfx.me.num_pipe_per_me; j++) {
>> +                     for (k = 0; k < adev->gfx.me.num_queue_per_pipe;
>> k++) {
>> +                             drm_printf(p, "me %d, pipe %d, queue %d\n",
>> i, j, k);
>> +                             for (reg = 0; reg < reg_count; reg++) {
>> +                                     drm_printf(p, "%-50s \t 0x%08x\n",
>> +
>> gc_gfx_queue_reg_list_10[reg].reg_name,
>> +                                                adev-
>>> gfx.ipdump_gfx_queue[index + reg]);
>> +                             }
>> +                             index += reg_count;
>> +                     }
>> +             }
>> +     }
>>   }
>>
>>   static void gfx_v10_ip_dump(void *handle) @@ -9414,6 +9475,31 @@ static
>> void gfx_v10_ip_dump(void *handle)
>>        nv_grbm_select(adev, 0, 0, 0, 0);
>>        mutex_unlock(&adev->srbm_mutex);
>>        amdgpu_gfx_off_ctrl(adev, true);
>> +
>> +     /* dump gfx queue registers for all instances */
>> +     if (!adev->gfx.ipdump_gfx_queue)
>> +             return;
>> +
>> +     reg_count = ARRAY_SIZE(gc_gfx_queue_reg_list_10);
>> +     amdgpu_gfx_off_ctrl(adev, false);
>> +     mutex_lock(&adev->srbm_mutex);
>> +     for (i = 0; i < adev->gfx.me.num_me; i++) {
>> +             for (j = 0; j < adev->gfx.me.num_pipe_per_me; j++) {
>> +                     for (k = 0; k < adev->gfx.me.num_queue_per_pipe;
>> k++) {
>> +                             nv_grbm_select(adev, i, j, k, 0);
>> +
>> +                             for (reg = 0; reg < reg_count; reg++) {
>> +                                     adev->gfx.ipdump_gfx_queue[index +
>> reg] =
>> +
>>        RREG32(SOC15_REG_ENTRY_OFFSET(
>> +
>>        gc_gfx_queue_reg_list_10[reg]));
>> +                             }
>> +                             index += reg_count;
>> +                     }
>> +             }
>> +     }
> Does this one not need an msleep?

Since the registers are also less and also total no of times loop runs 
is lesser too. But we can add the msleep as it does not seems to add any 
delay in overall.

Regards
Sunil Khatri

>
> Alex
>
>> +     nv_grbm_select(adev, 0, 0, 0, 0);
>> +     mutex_unlock(&adev->srbm_mutex);
>> +     amdgpu_gfx_off_ctrl(adev, true);
>>   }
>>
>>   static const struct amd_ip_funcs gfx_v10_0_ip_funcs = {
>> --
>> 2.34.1

  reply	other threads:[~2024-05-16  3:40 UTC|newest]

Thread overview: 10+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2024-05-15 12:17 [PATCH v3 1/4] drm/amdgpu: update the ip_dump to ipdump_core Sunil Khatri
2024-05-15 12:17 ` [PATCH v3 2/4] drm/amdgpu: Add support to dump gfx10 cp registers Sunil Khatri
2024-05-15 20:10   ` Deucher, Alexander
2024-05-16  3:41     ` Khatri, Sunil
2024-05-15 12:17 ` [PATCH v3 3/4] drm/amdgpu: add support to dump gfx10 queue registers Sunil Khatri
2024-05-15 20:12   ` Deucher, Alexander
2024-05-16  3:40     ` Khatri, Sunil [this message]
2024-05-15 12:17 ` [PATCH v3 4/4] drm/amdgpu: add prints while ip registr dump Sunil Khatri
2024-05-15 20:07 ` [PATCH v3 1/4] drm/amdgpu: update the ip_dump to ipdump_core Deucher, Alexander
2024-05-16  3:38   ` Khatri, Sunil

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=62b9661e-9d70-4152-af4b-2fa9ddac49c4@amd.com \
    --to=sukhatri@amd.com \
    --cc=Alexander.Deucher@amd.com \
    --cc=Christian.Koenig@amd.com \
    --cc=Sunil.Khatri@amd.com \
    --cc=amd-gfx@lists.freedesktop.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).