From: "Deucher, Alexander" <Alexander.Deucher@amd.com>
To: "Khatri, Sunil" <Sunil.Khatri@amd.com>,
"Koenig, Christian" <Christian.Koenig@amd.com>
Cc: "amd-gfx@lists.freedesktop.org" <amd-gfx@lists.freedesktop.org>,
"Khatri, Sunil" <Sunil.Khatri@amd.com>
Subject: RE: [PATCH v3 3/4] drm/amdgpu: add support to dump gfx10 queue registers
Date: Wed, 15 May 2024 20:12:20 +0000 [thread overview]
Message-ID: <BL1PR12MB5144866E6DD7FD0EC9226EB9F7EC2@BL1PR12MB5144.namprd12.prod.outlook.com> (raw)
In-Reply-To: <20240515121754.1810181-3-sunil.khatri@amd.com>
[Public]
> -----Original Message-----
> From: Sunil Khatri <sunil.khatri@amd.com>
> Sent: Wednesday, May 15, 2024 8:18 AM
> To: Deucher, Alexander <Alexander.Deucher@amd.com>; Koenig, Christian
> <Christian.Koenig@amd.com>
> Cc: amd-gfx@lists.freedesktop.org; Khatri, Sunil <Sunil.Khatri@amd.com>
> Subject: [PATCH v3 3/4] drm/amdgpu: add support to dump gfx10 queue
> registers
>
> Add gfx queue register for all instances in ip dump for gfx10.
>
> Signed-off-by: Sunil Khatri <sunil.khatri@amd.com>
> ---
> drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h | 1 +
> drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c | 86
> +++++++++++++++++++++++++
> 2 files changed, 87 insertions(+)
>
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
> b/drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
> index d96873c154ed..54232066cd3b 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
> @@ -437,6 +437,7 @@ struct amdgpu_gfx {
> /* IP reg dump */
> uint32_t *ipdump_core;
> uint32_t *ipdump_cp;
> + uint32_t *ipdump_gfx_queue;
I'd call this ip_dump_gfx or ip_dump_gfx_queues to better align with that it stores.
> };
>
> struct amdgpu_gfx_ras_reg_entry {
> diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
> b/drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
> index daf9a3571183..5b8132ecc039 100644
> --- a/drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
> +++ b/drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
> @@ -424,6 +424,33 @@ static const struct amdgpu_hwip_reg_entry
> gc_cp_reg_list_10[] = {
> SOC15_REG_ENTRY_STR(GC, 0, mmCP_HQD_DEQUEUE_STATUS) };
>
> +static const struct amdgpu_hwip_reg_entry gc_gfx_queue_reg_list_10[] = {
> + /* gfx queue registers */
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_ACTIVE),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_QUEUE_PRIORITY),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_BASE),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_BASE_HI),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_OFFSET),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_CSMD_RPTR),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_WPTR),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_WPTR_HI),
> + SOC15_REG_ENTRY_STR(GC, 0,
> mmCP_GFX_HQD_DEQUEUE_REQUEST),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_MAPPED),
> + SOC15_REG_ENTRY_STR(GC, 0,
> mmCP_GFX_HQD_QUE_MGR_CONTROL),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_HQ_CONTROL0),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_HQ_STATUS0),
> + SOC15_REG_ENTRY_STR(GC, 0,
> mmCP_GFX_HQD_CE_WPTR_POLL_ADDR_LO),
> + SOC15_REG_ENTRY_STR(GC, 0,
> mmCP_GFX_HQD_CE_WPTR_POLL_ADDR_HI),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_CE_OFFSET),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_CE_CSMD_RPTR),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_CE_WPTR),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_HQD_CE_WPTR_HI),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_MQD_BASE_ADDR),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_GFX_MQD_BASE_ADDR_HI),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_RB_WPTR_POLL_ADDR_LO),
> + SOC15_REG_ENTRY_STR(GC, 0, mmCP_RB_WPTR_POLL_ADDR_HI) };
> +
> static const struct soc15_reg_golden golden_settings_gc_10_1[] = {
> SOC15_REG_GOLDEN_VALUE(GC, 0, mmCB_HW_CONTROL_4,
> 0xffffffff, 0x00400014),
> SOC15_REG_GOLDEN_VALUE(GC, 0, mmCGTT_CPF_CLK_CTRL,
> 0xfcff8fff, 0xf8000100), @@ -4664,6 +4691,19 @@ static void
> gfx_v10_0_alloc_ip_dump(struct amdgpu_device *adev)
> } else {
> adev->gfx.ipdump_cp = ptr;
> }
> +
> + /* Allocate memory for gfx cp queue registers for all the instances */
> + reg_count = ARRAY_SIZE(gc_gfx_queue_reg_list_10);
> + inst = adev->gfx.me.num_me * adev->gfx.me.num_pipe_per_me *
> + adev->gfx.me.num_queue_per_pipe;
> +
> + ptr = kcalloc(reg_count * inst, sizeof(uint32_t), GFP_KERNEL);
> + if (ptr == NULL) {
> + DRM_ERROR("Failed to allocate memory for GFX CP IP
> Dump\n");
> + adev->gfx.ipdump_gfx_queue = NULL;
> + } else {
> + adev->gfx.ipdump_gfx_queue = ptr;
> + }
> }
>
> static int gfx_v10_0_sw_init(void *handle) @@ -4874,6 +4914,7 @@ static
> int gfx_v10_0_sw_fini(void *handle)
>
> kfree(adev->gfx.ipdump_core);
> kfree(adev->gfx.ipdump_cp);
> + kfree(adev->gfx.ipdump_gfx_queue);
>
> return 0;
> }
> @@ -9368,6 +9409,26 @@ static void gfx_v10_ip_print(void *handle, struct
> drm_printer *p)
> }
> }
> }
> +
> + /* print gfx queue registers for all instances */
> + if (!adev->gfx.ipdump_gfx_queue)
> + return;
> +
> + reg_count = ARRAY_SIZE(gc_gfx_queue_reg_list_10);
> +
> + for (i = 0; i < adev->gfx.me.num_me; i++) {
> + for (j = 0; j < adev->gfx.me.num_pipe_per_me; j++) {
> + for (k = 0; k < adev->gfx.me.num_queue_per_pipe;
> k++) {
> + drm_printf(p, "me %d, pipe %d, queue %d\n",
> i, j, k);
> + for (reg = 0; reg < reg_count; reg++) {
> + drm_printf(p, "%-50s \t 0x%08x\n",
> +
> gc_gfx_queue_reg_list_10[reg].reg_name,
> + adev-
> >gfx.ipdump_gfx_queue[index + reg]);
> + }
> + index += reg_count;
> + }
> + }
> + }
> }
>
> static void gfx_v10_ip_dump(void *handle) @@ -9414,6 +9475,31 @@ static
> void gfx_v10_ip_dump(void *handle)
> nv_grbm_select(adev, 0, 0, 0, 0);
> mutex_unlock(&adev->srbm_mutex);
> amdgpu_gfx_off_ctrl(adev, true);
> +
> + /* dump gfx queue registers for all instances */
> + if (!adev->gfx.ipdump_gfx_queue)
> + return;
> +
> + reg_count = ARRAY_SIZE(gc_gfx_queue_reg_list_10);
> + amdgpu_gfx_off_ctrl(adev, false);
> + mutex_lock(&adev->srbm_mutex);
> + for (i = 0; i < adev->gfx.me.num_me; i++) {
> + for (j = 0; j < adev->gfx.me.num_pipe_per_me; j++) {
> + for (k = 0; k < adev->gfx.me.num_queue_per_pipe;
> k++) {
> + nv_grbm_select(adev, i, j, k, 0);
> +
> + for (reg = 0; reg < reg_count; reg++) {
> + adev->gfx.ipdump_gfx_queue[index +
> reg] =
> +
> RREG32(SOC15_REG_ENTRY_OFFSET(
> +
> gc_gfx_queue_reg_list_10[reg]));
> + }
> + index += reg_count;
> + }
> + }
> + }
Does this one not need an msleep?
Alex
> + nv_grbm_select(adev, 0, 0, 0, 0);
> + mutex_unlock(&adev->srbm_mutex);
> + amdgpu_gfx_off_ctrl(adev, true);
> }
>
> static const struct amd_ip_funcs gfx_v10_0_ip_funcs = {
> --
> 2.34.1
next prev parent reply other threads:[~2024-05-15 20:12 UTC|newest]
Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-05-15 12:17 [PATCH v3 1/4] drm/amdgpu: update the ip_dump to ipdump_core Sunil Khatri
2024-05-15 12:17 ` [PATCH v3 2/4] drm/amdgpu: Add support to dump gfx10 cp registers Sunil Khatri
2024-05-15 20:10 ` Deucher, Alexander
2024-05-16 3:41 ` Khatri, Sunil
2024-05-15 12:17 ` [PATCH v3 3/4] drm/amdgpu: add support to dump gfx10 queue registers Sunil Khatri
2024-05-15 20:12 ` Deucher, Alexander [this message]
2024-05-16 3:40 ` Khatri, Sunil
2024-05-15 12:17 ` [PATCH v3 4/4] drm/amdgpu: add prints while ip registr dump Sunil Khatri
2024-05-15 20:07 ` [PATCH v3 1/4] drm/amdgpu: update the ip_dump to ipdump_core Deucher, Alexander
2024-05-16 3:38 ` Khatri, Sunil
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=BL1PR12MB5144866E6DD7FD0EC9226EB9F7EC2@BL1PR12MB5144.namprd12.prod.outlook.com \
--to=alexander.deucher@amd.com \
--cc=Christian.Koenig@amd.com \
--cc=Sunil.Khatri@amd.com \
--cc=amd-gfx@lists.freedesktop.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).