qemu-ppc
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [PATCH v2 11/14] ppc/xive2: Check crowd backlog when scanning group


From: Nicholas Piggin
Subject: Re: [PATCH v2 11/14] ppc/xive2: Check crowd backlog when scanning group backlog
Date: Mon, 10 Mar 2025 17:32:30 +1000

On Tue Dec 10, 2024 at 10:05 AM AEST, Michael Kowal wrote:
> From: Frederic Barrat <fbarrat@linux.ibm.com>
>
> When processing a backlog scan for group interrupts, also take
> into account crowd interrupts.
>
> Signed-off-by: Frederic Barrat <fbarrat@linux.ibm.com>
> Signed-off-by: Michael Kowal <kowal@linux.ibm.com>

Reviewed-by: Nicholas Piggin <npiggin@gmail.com>

> ---
>  include/hw/ppc/xive2_regs.h |  4 ++
>  hw/intc/xive2.c             | 82 +++++++++++++++++++++++++------------
>  2 files changed, 60 insertions(+), 26 deletions(-)
>
> diff --git a/include/hw/ppc/xive2_regs.h b/include/hw/ppc/xive2_regs.h
> index 9bcf7a8a6f..b11395c563 100644
> --- a/include/hw/ppc/xive2_regs.h
> +++ b/include/hw/ppc/xive2_regs.h
> @@ -236,4 +236,8 @@ void xive2_nvgc_pic_print_info(Xive2Nvgc *nvgc, uint32_t 
> nvgc_idx,
>  #define NVx_BACKLOG_OP            PPC_BITMASK(52, 53)
>  #define NVx_BACKLOG_PRIO          PPC_BITMASK(57, 59)
>  
> +/* split the 6-bit crowd/group level */
> +#define NVx_CROWD_LVL(level)      ((level >> 4) & 0b11)
> +#define NVx_GROUP_LVL(level)      (level & 0b1111)
> +
>  #endif /* PPC_XIVE2_REGS_H */
> diff --git a/hw/intc/xive2.c b/hw/intc/xive2.c
> index 20d63e8f6e..c29d8e4831 100644
> --- a/hw/intc/xive2.c
> +++ b/hw/intc/xive2.c
> @@ -366,6 +366,35 @@ static void xive2_end_enqueue(Xive2End *end, uint32_t 
> data)
>      end->w1 = xive_set_field32(END2_W1_PAGE_OFF, end->w1, qindex);
>  }
>  
> +static void xive2_pgofnext(uint8_t *nvgc_blk, uint32_t *nvgc_idx,
> +                           uint8_t next_level)
> +{
> +    uint32_t mask, next_idx;
> +    uint8_t next_blk;
> +
> +    /*
> +     * Adjust the block and index of a VP for the next group/crowd
> +     * size (PGofFirst/PGofNext field in the NVP and NVGC structures).
> +     *
> +     * The 6-bit group level is split into a 2-bit crowd and 4-bit
> +     * group levels. Encoding is similar. However, we don't support
> +     * crowd size of 8. So a crowd level of 0b11 is bumped to a crowd
> +     * size of 16.
> +     */
> +    next_blk = NVx_CROWD_LVL(next_level);
> +    if (next_blk == 3) {
> +        next_blk = 4;
> +    }
> +    mask = (1 << next_blk) - 1;
> +    *nvgc_blk &= ~mask;
> +    *nvgc_blk |= mask >> 1;
> +
> +    next_idx = NVx_GROUP_LVL(next_level);
> +    mask = (1 << next_idx) - 1;
> +    *nvgc_idx &= ~mask;
> +    *nvgc_idx |= mask >> 1;
> +}
> +
>  /*
>   * Scan the group chain and return the highest priority and group
>   * level of pending group interrupts.
> @@ -376,29 +405,28 @@ static uint8_t 
> xive2_presenter_backlog_scan(XivePresenter *xptr,
>                                              uint8_t *out_level)
>  {
>      Xive2Router *xrtr = XIVE2_ROUTER(xptr);
> -    uint32_t nvgc_idx, mask;
> +    uint32_t nvgc_idx;
>      uint32_t current_level, count;
> -    uint8_t prio;
> +    uint8_t nvgc_blk, prio;
>      Xive2Nvgc nvgc;
>  
>      for (prio = 0; prio <= XIVE_PRIORITY_MAX; prio++) {
> -        current_level = first_group & 0xF;
> +        current_level = first_group & 0x3F;
> +        nvgc_blk = nvp_blk;
> +        nvgc_idx = nvp_idx;
>  
>          while (current_level) {
> -            mask = (1 << current_level) - 1;
> -            nvgc_idx = nvp_idx & ~mask;
> -            nvgc_idx |= mask >> 1;
> -            qemu_log("fxb %s checking backlog for prio %d group idx %x\n",
> -                     __func__, prio, nvgc_idx);
> -
> -            if (xive2_router_get_nvgc(xrtr, false, nvp_blk, nvgc_idx, 
> &nvgc)) {
> -                qemu_log_mask(LOG_GUEST_ERROR, "XIVE: No NVG %x/%x\n",
> -                              nvp_blk, nvgc_idx);
> +            xive2_pgofnext(&nvgc_blk, &nvgc_idx, current_level);
> +
> +            if (xive2_router_get_nvgc(xrtr, NVx_CROWD_LVL(current_level),
> +                                      nvgc_blk, nvgc_idx, &nvgc)) {
> +                qemu_log_mask(LOG_GUEST_ERROR, "XIVE: No NVGC %x/%x\n",
> +                              nvgc_blk, nvgc_idx);
>                  return 0xFF;
>              }
>              if (!xive2_nvgc_is_valid(&nvgc)) {
> -                qemu_log_mask(LOG_GUEST_ERROR, "XIVE: Invalid NVG %x/%x\n",
> -                              nvp_blk, nvgc_idx);
> +                qemu_log_mask(LOG_GUEST_ERROR, "XIVE: Invalid NVGC %x/%x\n",
> +                              nvgc_blk, nvgc_idx);
>                  return 0xFF;
>              }
>  
> @@ -407,7 +435,7 @@ static uint8_t xive2_presenter_backlog_scan(XivePresenter 
> *xptr,
>                  *out_level = current_level;
>                  return prio;
>              }
> -            current_level = xive_get_field32(NVGC2_W0_PGONEXT, nvgc.w0) & 
> 0xF;
> +            current_level = xive_get_field32(NVGC2_W0_PGONEXT, nvgc.w0) & 
> 0x3F;
>          }
>      }
>      return 0xFF;
> @@ -419,22 +447,23 @@ static void xive2_presenter_backlog_decr(XivePresenter 
> *xptr,
>                                           uint8_t group_level)
>  {
>      Xive2Router *xrtr = XIVE2_ROUTER(xptr);
> -    uint32_t nvgc_idx, mask, count;
> +    uint32_t nvgc_idx, count;
> +    uint8_t nvgc_blk;
>      Xive2Nvgc nvgc;
>  
> -    group_level &= 0xF;
> -    mask = (1 << group_level) - 1;
> -    nvgc_idx = nvp_idx & ~mask;
> -    nvgc_idx |= mask >> 1;
> +    nvgc_blk = nvp_blk;
> +    nvgc_idx = nvp_idx;
> +    xive2_pgofnext(&nvgc_blk, &nvgc_idx, group_level);
>  
> -    if (xive2_router_get_nvgc(xrtr, false, nvp_blk, nvgc_idx, &nvgc)) {
> -        qemu_log_mask(LOG_GUEST_ERROR, "XIVE: No NVG %x/%x\n",
> -                      nvp_blk, nvgc_idx);
> +    if (xive2_router_get_nvgc(xrtr, NVx_CROWD_LVL(group_level),
> +                              nvgc_blk, nvgc_idx, &nvgc)) {
> +        qemu_log_mask(LOG_GUEST_ERROR, "XIVE: No NVGC %x/%x\n",
> +                      nvgc_blk, nvgc_idx);
>          return;
>      }
>      if (!xive2_nvgc_is_valid(&nvgc)) {
> -        qemu_log_mask(LOG_GUEST_ERROR, "XIVE: Invalid NVG %x/%x\n",
> -                      nvp_blk, nvgc_idx);
> +        qemu_log_mask(LOG_GUEST_ERROR, "XIVE: Invalid NVGC %x/%x\n",
> +                      nvgc_blk, nvgc_idx);
>          return;
>      }
>      count = xive2_nvgc_get_backlog(&nvgc, group_prio);
> @@ -442,7 +471,8 @@ static void xive2_presenter_backlog_decr(XivePresenter 
> *xptr,
>          return;
>      }
>      xive2_nvgc_set_backlog(&nvgc, group_prio, count - 1);
> -    xive2_router_write_nvgc(xrtr, false, nvp_blk, nvgc_idx, &nvgc);
> +    xive2_router_write_nvgc(xrtr, NVx_CROWD_LVL(group_level),
> +                            nvgc_blk, nvgc_idx, &nvgc);
>  }
>  
>  /*




reply via email to

[Prev in Thread] Current Thread [Next in Thread]