[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Qemu-devel] [PATCH] ppc/xive: introduce helpers to access the XIVE
From: |
Cédric Le Goater |
Subject: |
Re: [Qemu-devel] [PATCH] ppc/xive: introduce helpers to access the XIVE structures |
Date: |
Wed, 19 Dec 2018 07:46:01 +0100 |
User-agent: |
Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.3.1 |
On 12/19/18 5:53 AM, David Gibson wrote:
> On Tue, Dec 18, 2018 at 10:38:03PM +0100, Cédric Le Goater wrote:
>> The XIVE internal structures (EAS, END, NVT) are architected to be Big
>> Endian. The XIVE models, today, access the different fields of these
>> structures using wrappers around the GETFIELD/SETFIELD macros. The
>> definitions of these macros depend on the host which seems unnecessary
>> for guest only values.
>>
>> Remove GETFIELD/SETFIELD and replace them with explicit XIVE handlers
>> doing the endianess conversion.
>>
>> Signed-off-by: Cédric Le Goater <address@hidden>
>
> So, rather than applying this as is, I've distributed its pieces and
> folded them into the existing XIVE patches, so as not to break
> bisection. I've also eliminated the existing GETFIELD and SETFIELD
> macros.
Yes that is better.
We will need similar routine for the PHB models, so we might introduce
a common set when time comes.
Thanks,
C.
>> ---
>> include/hw/ppc/xive_regs.h | 28 +++++++++++++---
>> target/ppc/cpu.h | 12 -------
>> hw/intc/spapr_xive.c | 48 +++++++++++++--------------
>> hw/intc/xive.c | 68 +++++++++++++++++++-------------------
>> 4 files changed, 82 insertions(+), 74 deletions(-)
>>
>> diff --git a/include/hw/ppc/xive_regs.h b/include/hw/ppc/xive_regs.h
>> index 85557e730cd8..78875b156980 100644
>> --- a/include/hw/ppc/xive_regs.h
>> +++ b/include/hw/ppc/xive_regs.h
>> @@ -126,8 +126,18 @@ typedef struct XiveEAS {
>> #define xive_eas_is_valid(eas) (be64_to_cpu((eas)->w) & EAS_VALID)
>> #define xive_eas_is_masked(eas) (be64_to_cpu((eas)->w) & EAS_MASKED)
>>
>> -#define GETFIELD_BE64(m, v) GETFIELD(m, be64_to_cpu(v))
>> -#define SETFIELD_BE64(m, v, val) cpu_to_be64(SETFIELD(m, be64_to_cpu(v),
>> val))
>> +static inline uint64_t xive_get_field64(uint64_t mask, uint64_t word)
>> +{
>> + return (be64_to_cpu(word) & mask) >> ctz64(mask);
>> +}
>> +
>> +static inline uint64_t xive_set_field64(uint64_t mask, uint64_t word,
>> + uint64_t value)
>> +{
>> + uint64_t tmp =
>> + (be64_to_cpu(word) & ~mask) | ((value << ctz64(mask)) & mask);
>> + return cpu_to_be64(tmp);
>> +}
>>
>> /* Event Notification Descriptor (END) */
>> typedef struct XiveEND {
>> @@ -183,8 +193,18 @@ typedef struct XiveEND {
>> #define xive_end_is_backlog(end) (be32_to_cpu((end)->w0) & END_W0_BACKLOG)
>> #define xive_end_is_escalate(end) (be32_to_cpu((end)->w0) &
>> END_W0_ESCALATE_CTL)
>>
>> -#define GETFIELD_BE32(m, v) GETFIELD(m, be32_to_cpu(v))
>> -#define SETFIELD_BE32(m, v, val) cpu_to_be32(SETFIELD(m, be32_to_cpu(v),
>> val))
>> +static inline uint32_t xive_get_field32(uint32_t mask, uint32_t word)
>> +{
>> + return (be32_to_cpu(word) & mask) >> ctz32(mask);
>> +}
>> +
>> +static inline uint32_t xive_set_field32(uint32_t mask, uint32_t word,
>> + uint32_t value)
>> +{
>> + uint32_t tmp =
>> + (be32_to_cpu(word) & ~mask) | ((value << ctz32(mask)) & mask);
>> + return cpu_to_be32(tmp);
>> +}
>>
>> /* Notification Virtual Target (NVT) */
>> typedef struct XiveNVT {
>> diff --git a/target/ppc/cpu.h b/target/ppc/cpu.h
>> index 527181c0f09f..d5f99f1fc7b9 100644
>> --- a/target/ppc/cpu.h
>> +++ b/target/ppc/cpu.h
>> @@ -78,18 +78,6 @@
>> PPC_BIT32(bs))
>> #define PPC_BITMASK8(bs, be) ((PPC_BIT8(bs) - PPC_BIT8(be)) |
>> PPC_BIT8(bs))
>>
>> -#if HOST_LONG_BITS == 32
>> -# define MASK_TO_LSH(m) (__builtin_ffsll(m) - 1)
>> -#elif HOST_LONG_BITS == 64
>> -# define MASK_TO_LSH(m) (__builtin_ffsl(m) - 1)
>> -#else
>> -# error Unknown sizeof long
>> -#endif
>> -
>> -#define GETFIELD(m, v) (((v) & (m)) >> MASK_TO_LSH(m))
>> -#define SETFIELD(m, v, val) \
>> - (((v) & ~(m)) | ((((typeof(v))(val)) << MASK_TO_LSH(m)) & (m)))
>> -
>>
>> /*****************************************************************************/
>> /* Exception vectors definitions
>> */
>> enum {
>> diff --git a/hw/intc/spapr_xive.c b/hw/intc/spapr_xive.c
>> index 3ceabe668b16..87424de26c1c 100644
>> --- a/hw/intc/spapr_xive.c
>> +++ b/hw/intc/spapr_xive.c
>> @@ -119,12 +119,12 @@ static int spapr_xive_target_to_end(uint32_t target,
>> uint8_t prio,
>> static void spapr_xive_end_pic_print_info(sPAPRXive *xive, XiveEND *end,
>> Monitor *mon)
>> {
>> - uint32_t qindex = GETFIELD_BE32(END_W1_PAGE_OFF, end->w1);
>> - uint32_t qgen = GETFIELD_BE32(END_W1_GENERATION, end->w1);
>> - uint32_t qsize = GETFIELD_BE32(END_W0_QSIZE, end->w0);
>> + uint32_t qindex = xive_get_field32(END_W1_PAGE_OFF, end->w1);
>> + uint32_t qgen = xive_get_field32(END_W1_GENERATION, end->w1);
>> + uint32_t qsize = xive_get_field32(END_W0_QSIZE, end->w0);
>> uint32_t qentries = 1 << (qsize + 10);
>> - uint32_t nvt = GETFIELD_BE32(END_W6_NVT_INDEX, end->w6);
>> - uint8_t priority = GETFIELD_BE32(END_W7_F0_PRIORITY, end->w7);
>> + uint32_t nvt = xive_get_field32(END_W6_NVT_INDEX, end->w6);
>> + uint8_t priority = xive_get_field32(END_W7_F0_PRIORITY, end->w7);
>>
>> monitor_printf(mon, "%3d/%d % 6d/%5d ^%d",
>> spapr_xive_nvt_to_target(0, nvt),
>> @@ -155,10 +155,10 @@ void spapr_xive_pic_print_info(sPAPRXive *xive,
>> Monitor *mon)
>> pq & XIVE_ESB_VAL_Q ? 'Q' : '-',
>> xsrc->status[i] & XIVE_STATUS_ASSERTED ? 'A' : ' ',
>> xive_eas_is_masked(eas) ? "M" : " ",
>> - (int) GETFIELD_BE64(EAS_END_DATA, eas->w));
>> + (int) xive_get_field64(EAS_END_DATA, eas->w));
>>
>> if (!xive_eas_is_masked(eas)) {
>> - uint32_t end_idx = GETFIELD_BE64(EAS_END_INDEX, eas->w);
>> + uint32_t end_idx = xive_get_field64(EAS_END_INDEX, eas->w);
>> XiveEND *end;
>>
>> assert(end_idx < xive->nr_ends);
>> @@ -741,11 +741,11 @@ static target_ulong h_int_set_source_config(PowerPCCPU
>> *cpu,
>> return H_P3;
>> }
>>
>> - new_eas.w = SETFIELD_BE64(EAS_END_BLOCK, new_eas.w, end_blk);
>> - new_eas.w = SETFIELD_BE64(EAS_END_INDEX, new_eas.w, end_idx);
>> + new_eas.w = xive_set_field64(EAS_END_BLOCK, new_eas.w, end_blk);
>> + new_eas.w = xive_set_field64(EAS_END_INDEX, new_eas.w, end_idx);
>>
>> if (flags & SPAPR_XIVE_SRC_SET_EISN) {
>> - new_eas.w = SETFIELD_BE64(EAS_END_DATA, new_eas.w, eisn);
>> + new_eas.w = xive_set_field64(EAS_END_DATA, new_eas.w, eisn);
>> }
>>
>> out:
>> @@ -810,22 +810,22 @@ static target_ulong h_int_get_source_config(PowerPCCPU
>> *cpu,
>> }
>>
>> /* EAS_END_BLOCK is unused on sPAPR */
>> - end_idx = GETFIELD_BE64(EAS_END_INDEX, eas.w);
>> + end_idx = xive_get_field64(EAS_END_INDEX, eas.w);
>>
>> assert(end_idx < xive->nr_ends);
>> end = &xive->endt[end_idx];
>>
>> - nvt_blk = GETFIELD_BE32(END_W6_NVT_BLOCK, end->w6);
>> - nvt_idx = GETFIELD_BE32(END_W6_NVT_INDEX, end->w6);
>> + nvt_blk = xive_get_field32(END_W6_NVT_BLOCK, end->w6);
>> + nvt_idx = xive_get_field32(END_W6_NVT_INDEX, end->w6);
>> args[0] = spapr_xive_nvt_to_target(nvt_blk, nvt_idx);
>>
>> if (xive_eas_is_masked(&eas)) {
>> args[1] = 0xff;
>> } else {
>> - args[1] = GETFIELD_BE32(END_W7_F0_PRIORITY, end->w7);
>> + args[1] = xive_get_field32(END_W7_F0_PRIORITY, end->w7);
>> }
>>
>> - args[2] = GETFIELD_BE64(EAS_END_DATA, eas.w);
>> + args[2] = xive_get_field64(EAS_END_DATA, eas.w);
>>
>> return H_SUCCESS;
>> }
>> @@ -894,7 +894,7 @@ static target_ulong h_int_get_queue_info(PowerPCCPU *cpu,
>>
>> args[0] = xive->end_base + (1ull << (end_xsrc->esb_shift + 1)) *
>> end_idx;
>> if (xive_end_is_enqueue(end)) {
>> - args[1] = GETFIELD_BE32(END_W0_QSIZE, end->w0) + 12;
>> + args[1] = xive_get_field32(END_W0_QSIZE, end->w0) + 12;
>> } else {
>> args[1] = 0;
>> }
>> @@ -987,7 +987,7 @@ static target_ulong h_int_set_queue_config(PowerPCCPU
>> *cpu,
>> end.w2 = cpu_to_be32((qpage >> 32) & 0x0fffffff);
>> end.w3 = cpu_to_be32(qpage & 0xffffffff);
>> end.w0 |= cpu_to_be32(END_W0_ENQUEUE);
>> - end.w0 = SETFIELD_BE32(END_W0_QSIZE, end.w0, qsize - 12);
>> + end.w0 = xive_set_field32(END_W0_QSIZE, end.w0, qsize - 12);
>> break;
>> case 0:
>> /* reset queue and disable queueing */
>> @@ -1026,9 +1026,9 @@ static target_ulong h_int_set_queue_config(PowerPCCPU
>> *cpu,
>> /* Ensure the priority and target are correctly set (they will not
>> * be right after allocation)
>> */
>> - end.w6 = SETFIELD_BE32(END_W6_NVT_BLOCK, 0ul, nvt_blk) |
>> - SETFIELD_BE32(END_W6_NVT_INDEX, 0ul, nvt_idx);
>> - end.w7 = SETFIELD_BE32(END_W7_F0_PRIORITY, 0ul, priority);
>> + end.w6 = xive_set_field32(END_W6_NVT_BLOCK, 0ul, nvt_blk) |
>> + xive_set_field32(END_W6_NVT_INDEX, 0ul, nvt_idx);
>> + end.w7 = xive_set_field32(END_W7_F0_PRIORITY, 0ul, priority);
>>
>> if (flags & SPAPR_XIVE_END_ALWAYS_NOTIFY) {
>> end.w0 |= cpu_to_be32(END_W0_UCOND_NOTIFY);
>> @@ -1040,7 +1040,7 @@ static target_ulong h_int_set_queue_config(PowerPCCPU
>> *cpu,
>> * offset counter starts at 0.
>> */
>> end.w1 = cpu_to_be32(END_W1_GENERATION) |
>> - SETFIELD_BE32(END_W1_PAGE_OFF, 0ul, 0ul);
>> + xive_set_field32(END_W1_PAGE_OFF, 0ul, 0ul);
>> end.w0 |= cpu_to_be32(END_W0_VALID);
>>
>> /* TODO: issue syncs required to ensure all in-flight interrupts
>> @@ -1132,7 +1132,7 @@ static target_ulong h_int_get_queue_config(PowerPCCPU
>> *cpu,
>> if (xive_end_is_enqueue(end)) {
>> args[1] = (uint64_t) be32_to_cpu(end->w2 & 0x0fffffff) << 32
>> | be32_to_cpu(end->w3);
>> - args[2] = GETFIELD_BE32(END_W0_QSIZE, end->w0) + 12;
>> + args[2] = xive_get_field32(END_W0_QSIZE, end->w0) + 12;
>> } else {
>> args[1] = 0;
>> args[2] = 0;
>> @@ -1141,10 +1141,10 @@ static target_ulong
>> h_int_get_queue_config(PowerPCCPU *cpu,
>> /* TODO: do we need any locking on the END ? */
>> if (flags & SPAPR_XIVE_END_DEBUG) {
>> /* Load the event queue generation number into the return flags */
>> - args[0] |= (uint64_t)GETFIELD_BE32(END_W1_GENERATION, end->w1) <<
>> 62;
>> + args[0] |= (uint64_t)xive_get_field32(END_W1_GENERATION, end->w1)
>> << 62;
>>
>> /* Load R7 with the event queue offset counter */
>> - args[3] = GETFIELD_BE32(END_W1_PAGE_OFF, end->w1);
>> + args[3] = xive_get_field32(END_W1_PAGE_OFF, end->w1);
>> } else {
>> args[3] = 0;
>> }
>> diff --git a/hw/intc/xive.c b/hw/intc/xive.c
>> index 53d2f191e8a3..f8510e426f63 100644
>> --- a/hw/intc/xive.c
>> +++ b/hw/intc/xive.c
>> @@ -982,8 +982,8 @@ void xive_end_queue_pic_print_info(XiveEND *end,
>> uint32_t width, Monitor *mon)
>> {
>> uint64_t qaddr_base = (uint64_t) be32_to_cpu(end->w2 & 0x0fffffff) << 32
>> | be32_to_cpu(end->w3);
>> - uint32_t qsize = GETFIELD_BE32(END_W0_QSIZE, end->w0);
>> - uint32_t qindex = GETFIELD_BE32(END_W1_PAGE_OFF, end->w1);
>> + uint32_t qsize = xive_get_field32(END_W0_QSIZE, end->w0);
>> + uint32_t qindex = xive_get_field32(END_W1_PAGE_OFF, end->w1);
>> uint32_t qentries = 1 << (qsize + 10);
>> int i;
>>
>> @@ -1012,13 +1012,13 @@ void xive_end_pic_print_info(XiveEND *end, uint32_t
>> end_idx, Monitor *mon)
>> {
>> uint64_t qaddr_base = (uint64_t) be32_to_cpu(end->w2 & 0x0fffffff) << 32
>> | be32_to_cpu(end->w3);
>> - uint32_t qindex = GETFIELD_BE32(END_W1_PAGE_OFF, end->w1);
>> - uint32_t qgen = GETFIELD_BE32(END_W1_GENERATION, end->w1);
>> - uint32_t qsize = GETFIELD_BE32(END_W0_QSIZE, end->w0);
>> + uint32_t qindex = xive_get_field32(END_W1_PAGE_OFF, end->w1);
>> + uint32_t qgen = xive_get_field32(END_W1_GENERATION, end->w1);
>> + uint32_t qsize = xive_get_field32(END_W0_QSIZE, end->w0);
>> uint32_t qentries = 1 << (qsize + 10);
>>
>> - uint32_t nvt = GETFIELD_BE32(END_W6_NVT_INDEX, end->w6);
>> - uint8_t priority = GETFIELD_BE32(END_W7_F0_PRIORITY, end->w7);
>> + uint32_t nvt = xive_get_field32(END_W6_NVT_INDEX, end->w6);
>> + uint8_t priority = xive_get_field32(END_W7_F0_PRIORITY, end->w7);
>>
>> if (!xive_end_is_valid(end)) {
>> return;
>> @@ -1041,9 +1041,9 @@ static void xive_end_enqueue(XiveEND *end, uint32_t
>> data)
>> {
>> uint64_t qaddr_base = (uint64_t) be32_to_cpu(end->w2 & 0x0fffffff) << 32
>> | be32_to_cpu(end->w3);
>> - uint32_t qsize = GETFIELD_BE32(END_W0_QSIZE, end->w0);
>> - uint32_t qindex = GETFIELD_BE32(END_W1_PAGE_OFF, end->w1);
>> - uint32_t qgen = GETFIELD_BE32(END_W1_GENERATION, end->w1);
>> + uint32_t qsize = xive_get_field32(END_W0_QSIZE, end->w0);
>> + uint32_t qindex = xive_get_field32(END_W1_PAGE_OFF, end->w1);
>> + uint32_t qgen = xive_get_field32(END_W1_GENERATION, end->w1);
>>
>> uint64_t qaddr = qaddr_base + (qindex << 2);
>> uint32_t qdata = cpu_to_be32((qgen << 31) | (data & 0x7fffffff));
>> @@ -1058,9 +1058,9 @@ static void xive_end_enqueue(XiveEND *end, uint32_t
>> data)
>> qindex = (qindex + 1) & (qentries - 1);
>> if (qindex == 0) {
>> qgen ^= 1;
>> - end->w1 = SETFIELD_BE32(END_W1_GENERATION, end->w1, qgen);
>> + end->w1 = xive_set_field32(END_W1_GENERATION, end->w1, qgen);
>> }
>> - end->w1 = SETFIELD_BE32(END_W1_PAGE_OFF, end->w1, qindex);
>> + end->w1 = xive_set_field32(END_W1_PAGE_OFF, end->w1, qindex);
>> }
>>
>> /*
>> @@ -1139,13 +1139,13 @@ static int xive_presenter_tctx_match(XiveTCTX *tctx,
>> uint8_t format,
>>
>> /* HV POOL ring */
>> if ((be32_to_cpu(qw2w2) & TM_QW2W2_VP) &&
>> - cam == GETFIELD_BE32(TM_QW2W2_POOL_CAM, qw2w2)) {
>> + cam == xive_get_field32(TM_QW2W2_POOL_CAM, qw2w2)) {
>> return TM_QW2_HV_POOL;
>> }
>>
>> /* OS ring */
>> if ((be32_to_cpu(qw1w2) & TM_QW1W2_VO) &&
>> - cam == GETFIELD_BE32(TM_QW1W2_OS_CAM, qw1w2)) {
>> + cam == xive_get_field32(TM_QW1W2_OS_CAM, qw1w2)) {
>> return TM_QW1_OS;
>> }
>> } else {
>> @@ -1153,9 +1153,9 @@ static int xive_presenter_tctx_match(XiveTCTX *tctx,
>> uint8_t format,
>>
>> /* USER ring */
>> if ((be32_to_cpu(qw1w2) & TM_QW1W2_VO) &&
>> - (cam == GETFIELD_BE32(TM_QW1W2_OS_CAM, qw1w2)) &&
>> + (cam == xive_get_field32(TM_QW1W2_OS_CAM, qw1w2)) &&
>> (be32_to_cpu(qw0w2) & TM_QW0W2_VU) &&
>> - (logic_serv == GETFIELD_BE32(TM_QW0W2_LOGIC_SERV, qw0w2))) {
>> + (logic_serv == xive_get_field32(TM_QW0W2_LOGIC_SERV, qw0w2))) {
>> return TM_QW0_USER;
>> }
>> }
>> @@ -1313,8 +1313,8 @@ static void xive_router_end_notify(XiveRouter *xrtr,
>> uint8_t end_blk,
>> * F=1 : User level Event-Based Branch (EBB) notification, no
>> * priority
>> */
>> - format = GETFIELD_BE32(END_W6_FORMAT_BIT, end.w6);
>> - priority = GETFIELD_BE32(END_W7_F0_PRIORITY, end.w7);
>> + format = xive_get_field32(END_W6_FORMAT_BIT, end.w6);
>> + priority = xive_get_field32(END_W7_F0_PRIORITY, end.w7);
>>
>> /* The END is masked */
>> if (format == 0 && priority == 0xff) {
>> @@ -1326,11 +1326,11 @@ static void xive_router_end_notify(XiveRouter *xrtr,
>> uint8_t end_blk,
>> * even futher coalescing in the Router
>> */
>> if (!xive_end_is_notify(&end)) {
>> - uint8_t pq = GETFIELD_BE32(END_W1_ESn, end.w1);
>> + uint8_t pq = xive_get_field32(END_W1_ESn, end.w1);
>> bool notify = xive_esb_trigger(&pq);
>>
>> - if (pq != GETFIELD_BE32(END_W1_ESn, end.w1)) {
>> - end.w1 = SETFIELD_BE32(END_W1_ESn, end.w1, pq);
>> + if (pq != xive_get_field32(END_W1_ESn, end.w1)) {
>> + end.w1 = xive_set_field32(END_W1_ESn, end.w1, pq);
>> xive_router_write_end(xrtr, end_blk, end_idx, &end, 1);
>> }
>>
>> @@ -1344,11 +1344,11 @@ static void xive_router_end_notify(XiveRouter *xrtr,
>> uint8_t end_blk,
>> * Follows IVPE notification
>> */
>> xive_presenter_notify(xrtr, format,
>> - GETFIELD_BE32(END_W6_NVT_BLOCK, end.w6),
>> - GETFIELD_BE32(END_W6_NVT_INDEX, end.w6),
>> - GETFIELD_BE32(END_W7_F0_IGNORE, end.w7),
>> + xive_get_field32(END_W6_NVT_BLOCK, end.w6),
>> + xive_get_field32(END_W6_NVT_INDEX, end.w6),
>> + xive_get_field32(END_W7_F0_IGNORE, end.w7),
>> priority,
>> - GETFIELD_BE32(END_W7_F1_LOG_SERVER_ID, end.w7));
>> + xive_get_field32(END_W7_F1_LOG_SERVER_ID,
>> end.w7));
>>
>> /* TODO: Auto EOI. */
>> }
>> @@ -1385,9 +1385,9 @@ static void xive_router_notify(XiveNotifier *xn,
>> uint32_t lisn)
>> * The event trigger becomes an END trigger
>> */
>> xive_router_end_notify(xrtr,
>> - GETFIELD_BE64(EAS_END_BLOCK, eas.w),
>> - GETFIELD_BE64(EAS_END_INDEX, eas.w),
>> - GETFIELD_BE64(EAS_END_DATA, eas.w));
>> + xive_get_field64(EAS_END_BLOCK, eas.w),
>> + xive_get_field64(EAS_END_INDEX, eas.w),
>> + xive_get_field64(EAS_END_DATA, eas.w));
>> }
>>
>> static void xive_router_class_init(ObjectClass *klass, void *data)
>> @@ -1419,9 +1419,9 @@ void xive_eas_pic_print_info(XiveEAS *eas, uint32_t
>> lisn, Monitor *mon)
>>
>> monitor_printf(mon, " %08x %s end:%02x/%04x data:%08x\n",
>> lisn, xive_eas_is_masked(eas) ? "M" : " ",
>> - (uint8_t) GETFIELD_BE64(EAS_END_BLOCK, eas->w),
>> - (uint32_t) GETFIELD_BE64(EAS_END_INDEX, eas->w),
>> - (uint32_t) GETFIELD_BE64(EAS_END_DATA, eas->w));
>> + (uint8_t) xive_get_field64(EAS_END_BLOCK, eas->w),
>> + (uint32_t) xive_get_field64(EAS_END_INDEX, eas->w),
>> + (uint32_t) xive_get_field64(EAS_END_DATA, eas->w));
>> }
>>
>> /*
>> @@ -1454,7 +1454,7 @@ static uint64_t xive_end_source_read(void *opaque,
>> hwaddr addr, unsigned size)
>> }
>>
>> end_esmask = addr_is_even(addr, xsrc->esb_shift) ? END_W1_ESn :
>> END_W1_ESe;
>> - pq = GETFIELD_BE32(end_esmask, end.w1);
>> + pq = xive_get_field32(end_esmask, end.w1);
>>
>> switch (offset) {
>> case XIVE_ESB_LOAD_EOI ... XIVE_ESB_LOAD_EOI + 0x7FF:
>> @@ -1479,8 +1479,8 @@ static uint64_t xive_end_source_read(void *opaque,
>> hwaddr addr, unsigned size)
>> return -1;
>> }
>>
>> - if (pq != GETFIELD_BE32(end_esmask, end.w1)) {
>> - end.w1 = SETFIELD_BE32(end_esmask, end.w1, pq);
>> + if (pq != xive_get_field32(end_esmask, end.w1)) {
>> + end.w1 = xive_set_field32(end_esmask, end.w1, pq);
>> xive_router_write_end(xsrc->xrtr, end_blk, end_idx, &end, 1);
>> }
>>
>