[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Qemu-ppc] [Qemu-devel] [PATCH] PPC: Fix compilation with TCG debug
From: |
Alexander Graf |
Subject: |
Re: [Qemu-ppc] [Qemu-devel] [PATCH] PPC: Fix compilation with TCG debug |
Date: |
Sun, 22 Dec 2013 18:16:44 +0100 |
On 22.12.2013, at 17:37, Aurelien Jarno <address@hidden> wrote:
> On Fri, Dec 20, 2013 at 11:01:50AM +0100, Alexander Graf wrote:
>> The recent VSX patches broken compilation of QEMU when configurated
>> with --enable-debug, as it was treating "target long" TCG variables
>> as "i64" which is not true for 32bit targets.
>>
>> This patch fixes all the places that the compiler has found to use
>> the correct variable type and if necessary manually cast.
>>
>> Reported-by: Stefan Weil <address@hidden>
>> Signed-off-by: Alexander Graf <address@hidden>
>> ---
>> target-ppc/translate.c | 143
>> ++++++++++++++++++++++++++++---------------------
>> 1 file changed, 81 insertions(+), 62 deletions(-)
>>
>> diff --git a/target-ppc/translate.c b/target-ppc/translate.c
>> index ea58dc9..c5c1108 100644
>> --- a/target-ppc/translate.c
>> +++ b/target-ppc/translate.c
>> @@ -2567,6 +2567,14 @@ static inline void gen_qemu_ld32u(DisasContext *ctx,
>> TCGv arg1, TCGv arg2)
>> }
>> }
>>
>> +static void gen_qemu_ld32u_i64(DisasContext *ctx, TCGv_i64 val, TCGv addr)
>> +{
>> + TCGv tmp = tcg_temp_new();
>> + gen_qemu_ld32u(ctx, tmp, addr);
>> + tcg_gen_extu_tl_i64(val, tmp);
>> + tcg_temp_free(tmp);
>> +}
>> +
>> static inline void gen_qemu_ld32s(DisasContext *ctx, TCGv arg1, TCGv arg2)
>> {
>> if (unlikely(ctx->le_mode)) {
>> @@ -2616,6 +2624,14 @@ static inline void gen_qemu_st32(DisasContext *ctx,
>> TCGv arg1, TCGv arg2)
>> }
>> }
>>
>> +static void gen_qemu_st32_i64(DisasContext *ctx, TCGv_i64 val, TCGv addr)
>> +{
>> + TCGv tmp = tcg_temp_new();
>> + tcg_gen_trunc_i64_tl(tmp, val);
>> + gen_qemu_st32(ctx, tmp, addr);
>> + tcg_temp_free(tmp);
>> +}
>> +
>> static inline void gen_qemu_st64(DisasContext *ctx, TCGv_i64 arg1, TCGv arg2)
>> {
>> if (unlikely(ctx->le_mode)) {
>> @@ -7048,13 +7064,14 @@ static void gen_lxvdsx(DisasContext *ctx)
>> EA = tcg_temp_new();
>> gen_addr_reg_index(ctx, EA);
>> gen_qemu_ld64(ctx, cpu_vsrh(xT(ctx->opcode)), EA);
>> - tcg_gen_mov_tl(cpu_vsrl(xT(ctx->opcode)), cpu_vsrh(xT(ctx->opcode)));
>> + tcg_gen_mov_i64(cpu_vsrl(xT(ctx->opcode)), cpu_vsrh(xT(ctx->opcode)));
>> tcg_temp_free(EA);
>> }
>>
>> static void gen_lxvw4x(DisasContext *ctx)
>> {
>> - TCGv EA, tmp;
>> + TCGv EA;
>> + TCGv_i64 tmp;
>> TCGv_i64 xth = cpu_vsrh(xT(ctx->opcode));
>> TCGv_i64 xtl = cpu_vsrl(xT(ctx->opcode));
>> if (unlikely(!ctx->vsx_enabled)) {
>> @@ -7063,21 +7080,22 @@ static void gen_lxvw4x(DisasContext *ctx)
>> }
>> gen_set_access_type(ctx, ACCESS_INT);
>> EA = tcg_temp_new();
>> - tmp = tcg_temp_new();
>> + tmp = tcg_temp_new_i64();
>> +
>> gen_addr_reg_index(ctx, EA);
>> - gen_qemu_ld32u(ctx, tmp, EA);
>> + gen_qemu_ld32u_i64(ctx, tmp, EA);
>> tcg_gen_addi_tl(EA, EA, 4);
>> - gen_qemu_ld32u(ctx, xth, EA);
>> + gen_qemu_ld32u_i64(ctx, xth, EA);
>> tcg_gen_deposit_i64(xth, xth, tmp, 32, 32);
>>
>> tcg_gen_addi_tl(EA, EA, 4);
>> - gen_qemu_ld32u(ctx, tmp, EA);
>> + gen_qemu_ld32u_i64(ctx, tmp, EA);
>> tcg_gen_addi_tl(EA, EA, 4);
>> - gen_qemu_ld32u(ctx, xtl, EA);
>> + gen_qemu_ld32u_i64(ctx, xtl, EA);
>> tcg_gen_deposit_i64(xtl, xtl, tmp, 32, 32);
>>
>> tcg_temp_free(EA);
>> - tcg_temp_free(tmp);
>> + tcg_temp_free_i64(tmp);
>> }
>>
>> static void gen_stxsdx(DisasContext *ctx)
>> @@ -7112,7 +7130,8 @@ static void gen_stxvd2x(DisasContext *ctx)
>>
>> static void gen_stxvw4x(DisasContext *ctx)
>> {
>> - TCGv EA, tmp;
>> + TCGv_i64 tmp;
>> + TCGv EA;
>> if (unlikely(!ctx->vsx_enabled)) {
>> gen_exception(ctx, POWERPC_EXCP_VSXU);
>> return;
>> @@ -7120,21 +7139,21 @@ static void gen_stxvw4x(DisasContext *ctx)
>> gen_set_access_type(ctx, ACCESS_INT);
>> EA = tcg_temp_new();
>> gen_addr_reg_index(ctx, EA);
>> - tmp = tcg_temp_new();
>> + tmp = tcg_temp_new_i64();
>>
>> tcg_gen_shri_i64(tmp, cpu_vsrh(xS(ctx->opcode)), 32);
>> - gen_qemu_st32(ctx, tmp, EA);
>> + gen_qemu_st32_i64(ctx, tmp, EA);
>> tcg_gen_addi_tl(EA, EA, 4);
>> - gen_qemu_st32(ctx, cpu_vsrh(xS(ctx->opcode)), EA);
>> + gen_qemu_st32_i64(ctx, cpu_vsrh(xS(ctx->opcode)), EA);
>>
>> tcg_gen_shri_i64(tmp, cpu_vsrl(xS(ctx->opcode)), 32);
>> tcg_gen_addi_tl(EA, EA, 4);
>> - gen_qemu_st32(ctx, tmp, EA);
>> + gen_qemu_st32_i64(ctx, tmp, EA);
>> tcg_gen_addi_tl(EA, EA, 4);
>> - gen_qemu_st32(ctx, cpu_vsrl(xS(ctx->opcode)), EA);
>> + gen_qemu_st32_i64(ctx, cpu_vsrl(xS(ctx->opcode)), EA);
>>
>> tcg_temp_free(EA);
>> - tcg_temp_free(tmp);
>> + tcg_temp_free_i64(tmp);
>> }
>>
>> static void gen_xxpermdi(DisasContext *ctx)
>> @@ -7171,8 +7190,8 @@ static void glue(gen_, name)(DisasContext * ctx)
>> \
>> gen_exception(ctx, POWERPC_EXCP_VSXU); \
>> return; \
>> } \
>> - xb = tcg_temp_new(); \
>> - sgm = tcg_temp_new(); \
>> + xb = tcg_temp_new_i64(); \
>> + sgm = tcg_temp_new_i64(); \
>> tcg_gen_mov_i64(xb, cpu_vsrh(xB(ctx->opcode))); \
>> tcg_gen_movi_i64(sgm, sgn_mask); \
>> switch (op) { \
>> @@ -7189,18 +7208,18 @@ static void glue(gen_, name)(DisasContext * ctx)
>> \
>> break; \
>> } \
>> case OP_CPSGN: { \
>> - TCGv_i64 xa = tcg_temp_new(); \
>> + TCGv_i64 xa = tcg_temp_new_i64(); \
>> tcg_gen_mov_i64(xa, cpu_vsrh(xA(ctx->opcode))); \
>> tcg_gen_and_i64(xa, xa, sgm); \
>> tcg_gen_andc_i64(xb, xb, sgm); \
>> tcg_gen_or_i64(xb, xb, xa); \
>> - tcg_temp_free(xa); \
>> + tcg_temp_free_i64(xa); \
>> break; \
>> } \
>> } \
>> tcg_gen_mov_i64(cpu_vsrh(xT(ctx->opcode)), xb); \
>> - tcg_temp_free(xb); \
>> - tcg_temp_free(sgm); \
>> + tcg_temp_free_i64(xb); \
>> + tcg_temp_free_i64(sgm); \
>> }
>>
>> VSX_SCALAR_MOVE(xsabsdp, OP_ABS, SGN_MASK_DP)
>> @@ -7216,9 +7235,9 @@ static void glue(gen_, name)(DisasContext * ctx)
>> \
>> gen_exception(ctx, POWERPC_EXCP_VSXU); \
>> return; \
>> } \
>> - xbh = tcg_temp_new(); \
>> - xbl = tcg_temp_new(); \
>> - sgm = tcg_temp_new(); \
>> + xbh = tcg_temp_new_i64(); \
>> + xbl = tcg_temp_new_i64(); \
>> + sgm = tcg_temp_new_i64(); \
>> tcg_gen_mov_i64(xbh, cpu_vsrh(xB(ctx->opcode))); \
>> tcg_gen_mov_i64(xbl, cpu_vsrl(xB(ctx->opcode))); \
>> tcg_gen_movi_i64(sgm, sgn_mask); \
>> @@ -7239,8 +7258,8 @@ static void glue(gen_, name)(DisasContext * ctx)
>> \
>> break; \
>> } \
>> case OP_CPSGN: { \
>> - TCGv_i64 xah = tcg_temp_new(); \
>> - TCGv_i64 xal = tcg_temp_new(); \
>> + TCGv_i64 xah = tcg_temp_new_i64(); \
>> + TCGv_i64 xal = tcg_temp_new_i64(); \
>> tcg_gen_mov_i64(xah, cpu_vsrh(xA(ctx->opcode))); \
>> tcg_gen_mov_i64(xal, cpu_vsrl(xA(ctx->opcode))); \
>> tcg_gen_and_i64(xah, xah, sgm); \
>> @@ -7249,16 +7268,16 @@ static void glue(gen_, name)(DisasContext * ctx)
>> \
>> tcg_gen_andc_i64(xbl, xbl, sgm); \
>> tcg_gen_or_i64(xbh, xbh, xah); \
>> tcg_gen_or_i64(xbl, xbl, xal); \
>> - tcg_temp_free(xah); \
>> - tcg_temp_free(xal); \
>> + tcg_temp_free_i64(xah); \
>> + tcg_temp_free_i64(xal); \
>> break; \
>> } \
>> } \
>> tcg_gen_mov_i64(cpu_vsrh(xT(ctx->opcode)), xbh); \
>> tcg_gen_mov_i64(cpu_vsrl(xT(ctx->opcode)), xbl); \
>> - tcg_temp_free(xbh); \
>> - tcg_temp_free(xbl); \
>> - tcg_temp_free(sgm); \
>> + tcg_temp_free_i64(xbh); \
>> + tcg_temp_free_i64(xbl); \
>> + tcg_temp_free_i64(sgm); \
>> }
>>
>> VSX_VECTOR_MOVE(xvabsdp, OP_ABS, SGN_MASK_DP)
>> @@ -7284,11 +7303,11 @@ static void glue(gen_, name)(DisasContext * ctx)
>> \
>> cpu_vsrl(xB(ctx->opcode))); \
>> }
>>
>> -VSX_LOGICAL(xxland, tcg_gen_and_tl)
>> -VSX_LOGICAL(xxlandc, tcg_gen_andc_tl)
>> -VSX_LOGICAL(xxlor, tcg_gen_or_tl)
>> -VSX_LOGICAL(xxlxor, tcg_gen_xor_tl)
>> -VSX_LOGICAL(xxlnor, tcg_gen_nor_tl)
>> +VSX_LOGICAL(xxland, tcg_gen_and_i64)
>> +VSX_LOGICAL(xxlandc, tcg_gen_andc_i64)
>> +VSX_LOGICAL(xxlor, tcg_gen_or_i64)
>> +VSX_LOGICAL(xxlxor, tcg_gen_xor_i64)
>> +VSX_LOGICAL(xxlnor, tcg_gen_nor_i64)
>>
>> #define VSX_XXMRG(name, high) \
>> static void glue(gen_, name)(DisasContext * ctx) \
>> @@ -7298,10 +7317,10 @@ static void glue(gen_, name)(DisasContext * ctx)
>> \
>> gen_exception(ctx, POWERPC_EXCP_VSXU); \
>> return; \
>> } \
>> - a0 = tcg_temp_new(); \
>> - a1 = tcg_temp_new(); \
>> - b0 = tcg_temp_new(); \
>> - b1 = tcg_temp_new(); \
>> + a0 = tcg_temp_new_i64(); \
>> + a1 = tcg_temp_new_i64(); \
>> + b0 = tcg_temp_new_i64(); \
>> + b1 = tcg_temp_new_i64(); \
>> if (high) { \
>> tcg_gen_mov_i64(a0, cpu_vsrh(xA(ctx->opcode))); \
>> tcg_gen_mov_i64(a1, cpu_vsrh(xA(ctx->opcode))); \
>> @@ -7319,10 +7338,10 @@ static void glue(gen_, name)(DisasContext * ctx)
>> \
>> b0, a0, 32, 32); \
>> tcg_gen_deposit_i64(cpu_vsrl(xT(ctx->opcode)), \
>> b1, a1, 32, 32); \
>> - tcg_temp_free(a0); \
>> - tcg_temp_free(a1); \
>> - tcg_temp_free(b0); \
>> - tcg_temp_free(b1); \
>> + tcg_temp_free_i64(a0); \
>> + tcg_temp_free_i64(a1); \
>> + tcg_temp_free_i64(b0); \
>> + tcg_temp_free_i64(b1); \
>> }
>>
>> VSX_XXMRG(xxmrghw, 1)
>> @@ -7335,9 +7354,9 @@ static void gen_xxsel(DisasContext * ctx)
>> gen_exception(ctx, POWERPC_EXCP_VSXU);
>> return;
>> }
>> - a = tcg_temp_new();
>> - b = tcg_temp_new();
>> - c = tcg_temp_new();
>> + a = tcg_temp_new_i64();
>> + b = tcg_temp_new_i64();
>> + c = tcg_temp_new_i64();
>>
>> tcg_gen_mov_i64(a, cpu_vsrh(xA(ctx->opcode)));
>> tcg_gen_mov_i64(b, cpu_vsrh(xB(ctx->opcode)));
>> @@ -7355,9 +7374,9 @@ static void gen_xxsel(DisasContext * ctx)
>> tcg_gen_andc_i64(a, a, c);
>> tcg_gen_or_i64(cpu_vsrl(xT(ctx->opcode)), a, b);
>>
>> - tcg_temp_free(a);
>> - tcg_temp_free(b);
>> - tcg_temp_free(c);
>> + tcg_temp_free_i64(a);
>> + tcg_temp_free_i64(b);
>> + tcg_temp_free_i64(c);
>> }
>>
>> static void gen_xxspltw(DisasContext *ctx)
>> @@ -7372,8 +7391,8 @@ static void gen_xxspltw(DisasContext *ctx)
>> return;
>> }
>>
>> - b = tcg_temp_new();
>> - b2 = tcg_temp_new();
>> + b = tcg_temp_new_i64();
>> + b2 = tcg_temp_new_i64();
>>
>> if (UIM(ctx->opcode) & 1) {
>> tcg_gen_ext32u_i64(b, vsr);
>> @@ -7385,8 +7404,8 @@ static void gen_xxspltw(DisasContext *ctx)
>> tcg_gen_or_i64(cpu_vsrh(xT(ctx->opcode)), b, b2);
>> tcg_gen_mov_i64(cpu_vsrl(xT(ctx->opcode)), cpu_vsrh(xT(ctx->opcode)));
>>
>> - tcg_temp_free(b);
>> - tcg_temp_free(b2);
>> + tcg_temp_free_i64(b);
>> + tcg_temp_free_i64(b2);
>> }
>>
>> static void gen_xxsldwi(DisasContext *ctx)
>> @@ -7396,8 +7415,8 @@ static void gen_xxsldwi(DisasContext *ctx)
>> gen_exception(ctx, POWERPC_EXCP_VSXU);
>> return;
>> }
>> - xth = tcg_temp_new();
>> - xtl = tcg_temp_new();
>> + xth = tcg_temp_new_i64();
>> + xtl = tcg_temp_new_i64();
>>
>> switch (SHW(ctx->opcode)) {
>> case 0: {
>> @@ -7406,7 +7425,7 @@ static void gen_xxsldwi(DisasContext *ctx)
>> break;
>> }
>> case 1: {
>> - TCGv_i64 t0 = tcg_temp_new();
>> + TCGv_i64 t0 = tcg_temp_new_i64();
>> tcg_gen_mov_i64(xth, cpu_vsrh(xA(ctx->opcode)));
>> tcg_gen_shli_i64(xth, xth, 32);
>> tcg_gen_mov_i64(t0, cpu_vsrl(xA(ctx->opcode)));
>> @@ -7417,7 +7436,7 @@ static void gen_xxsldwi(DisasContext *ctx)
>> tcg_gen_mov_i64(t0, cpu_vsrh(xB(ctx->opcode)));
>> tcg_gen_shri_i64(t0, t0, 32);
>> tcg_gen_or_i64(xtl, xtl, t0);
>> - tcg_temp_free(t0);
>> + tcg_temp_free_i64(t0);
>> break;
>> }
>> case 2: {
>> @@ -7426,7 +7445,7 @@ static void gen_xxsldwi(DisasContext *ctx)
>> break;
>> }
>> case 3: {
>> - TCGv_i64 t0 = tcg_temp_new();
>> + TCGv_i64 t0 = tcg_temp_new_i64();
>> tcg_gen_mov_i64(xth, cpu_vsrl(xA(ctx->opcode)));
>> tcg_gen_shli_i64(xth, xth, 32);
>> tcg_gen_mov_i64(t0, cpu_vsrh(xB(ctx->opcode)));
>> @@ -7437,7 +7456,7 @@ static void gen_xxsldwi(DisasContext *ctx)
>> tcg_gen_mov_i64(t0, cpu_vsrl(xB(ctx->opcode)));
>> tcg_gen_shri_i64(t0, t0, 32);
>> tcg_gen_or_i64(xtl, xtl, t0);
>> - tcg_temp_free(t0);
>> + tcg_temp_free_i64(t0);
>> break;
>> }
>> }
>> @@ -7445,8 +7464,8 @@ static void gen_xxsldwi(DisasContext *ctx)
>> tcg_gen_mov_i64(cpu_vsrh(xT(ctx->opcode)), xth);
>> tcg_gen_mov_i64(cpu_vsrl(xT(ctx->opcode)), xtl);
>>
>> - tcg_temp_free(xth);
>> - tcg_temp_free(xtl);
>> + tcg_temp_free_i64(xth);
>> + tcg_temp_free_i64(xtl);
>> }
>>
>
> This takes a different approach for the load/store part than in my
> patch, but it is also good. So:
>
> Reviewed-by: Aurelien Jarno <address@hidden>
Please apply it to the tree straight ahead then :).
Alex