qemu-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Qemu-devel] [PATCH v6 31/35] target/arm: Implement SVE fp complex m


From: Alex Bennée
Subject: Re: [Qemu-devel] [PATCH v6 31/35] target/arm: Implement SVE fp complex multiply add (indexed)
Date: Thu, 28 Jun 2018 14:55:11 +0100
User-agent: mu4e 1.1.0; emacs 26.1.50

Richard Henderson <address@hidden> writes:

> Enhance the existing helpers to support SVE, which takes the
> index from each 128-bit segment.  The change has no effect
> for AdvSIMD, since there is only one such segment.
>
> Signed-off-by: Richard Henderson <address@hidden>

Reviewed-by: Alex Bennée <address@hidden>

> ---
>  target/arm/translate-sve.c | 23 ++++++++++++++++++
>  target/arm/vec_helper.c    | 50 +++++++++++++++++++++++---------------
>  target/arm/sve.decode      |  6 +++++
>  3 files changed, 59 insertions(+), 20 deletions(-)
>
> diff --git a/target/arm/translate-sve.c b/target/arm/translate-sve.c
> index 7ce3222158..4f2152fb70 100644
> --- a/target/arm/translate-sve.c
> +++ b/target/arm/translate-sve.c
> @@ -4005,6 +4005,29 @@ static bool trans_FCMLA_zpzzz(DisasContext *s,
>      return true;
>  }
>
> +static bool trans_FCMLA_zzxz(DisasContext *s, arg_FCMLA_zzxz *a, uint32_t 
> insn)
> +{
> +    static gen_helper_gvec_3_ptr * const fns[2] = {
> +        gen_helper_gvec_fcmlah_idx,
> +        gen_helper_gvec_fcmlas_idx,
> +    };
> +
> +    tcg_debug_assert(a->esz == 1 || a->esz == 2);
> +    tcg_debug_assert(a->rd == a->ra);
> +    if (sve_access_check(s)) {
> +        unsigned vsz = vec_full_reg_size(s);
> +        TCGv_ptr status = get_fpstatus_ptr(a->esz == MO_16);
> +        tcg_gen_gvec_3_ptr(vec_full_reg_offset(s, a->rd),
> +                           vec_full_reg_offset(s, a->rn),
> +                           vec_full_reg_offset(s, a->rm),
> +                           status, vsz, vsz,
> +                           a->index * 4 + a->rot,
> +                           fns[a->esz - 1]);
> +        tcg_temp_free_ptr(status);
> +    }
> +    return true;
> +}
> +
>  /*
>   *** SVE Floating Point Unary Operations Prediated Group
>   */
> diff --git a/target/arm/vec_helper.c b/target/arm/vec_helper.c
> index 8f2dc4b989..db5aeb9f24 100644
> --- a/target/arm/vec_helper.c
> +++ b/target/arm/vec_helper.c
> @@ -319,22 +319,27 @@ void HELPER(gvec_fcmlah_idx)(void *vd, void *vn, void 
> *vm,
>      uint32_t neg_imag = extract32(desc, SIMD_DATA_SHIFT + 1, 1);
>      intptr_t index = extract32(desc, SIMD_DATA_SHIFT + 2, 2);
>      uint32_t neg_real = flip ^ neg_imag;
> -    uintptr_t i;
> -    float16 e1 = m[H2(2 * index + flip)];
> -    float16 e3 = m[H2(2 * index + 1 - flip)];
> +    intptr_t elements = opr_sz / sizeof(float16);
> +    intptr_t eltspersegment = 16 / sizeof(float16);
> +    intptr_t i, j;
>
>      /* Shift boolean to the sign bit so we can xor to negate.  */
>      neg_real <<= 15;
>      neg_imag <<= 15;
> -    e1 ^= neg_real;
> -    e3 ^= neg_imag;
>
> -    for (i = 0; i < opr_sz / 2; i += 2) {
> -        float16 e2 = n[H2(i + flip)];
> -        float16 e4 = e2;
> +    for (i = 0; i < elements; i += eltspersegment) {
> +        float16 mr = m[H2(i + 2 * index + 0)];
> +        float16 mi = m[H2(i + 2 * index + 1)];
> +        float16 e1 = neg_real ^ (flip ? mi : mr);
> +        float16 e3 = neg_imag ^ (flip ? mr : mi);
>
> -        d[H2(i)] = float16_muladd(e2, e1, d[H2(i)], 0, fpst);
> -        d[H2(i + 1)] = float16_muladd(e4, e3, d[H2(i + 1)], 0, fpst);
> +        for (j = i; j < i + eltspersegment; j += 2) {
> +            float16 e2 = n[H2(j + flip)];
> +            float16 e4 = e2;
> +
> +            d[H2(j)] = float16_muladd(e2, e1, d[H2(j)], 0, fpst);
> +            d[H2(j + 1)] = float16_muladd(e4, e3, d[H2(j + 1)], 0, fpst);
> +        }
>      }
>      clear_tail(d, opr_sz, simd_maxsz(desc));
>  }
> @@ -380,22 +385,27 @@ void HELPER(gvec_fcmlas_idx)(void *vd, void *vn, void 
> *vm,
>      uint32_t neg_imag = extract32(desc, SIMD_DATA_SHIFT + 1, 1);
>      intptr_t index = extract32(desc, SIMD_DATA_SHIFT + 2, 2);
>      uint32_t neg_real = flip ^ neg_imag;
> -    uintptr_t i;
> -    float32 e1 = m[H4(2 * index + flip)];
> -    float32 e3 = m[H4(2 * index + 1 - flip)];
> +    intptr_t elements = opr_sz / sizeof(float32);
> +    intptr_t eltspersegment = 16 / sizeof(float32);
> +    intptr_t i, j;
>
>      /* Shift boolean to the sign bit so we can xor to negate.  */
>      neg_real <<= 31;
>      neg_imag <<= 31;
> -    e1 ^= neg_real;
> -    e3 ^= neg_imag;
>
> -    for (i = 0; i < opr_sz / 4; i += 2) {
> -        float32 e2 = n[H4(i + flip)];
> -        float32 e4 = e2;
> +    for (i = 0; i < elements; i += eltspersegment) {
> +        float32 mr = m[H4(i + 2 * index + 0)];
> +        float32 mi = m[H4(i + 2 * index + 1)];
> +        float32 e1 = neg_real ^ (flip ? mi : mr);
> +        float32 e3 = neg_imag ^ (flip ? mr : mi);
>
> -        d[H4(i)] = float32_muladd(e2, e1, d[H4(i)], 0, fpst);
> -        d[H4(i + 1)] = float32_muladd(e4, e3, d[H4(i + 1)], 0, fpst);
> +        for (j = i; j < i + eltspersegment; j += 2) {
> +            float32 e2 = n[H4(j + flip)];
> +            float32 e4 = e2;
> +
> +            d[H4(j)] = float32_muladd(e2, e1, d[H4(j)], 0, fpst);
> +            d[H4(j + 1)] = float32_muladd(e4, e3, d[H4(j + 1)], 0, fpst);
> +        }
>      }
>      clear_tail(d, opr_sz, simd_maxsz(desc));
>  }
> diff --git a/target/arm/sve.decode b/target/arm/sve.decode
> index e342cfdf14..62365ed90f 100644
> --- a/target/arm/sve.decode
> +++ b/target/arm/sve.decode
> @@ -733,6 +733,12 @@ FCADD           01100100 esz:2 00000 rot:1 100 pg:3 rm:5 
> rd:5 \
>  FCMLA_zpzzz     01100100 esz:2 0 rm:5 0 rot:2 pg:3 rn:5 rd:5 \
>                  ra=%reg_movprfx
>
> +# SVE floating-point complex multiply-add (indexed)
> +FCMLA_zzxz      01100100 10 1 index:2 rm:3 0001 rot:2 rn:5 rd:5 \
> +                ra=%reg_movprfx esz=1
> +FCMLA_zzxz      01100100 11 1 index:1 rm:4 0001 rot:2 rn:5 rd:5 \
> +                ra=%reg_movprfx esz=2
> +
>  ### SVE FP Multiply-Add Indexed Group
>
>  # SVE floating-point multiply-add (indexed)


--
Alex Bennée



reply via email to

[Prev in Thread] Current Thread [Next in Thread]