diff options
Diffstat (limited to 'lib/gcc/aarch64-linux-android/4.9.1')
30 files changed, 7786 insertions, 7564 deletions
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/crtbegin.o b/lib/gcc/aarch64-linux-android/4.9.1/crtbegin.o Binary files differindex ac99ac3..cf40b92 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/crtbegin.o +++ b/lib/gcc/aarch64-linux-android/4.9.1/crtbegin.o diff --git a/lib/gcc/aarch64-linux-android/4.9.1/crtbeginS.o b/lib/gcc/aarch64-linux-android/4.9.1/crtbeginS.o Binary files differindex 09d432c..2fd7372 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/crtbeginS.o +++ b/lib/gcc/aarch64-linux-android/4.9.1/crtbeginS.o diff --git a/lib/gcc/aarch64-linux-android/4.9.1/crtbeginT.o b/lib/gcc/aarch64-linux-android/4.9.1/crtbeginT.o Binary files differindex ac99ac3..cf40b92 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/crtbeginT.o +++ b/lib/gcc/aarch64-linux-android/4.9.1/crtbeginT.o diff --git a/lib/gcc/aarch64-linux-android/4.9.1/crtend.o b/lib/gcc/aarch64-linux-android/4.9.1/crtend.o Binary files differindex 602fa94..26d17b0 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/crtend.o +++ b/lib/gcc/aarch64-linux-android/4.9.1/crtend.o diff --git a/lib/gcc/aarch64-linux-android/4.9.1/crtendS.o b/lib/gcc/aarch64-linux-android/4.9.1/crtendS.o Binary files differindex 602fa94..26d17b0 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/crtendS.o +++ b/lib/gcc/aarch64-linux-android/4.9.1/crtendS.o diff --git a/lib/gcc/aarch64-linux-android/4.9.1/include/arm_acle.h b/lib/gcc/aarch64-linux-android/4.9.1/include/arm_acle.h new file mode 100644 index 0000000..2e74696 --- /dev/null +++ b/lib/gcc/aarch64-linux-android/4.9.1/include/arm_acle.h @@ -0,0 +1,90 @@ +/* AArch64 Non-NEON ACLE intrinsics include file. + + Copyright (C) 2014 Free Software Foundation, Inc. + Contributed by ARM Ltd. + + This file is part of GCC. + + GCC is free software; you can redistribute it and/or modify it + under the terms of the GNU General Public License as published + by the Free Software Foundation; either version 3, or (at your + option) any later version. + + GCC is distributed in the hope that it will be useful, but WITHOUT + ANY WARRANTY; without even the implied warranty of MERCHANTABILITY + or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public + License for more details. + + Under Section 7 of GPL version 3, you are granted additional + permissions described in the GCC Runtime Library Exception, version + 3.1, as published by the Free Software Foundation. + + You should have received a copy of the GNU General Public License and + a copy of the GCC Runtime Library Exception along with this program; + see the files COPYING3 and COPYING.RUNTIME respectively. If not, see + <http://www.gnu.org/licenses/>. */ + +#ifndef _GCC_ARM_ACLE_H +#define _GCC_ARM_ACLE_H + +#include <stdint.h> +#ifdef __cplusplus +extern "C" { +#endif + +#ifdef __ARM_FEATURE_CRC32 +__extension__ static __inline uint32_t __attribute__ ((__always_inline__)) +__crc32b (uint32_t __a, uint8_t __b) +{ + return __builtin_aarch64_crc32b (__a, __b); +} + +__extension__ static __inline uint32_t __attribute__ ((__always_inline__)) +__crc32cb (uint32_t __a, uint8_t __b) +{ + return __builtin_aarch64_crc32cb (__a, __b); +} + +__extension__ static __inline uint32_t __attribute__ ((__always_inline__)) +__crc32ch (uint32_t __a, uint16_t __b) +{ + return __builtin_aarch64_crc32ch (__a, __b); +} + +__extension__ static __inline uint32_t __attribute__ ((__always_inline__)) +__crc32cw (uint32_t __a, uint32_t __b) +{ + return __builtin_aarch64_crc32cw (__a, __b); +} + +__extension__ static __inline uint32_t __attribute__ ((__always_inline__)) +__crc32cd (uint32_t __a, uint64_t __b) +{ + return __builtin_aarch64_crc32cx (__a, __b); +} + +__extension__ static __inline uint32_t __attribute__ ((__always_inline__)) +__crc32h (uint32_t __a, uint16_t __b) +{ + return __builtin_aarch64_crc32h (__a, __b); +} + +__extension__ static __inline uint32_t __attribute__ ((__always_inline__)) +__crc32w (uint32_t __a, uint32_t __b) +{ + return __builtin_aarch64_crc32w (__a, __b); +} + +__extension__ static __inline uint32_t __attribute__ ((__always_inline__)) +__crc32d (uint32_t __a, uint64_t __b) +{ + return __builtin_aarch64_crc32x (__a, __b); +} + +#endif + +#ifdef __cplusplus +} +#endif + +#endif diff --git a/lib/gcc/aarch64-linux-android/4.9.1/include/arm_neon.h b/lib/gcc/aarch64-linux-android/4.9.1/include/arm_neon.h index d4424df..dd6f43c 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/include/arm_neon.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/include/arm_neon.h @@ -2119,29 +2119,26 @@ vqadd_s64 (int64x1_t __a, int64x1_t __b) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vqadd_u8 (uint8x8_t __a, uint8x8_t __b) { - return (uint8x8_t) __builtin_aarch64_uqaddv8qi ((int8x8_t) __a, - (int8x8_t) __b); + return __builtin_aarch64_uqaddv8qi_uuu (__a, __b); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vqadd_u16 (uint16x4_t __a, uint16x4_t __b) { - return (uint16x4_t) __builtin_aarch64_uqaddv4hi ((int16x4_t) __a, - (int16x4_t) __b); + return __builtin_aarch64_uqaddv4hi_uuu (__a, __b); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vqadd_u32 (uint32x2_t __a, uint32x2_t __b) { - return (uint32x2_t) __builtin_aarch64_uqaddv2si ((int32x2_t) __a, - (int32x2_t) __b); + return __builtin_aarch64_uqaddv2si_uuu (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vqadd_u64 (uint64x1_t __a, uint64x1_t __b) { - return (uint64x1_t) __builtin_aarch64_uqadddi ((int64x1_t) __a, - (int64x1_t) __b); + return (uint64x1_t) __builtin_aarch64_uqadddi_uuu ((uint64_t) __a, + (uint64_t) __b); } __extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) @@ -2171,29 +2168,25 @@ vqaddq_s64 (int64x2_t __a, int64x2_t __b) __extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) vqaddq_u8 (uint8x16_t __a, uint8x16_t __b) { - return (uint8x16_t) __builtin_aarch64_uqaddv16qi ((int8x16_t) __a, - (int8x16_t) __b); + return __builtin_aarch64_uqaddv16qi_uuu (__a, __b); } __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vqaddq_u16 (uint16x8_t __a, uint16x8_t __b) { - return (uint16x8_t) __builtin_aarch64_uqaddv8hi ((int16x8_t) __a, - (int16x8_t) __b); + return __builtin_aarch64_uqaddv8hi_uuu (__a, __b); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vqaddq_u32 (uint32x4_t __a, uint32x4_t __b) { - return (uint32x4_t) __builtin_aarch64_uqaddv4si ((int32x4_t) __a, - (int32x4_t) __b); + return __builtin_aarch64_uqaddv4si_uuu (__a, __b); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vqaddq_u64 (uint64x2_t __a, uint64x2_t __b) { - return (uint64x2_t) __builtin_aarch64_uqaddv2di ((int64x2_t) __a, - (int64x2_t) __b); + return __builtin_aarch64_uqaddv2di_uuu (__a, __b); } __extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) @@ -2223,29 +2216,26 @@ vqsub_s64 (int64x1_t __a, int64x1_t __b) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vqsub_u8 (uint8x8_t __a, uint8x8_t __b) { - return (uint8x8_t) __builtin_aarch64_uqsubv8qi ((int8x8_t) __a, - (int8x8_t) __b); + return __builtin_aarch64_uqsubv8qi_uuu (__a, __b); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vqsub_u16 (uint16x4_t __a, uint16x4_t __b) { - return (uint16x4_t) __builtin_aarch64_uqsubv4hi ((int16x4_t) __a, - (int16x4_t) __b); + return __builtin_aarch64_uqsubv4hi_uuu (__a, __b); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vqsub_u32 (uint32x2_t __a, uint32x2_t __b) { - return (uint32x2_t) __builtin_aarch64_uqsubv2si ((int32x2_t) __a, - (int32x2_t) __b); + return __builtin_aarch64_uqsubv2si_uuu (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vqsub_u64 (uint64x1_t __a, uint64x1_t __b) { - return (uint64x1_t) __builtin_aarch64_uqsubdi ((int64x1_t) __a, - (int64x1_t) __b); + return (uint64x1_t) __builtin_aarch64_uqsubdi_uuu ((uint64_t) __a, + (uint64_t) __b); } __extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) @@ -2275,29 +2265,25 @@ vqsubq_s64 (int64x2_t __a, int64x2_t __b) __extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) vqsubq_u8 (uint8x16_t __a, uint8x16_t __b) { - return (uint8x16_t) __builtin_aarch64_uqsubv16qi ((int8x16_t) __a, - (int8x16_t) __b); + return __builtin_aarch64_uqsubv16qi_uuu (__a, __b); } __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vqsubq_u16 (uint16x8_t __a, uint16x8_t __b) { - return (uint16x8_t) __builtin_aarch64_uqsubv8hi ((int16x8_t) __a, - (int16x8_t) __b); + return __builtin_aarch64_uqsubv8hi_uuu (__a, __b); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vqsubq_u32 (uint32x4_t __a, uint32x4_t __b) { - return (uint32x4_t) __builtin_aarch64_uqsubv4si ((int32x4_t) __a, - (int32x4_t) __b); + return __builtin_aarch64_uqsubv4si_uuu (__a, __b); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vqsubq_u64 (uint64x2_t __a, uint64x2_t __b) { - return (uint64x2_t) __builtin_aarch64_uqsubv2di ((int64x2_t) __a, - (int64x2_t) __b); + return __builtin_aarch64_uqsubv2di_uuu (__a, __b); } __extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) @@ -5661,318 +5647,6 @@ vcvtxd_f32_f64 (float64_t a) return result; } -#define vext_f32(a, b, c) \ - __extension__ \ - ({ \ - float32x2_t b_ = (b); \ - float32x2_t a_ = (a); \ - float32x2_t result; \ - __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*4" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vext_f64(a, b, c) \ - __extension__ \ - ({ \ - float64x1_t b_ = (b); \ - float64x1_t a_ = (a); \ - float64x1_t result; \ - __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*8" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vext_p8(a, b, c) \ - __extension__ \ - ({ \ - poly8x8_t b_ = (b); \ - poly8x8_t a_ = (a); \ - poly8x8_t result; \ - __asm__ ("ext %0.8b,%1.8b,%2.8b,%3" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vext_p16(a, b, c) \ - __extension__ \ - ({ \ - poly16x4_t b_ = (b); \ - poly16x4_t a_ = (a); \ - poly16x4_t result; \ - __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*2" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vext_s8(a, b, c) \ - __extension__ \ - ({ \ - int8x8_t b_ = (b); \ - int8x8_t a_ = (a); \ - int8x8_t result; \ - __asm__ ("ext %0.8b,%1.8b,%2.8b,%3" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vext_s16(a, b, c) \ - __extension__ \ - ({ \ - int16x4_t b_ = (b); \ - int16x4_t a_ = (a); \ - int16x4_t result; \ - __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*2" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vext_s32(a, b, c) \ - __extension__ \ - ({ \ - int32x2_t b_ = (b); \ - int32x2_t a_ = (a); \ - int32x2_t result; \ - __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*4" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vext_s64(a, b, c) \ - __extension__ \ - ({ \ - int64x1_t b_ = (b); \ - int64x1_t a_ = (a); \ - int64x1_t result; \ - __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*8" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vext_u8(a, b, c) \ - __extension__ \ - ({ \ - uint8x8_t b_ = (b); \ - uint8x8_t a_ = (a); \ - uint8x8_t result; \ - __asm__ ("ext %0.8b,%1.8b,%2.8b,%3" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vext_u16(a, b, c) \ - __extension__ \ - ({ \ - uint16x4_t b_ = (b); \ - uint16x4_t a_ = (a); \ - uint16x4_t result; \ - __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*2" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vext_u32(a, b, c) \ - __extension__ \ - ({ \ - uint32x2_t b_ = (b); \ - uint32x2_t a_ = (a); \ - uint32x2_t result; \ - __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*4" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vext_u64(a, b, c) \ - __extension__ \ - ({ \ - uint64x1_t b_ = (b); \ - uint64x1_t a_ = (a); \ - uint64x1_t result; \ - __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*8" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vextq_f32(a, b, c) \ - __extension__ \ - ({ \ - float32x4_t b_ = (b); \ - float32x4_t a_ = (a); \ - float32x4_t result; \ - __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*4" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vextq_f64(a, b, c) \ - __extension__ \ - ({ \ - float64x2_t b_ = (b); \ - float64x2_t a_ = (a); \ - float64x2_t result; \ - __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*8" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vextq_p8(a, b, c) \ - __extension__ \ - ({ \ - poly8x16_t b_ = (b); \ - poly8x16_t a_ = (a); \ - poly8x16_t result; \ - __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vextq_p16(a, b, c) \ - __extension__ \ - ({ \ - poly16x8_t b_ = (b); \ - poly16x8_t a_ = (a); \ - poly16x8_t result; \ - __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*2" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vextq_s8(a, b, c) \ - __extension__ \ - ({ \ - int8x16_t b_ = (b); \ - int8x16_t a_ = (a); \ - int8x16_t result; \ - __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vextq_s16(a, b, c) \ - __extension__ \ - ({ \ - int16x8_t b_ = (b); \ - int16x8_t a_ = (a); \ - int16x8_t result; \ - __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*2" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vextq_s32(a, b, c) \ - __extension__ \ - ({ \ - int32x4_t b_ = (b); \ - int32x4_t a_ = (a); \ - int32x4_t result; \ - __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*4" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vextq_s64(a, b, c) \ - __extension__ \ - ({ \ - int64x2_t b_ = (b); \ - int64x2_t a_ = (a); \ - int64x2_t result; \ - __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*8" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vextq_u8(a, b, c) \ - __extension__ \ - ({ \ - uint8x16_t b_ = (b); \ - uint8x16_t a_ = (a); \ - uint8x16_t result; \ - __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vextq_u16(a, b, c) \ - __extension__ \ - ({ \ - uint16x8_t b_ = (b); \ - uint16x8_t a_ = (a); \ - uint16x8_t result; \ - __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*2" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vextq_u32(a, b, c) \ - __extension__ \ - ({ \ - uint32x4_t b_ = (b); \ - uint32x4_t a_ = (a); \ - uint32x4_t result; \ - __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*4" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - -#define vextq_u64(a, b, c) \ - __extension__ \ - ({ \ - uint64x2_t b_ = (b); \ - uint64x2_t a_ = (a); \ - uint64x2_t result; \ - __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*8" \ - : "=w"(result) \ - : "w"(a_), "w"(b_), "i"(c) \ - : /* No clobbers */); \ - result; \ - }) - __extension__ static __inline float32x2_t __attribute__ ((__always_inline__)) vfma_f32 (float32x2_t a, float32x2_t b, float32x2_t c) { @@ -10875,402 +10549,6 @@ vrecpeq_u32 (uint32x4_t a) return result; } -__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) -vrev16_p8 (poly8x8_t a) -{ - poly8x8_t result; - __asm__ ("rev16 %0.8b,%1.8b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) -vrev16_s8 (int8x8_t a) -{ - int8x8_t result; - __asm__ ("rev16 %0.8b,%1.8b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) -vrev16_u8 (uint8x8_t a) -{ - uint8x8_t result; - __asm__ ("rev16 %0.8b,%1.8b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) -vrev16q_p8 (poly8x16_t a) -{ - poly8x16_t result; - __asm__ ("rev16 %0.16b,%1.16b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) -vrev16q_s8 (int8x16_t a) -{ - int8x16_t result; - __asm__ ("rev16 %0.16b,%1.16b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) -vrev16q_u8 (uint8x16_t a) -{ - uint8x16_t result; - __asm__ ("rev16 %0.16b,%1.16b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) -vrev32_p8 (poly8x8_t a) -{ - poly8x8_t result; - __asm__ ("rev32 %0.8b,%1.8b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__)) -vrev32_p16 (poly16x4_t a) -{ - poly16x4_t result; - __asm__ ("rev32 %0.4h,%1.4h" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) -vrev32_s8 (int8x8_t a) -{ - int8x8_t result; - __asm__ ("rev32 %0.8b,%1.8b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) -vrev32_s16 (int16x4_t a) -{ - int16x4_t result; - __asm__ ("rev32 %0.4h,%1.4h" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) -vrev32_u8 (uint8x8_t a) -{ - uint8x8_t result; - __asm__ ("rev32 %0.8b,%1.8b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) -vrev32_u16 (uint16x4_t a) -{ - uint16x4_t result; - __asm__ ("rev32 %0.4h,%1.4h" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) -vrev32q_p8 (poly8x16_t a) -{ - poly8x16_t result; - __asm__ ("rev32 %0.16b,%1.16b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__)) -vrev32q_p16 (poly16x8_t a) -{ - poly16x8_t result; - __asm__ ("rev32 %0.8h,%1.8h" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) -vrev32q_s8 (int8x16_t a) -{ - int8x16_t result; - __asm__ ("rev32 %0.16b,%1.16b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) -vrev32q_s16 (int16x8_t a) -{ - int16x8_t result; - __asm__ ("rev32 %0.8h,%1.8h" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) -vrev32q_u8 (uint8x16_t a) -{ - uint8x16_t result; - __asm__ ("rev32 %0.16b,%1.16b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) -vrev32q_u16 (uint16x8_t a) -{ - uint16x8_t result; - __asm__ ("rev32 %0.8h,%1.8h" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline float32x2_t __attribute__ ((__always_inline__)) -vrev64_f32 (float32x2_t a) -{ - float32x2_t result; - __asm__ ("rev64 %0.2s,%1.2s" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) -vrev64_p8 (poly8x8_t a) -{ - poly8x8_t result; - __asm__ ("rev64 %0.8b,%1.8b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__)) -vrev64_p16 (poly16x4_t a) -{ - poly16x4_t result; - __asm__ ("rev64 %0.4h,%1.4h" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) -vrev64_s8 (int8x8_t a) -{ - int8x8_t result; - __asm__ ("rev64 %0.8b,%1.8b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) -vrev64_s16 (int16x4_t a) -{ - int16x4_t result; - __asm__ ("rev64 %0.4h,%1.4h" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int32x2_t __attribute__ ((__always_inline__)) -vrev64_s32 (int32x2_t a) -{ - int32x2_t result; - __asm__ ("rev64 %0.2s,%1.2s" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) -vrev64_u8 (uint8x8_t a) -{ - uint8x8_t result; - __asm__ ("rev64 %0.8b,%1.8b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) -vrev64_u16 (uint16x4_t a) -{ - uint16x4_t result; - __asm__ ("rev64 %0.4h,%1.4h" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) -vrev64_u32 (uint32x2_t a) -{ - uint32x2_t result; - __asm__ ("rev64 %0.2s,%1.2s" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline float32x4_t __attribute__ ((__always_inline__)) -vrev64q_f32 (float32x4_t a) -{ - float32x4_t result; - __asm__ ("rev64 %0.4s,%1.4s" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) -vrev64q_p8 (poly8x16_t a) -{ - poly8x16_t result; - __asm__ ("rev64 %0.16b,%1.16b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__)) -vrev64q_p16 (poly16x8_t a) -{ - poly16x8_t result; - __asm__ ("rev64 %0.8h,%1.8h" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) -vrev64q_s8 (int8x16_t a) -{ - int8x16_t result; - __asm__ ("rev64 %0.16b,%1.16b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) -vrev64q_s16 (int16x8_t a) -{ - int16x8_t result; - __asm__ ("rev64 %0.8h,%1.8h" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) -vrev64q_s32 (int32x4_t a) -{ - int32x4_t result; - __asm__ ("rev64 %0.4s,%1.4s" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) -vrev64q_u8 (uint8x16_t a) -{ - uint8x16_t result; - __asm__ ("rev64 %0.16b,%1.16b" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) -vrev64q_u16 (uint16x8_t a) -{ - uint16x8_t result; - __asm__ ("rev64 %0.8h,%1.8h" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) -vrev64q_u32 (uint32x4_t a) -{ - uint32x4_t result; - __asm__ ("rev64 %0.4s,%1.4s" - : "=w"(result) - : "w"(a) - : /* No clobbers */); - return result; -} - #define vrshrn_high_n_s16(a, b, c) \ __extension__ \ ({ \ @@ -12694,468 +11972,6 @@ vsubhn_u64 (uint64x2_t a, uint64x2_t b) return result; } -__extension__ static __inline float32x2_t __attribute__ ((__always_inline__)) -vtrn1_f32 (float32x2_t a, float32x2_t b) -{ - float32x2_t result; - __asm__ ("trn1 %0.2s,%1.2s,%2.2s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) -vtrn1_p8 (poly8x8_t a, poly8x8_t b) -{ - poly8x8_t result; - __asm__ ("trn1 %0.8b,%1.8b,%2.8b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__)) -vtrn1_p16 (poly16x4_t a, poly16x4_t b) -{ - poly16x4_t result; - __asm__ ("trn1 %0.4h,%1.4h,%2.4h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) -vtrn1_s8 (int8x8_t a, int8x8_t b) -{ - int8x8_t result; - __asm__ ("trn1 %0.8b,%1.8b,%2.8b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) -vtrn1_s16 (int16x4_t a, int16x4_t b) -{ - int16x4_t result; - __asm__ ("trn1 %0.4h,%1.4h,%2.4h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int32x2_t __attribute__ ((__always_inline__)) -vtrn1_s32 (int32x2_t a, int32x2_t b) -{ - int32x2_t result; - __asm__ ("trn1 %0.2s,%1.2s,%2.2s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) -vtrn1_u8 (uint8x8_t a, uint8x8_t b) -{ - uint8x8_t result; - __asm__ ("trn1 %0.8b,%1.8b,%2.8b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) -vtrn1_u16 (uint16x4_t a, uint16x4_t b) -{ - uint16x4_t result; - __asm__ ("trn1 %0.4h,%1.4h,%2.4h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) -vtrn1_u32 (uint32x2_t a, uint32x2_t b) -{ - uint32x2_t result; - __asm__ ("trn1 %0.2s,%1.2s,%2.2s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline float32x4_t __attribute__ ((__always_inline__)) -vtrn1q_f32 (float32x4_t a, float32x4_t b) -{ - float32x4_t result; - __asm__ ("trn1 %0.4s,%1.4s,%2.4s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline float64x2_t __attribute__ ((__always_inline__)) -vtrn1q_f64 (float64x2_t a, float64x2_t b) -{ - float64x2_t result; - __asm__ ("trn1 %0.2d,%1.2d,%2.2d" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) -vtrn1q_p8 (poly8x16_t a, poly8x16_t b) -{ - poly8x16_t result; - __asm__ ("trn1 %0.16b,%1.16b,%2.16b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__)) -vtrn1q_p16 (poly16x8_t a, poly16x8_t b) -{ - poly16x8_t result; - __asm__ ("trn1 %0.8h,%1.8h,%2.8h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) -vtrn1q_s8 (int8x16_t a, int8x16_t b) -{ - int8x16_t result; - __asm__ ("trn1 %0.16b,%1.16b,%2.16b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) -vtrn1q_s16 (int16x8_t a, int16x8_t b) -{ - int16x8_t result; - __asm__ ("trn1 %0.8h,%1.8h,%2.8h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) -vtrn1q_s32 (int32x4_t a, int32x4_t b) -{ - int32x4_t result; - __asm__ ("trn1 %0.4s,%1.4s,%2.4s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) -vtrn1q_s64 (int64x2_t a, int64x2_t b) -{ - int64x2_t result; - __asm__ ("trn1 %0.2d,%1.2d,%2.2d" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) -vtrn1q_u8 (uint8x16_t a, uint8x16_t b) -{ - uint8x16_t result; - __asm__ ("trn1 %0.16b,%1.16b,%2.16b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) -vtrn1q_u16 (uint16x8_t a, uint16x8_t b) -{ - uint16x8_t result; - __asm__ ("trn1 %0.8h,%1.8h,%2.8h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) -vtrn1q_u32 (uint32x4_t a, uint32x4_t b) -{ - uint32x4_t result; - __asm__ ("trn1 %0.4s,%1.4s,%2.4s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) -vtrn1q_u64 (uint64x2_t a, uint64x2_t b) -{ - uint64x2_t result; - __asm__ ("trn1 %0.2d,%1.2d,%2.2d" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline float32x2_t __attribute__ ((__always_inline__)) -vtrn2_f32 (float32x2_t a, float32x2_t b) -{ - float32x2_t result; - __asm__ ("trn2 %0.2s,%1.2s,%2.2s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) -vtrn2_p8 (poly8x8_t a, poly8x8_t b) -{ - poly8x8_t result; - __asm__ ("trn2 %0.8b,%1.8b,%2.8b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__)) -vtrn2_p16 (poly16x4_t a, poly16x4_t b) -{ - poly16x4_t result; - __asm__ ("trn2 %0.4h,%1.4h,%2.4h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) -vtrn2_s8 (int8x8_t a, int8x8_t b) -{ - int8x8_t result; - __asm__ ("trn2 %0.8b,%1.8b,%2.8b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) -vtrn2_s16 (int16x4_t a, int16x4_t b) -{ - int16x4_t result; - __asm__ ("trn2 %0.4h,%1.4h,%2.4h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int32x2_t __attribute__ ((__always_inline__)) -vtrn2_s32 (int32x2_t a, int32x2_t b) -{ - int32x2_t result; - __asm__ ("trn2 %0.2s,%1.2s,%2.2s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) -vtrn2_u8 (uint8x8_t a, uint8x8_t b) -{ - uint8x8_t result; - __asm__ ("trn2 %0.8b,%1.8b,%2.8b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) -vtrn2_u16 (uint16x4_t a, uint16x4_t b) -{ - uint16x4_t result; - __asm__ ("trn2 %0.4h,%1.4h,%2.4h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) -vtrn2_u32 (uint32x2_t a, uint32x2_t b) -{ - uint32x2_t result; - __asm__ ("trn2 %0.2s,%1.2s,%2.2s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline float32x4_t __attribute__ ((__always_inline__)) -vtrn2q_f32 (float32x4_t a, float32x4_t b) -{ - float32x4_t result; - __asm__ ("trn2 %0.4s,%1.4s,%2.4s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline float64x2_t __attribute__ ((__always_inline__)) -vtrn2q_f64 (float64x2_t a, float64x2_t b) -{ - float64x2_t result; - __asm__ ("trn2 %0.2d,%1.2d,%2.2d" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) -vtrn2q_p8 (poly8x16_t a, poly8x16_t b) -{ - poly8x16_t result; - __asm__ ("trn2 %0.16b,%1.16b,%2.16b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__)) -vtrn2q_p16 (poly16x8_t a, poly16x8_t b) -{ - poly16x8_t result; - __asm__ ("trn2 %0.8h,%1.8h,%2.8h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) -vtrn2q_s8 (int8x16_t a, int8x16_t b) -{ - int8x16_t result; - __asm__ ("trn2 %0.16b,%1.16b,%2.16b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) -vtrn2q_s16 (int16x8_t a, int16x8_t b) -{ - int16x8_t result; - __asm__ ("trn2 %0.8h,%1.8h,%2.8h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) -vtrn2q_s32 (int32x4_t a, int32x4_t b) -{ - int32x4_t result; - __asm__ ("trn2 %0.4s,%1.4s,%2.4s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) -vtrn2q_s64 (int64x2_t a, int64x2_t b) -{ - int64x2_t result; - __asm__ ("trn2 %0.2d,%1.2d,%2.2d" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) -vtrn2q_u8 (uint8x16_t a, uint8x16_t b) -{ - uint8x16_t result; - __asm__ ("trn2 %0.16b,%1.16b,%2.16b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) -vtrn2q_u16 (uint16x8_t a, uint16x8_t b) -{ - uint16x8_t result; - __asm__ ("trn2 %0.8h,%1.8h,%2.8h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) -vtrn2q_u32 (uint32x4_t a, uint32x4_t b) -{ - uint32x4_t result; - __asm__ ("trn2 %0.4s,%1.4s,%2.4s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) -vtrn2q_u64 (uint64x2_t a, uint64x2_t b) -{ - uint64x2_t result; - __asm__ ("trn2 %0.2d,%1.2d,%2.2d" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vtst_p8 (poly8x8_t a, poly8x8_t b) { @@ -13199,467 +12015,6 @@ vtstq_p16 (poly16x8_t a, poly16x8_t b) : /* No clobbers */); return result; } -__extension__ static __inline float32x2_t __attribute__ ((__always_inline__)) -vuzp1_f32 (float32x2_t a, float32x2_t b) -{ - float32x2_t result; - __asm__ ("uzp1 %0.2s,%1.2s,%2.2s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) -vuzp1_p8 (poly8x8_t a, poly8x8_t b) -{ - poly8x8_t result; - __asm__ ("uzp1 %0.8b,%1.8b,%2.8b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__)) -vuzp1_p16 (poly16x4_t a, poly16x4_t b) -{ - poly16x4_t result; - __asm__ ("uzp1 %0.4h,%1.4h,%2.4h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) -vuzp1_s8 (int8x8_t a, int8x8_t b) -{ - int8x8_t result; - __asm__ ("uzp1 %0.8b,%1.8b,%2.8b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) -vuzp1_s16 (int16x4_t a, int16x4_t b) -{ - int16x4_t result; - __asm__ ("uzp1 %0.4h,%1.4h,%2.4h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int32x2_t __attribute__ ((__always_inline__)) -vuzp1_s32 (int32x2_t a, int32x2_t b) -{ - int32x2_t result; - __asm__ ("uzp1 %0.2s,%1.2s,%2.2s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) -vuzp1_u8 (uint8x8_t a, uint8x8_t b) -{ - uint8x8_t result; - __asm__ ("uzp1 %0.8b,%1.8b,%2.8b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) -vuzp1_u16 (uint16x4_t a, uint16x4_t b) -{ - uint16x4_t result; - __asm__ ("uzp1 %0.4h,%1.4h,%2.4h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) -vuzp1_u32 (uint32x2_t a, uint32x2_t b) -{ - uint32x2_t result; - __asm__ ("uzp1 %0.2s,%1.2s,%2.2s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline float32x4_t __attribute__ ((__always_inline__)) -vuzp1q_f32 (float32x4_t a, float32x4_t b) -{ - float32x4_t result; - __asm__ ("uzp1 %0.4s,%1.4s,%2.4s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline float64x2_t __attribute__ ((__always_inline__)) -vuzp1q_f64 (float64x2_t a, float64x2_t b) -{ - float64x2_t result; - __asm__ ("uzp1 %0.2d,%1.2d,%2.2d" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) -vuzp1q_p8 (poly8x16_t a, poly8x16_t b) -{ - poly8x16_t result; - __asm__ ("uzp1 %0.16b,%1.16b,%2.16b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__)) -vuzp1q_p16 (poly16x8_t a, poly16x8_t b) -{ - poly16x8_t result; - __asm__ ("uzp1 %0.8h,%1.8h,%2.8h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) -vuzp1q_s8 (int8x16_t a, int8x16_t b) -{ - int8x16_t result; - __asm__ ("uzp1 %0.16b,%1.16b,%2.16b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) -vuzp1q_s16 (int16x8_t a, int16x8_t b) -{ - int16x8_t result; - __asm__ ("uzp1 %0.8h,%1.8h,%2.8h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) -vuzp1q_s32 (int32x4_t a, int32x4_t b) -{ - int32x4_t result; - __asm__ ("uzp1 %0.4s,%1.4s,%2.4s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) -vuzp1q_s64 (int64x2_t a, int64x2_t b) -{ - int64x2_t result; - __asm__ ("uzp1 %0.2d,%1.2d,%2.2d" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) -vuzp1q_u8 (uint8x16_t a, uint8x16_t b) -{ - uint8x16_t result; - __asm__ ("uzp1 %0.16b,%1.16b,%2.16b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) -vuzp1q_u16 (uint16x8_t a, uint16x8_t b) -{ - uint16x8_t result; - __asm__ ("uzp1 %0.8h,%1.8h,%2.8h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) -vuzp1q_u32 (uint32x4_t a, uint32x4_t b) -{ - uint32x4_t result; - __asm__ ("uzp1 %0.4s,%1.4s,%2.4s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) -vuzp1q_u64 (uint64x2_t a, uint64x2_t b) -{ - uint64x2_t result; - __asm__ ("uzp1 %0.2d,%1.2d,%2.2d" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline float32x2_t __attribute__ ((__always_inline__)) -vuzp2_f32 (float32x2_t a, float32x2_t b) -{ - float32x2_t result; - __asm__ ("uzp2 %0.2s,%1.2s,%2.2s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) -vuzp2_p8 (poly8x8_t a, poly8x8_t b) -{ - poly8x8_t result; - __asm__ ("uzp2 %0.8b,%1.8b,%2.8b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__)) -vuzp2_p16 (poly16x4_t a, poly16x4_t b) -{ - poly16x4_t result; - __asm__ ("uzp2 %0.4h,%1.4h,%2.4h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) -vuzp2_s8 (int8x8_t a, int8x8_t b) -{ - int8x8_t result; - __asm__ ("uzp2 %0.8b,%1.8b,%2.8b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) -vuzp2_s16 (int16x4_t a, int16x4_t b) -{ - int16x4_t result; - __asm__ ("uzp2 %0.4h,%1.4h,%2.4h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int32x2_t __attribute__ ((__always_inline__)) -vuzp2_s32 (int32x2_t a, int32x2_t b) -{ - int32x2_t result; - __asm__ ("uzp2 %0.2s,%1.2s,%2.2s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) -vuzp2_u8 (uint8x8_t a, uint8x8_t b) -{ - uint8x8_t result; - __asm__ ("uzp2 %0.8b,%1.8b,%2.8b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) -vuzp2_u16 (uint16x4_t a, uint16x4_t b) -{ - uint16x4_t result; - __asm__ ("uzp2 %0.4h,%1.4h,%2.4h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) -vuzp2_u32 (uint32x2_t a, uint32x2_t b) -{ - uint32x2_t result; - __asm__ ("uzp2 %0.2s,%1.2s,%2.2s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline float32x4_t __attribute__ ((__always_inline__)) -vuzp2q_f32 (float32x4_t a, float32x4_t b) -{ - float32x4_t result; - __asm__ ("uzp2 %0.4s,%1.4s,%2.4s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline float64x2_t __attribute__ ((__always_inline__)) -vuzp2q_f64 (float64x2_t a, float64x2_t b) -{ - float64x2_t result; - __asm__ ("uzp2 %0.2d,%1.2d,%2.2d" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) -vuzp2q_p8 (poly8x16_t a, poly8x16_t b) -{ - poly8x16_t result; - __asm__ ("uzp2 %0.16b,%1.16b,%2.16b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__)) -vuzp2q_p16 (poly16x8_t a, poly16x8_t b) -{ - poly16x8_t result; - __asm__ ("uzp2 %0.8h,%1.8h,%2.8h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) -vuzp2q_s8 (int8x16_t a, int8x16_t b) -{ - int8x16_t result; - __asm__ ("uzp2 %0.16b,%1.16b,%2.16b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) -vuzp2q_s16 (int16x8_t a, int16x8_t b) -{ - int16x8_t result; - __asm__ ("uzp2 %0.8h,%1.8h,%2.8h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) -vuzp2q_s32 (int32x4_t a, int32x4_t b) -{ - int32x4_t result; - __asm__ ("uzp2 %0.4s,%1.4s,%2.4s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) -vuzp2q_s64 (int64x2_t a, int64x2_t b) -{ - int64x2_t result; - __asm__ ("uzp2 %0.2d,%1.2d,%2.2d" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) -vuzp2q_u8 (uint8x16_t a, uint8x16_t b) -{ - uint8x16_t result; - __asm__ ("uzp2 %0.16b,%1.16b,%2.16b" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) -vuzp2q_u16 (uint16x8_t a, uint16x8_t b) -{ - uint16x8_t result; - __asm__ ("uzp2 %0.8h,%1.8h,%2.8h" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) -vuzp2q_u32 (uint32x4_t a, uint32x4_t b) -{ - uint32x4_t result; - __asm__ ("uzp2 %0.4s,%1.4s,%2.4s" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} - -__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) -vuzp2q_u64 (uint64x2_t a, uint64x2_t b) -{ - uint64x2_t result; - __asm__ ("uzp2 %0.2d,%1.2d,%2.2d" - : "=w"(result) - : "w"(a), "w"(b) - : /* No clobbers */); - return result; -} /* End of temporary inline asm implementations. */ @@ -18367,6 +16722,292 @@ vdupd_laneq_u64 (uint64x2_t __a, const int __b) return __aarch64_vgetq_lane_u64 (__a, __b); } +/* vext */ + +__extension__ static __inline float32x2_t __attribute__ ((__always_inline__)) +vext_f32 (float32x2_t __a, float32x2_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 2); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint32x2_t) {2-__c, 3-__c}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {__c, __c+1}); +#endif +} + +__extension__ static __inline float64x1_t __attribute__ ((__always_inline__)) +vext_f64 (float64x1_t __a, float64x1_t __b, __const int __c) +{ + /* The only possible index to the assembler instruction returns element 0. */ + __builtin_aarch64_im_lane_boundsi (__c, 1); + return __a; +} +__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) +vext_p8 (poly8x8_t __a, poly8x8_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 8); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint8x8_t) + {8-__c, 9-__c, 10-__c, 11-__c, 12-__c, 13-__c, 14-__c, 15-__c}); +#else + return __builtin_shuffle (__a, __b, + (uint8x8_t) {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7}); +#endif +} + +__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__)) +vext_p16 (poly16x4_t __a, poly16x4_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 4); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, + (uint16x4_t) {4-__c, 5-__c, 6-__c, 7-__c}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {__c, __c+1, __c+2, __c+3}); +#endif +} + +__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) +vext_s8 (int8x8_t __a, int8x8_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 8); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint8x8_t) + {8-__c, 9-__c, 10-__c, 11-__c, 12-__c, 13-__c, 14-__c, 15-__c}); +#else + return __builtin_shuffle (__a, __b, + (uint8x8_t) {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7}); +#endif +} + +__extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) +vext_s16 (int16x4_t __a, int16x4_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 4); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, + (uint16x4_t) {4-__c, 5-__c, 6-__c, 7-__c}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {__c, __c+1, __c+2, __c+3}); +#endif +} + +__extension__ static __inline int32x2_t __attribute__ ((__always_inline__)) +vext_s32 (int32x2_t __a, int32x2_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 2); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint32x2_t) {2-__c, 3-__c}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {__c, __c+1}); +#endif +} + +__extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) +vext_s64 (int64x1_t __a, int64x1_t __b, __const int __c) +{ + /* The only possible index to the assembler instruction returns element 0. */ + __builtin_aarch64_im_lane_boundsi (__c, 1); + return __a; +} + +__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) +vext_u8 (uint8x8_t __a, uint8x8_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 8); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint8x8_t) + {8-__c, 9-__c, 10-__c, 11-__c, 12-__c, 13-__c, 14-__c, 15-__c}); +#else + return __builtin_shuffle (__a, __b, + (uint8x8_t) {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7}); +#endif +} + +__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) +vext_u16 (uint16x4_t __a, uint16x4_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 4); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, + (uint16x4_t) {4-__c, 5-__c, 6-__c, 7-__c}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {__c, __c+1, __c+2, __c+3}); +#endif +} + +__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) +vext_u32 (uint32x2_t __a, uint32x2_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 2); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint32x2_t) {2-__c, 3-__c}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {__c, __c+1}); +#endif +} + +__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) +vext_u64 (uint64x1_t __a, uint64x1_t __b, __const int __c) +{ + /* The only possible index to the assembler instruction returns element 0. */ + __builtin_aarch64_im_lane_boundsi (__c, 1); + return __a; +} + +__extension__ static __inline float32x4_t __attribute__ ((__always_inline__)) +vextq_f32 (float32x4_t __a, float32x4_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 4); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, + (uint32x4_t) {4-__c, 5-__c, 6-__c, 7-__c}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {__c, __c+1, __c+2, __c+3}); +#endif +} + +__extension__ static __inline float64x2_t __attribute__ ((__always_inline__)) +vextq_f64 (float64x2_t __a, float64x2_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 2); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint64x2_t) {2-__c, 3-__c}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {__c, __c+1}); +#endif +} + +__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) +vextq_p8 (poly8x16_t __a, poly8x16_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 16); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint8x16_t) + {16-__c, 17-__c, 18-__c, 19-__c, 20-__c, 21-__c, 22-__c, 23-__c, + 24-__c, 25-__c, 26-__c, 27-__c, 28-__c, 29-__c, 30-__c, 31-__c}); +#else + return __builtin_shuffle (__a, __b, (uint8x16_t) + {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7, + __c+8, __c+9, __c+10, __c+11, __c+12, __c+13, __c+14, __c+15}); +#endif +} + +__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__)) +vextq_p16 (poly16x8_t __a, poly16x8_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 8); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint16x8_t) + {8-__c, 9-__c, 10-__c, 11-__c, 12-__c, 13-__c, 14-__c, 15-__c}); +#else + return __builtin_shuffle (__a, __b, + (uint16x8_t) {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7}); +#endif +} + +__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) +vextq_s8 (int8x16_t __a, int8x16_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 16); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint8x16_t) + {16-__c, 17-__c, 18-__c, 19-__c, 20-__c, 21-__c, 22-__c, 23-__c, + 24-__c, 25-__c, 26-__c, 27-__c, 28-__c, 29-__c, 30-__c, 31-__c}); +#else + return __builtin_shuffle (__a, __b, (uint8x16_t) + {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7, + __c+8, __c+9, __c+10, __c+11, __c+12, __c+13, __c+14, __c+15}); +#endif +} + +__extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) +vextq_s16 (int16x8_t __a, int16x8_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 8); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint16x8_t) + {8-__c, 9-__c, 10-__c, 11-__c, 12-__c, 13-__c, 14-__c, 15-__c}); +#else + return __builtin_shuffle (__a, __b, + (uint16x8_t) {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7}); +#endif +} + +__extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) +vextq_s32 (int32x4_t __a, int32x4_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 4); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, + (uint32x4_t) {4-__c, 5-__c, 6-__c, 7-__c}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {__c, __c+1, __c+2, __c+3}); +#endif +} + +__extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) +vextq_s64 (int64x2_t __a, int64x2_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 2); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint64x2_t) {2-__c, 3-__c}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {__c, __c+1}); +#endif +} + +__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) +vextq_u8 (uint8x16_t __a, uint8x16_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 16); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint8x16_t) + {16-__c, 17-__c, 18-__c, 19-__c, 20-__c, 21-__c, 22-__c, 23-__c, + 24-__c, 25-__c, 26-__c, 27-__c, 28-__c, 29-__c, 30-__c, 31-__c}); +#else + return __builtin_shuffle (__a, __b, (uint8x16_t) + {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7, + __c+8, __c+9, __c+10, __c+11, __c+12, __c+13, __c+14, __c+15}); +#endif +} + +__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) +vextq_u16 (uint16x8_t __a, uint16x8_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 8); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint16x8_t) + {8-__c, 9-__c, 10-__c, 11-__c, 12-__c, 13-__c, 14-__c, 15-__c}); +#else + return __builtin_shuffle (__a, __b, + (uint16x8_t) {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7}); +#endif +} + +__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) +vextq_u32 (uint32x4_t __a, uint32x4_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 4); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, + (uint32x4_t) {4-__c, 5-__c, 6-__c, 7-__c}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {__c, __c+1, __c+2, __c+3}); +#endif +} + +__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) +vextq_u64 (uint64x2_t __a, uint64x2_t __b, __const int __c) +{ + __builtin_aarch64_im_lane_boundsi (__c, 2); +#ifdef __AARCH64EB__ + return __builtin_shuffle (__b, __a, (uint64x2_t) {2-__c, 3-__c}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {__c, __c+1}); +#endif +} + /* vfma_lane */ __extension__ static __inline float32x2_t __attribute__ ((__always_inline__)) @@ -20856,25 +19497,26 @@ vqaddd_s64 (int64x1_t __a, int64x1_t __b) __extension__ static __inline uint8x1_t __attribute__ ((__always_inline__)) vqaddb_u8 (uint8x1_t __a, uint8x1_t __b) { - return (uint8x1_t) __builtin_aarch64_uqaddqi (__a, __b); + return (uint8x1_t) __builtin_aarch64_uqaddqi_uuu (__a, __b); } __extension__ static __inline uint16x1_t __attribute__ ((__always_inline__)) vqaddh_u16 (uint16x1_t __a, uint16x1_t __b) { - return (uint16x1_t) __builtin_aarch64_uqaddhi (__a, __b); + return (uint16x1_t) __builtin_aarch64_uqaddhi_uuu (__a, __b); } __extension__ static __inline uint32x1_t __attribute__ ((__always_inline__)) vqadds_u32 (uint32x1_t __a, uint32x1_t __b) { - return (uint32x1_t) __builtin_aarch64_uqaddsi (__a, __b); + return (uint32x1_t) __builtin_aarch64_uqaddsi_uuu (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vqaddd_u64 (uint64x1_t __a, uint64x1_t __b) { - return (uint64x1_t) __builtin_aarch64_uqadddi (__a, __b); + return (uint64x1_t) __builtin_aarch64_uqadddi_uuu ((uint64_t) __a, + (uint64_t) __b); } /* vqdmlal */ @@ -20892,7 +19534,7 @@ vqdmlal_high_s16 (int32x4_t __a, int16x8_t __b, int16x8_t __c) } __extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) -vqdmlal_high_lane_s16 (int32x4_t __a, int16x8_t __b, int16x8_t __c, +vqdmlal_high_lane_s16 (int32x4_t __a, int16x8_t __b, int16x4_t __c, int const __d) { return __builtin_aarch64_sqdmlal2_lanev8hi (__a, __b, __c, __d); @@ -20914,8 +19556,7 @@ vqdmlal_high_n_s16 (int32x4_t __a, int16x8_t __b, int16_t __c) __extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) vqdmlal_lane_s16 (int32x4_t __a, int16x4_t __b, int16x4_t __c, int const __d) { - int16x8_t __tmp = vcombine_s16 (__c, vcreate_s16 (__AARCH64_INT64_C (0))); - return __builtin_aarch64_sqdmlal_lanev4hi (__a, __b, __tmp, __d); + return __builtin_aarch64_sqdmlal_lanev4hi (__a, __b, __c, __d); } __extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) @@ -20943,7 +19584,7 @@ vqdmlal_high_s32 (int64x2_t __a, int32x4_t __b, int32x4_t __c) } __extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) -vqdmlal_high_lane_s32 (int64x2_t __a, int32x4_t __b, int32x4_t __c, +vqdmlal_high_lane_s32 (int64x2_t __a, int32x4_t __b, int32x2_t __c, int const __d) { return __builtin_aarch64_sqdmlal2_lanev4si (__a, __b, __c, __d); @@ -20965,8 +19606,7 @@ vqdmlal_high_n_s32 (int64x2_t __a, int32x4_t __b, int32_t __c) __extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) vqdmlal_lane_s32 (int64x2_t __a, int32x2_t __b, int32x2_t __c, int const __d) { - int32x4_t __tmp = vcombine_s32 (__c, vcreate_s32 (__AARCH64_INT64_C (0))); - return __builtin_aarch64_sqdmlal_lanev2si (__a, __b, __tmp, __d); + return __builtin_aarch64_sqdmlal_lanev2si (__a, __b, __c, __d); } __extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) @@ -20988,7 +19628,7 @@ vqdmlalh_s16 (int32x1_t __a, int16x1_t __b, int16x1_t __c) } __extension__ static __inline int32x1_t __attribute__ ((__always_inline__)) -vqdmlalh_lane_s16 (int32x1_t __a, int16x1_t __b, int16x8_t __c, const int __d) +vqdmlalh_lane_s16 (int32x1_t __a, int16x1_t __b, int16x4_t __c, const int __d) { return __builtin_aarch64_sqdmlal_lanehi (__a, __b, __c, __d); } @@ -21000,7 +19640,7 @@ vqdmlals_s32 (int64x1_t __a, int32x1_t __b, int32x1_t __c) } __extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) -vqdmlals_lane_s32 (int64x1_t __a, int32x1_t __b, int32x4_t __c, const int __d) +vqdmlals_lane_s32 (int64x1_t __a, int32x1_t __b, int32x2_t __c, const int __d) { return __builtin_aarch64_sqdmlal_lanesi (__a, __b, __c, __d); } @@ -21020,7 +19660,7 @@ vqdmlsl_high_s16 (int32x4_t __a, int16x8_t __b, int16x8_t __c) } __extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) -vqdmlsl_high_lane_s16 (int32x4_t __a, int16x8_t __b, int16x8_t __c, +vqdmlsl_high_lane_s16 (int32x4_t __a, int16x8_t __b, int16x4_t __c, int const __d) { return __builtin_aarch64_sqdmlsl2_lanev8hi (__a, __b, __c, __d); @@ -21042,8 +19682,7 @@ vqdmlsl_high_n_s16 (int32x4_t __a, int16x8_t __b, int16_t __c) __extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) vqdmlsl_lane_s16 (int32x4_t __a, int16x4_t __b, int16x4_t __c, int const __d) { - int16x8_t __tmp = vcombine_s16 (__c, vcreate_s16 (__AARCH64_INT64_C (0))); - return __builtin_aarch64_sqdmlsl_lanev4hi (__a, __b, __tmp, __d); + return __builtin_aarch64_sqdmlsl_lanev4hi (__a, __b, __c, __d); } __extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) @@ -21071,7 +19710,7 @@ vqdmlsl_high_s32 (int64x2_t __a, int32x4_t __b, int32x4_t __c) } __extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) -vqdmlsl_high_lane_s32 (int64x2_t __a, int32x4_t __b, int32x4_t __c, +vqdmlsl_high_lane_s32 (int64x2_t __a, int32x4_t __b, int32x2_t __c, int const __d) { return __builtin_aarch64_sqdmlsl2_lanev4si (__a, __b, __c, __d); @@ -21093,8 +19732,7 @@ vqdmlsl_high_n_s32 (int64x2_t __a, int32x4_t __b, int32_t __c) __extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) vqdmlsl_lane_s32 (int64x2_t __a, int32x2_t __b, int32x2_t __c, int const __d) { - int32x4_t __tmp = vcombine_s32 (__c, vcreate_s32 (__AARCH64_INT64_C (0))); - return __builtin_aarch64_sqdmlsl_lanev2si (__a, __b, __tmp, __d); + return __builtin_aarch64_sqdmlsl_lanev2si (__a, __b, __c, __d); } __extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) @@ -21116,7 +19754,7 @@ vqdmlslh_s16 (int32x1_t __a, int16x1_t __b, int16x1_t __c) } __extension__ static __inline int32x1_t __attribute__ ((__always_inline__)) -vqdmlslh_lane_s16 (int32x1_t __a, int16x1_t __b, int16x8_t __c, const int __d) +vqdmlslh_lane_s16 (int32x1_t __a, int16x1_t __b, int16x4_t __c, const int __d) { return __builtin_aarch64_sqdmlsl_lanehi (__a, __b, __c, __d); } @@ -21128,7 +19766,7 @@ vqdmlsls_s32 (int64x1_t __a, int32x1_t __b, int32x1_t __c) } __extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) -vqdmlsls_lane_s32 (int64x1_t __a, int32x1_t __b, int32x4_t __c, const int __d) +vqdmlsls_lane_s32 (int64x1_t __a, int32x1_t __b, int32x2_t __c, const int __d) { return __builtin_aarch64_sqdmlsl_lanesi (__a, __b, __c, __d); } @@ -21166,7 +19804,7 @@ vqdmulhh_s16 (int16x1_t __a, int16x1_t __b) } __extension__ static __inline int16x1_t __attribute__ ((__always_inline__)) -vqdmulhh_lane_s16 (int16x1_t __a, int16x8_t __b, const int __c) +vqdmulhh_lane_s16 (int16x1_t __a, int16x4_t __b, const int __c) { return __builtin_aarch64_sqdmulh_lanehi (__a, __b, __c); } @@ -21178,7 +19816,7 @@ vqdmulhs_s32 (int32x1_t __a, int32x1_t __b) } __extension__ static __inline int32x1_t __attribute__ ((__always_inline__)) -vqdmulhs_lane_s32 (int32x1_t __a, int32x4_t __b, const int __c) +vqdmulhs_lane_s32 (int32x1_t __a, int32x2_t __b, const int __c) { return __builtin_aarch64_sqdmulh_lanesi (__a, __b, __c); } @@ -21198,7 +19836,7 @@ vqdmull_high_s16 (int16x8_t __a, int16x8_t __b) } __extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) -vqdmull_high_lane_s16 (int16x8_t __a, int16x8_t __b, int const __c) +vqdmull_high_lane_s16 (int16x8_t __a, int16x4_t __b, int const __c) { return __builtin_aarch64_sqdmull2_lanev8hi (__a, __b,__c); } @@ -21218,8 +19856,7 @@ vqdmull_high_n_s16 (int16x8_t __a, int16_t __b) __extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) vqdmull_lane_s16 (int16x4_t __a, int16x4_t __b, int const __c) { - int16x8_t __tmp = vcombine_s16 (__b, vcreate_s16 (__AARCH64_INT64_C (0))); - return __builtin_aarch64_sqdmull_lanev4hi (__a, __tmp, __c); + return __builtin_aarch64_sqdmull_lanev4hi (__a, __b, __c); } __extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) @@ -21247,7 +19884,7 @@ vqdmull_high_s32 (int32x4_t __a, int32x4_t __b) } __extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) -vqdmull_high_lane_s32 (int32x4_t __a, int32x4_t __b, int const __c) +vqdmull_high_lane_s32 (int32x4_t __a, int32x2_t __b, int const __c) { return __builtin_aarch64_sqdmull2_lanev4si (__a, __b, __c); } @@ -21267,8 +19904,7 @@ vqdmull_high_n_s32 (int32x4_t __a, int32_t __b) __extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) vqdmull_lane_s32 (int32x2_t __a, int32x2_t __b, int const __c) { - int32x4_t __tmp = vcombine_s32 (__b, vcreate_s32 (__AARCH64_INT64_C (0))); - return __builtin_aarch64_sqdmull_lanev2si (__a, __tmp, __c); + return __builtin_aarch64_sqdmull_lanev2si (__a, __b, __c); } __extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) @@ -21290,7 +19926,7 @@ vqdmullh_s16 (int16x1_t __a, int16x1_t __b) } __extension__ static __inline int32x1_t __attribute__ ((__always_inline__)) -vqdmullh_lane_s16 (int16x1_t __a, int16x8_t __b, const int __c) +vqdmullh_lane_s16 (int16x1_t __a, int16x4_t __b, const int __c) { return __builtin_aarch64_sqdmull_lanehi (__a, __b, __c); } @@ -21302,7 +19938,7 @@ vqdmulls_s32 (int32x1_t __a, int32x1_t __b) } __extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) -vqdmulls_lane_s32 (int32x1_t __a, int32x4_t __b, const int __c) +vqdmulls_lane_s32 (int32x1_t __a, int32x2_t __b, const int __c) { return __builtin_aarch64_sqdmull_lanesi (__a, __b, __c); } @@ -21484,7 +20120,7 @@ vqrdmulhh_s16 (int16x1_t __a, int16x1_t __b) } __extension__ static __inline int16x1_t __attribute__ ((__always_inline__)) -vqrdmulhh_lane_s16 (int16x1_t __a, int16x8_t __b, const int __c) +vqrdmulhh_lane_s16 (int16x1_t __a, int16x4_t __b, const int __c) { return __builtin_aarch64_sqrdmulh_lanehi (__a, __b, __c); } @@ -21496,7 +20132,7 @@ vqrdmulhs_s32 (int32x1_t __a, int32x1_t __b) } __extension__ static __inline int32x1_t __attribute__ ((__always_inline__)) -vqrdmulhs_lane_s32 (int32x1_t __a, int32x4_t __b, const int __c) +vqrdmulhs_lane_s32 (int32x1_t __a, int32x2_t __b, const int __c) { return __builtin_aarch64_sqrdmulh_lanesi (__a, __b, __c); } @@ -21530,25 +20166,25 @@ vqrshl_s64 (int64x1_t __a, int64x1_t __b) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vqrshl_u8 (uint8x8_t __a, int8x8_t __b) { - return (uint8x8_t) __builtin_aarch64_uqrshlv8qi ((int8x8_t) __a, __b); + return __builtin_aarch64_uqrshlv8qi_uus ( __a, __b); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vqrshl_u16 (uint16x4_t __a, int16x4_t __b) { - return (uint16x4_t) __builtin_aarch64_uqrshlv4hi ((int16x4_t) __a, __b); + return __builtin_aarch64_uqrshlv4hi_uus ( __a, __b); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vqrshl_u32 (uint32x2_t __a, int32x2_t __b) { - return (uint32x2_t) __builtin_aarch64_uqrshlv2si ((int32x2_t) __a, __b); + return __builtin_aarch64_uqrshlv2si_uus ( __a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vqrshl_u64 (uint64x1_t __a, int64x1_t __b) { - return (uint64x1_t) __builtin_aarch64_uqrshldi ((int64x1_t) __a, __b); + return __builtin_aarch64_uqrshldi_uus ( __a, __b); } __extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) @@ -21578,25 +20214,25 @@ vqrshlq_s64 (int64x2_t __a, int64x2_t __b) __extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) vqrshlq_u8 (uint8x16_t __a, int8x16_t __b) { - return (uint8x16_t) __builtin_aarch64_uqrshlv16qi ((int8x16_t) __a, __b); + return __builtin_aarch64_uqrshlv16qi_uus ( __a, __b); } __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vqrshlq_u16 (uint16x8_t __a, int16x8_t __b) { - return (uint16x8_t) __builtin_aarch64_uqrshlv8hi ((int16x8_t) __a, __b); + return __builtin_aarch64_uqrshlv8hi_uus ( __a, __b); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vqrshlq_u32 (uint32x4_t __a, int32x4_t __b) { - return (uint32x4_t) __builtin_aarch64_uqrshlv4si ((int32x4_t) __a, __b); + return __builtin_aarch64_uqrshlv4si_uus ( __a, __b); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vqrshlq_u64 (uint64x2_t __a, int64x2_t __b) { - return (uint64x2_t) __builtin_aarch64_uqrshlv2di ((int64x2_t) __a, __b); + return __builtin_aarch64_uqrshlv2di_uus ( __a, __b); } __extension__ static __inline int8x1_t __attribute__ ((__always_inline__)) @@ -21626,25 +20262,25 @@ vqrshld_s64 (int64x1_t __a, int64x1_t __b) __extension__ static __inline uint8x1_t __attribute__ ((__always_inline__)) vqrshlb_u8 (uint8x1_t __a, uint8x1_t __b) { - return (uint8x1_t) __builtin_aarch64_uqrshlqi (__a, __b); + return __builtin_aarch64_uqrshlqi_uus (__a, __b); } __extension__ static __inline uint16x1_t __attribute__ ((__always_inline__)) vqrshlh_u16 (uint16x1_t __a, uint16x1_t __b) { - return (uint16x1_t) __builtin_aarch64_uqrshlhi (__a, __b); + return __builtin_aarch64_uqrshlhi_uus (__a, __b); } __extension__ static __inline uint32x1_t __attribute__ ((__always_inline__)) vqrshls_u32 (uint32x1_t __a, uint32x1_t __b) { - return (uint32x1_t) __builtin_aarch64_uqrshlsi (__a, __b); + return __builtin_aarch64_uqrshlsi_uus (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vqrshld_u64 (uint64x1_t __a, uint64x1_t __b) { - return (uint64x1_t) __builtin_aarch64_uqrshldi (__a, __b); + return __builtin_aarch64_uqrshldi_uus (__a, __b); } /* vqrshrn */ @@ -21670,19 +20306,19 @@ vqrshrn_n_s64 (int64x2_t __a, const int __b) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vqrshrn_n_u16 (uint16x8_t __a, const int __b) { - return (uint8x8_t) __builtin_aarch64_uqrshrn_nv8hi ((int16x8_t) __a, __b); + return __builtin_aarch64_uqrshrn_nv8hi_uus ( __a, __b); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vqrshrn_n_u32 (uint32x4_t __a, const int __b) { - return (uint16x4_t) __builtin_aarch64_uqrshrn_nv4si ((int32x4_t) __a, __b); + return __builtin_aarch64_uqrshrn_nv4si_uus ( __a, __b); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vqrshrn_n_u64 (uint64x2_t __a, const int __b) { - return (uint32x2_t) __builtin_aarch64_uqrshrn_nv2di ((int64x2_t) __a, __b); + return __builtin_aarch64_uqrshrn_nv2di_uus ( __a, __b); } __extension__ static __inline int8x1_t __attribute__ ((__always_inline__)) @@ -21706,19 +20342,19 @@ vqrshrnd_n_s64 (int64x1_t __a, const int __b) __extension__ static __inline uint8x1_t __attribute__ ((__always_inline__)) vqrshrnh_n_u16 (uint16x1_t __a, const int __b) { - return (uint8x1_t) __builtin_aarch64_uqrshrn_nhi (__a, __b); + return __builtin_aarch64_uqrshrn_nhi_uus (__a, __b); } __extension__ static __inline uint16x1_t __attribute__ ((__always_inline__)) vqrshrns_n_u32 (uint32x1_t __a, const int __b) { - return (uint16x1_t) __builtin_aarch64_uqrshrn_nsi (__a, __b); + return __builtin_aarch64_uqrshrn_nsi_uus (__a, __b); } __extension__ static __inline uint32x1_t __attribute__ ((__always_inline__)) vqrshrnd_n_u64 (uint64x1_t __a, const int __b) { - return (uint32x1_t) __builtin_aarch64_uqrshrn_ndi (__a, __b); + return __builtin_aarch64_uqrshrn_ndi_uus (__a, __b); } /* vqrshrun */ @@ -21788,25 +20424,25 @@ vqshl_s64 (int64x1_t __a, int64x1_t __b) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vqshl_u8 (uint8x8_t __a, int8x8_t __b) { - return (uint8x8_t) __builtin_aarch64_uqshlv8qi ((int8x8_t) __a, __b); + return __builtin_aarch64_uqshlv8qi_uus ( __a, __b); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vqshl_u16 (uint16x4_t __a, int16x4_t __b) { - return (uint16x4_t) __builtin_aarch64_uqshlv4hi ((int16x4_t) __a, __b); + return __builtin_aarch64_uqshlv4hi_uus ( __a, __b); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vqshl_u32 (uint32x2_t __a, int32x2_t __b) { - return (uint32x2_t) __builtin_aarch64_uqshlv2si ((int32x2_t) __a, __b); + return __builtin_aarch64_uqshlv2si_uus ( __a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vqshl_u64 (uint64x1_t __a, int64x1_t __b) { - return (uint64x1_t) __builtin_aarch64_uqshldi ((int64x1_t) __a, __b); + return __builtin_aarch64_uqshldi_uus ( __a, __b); } __extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) @@ -21836,25 +20472,25 @@ vqshlq_s64 (int64x2_t __a, int64x2_t __b) __extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) vqshlq_u8 (uint8x16_t __a, int8x16_t __b) { - return (uint8x16_t) __builtin_aarch64_uqshlv16qi ((int8x16_t) __a, __b); + return __builtin_aarch64_uqshlv16qi_uus ( __a, __b); } __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vqshlq_u16 (uint16x8_t __a, int16x8_t __b) { - return (uint16x8_t) __builtin_aarch64_uqshlv8hi ((int16x8_t) __a, __b); + return __builtin_aarch64_uqshlv8hi_uus ( __a, __b); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vqshlq_u32 (uint32x4_t __a, int32x4_t __b) { - return (uint32x4_t) __builtin_aarch64_uqshlv4si ((int32x4_t) __a, __b); + return __builtin_aarch64_uqshlv4si_uus ( __a, __b); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vqshlq_u64 (uint64x2_t __a, int64x2_t __b) { - return (uint64x2_t) __builtin_aarch64_uqshlv2di ((int64x2_t) __a, __b); + return __builtin_aarch64_uqshlv2di_uus ( __a, __b); } __extension__ static __inline int8x1_t __attribute__ ((__always_inline__)) @@ -21884,25 +20520,25 @@ vqshld_s64 (int64x1_t __a, int64x1_t __b) __extension__ static __inline uint8x1_t __attribute__ ((__always_inline__)) vqshlb_u8 (uint8x1_t __a, uint8x1_t __b) { - return (uint8x1_t) __builtin_aarch64_uqshlqi (__a, __b); + return __builtin_aarch64_uqshlqi_uus (__a, __b); } __extension__ static __inline uint16x1_t __attribute__ ((__always_inline__)) vqshlh_u16 (uint16x1_t __a, uint16x1_t __b) { - return (uint16x1_t) __builtin_aarch64_uqshlhi (__a, __b); + return __builtin_aarch64_uqshlhi_uus (__a, __b); } __extension__ static __inline uint32x1_t __attribute__ ((__always_inline__)) vqshls_u32 (uint32x1_t __a, uint32x1_t __b) { - return (uint32x1_t) __builtin_aarch64_uqshlsi (__a, __b); + return __builtin_aarch64_uqshlsi_uus (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vqshld_u64 (uint64x1_t __a, uint64x1_t __b) { - return (uint64x1_t) __builtin_aarch64_uqshldi (__a, __b); + return __builtin_aarch64_uqshldi_uus (__a, __b); } __extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) @@ -21932,25 +20568,25 @@ vqshl_n_s64 (int64x1_t __a, const int __b) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vqshl_n_u8 (uint8x8_t __a, const int __b) { - return (uint8x8_t) __builtin_aarch64_uqshl_nv8qi ((int8x8_t) __a, __b); + return __builtin_aarch64_uqshl_nv8qi_uus (__a, __b); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vqshl_n_u16 (uint16x4_t __a, const int __b) { - return (uint16x4_t) __builtin_aarch64_uqshl_nv4hi ((int16x4_t) __a, __b); + return __builtin_aarch64_uqshl_nv4hi_uus (__a, __b); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vqshl_n_u32 (uint32x2_t __a, const int __b) { - return (uint32x2_t) __builtin_aarch64_uqshl_nv2si ((int32x2_t) __a, __b); + return __builtin_aarch64_uqshl_nv2si_uus (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vqshl_n_u64 (uint64x1_t __a, const int __b) { - return (uint64x1_t) __builtin_aarch64_uqshl_ndi ((int64x1_t) __a, __b); + return __builtin_aarch64_uqshl_ndi_uus (__a, __b); } __extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) @@ -21980,25 +20616,25 @@ vqshlq_n_s64 (int64x2_t __a, const int __b) __extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) vqshlq_n_u8 (uint8x16_t __a, const int __b) { - return (uint8x16_t) __builtin_aarch64_uqshl_nv16qi ((int8x16_t) __a, __b); + return __builtin_aarch64_uqshl_nv16qi_uus (__a, __b); } __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vqshlq_n_u16 (uint16x8_t __a, const int __b) { - return (uint16x8_t) __builtin_aarch64_uqshl_nv8hi ((int16x8_t) __a, __b); + return __builtin_aarch64_uqshl_nv8hi_uus (__a, __b); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vqshlq_n_u32 (uint32x4_t __a, const int __b) { - return (uint32x4_t) __builtin_aarch64_uqshl_nv4si ((int32x4_t) __a, __b); + return __builtin_aarch64_uqshl_nv4si_uus (__a, __b); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vqshlq_n_u64 (uint64x2_t __a, const int __b) { - return (uint64x2_t) __builtin_aarch64_uqshl_nv2di ((int64x2_t) __a, __b); + return __builtin_aarch64_uqshl_nv2di_uus (__a, __b); } __extension__ static __inline int8x1_t __attribute__ ((__always_inline__)) @@ -22028,25 +20664,25 @@ vqshld_n_s64 (int64x1_t __a, const int __b) __extension__ static __inline uint8x1_t __attribute__ ((__always_inline__)) vqshlb_n_u8 (uint8x1_t __a, const int __b) { - return (uint8x1_t) __builtin_aarch64_uqshl_nqi (__a, __b); + return __builtin_aarch64_uqshl_nqi_uus (__a, __b); } __extension__ static __inline uint16x1_t __attribute__ ((__always_inline__)) vqshlh_n_u16 (uint16x1_t __a, const int __b) { - return (uint16x1_t) __builtin_aarch64_uqshl_nhi (__a, __b); + return __builtin_aarch64_uqshl_nhi_uus (__a, __b); } __extension__ static __inline uint32x1_t __attribute__ ((__always_inline__)) vqshls_n_u32 (uint32x1_t __a, const int __b) { - return (uint32x1_t) __builtin_aarch64_uqshl_nsi (__a, __b); + return __builtin_aarch64_uqshl_nsi_uus (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vqshld_n_u64 (uint64x1_t __a, const int __b) { - return (uint64x1_t) __builtin_aarch64_uqshl_ndi (__a, __b); + return __builtin_aarch64_uqshl_ndi_uus (__a, __b); } /* vqshlu */ @@ -22054,73 +20690,73 @@ vqshld_n_u64 (uint64x1_t __a, const int __b) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vqshlu_n_s8 (int8x8_t __a, const int __b) { - return (uint8x8_t) __builtin_aarch64_sqshlu_nv8qi (__a, __b); + return __builtin_aarch64_sqshlu_nv8qi_uss (__a, __b); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vqshlu_n_s16 (int16x4_t __a, const int __b) { - return (uint16x4_t) __builtin_aarch64_sqshlu_nv4hi (__a, __b); + return __builtin_aarch64_sqshlu_nv4hi_uss (__a, __b); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vqshlu_n_s32 (int32x2_t __a, const int __b) { - return (uint32x2_t) __builtin_aarch64_sqshlu_nv2si (__a, __b); + return __builtin_aarch64_sqshlu_nv2si_uss (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vqshlu_n_s64 (int64x1_t __a, const int __b) { - return (uint64x1_t) __builtin_aarch64_sqshlu_ndi (__a, __b); + return __builtin_aarch64_sqshlu_ndi_uss (__a, __b); } __extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) vqshluq_n_s8 (int8x16_t __a, const int __b) { - return (uint8x16_t) __builtin_aarch64_sqshlu_nv16qi (__a, __b); + return __builtin_aarch64_sqshlu_nv16qi_uss (__a, __b); } __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vqshluq_n_s16 (int16x8_t __a, const int __b) { - return (uint16x8_t) __builtin_aarch64_sqshlu_nv8hi (__a, __b); + return __builtin_aarch64_sqshlu_nv8hi_uss (__a, __b); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vqshluq_n_s32 (int32x4_t __a, const int __b) { - return (uint32x4_t) __builtin_aarch64_sqshlu_nv4si (__a, __b); + return __builtin_aarch64_sqshlu_nv4si_uss (__a, __b); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vqshluq_n_s64 (int64x2_t __a, const int __b) { - return (uint64x2_t) __builtin_aarch64_sqshlu_nv2di (__a, __b); + return __builtin_aarch64_sqshlu_nv2di_uss (__a, __b); } __extension__ static __inline int8x1_t __attribute__ ((__always_inline__)) vqshlub_n_s8 (int8x1_t __a, const int __b) { - return (int8x1_t) __builtin_aarch64_sqshlu_nqi (__a, __b); + return (int8x1_t) __builtin_aarch64_sqshlu_nqi_uss (__a, __b); } __extension__ static __inline int16x1_t __attribute__ ((__always_inline__)) vqshluh_n_s16 (int16x1_t __a, const int __b) { - return (int16x1_t) __builtin_aarch64_sqshlu_nhi (__a, __b); + return (int16x1_t) __builtin_aarch64_sqshlu_nhi_uss (__a, __b); } __extension__ static __inline int32x1_t __attribute__ ((__always_inline__)) vqshlus_n_s32 (int32x1_t __a, const int __b) { - return (int32x1_t) __builtin_aarch64_sqshlu_nsi (__a, __b); + return (int32x1_t) __builtin_aarch64_sqshlu_nsi_uss (__a, __b); } __extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) vqshlud_n_s64 (int64x1_t __a, const int __b) { - return (int64x1_t) __builtin_aarch64_sqshlu_ndi (__a, __b); + return (int64x1_t) __builtin_aarch64_sqshlu_ndi_uss (__a, __b); } /* vqshrn */ @@ -22146,19 +20782,19 @@ vqshrn_n_s64 (int64x2_t __a, const int __b) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vqshrn_n_u16 (uint16x8_t __a, const int __b) { - return (uint8x8_t) __builtin_aarch64_uqshrn_nv8hi ((int16x8_t) __a, __b); + return __builtin_aarch64_uqshrn_nv8hi_uus ( __a, __b); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vqshrn_n_u32 (uint32x4_t __a, const int __b) { - return (uint16x4_t) __builtin_aarch64_uqshrn_nv4si ((int32x4_t) __a, __b); + return __builtin_aarch64_uqshrn_nv4si_uus ( __a, __b); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vqshrn_n_u64 (uint64x2_t __a, const int __b) { - return (uint32x2_t) __builtin_aarch64_uqshrn_nv2di ((int64x2_t) __a, __b); + return __builtin_aarch64_uqshrn_nv2di_uus ( __a, __b); } __extension__ static __inline int8x1_t __attribute__ ((__always_inline__)) @@ -22182,19 +20818,19 @@ vqshrnd_n_s64 (int64x1_t __a, const int __b) __extension__ static __inline uint8x1_t __attribute__ ((__always_inline__)) vqshrnh_n_u16 (uint16x1_t __a, const int __b) { - return (uint8x1_t) __builtin_aarch64_uqshrn_nhi (__a, __b); + return __builtin_aarch64_uqshrn_nhi_uus (__a, __b); } __extension__ static __inline uint16x1_t __attribute__ ((__always_inline__)) vqshrns_n_u32 (uint32x1_t __a, const int __b) { - return (uint16x1_t) __builtin_aarch64_uqshrn_nsi (__a, __b); + return __builtin_aarch64_uqshrn_nsi_uus (__a, __b); } __extension__ static __inline uint32x1_t __attribute__ ((__always_inline__)) vqshrnd_n_u64 (uint64x1_t __a, const int __b) { - return (uint32x1_t) __builtin_aarch64_uqshrn_ndi (__a, __b); + return __builtin_aarch64_uqshrn_ndi_uus (__a, __b); } /* vqshrun */ @@ -22264,25 +20900,26 @@ vqsubd_s64 (int64x1_t __a, int64x1_t __b) __extension__ static __inline uint8x1_t __attribute__ ((__always_inline__)) vqsubb_u8 (uint8x1_t __a, uint8x1_t __b) { - return (uint8x1_t) __builtin_aarch64_uqsubqi (__a, __b); + return (uint8x1_t) __builtin_aarch64_uqsubqi_uuu (__a, __b); } __extension__ static __inline uint16x1_t __attribute__ ((__always_inline__)) vqsubh_u16 (uint16x1_t __a, uint16x1_t __b) { - return (uint16x1_t) __builtin_aarch64_uqsubhi (__a, __b); + return (uint16x1_t) __builtin_aarch64_uqsubhi_uuu (__a, __b); } __extension__ static __inline uint32x1_t __attribute__ ((__always_inline__)) vqsubs_u32 (uint32x1_t __a, uint32x1_t __b) { - return (uint32x1_t) __builtin_aarch64_uqsubsi (__a, __b); + return (uint32x1_t) __builtin_aarch64_uqsubsi_uuu (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vqsubd_u64 (uint64x1_t __a, uint64x1_t __b) { - return (uint64x1_t) __builtin_aarch64_uqsubdi (__a, __b); + return (uint64x1_t) __builtin_aarch64_uqsubdi_uuu ((uint64_t) __a, + (uint64_t) __b); } /* vrecpe */ @@ -22363,6 +21000,234 @@ vrecpxd_f64 (float64_t __a) return __builtin_aarch64_frecpxdf (__a); } + +/* vrev */ + +__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) +vrev16_p8 (poly8x8_t a) +{ + return __builtin_shuffle (a, (uint8x8_t) { 1, 0, 3, 2, 5, 4, 7, 6 }); +} + +__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) +vrev16_s8 (int8x8_t a) +{ + return __builtin_shuffle (a, (uint8x8_t) { 1, 0, 3, 2, 5, 4, 7, 6 }); +} + +__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) +vrev16_u8 (uint8x8_t a) +{ + return __builtin_shuffle (a, (uint8x8_t) { 1, 0, 3, 2, 5, 4, 7, 6 }); +} + +__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) +vrev16q_p8 (poly8x16_t a) +{ + return __builtin_shuffle (a, + (uint8x16_t) { 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14 }); +} + +__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) +vrev16q_s8 (int8x16_t a) +{ + return __builtin_shuffle (a, + (uint8x16_t) { 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14 }); +} + +__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) +vrev16q_u8 (uint8x16_t a) +{ + return __builtin_shuffle (a, + (uint8x16_t) { 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14 }); +} + +__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) +vrev32_p8 (poly8x8_t a) +{ + return __builtin_shuffle (a, (uint8x8_t) { 3, 2, 1, 0, 7, 6, 5, 4 }); +} + +__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__)) +vrev32_p16 (poly16x4_t a) +{ + return __builtin_shuffle (a, (uint16x4_t) { 1, 0, 3, 2 }); +} + +__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) +vrev32_s8 (int8x8_t a) +{ + return __builtin_shuffle (a, (uint8x8_t) { 3, 2, 1, 0, 7, 6, 5, 4 }); +} + +__extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) +vrev32_s16 (int16x4_t a) +{ + return __builtin_shuffle (a, (uint16x4_t) { 1, 0, 3, 2 }); +} + +__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) +vrev32_u8 (uint8x8_t a) +{ + return __builtin_shuffle (a, (uint8x8_t) { 3, 2, 1, 0, 7, 6, 5, 4 }); +} + +__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) +vrev32_u16 (uint16x4_t a) +{ + return __builtin_shuffle (a, (uint16x4_t) { 1, 0, 3, 2 }); +} + +__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) +vrev32q_p8 (poly8x16_t a) +{ + return __builtin_shuffle (a, + (uint8x16_t) { 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12 }); +} + +__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__)) +vrev32q_p16 (poly16x8_t a) +{ + return __builtin_shuffle (a, (uint16x8_t) { 1, 0, 3, 2, 5, 4, 7, 6 }); +} + +__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) +vrev32q_s8 (int8x16_t a) +{ + return __builtin_shuffle (a, + (uint8x16_t) { 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12 }); +} + +__extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) +vrev32q_s16 (int16x8_t a) +{ + return __builtin_shuffle (a, (uint16x8_t) { 1, 0, 3, 2, 5, 4, 7, 6 }); +} + +__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) +vrev32q_u8 (uint8x16_t a) +{ + return __builtin_shuffle (a, + (uint8x16_t) { 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12 }); +} + +__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) +vrev32q_u16 (uint16x8_t a) +{ + return __builtin_shuffle (a, (uint16x8_t) { 1, 0, 3, 2, 5, 4, 7, 6 }); +} + +__extension__ static __inline float32x2_t __attribute__ ((__always_inline__)) +vrev64_f32 (float32x2_t a) +{ + return __builtin_shuffle (a, (uint32x2_t) { 1, 0 }); +} + +__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) +vrev64_p8 (poly8x8_t a) +{ + return __builtin_shuffle (a, (uint8x8_t) { 7, 6, 5, 4, 3, 2, 1, 0 }); +} + +__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__)) +vrev64_p16 (poly16x4_t a) +{ + return __builtin_shuffle (a, (uint16x4_t) { 3, 2, 1, 0 }); +} + +__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) +vrev64_s8 (int8x8_t a) +{ + return __builtin_shuffle (a, (uint8x8_t) { 7, 6, 5, 4, 3, 2, 1, 0 }); +} + +__extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) +vrev64_s16 (int16x4_t a) +{ + return __builtin_shuffle (a, (uint16x4_t) { 3, 2, 1, 0 }); +} + +__extension__ static __inline int32x2_t __attribute__ ((__always_inline__)) +vrev64_s32 (int32x2_t a) +{ + return __builtin_shuffle (a, (uint32x2_t) { 1, 0 }); +} + +__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) +vrev64_u8 (uint8x8_t a) +{ + return __builtin_shuffle (a, (uint8x8_t) { 7, 6, 5, 4, 3, 2, 1, 0 }); +} + +__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) +vrev64_u16 (uint16x4_t a) +{ + return __builtin_shuffle (a, (uint16x4_t) { 3, 2, 1, 0 }); +} + +__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) +vrev64_u32 (uint32x2_t a) +{ + return __builtin_shuffle (a, (uint32x2_t) { 1, 0 }); +} + +__extension__ static __inline float32x4_t __attribute__ ((__always_inline__)) +vrev64q_f32 (float32x4_t a) +{ + return __builtin_shuffle (a, (uint32x4_t) { 1, 0, 3, 2 }); +} + +__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) +vrev64q_p8 (poly8x16_t a) +{ + return __builtin_shuffle (a, + (uint8x16_t) { 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8 }); +} + +__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__)) +vrev64q_p16 (poly16x8_t a) +{ + return __builtin_shuffle (a, (uint16x8_t) { 3, 2, 1, 0, 7, 6, 5, 4 }); +} + +__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) +vrev64q_s8 (int8x16_t a) +{ + return __builtin_shuffle (a, + (uint8x16_t) { 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8 }); +} + +__extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) +vrev64q_s16 (int16x8_t a) +{ + return __builtin_shuffle (a, (uint16x8_t) { 3, 2, 1, 0, 7, 6, 5, 4 }); +} + +__extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) +vrev64q_s32 (int32x4_t a) +{ + return __builtin_shuffle (a, (uint32x4_t) { 1, 0, 3, 2 }); +} + +__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) +vrev64q_u8 (uint8x16_t a) +{ + return __builtin_shuffle (a, + (uint8x16_t) { 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8 }); +} + +__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) +vrev64q_u16 (uint16x8_t a) +{ + return __builtin_shuffle (a, (uint16x8_t) { 3, 2, 1, 0, 7, 6, 5, 4 }); +} + +__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) +vrev64q_u32 (uint32x4_t a) +{ + return __builtin_shuffle (a, (uint32x4_t) { 1, 0, 3, 2 }); +} + /* vrnd */ __extension__ static __inline float32x2_t __attribute__ ((__always_inline__)) @@ -22574,25 +21439,25 @@ vrshl_s64 (int64x1_t __a, int64x1_t __b) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vrshl_u8 (uint8x8_t __a, int8x8_t __b) { - return (uint8x8_t) __builtin_aarch64_urshlv8qi ((int8x8_t) __a, __b); + return __builtin_aarch64_urshlv8qi_uus (__a, __b); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vrshl_u16 (uint16x4_t __a, int16x4_t __b) { - return (uint16x4_t) __builtin_aarch64_urshlv4hi ((int16x4_t) __a, __b); + return __builtin_aarch64_urshlv4hi_uus (__a, __b); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vrshl_u32 (uint32x2_t __a, int32x2_t __b) { - return (uint32x2_t) __builtin_aarch64_urshlv2si ((int32x2_t) __a, __b); + return __builtin_aarch64_urshlv2si_uus (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vrshl_u64 (uint64x1_t __a, int64x1_t __b) { - return (uint64x1_t) __builtin_aarch64_urshldi ((int64x1_t) __a, __b); + return __builtin_aarch64_urshldi_uus (__a, __b); } __extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) @@ -22622,25 +21487,25 @@ vrshlq_s64 (int64x2_t __a, int64x2_t __b) __extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) vrshlq_u8 (uint8x16_t __a, int8x16_t __b) { - return (uint8x16_t) __builtin_aarch64_urshlv16qi ((int8x16_t) __a, __b); + return __builtin_aarch64_urshlv16qi_uus (__a, __b); } __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vrshlq_u16 (uint16x8_t __a, int16x8_t __b) { - return (uint16x8_t) __builtin_aarch64_urshlv8hi ((int16x8_t) __a, __b); + return __builtin_aarch64_urshlv8hi_uus (__a, __b); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vrshlq_u32 (uint32x4_t __a, int32x4_t __b) { - return (uint32x4_t) __builtin_aarch64_urshlv4si ((int32x4_t) __a, __b); + return __builtin_aarch64_urshlv4si_uus (__a, __b); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vrshlq_u64 (uint64x2_t __a, int64x2_t __b) { - return (uint64x2_t) __builtin_aarch64_urshlv2di ((int64x2_t) __a, __b); + return __builtin_aarch64_urshlv2di_uus (__a, __b); } __extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) @@ -22652,7 +21517,7 @@ vrshld_s64 (int64x1_t __a, int64x1_t __b) __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vrshld_u64 (uint64x1_t __a, uint64x1_t __b) { - return (uint64x1_t) __builtin_aarch64_urshldi (__a, __b); + return __builtin_aarch64_urshldi_uus (__a, __b); } /* vrshr */ @@ -22684,25 +21549,25 @@ vrshr_n_s64 (int64x1_t __a, const int __b) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vrshr_n_u8 (uint8x8_t __a, const int __b) { - return (uint8x8_t) __builtin_aarch64_urshr_nv8qi ((int8x8_t) __a, __b); + return __builtin_aarch64_urshr_nv8qi_uus (__a, __b); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vrshr_n_u16 (uint16x4_t __a, const int __b) { - return (uint16x4_t) __builtin_aarch64_urshr_nv4hi ((int16x4_t) __a, __b); + return __builtin_aarch64_urshr_nv4hi_uus (__a, __b); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vrshr_n_u32 (uint32x2_t __a, const int __b) { - return (uint32x2_t) __builtin_aarch64_urshr_nv2si ((int32x2_t) __a, __b); + return __builtin_aarch64_urshr_nv2si_uus (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vrshr_n_u64 (uint64x1_t __a, const int __b) { - return (uint64x1_t) __builtin_aarch64_urshr_ndi ((int64x1_t) __a, __b); + return __builtin_aarch64_urshr_ndi_uus (__a, __b); } __extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) @@ -22732,25 +21597,25 @@ vrshrq_n_s64 (int64x2_t __a, const int __b) __extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) vrshrq_n_u8 (uint8x16_t __a, const int __b) { - return (uint8x16_t) __builtin_aarch64_urshr_nv16qi ((int8x16_t) __a, __b); + return __builtin_aarch64_urshr_nv16qi_uus (__a, __b); } __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vrshrq_n_u16 (uint16x8_t __a, const int __b) { - return (uint16x8_t) __builtin_aarch64_urshr_nv8hi ((int16x8_t) __a, __b); + return __builtin_aarch64_urshr_nv8hi_uus (__a, __b); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vrshrq_n_u32 (uint32x4_t __a, const int __b) { - return (uint32x4_t) __builtin_aarch64_urshr_nv4si ((int32x4_t) __a, __b); + return __builtin_aarch64_urshr_nv4si_uus (__a, __b); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vrshrq_n_u64 (uint64x2_t __a, const int __b) { - return (uint64x2_t) __builtin_aarch64_urshr_nv2di ((int64x2_t) __a, __b); + return __builtin_aarch64_urshr_nv2di_uus (__a, __b); } __extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) @@ -22762,7 +21627,7 @@ vrshrd_n_s64 (int64x1_t __a, const int __b) __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vrshrd_n_u64 (uint64x1_t __a, const int __b) { - return (uint64x1_t) __builtin_aarch64_urshr_ndi (__a, __b); + return __builtin_aarch64_urshr_ndi_uus (__a, __b); } /* vrsra */ @@ -22794,29 +21659,25 @@ vrsra_n_s64 (int64x1_t __a, int64x1_t __b, const int __c) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vrsra_n_u8 (uint8x8_t __a, uint8x8_t __b, const int __c) { - return (uint8x8_t) __builtin_aarch64_ursra_nv8qi ((int8x8_t) __a, - (int8x8_t) __b, __c); + return __builtin_aarch64_ursra_nv8qi_uuus (__a, __b, __c); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vrsra_n_u16 (uint16x4_t __a, uint16x4_t __b, const int __c) { - return (uint16x4_t) __builtin_aarch64_ursra_nv4hi ((int16x4_t) __a, - (int16x4_t) __b, __c); + return __builtin_aarch64_ursra_nv4hi_uuus (__a, __b, __c); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vrsra_n_u32 (uint32x2_t __a, uint32x2_t __b, const int __c) { - return (uint32x2_t) __builtin_aarch64_ursra_nv2si ((int32x2_t) __a, - (int32x2_t) __b, __c); + return __builtin_aarch64_ursra_nv2si_uuus (__a, __b, __c); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vrsra_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c) { - return (uint64x1_t) __builtin_aarch64_ursra_ndi ((int64x1_t) __a, - (int64x1_t) __b, __c); + return __builtin_aarch64_ursra_ndi_uuus (__a, __b, __c); } __extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) @@ -22846,29 +21707,25 @@ vrsraq_n_s64 (int64x2_t __a, int64x2_t __b, const int __c) __extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) vrsraq_n_u8 (uint8x16_t __a, uint8x16_t __b, const int __c) { - return (uint8x16_t) __builtin_aarch64_ursra_nv16qi ((int8x16_t) __a, - (int8x16_t) __b, __c); + return __builtin_aarch64_ursra_nv16qi_uuus (__a, __b, __c); } __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vrsraq_n_u16 (uint16x8_t __a, uint16x8_t __b, const int __c) { - return (uint16x8_t) __builtin_aarch64_ursra_nv8hi ((int16x8_t) __a, - (int16x8_t) __b, __c); + return __builtin_aarch64_ursra_nv8hi_uuus (__a, __b, __c); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vrsraq_n_u32 (uint32x4_t __a, uint32x4_t __b, const int __c) { - return (uint32x4_t) __builtin_aarch64_ursra_nv4si ((int32x4_t) __a, - (int32x4_t) __b, __c); + return __builtin_aarch64_ursra_nv4si_uuus (__a, __b, __c); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vrsraq_n_u64 (uint64x2_t __a, uint64x2_t __b, const int __c) { - return (uint64x2_t) __builtin_aarch64_ursra_nv2di ((int64x2_t) __a, - (int64x2_t) __b, __c); + return __builtin_aarch64_ursra_nv2di_uuus (__a, __b, __c); } __extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) @@ -22880,7 +21737,7 @@ vrsrad_n_s64 (int64x1_t __a, int64x1_t __b, const int __c) __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vrsrad_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c) { - return (uint64x1_t) __builtin_aarch64_ursra_ndi (__a, __b, __c); + return __builtin_aarch64_ursra_ndi_uuus (__a, __b, __c); } #ifdef __ARM_FEATURE_CRYPTO @@ -23073,109 +21930,109 @@ vshld_n_u64 (uint64x1_t __a, const int __b) __extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) vshl_s8 (int8x8_t __a, int8x8_t __b) { - return (int8x8_t) __builtin_aarch64_sshlv8qi (__a, __b); + return __builtin_aarch64_sshlv8qi (__a, __b); } __extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) vshl_s16 (int16x4_t __a, int16x4_t __b) { - return (int16x4_t) __builtin_aarch64_sshlv4hi (__a, __b); + return __builtin_aarch64_sshlv4hi (__a, __b); } __extension__ static __inline int32x2_t __attribute__ ((__always_inline__)) vshl_s32 (int32x2_t __a, int32x2_t __b) { - return (int32x2_t) __builtin_aarch64_sshlv2si (__a, __b); + return __builtin_aarch64_sshlv2si (__a, __b); } __extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) vshl_s64 (int64x1_t __a, int64x1_t __b) { - return (int64x1_t) __builtin_aarch64_sshldi (__a, __b); + return __builtin_aarch64_sshldi (__a, __b); } __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vshl_u8 (uint8x8_t __a, int8x8_t __b) { - return (uint8x8_t) __builtin_aarch64_ushlv8qi ((int8x8_t) __a, __b); + return __builtin_aarch64_ushlv8qi_uus (__a, __b); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vshl_u16 (uint16x4_t __a, int16x4_t __b) { - return (uint16x4_t) __builtin_aarch64_ushlv4hi ((int16x4_t) __a, __b); + return __builtin_aarch64_ushlv4hi_uus (__a, __b); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vshl_u32 (uint32x2_t __a, int32x2_t __b) { - return (uint32x2_t) __builtin_aarch64_ushlv2si ((int32x2_t) __a, __b); + return __builtin_aarch64_ushlv2si_uus (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vshl_u64 (uint64x1_t __a, int64x1_t __b) { - return (uint64x1_t) __builtin_aarch64_ushldi ((int64x1_t) __a, __b); + return __builtin_aarch64_ushldi_uus (__a, __b); } __extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) vshlq_s8 (int8x16_t __a, int8x16_t __b) { - return (int8x16_t) __builtin_aarch64_sshlv16qi (__a, __b); + return __builtin_aarch64_sshlv16qi (__a, __b); } __extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) vshlq_s16 (int16x8_t __a, int16x8_t __b) { - return (int16x8_t) __builtin_aarch64_sshlv8hi (__a, __b); + return __builtin_aarch64_sshlv8hi (__a, __b); } __extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) vshlq_s32 (int32x4_t __a, int32x4_t __b) { - return (int32x4_t) __builtin_aarch64_sshlv4si (__a, __b); + return __builtin_aarch64_sshlv4si (__a, __b); } __extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) vshlq_s64 (int64x2_t __a, int64x2_t __b) { - return (int64x2_t) __builtin_aarch64_sshlv2di (__a, __b); + return __builtin_aarch64_sshlv2di (__a, __b); } __extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) vshlq_u8 (uint8x16_t __a, int8x16_t __b) { - return (uint8x16_t) __builtin_aarch64_ushlv16qi ((int8x16_t) __a, __b); + return __builtin_aarch64_ushlv16qi_uus (__a, __b); } __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vshlq_u16 (uint16x8_t __a, int16x8_t __b) { - return (uint16x8_t) __builtin_aarch64_ushlv8hi ((int16x8_t) __a, __b); + return __builtin_aarch64_ushlv8hi_uus (__a, __b); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vshlq_u32 (uint32x4_t __a, int32x4_t __b) { - return (uint32x4_t) __builtin_aarch64_ushlv4si ((int32x4_t) __a, __b); + return __builtin_aarch64_ushlv4si_uus (__a, __b); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vshlq_u64 (uint64x2_t __a, int64x2_t __b) { - return (uint64x2_t) __builtin_aarch64_ushlv2di ((int64x2_t) __a, __b); + return __builtin_aarch64_ushlv2di_uus (__a, __b); } __extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) vshld_s64 (int64x1_t __a, int64x1_t __b) { - return (int64x1_t) __builtin_aarch64_sshldi (__a, __b); + return __builtin_aarch64_sshldi (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vshld_u64 (uint64x1_t __a, uint64x1_t __b) { - return (uint64x1_t) __builtin_aarch64_ushldi (__a, __b); + return __builtin_aarch64_ushldi_uus (__a, __b); } __extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) @@ -23235,19 +22092,19 @@ vshll_n_s32 (int32x2_t __a, const int __b) __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vshll_n_u8 (uint8x8_t __a, const int __b) { - return (uint16x8_t) __builtin_aarch64_ushll_nv8qi ((int8x8_t) __a, __b); + return __builtin_aarch64_ushll_nv8qi_uus (__a, __b); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vshll_n_u16 (uint16x4_t __a, const int __b) { - return (uint32x4_t) __builtin_aarch64_ushll_nv4hi ((int16x4_t) __a, __b); + return __builtin_aarch64_ushll_nv4hi_uus (__a, __b); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vshll_n_u32 (uint32x2_t __a, const int __b) { - return (uint64x2_t) __builtin_aarch64_ushll_nv2si ((int32x2_t) __a, __b); + return __builtin_aarch64_ushll_nv2si_uus (__a, __b); } /* vshr */ @@ -23389,29 +22246,25 @@ vsli_n_s64 (int64x1_t __a, int64x1_t __b, const int __c) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vsli_n_u8 (uint8x8_t __a, uint8x8_t __b, const int __c) { - return (uint8x8_t) __builtin_aarch64_usli_nv8qi ((int8x8_t) __a, - (int8x8_t) __b, __c); + return __builtin_aarch64_usli_nv8qi_uuus (__a, __b, __c); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vsli_n_u16 (uint16x4_t __a, uint16x4_t __b, const int __c) { - return (uint16x4_t) __builtin_aarch64_usli_nv4hi ((int16x4_t) __a, - (int16x4_t) __b, __c); + return __builtin_aarch64_usli_nv4hi_uuus (__a, __b, __c); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vsli_n_u32 (uint32x2_t __a, uint32x2_t __b, const int __c) { - return (uint32x2_t) __builtin_aarch64_usli_nv2si ((int32x2_t) __a, - (int32x2_t) __b, __c); + return __builtin_aarch64_usli_nv2si_uuus (__a, __b, __c); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vsli_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c) { - return (uint64x1_t) __builtin_aarch64_usli_ndi ((int64x1_t) __a, - (int64x1_t) __b, __c); + return __builtin_aarch64_usli_ndi_uuus (__a, __b, __c); } __extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) @@ -23441,29 +22294,25 @@ vsliq_n_s64 (int64x2_t __a, int64x2_t __b, const int __c) __extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) vsliq_n_u8 (uint8x16_t __a, uint8x16_t __b, const int __c) { - return (uint8x16_t) __builtin_aarch64_usli_nv16qi ((int8x16_t) __a, - (int8x16_t) __b, __c); + return __builtin_aarch64_usli_nv16qi_uuus (__a, __b, __c); } __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vsliq_n_u16 (uint16x8_t __a, uint16x8_t __b, const int __c) { - return (uint16x8_t) __builtin_aarch64_usli_nv8hi ((int16x8_t) __a, - (int16x8_t) __b, __c); + return __builtin_aarch64_usli_nv8hi_uuus (__a, __b, __c); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vsliq_n_u32 (uint32x4_t __a, uint32x4_t __b, const int __c) { - return (uint32x4_t) __builtin_aarch64_usli_nv4si ((int32x4_t) __a, - (int32x4_t) __b, __c); + return __builtin_aarch64_usli_nv4si_uuus (__a, __b, __c); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vsliq_n_u64 (uint64x2_t __a, uint64x2_t __b, const int __c) { - return (uint64x2_t) __builtin_aarch64_usli_nv2di ((int64x2_t) __a, - (int64x2_t) __b, __c); + return __builtin_aarch64_usli_nv2di_uuus (__a, __b, __c); } __extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) @@ -23475,7 +22324,7 @@ vslid_n_s64 (int64x1_t __a, int64x1_t __b, const int __c) __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vslid_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c) { - return (uint64x1_t) __builtin_aarch64_usli_ndi (__a, __b, __c); + return __builtin_aarch64_usli_ndi_uuus (__a, __b, __c); } /* vsqadd */ @@ -23483,80 +22332,73 @@ vslid_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vsqadd_u8 (uint8x8_t __a, int8x8_t __b) { - return (uint8x8_t) __builtin_aarch64_usqaddv8qi ((int8x8_t) __a, - (int8x8_t) __b); + return __builtin_aarch64_usqaddv8qi_uus (__a, __b); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vsqadd_u16 (uint16x4_t __a, int16x4_t __b) { - return (uint16x4_t) __builtin_aarch64_usqaddv4hi ((int16x4_t) __a, - (int16x4_t) __b); + return __builtin_aarch64_usqaddv4hi_uus (__a, __b); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vsqadd_u32 (uint32x2_t __a, int32x2_t __b) { - return (uint32x2_t) __builtin_aarch64_usqaddv2si ((int32x2_t) __a, - (int32x2_t) __b); + return __builtin_aarch64_usqaddv2si_uus (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vsqadd_u64 (uint64x1_t __a, int64x1_t __b) { - return (uint64x1_t) __builtin_aarch64_usqadddi ((int64x1_t) __a, __b); + return __builtin_aarch64_usqadddi_uus (__a, __b); } __extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) vsqaddq_u8 (uint8x16_t __a, int8x16_t __b) { - return (uint8x16_t) __builtin_aarch64_usqaddv16qi ((int8x16_t) __a, - (int8x16_t) __b); + return __builtin_aarch64_usqaddv16qi_uus (__a, __b); } __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vsqaddq_u16 (uint16x8_t __a, int16x8_t __b) { - return (uint16x8_t) __builtin_aarch64_usqaddv8hi ((int16x8_t) __a, - (int16x8_t) __b); + return __builtin_aarch64_usqaddv8hi_uus (__a, __b); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vsqaddq_u32 (uint32x4_t __a, int32x4_t __b) { - return (uint32x4_t) __builtin_aarch64_usqaddv4si ((int32x4_t) __a, - (int32x4_t) __b); + return __builtin_aarch64_usqaddv4si_uus (__a, __b); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vsqaddq_u64 (uint64x2_t __a, int64x2_t __b) { - return (uint64x2_t) __builtin_aarch64_usqaddv2di ((int64x2_t) __a, - (int64x2_t) __b); + return __builtin_aarch64_usqaddv2di_uus (__a, __b); } __extension__ static __inline uint8x1_t __attribute__ ((__always_inline__)) vsqaddb_u8 (uint8x1_t __a, int8x1_t __b) { - return (uint8x1_t) __builtin_aarch64_usqaddqi ((int8x1_t) __a, __b); + return __builtin_aarch64_usqaddqi_uus (__a, __b); } __extension__ static __inline uint16x1_t __attribute__ ((__always_inline__)) vsqaddh_u16 (uint16x1_t __a, int16x1_t __b) { - return (uint16x1_t) __builtin_aarch64_usqaddhi ((int16x1_t) __a, __b); + return __builtin_aarch64_usqaddhi_uus (__a, __b); } __extension__ static __inline uint32x1_t __attribute__ ((__always_inline__)) vsqadds_u32 (uint32x1_t __a, int32x1_t __b) { - return (uint32x1_t) __builtin_aarch64_usqaddsi ((int32x1_t) __a, __b); + return __builtin_aarch64_usqaddsi_uus (__a, __b); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vsqaddd_u64 (uint64x1_t __a, int64x1_t __b) { - return (uint64x1_t) __builtin_aarch64_usqadddi ((int64x1_t) __a, __b); + return __builtin_aarch64_usqadddi_uus (__a, __b); } /* vsqrt */ @@ -23607,29 +22449,25 @@ vsra_n_s64 (int64x1_t __a, int64x1_t __b, const int __c) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vsra_n_u8 (uint8x8_t __a, uint8x8_t __b, const int __c) { - return (uint8x8_t) __builtin_aarch64_usra_nv8qi ((int8x8_t) __a, - (int8x8_t) __b, __c); + return __builtin_aarch64_usra_nv8qi_uuus (__a, __b, __c); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vsra_n_u16 (uint16x4_t __a, uint16x4_t __b, const int __c) { - return (uint16x4_t) __builtin_aarch64_usra_nv4hi ((int16x4_t) __a, - (int16x4_t) __b, __c); + return __builtin_aarch64_usra_nv4hi_uuus (__a, __b, __c); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vsra_n_u32 (uint32x2_t __a, uint32x2_t __b, const int __c) { - return (uint32x2_t) __builtin_aarch64_usra_nv2si ((int32x2_t) __a, - (int32x2_t) __b, __c); + return __builtin_aarch64_usra_nv2si_uuus (__a, __b, __c); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vsra_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c) { - return (uint64x1_t) __builtin_aarch64_usra_ndi ((int64x1_t) __a, - (int64x1_t) __b, __c); + return __builtin_aarch64_usra_ndi_uuus (__a, __b, __c); } __extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) @@ -23659,29 +22497,25 @@ vsraq_n_s64 (int64x2_t __a, int64x2_t __b, const int __c) __extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) vsraq_n_u8 (uint8x16_t __a, uint8x16_t __b, const int __c) { - return (uint8x16_t) __builtin_aarch64_usra_nv16qi ((int8x16_t) __a, - (int8x16_t) __b, __c); + return __builtin_aarch64_usra_nv16qi_uuus (__a, __b, __c); } __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vsraq_n_u16 (uint16x8_t __a, uint16x8_t __b, const int __c) { - return (uint16x8_t) __builtin_aarch64_usra_nv8hi ((int16x8_t) __a, - (int16x8_t) __b, __c); + return __builtin_aarch64_usra_nv8hi_uuus (__a, __b, __c); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vsraq_n_u32 (uint32x4_t __a, uint32x4_t __b, const int __c) { - return (uint32x4_t) __builtin_aarch64_usra_nv4si ((int32x4_t) __a, - (int32x4_t) __b, __c); + return __builtin_aarch64_usra_nv4si_uuus (__a, __b, __c); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vsraq_n_u64 (uint64x2_t __a, uint64x2_t __b, const int __c) { - return (uint64x2_t) __builtin_aarch64_usra_nv2di ((int64x2_t) __a, - (int64x2_t) __b, __c); + return __builtin_aarch64_usra_nv2di_uuus (__a, __b, __c); } __extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) @@ -23693,7 +22527,7 @@ vsrad_n_s64 (int64x1_t __a, int64x1_t __b, const int __c) __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vsrad_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c) { - return (uint64x1_t) __builtin_aarch64_usra_ndi (__a, __b, __c); + return __builtin_aarch64_usra_ndi_uuus (__a, __b, __c); } /* vsri */ @@ -23725,29 +22559,25 @@ vsri_n_s64 (int64x1_t __a, int64x1_t __b, const int __c) __extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) vsri_n_u8 (uint8x8_t __a, uint8x8_t __b, const int __c) { - return (uint8x8_t) __builtin_aarch64_usri_nv8qi ((int8x8_t) __a, - (int8x8_t) __b, __c); + return __builtin_aarch64_usri_nv8qi_uuus (__a, __b, __c); } __extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) vsri_n_u16 (uint16x4_t __a, uint16x4_t __b, const int __c) { - return (uint16x4_t) __builtin_aarch64_usri_nv4hi ((int16x4_t) __a, - (int16x4_t) __b, __c); + return __builtin_aarch64_usri_nv4hi_uuus (__a, __b, __c); } __extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) vsri_n_u32 (uint32x2_t __a, uint32x2_t __b, const int __c) { - return (uint32x2_t) __builtin_aarch64_usri_nv2si ((int32x2_t) __a, - (int32x2_t) __b, __c); + return __builtin_aarch64_usri_nv2si_uuus (__a, __b, __c); } __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vsri_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c) { - return (uint64x1_t) __builtin_aarch64_usri_ndi ((int64x1_t) __a, - (int64x1_t) __b, __c); + return __builtin_aarch64_usri_ndi_uuus (__a, __b, __c); } __extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) @@ -23777,29 +22607,25 @@ vsriq_n_s64 (int64x2_t __a, int64x2_t __b, const int __c) __extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) vsriq_n_u8 (uint8x16_t __a, uint8x16_t __b, const int __c) { - return (uint8x16_t) __builtin_aarch64_usri_nv16qi ((int8x16_t) __a, - (int8x16_t) __b, __c); + return __builtin_aarch64_usri_nv16qi_uuus (__a, __b, __c); } __extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) vsriq_n_u16 (uint16x8_t __a, uint16x8_t __b, const int __c) { - return (uint16x8_t) __builtin_aarch64_usri_nv8hi ((int16x8_t) __a, - (int16x8_t) __b, __c); + return __builtin_aarch64_usri_nv8hi_uuus (__a, __b, __c); } __extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) vsriq_n_u32 (uint32x4_t __a, uint32x4_t __b, const int __c) { - return (uint32x4_t) __builtin_aarch64_usri_nv4si ((int32x4_t) __a, - (int32x4_t) __b, __c); + return __builtin_aarch64_usri_nv4si_uuus (__a, __b, __c); } __extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) vsriq_n_u64 (uint64x2_t __a, uint64x2_t __b, const int __c) { - return (uint64x2_t) __builtin_aarch64_usri_nv2di ((int64x2_t) __a, - (int64x2_t) __b, __c); + return __builtin_aarch64_usri_nv2di_uuus (__a, __b, __c); } __extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) @@ -23811,7 +22637,7 @@ vsrid_n_s64 (int64x1_t __a, int64x1_t __b, const int __c) __extension__ static __inline uint64x1_t __attribute__ ((__always_inline__)) vsrid_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c) { - return (uint64x1_t) __builtin_aarch64_usri_ndi (__a, __b, __c); + return __builtin_aarch64_usri_ndi_uuus (__a, __b, __c); } /* vst1 */ @@ -24915,6 +23741,438 @@ vtbx3_p8 (poly8x8_t __r, poly8x8x3_t __tab, uint8x8_t __idx) /* vtrn */ +__extension__ static __inline float32x2_t __attribute__ ((__always_inline__)) +vtrn1_f32 (float32x2_t __a, float32x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x2_t) {3, 1}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {0, 2}); +#endif +} + +__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) +vtrn1_p8 (poly8x8_t __a, poly8x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint8x8_t) {9, 1, 11, 3, 13, 5, 15, 7}); +#else + return __builtin_shuffle (__a, __b, (uint8x8_t) {0, 8, 2, 10, 4, 12, 6, 14}); +#endif +} + +__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__)) +vtrn1_p16 (poly16x4_t __a, poly16x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x4_t) {5, 1, 7, 3}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {0, 4, 2, 6}); +#endif +} + +__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) +vtrn1_s8 (int8x8_t __a, int8x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint8x8_t) {9, 1, 11, 3, 13, 5, 15, 7}); +#else + return __builtin_shuffle (__a, __b, (uint8x8_t) {0, 8, 2, 10, 4, 12, 6, 14}); +#endif +} + +__extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) +vtrn1_s16 (int16x4_t __a, int16x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x4_t) {5, 1, 7, 3}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {0, 4, 2, 6}); +#endif +} + +__extension__ static __inline int32x2_t __attribute__ ((__always_inline__)) +vtrn1_s32 (int32x2_t __a, int32x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x2_t) {3, 1}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {0, 2}); +#endif +} + +__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) +vtrn1_u8 (uint8x8_t __a, uint8x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint8x8_t) {9, 1, 11, 3, 13, 5, 15, 7}); +#else + return __builtin_shuffle (__a, __b, (uint8x8_t) {0, 8, 2, 10, 4, 12, 6, 14}); +#endif +} + +__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) +vtrn1_u16 (uint16x4_t __a, uint16x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x4_t) {5, 1, 7, 3}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {0, 4, 2, 6}); +#endif +} + +__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) +vtrn1_u32 (uint32x2_t __a, uint32x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x2_t) {3, 1}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {0, 2}); +#endif +} + +__extension__ static __inline float32x4_t __attribute__ ((__always_inline__)) +vtrn1q_f32 (float32x4_t __a, float32x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x4_t) {5, 1, 7, 3}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {0, 4, 2, 6}); +#endif +} + +__extension__ static __inline float64x2_t __attribute__ ((__always_inline__)) +vtrn1q_f64 (float64x2_t __a, float64x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint64x2_t) {3, 1}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {0, 2}); +#endif +} + +__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) +vtrn1q_p8 (poly8x16_t __a, poly8x16_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, + (uint8x16_t) {17, 1, 19, 3, 21, 5, 23, 7, 25, 9, 27, 11, 29, 13, 31, 15}); +#else + return __builtin_shuffle (__a, __b, + (uint8x16_t) {0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30}); +#endif +} + +__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__)) +vtrn1q_p16 (poly16x8_t __a, poly16x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x8_t) {9, 1, 11, 3, 13, 5, 15, 7}); +#else + return __builtin_shuffle (__a, __b, (uint16x8_t) {0, 8, 2, 10, 4, 12, 6, 14}); +#endif +} + +__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) +vtrn1q_s8 (int8x16_t __a, int8x16_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, + (uint8x16_t) {17, 1, 19, 3, 21, 5, 23, 7, 25, 9, 27, 11, 29, 13, 31, 15}); +#else + return __builtin_shuffle (__a, __b, + (uint8x16_t) {0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30}); +#endif +} + +__extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) +vtrn1q_s16 (int16x8_t __a, int16x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x8_t) {9, 1, 11, 3, 13, 5, 15, 7}); +#else + return __builtin_shuffle (__a, __b, (uint16x8_t) {0, 8, 2, 10, 4, 12, 6, 14}); +#endif +} + +__extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) +vtrn1q_s32 (int32x4_t __a, int32x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x4_t) {5, 1, 7, 3}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {0, 4, 2, 6}); +#endif +} + +__extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) +vtrn1q_s64 (int64x2_t __a, int64x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint64x2_t) {3, 1}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {0, 2}); +#endif +} + +__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) +vtrn1q_u8 (uint8x16_t __a, uint8x16_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, + (uint8x16_t) {17, 1, 19, 3, 21, 5, 23, 7, 25, 9, 27, 11, 29, 13, 31, 15}); +#else + return __builtin_shuffle (__a, __b, + (uint8x16_t) {0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30}); +#endif +} + +__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) +vtrn1q_u16 (uint16x8_t __a, uint16x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x8_t) {9, 1, 11, 3, 13, 5, 15, 7}); +#else + return __builtin_shuffle (__a, __b, (uint16x8_t) {0, 8, 2, 10, 4, 12, 6, 14}); +#endif +} + +__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) +vtrn1q_u32 (uint32x4_t __a, uint32x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x4_t) {5, 1, 7, 3}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {0, 4, 2, 6}); +#endif +} + +__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) +vtrn1q_u64 (uint64x2_t __a, uint64x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint64x2_t) {3, 1}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {0, 2}); +#endif +} + +__extension__ static __inline float32x2_t __attribute__ ((__always_inline__)) +vtrn2_f32 (float32x2_t __a, float32x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x2_t) {2, 0}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {1, 3}); +#endif +} + +__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) +vtrn2_p8 (poly8x8_t __a, poly8x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint8x8_t) {8, 0, 10, 2, 12, 4, 14, 6}); +#else + return __builtin_shuffle (__a, __b, (uint8x8_t) {1, 9, 3, 11, 5, 13, 7, 15}); +#endif +} + +__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__)) +vtrn2_p16 (poly16x4_t __a, poly16x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x4_t) {4, 0, 6, 2}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {1, 5, 3, 7}); +#endif +} + +__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) +vtrn2_s8 (int8x8_t __a, int8x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint8x8_t) {8, 0, 10, 2, 12, 4, 14, 6}); +#else + return __builtin_shuffle (__a, __b, (uint8x8_t) {1, 9, 3, 11, 5, 13, 7, 15}); +#endif +} + +__extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) +vtrn2_s16 (int16x4_t __a, int16x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x4_t) {4, 0, 6, 2}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {1, 5, 3, 7}); +#endif +} + +__extension__ static __inline int32x2_t __attribute__ ((__always_inline__)) +vtrn2_s32 (int32x2_t __a, int32x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x2_t) {2, 0}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {1, 3}); +#endif +} + +__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) +vtrn2_u8 (uint8x8_t __a, uint8x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint8x8_t) {8, 0, 10, 2, 12, 4, 14, 6}); +#else + return __builtin_shuffle (__a, __b, (uint8x8_t) {1, 9, 3, 11, 5, 13, 7, 15}); +#endif +} + +__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) +vtrn2_u16 (uint16x4_t __a, uint16x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x4_t) {4, 0, 6, 2}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {1, 5, 3, 7}); +#endif +} + +__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) +vtrn2_u32 (uint32x2_t __a, uint32x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x2_t) {2, 0}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {1, 3}); +#endif +} + +__extension__ static __inline float32x4_t __attribute__ ((__always_inline__)) +vtrn2q_f32 (float32x4_t __a, float32x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x4_t) {4, 0, 6, 2}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {1, 5, 3, 7}); +#endif +} + +__extension__ static __inline float64x2_t __attribute__ ((__always_inline__)) +vtrn2q_f64 (float64x2_t __a, float64x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint64x2_t) {2, 0}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {1, 3}); +#endif +} + +__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) +vtrn2q_p8 (poly8x16_t __a, poly8x16_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, + (uint8x16_t) {16, 0, 18, 2, 20, 4, 22, 6, 24, 8, 26, 10, 28, 12, 30, 14}); +#else + return __builtin_shuffle (__a, __b, + (uint8x16_t) {1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31}); +#endif +} + +__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__)) +vtrn2q_p16 (poly16x8_t __a, poly16x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x8_t) {8, 0, 10, 2, 12, 4, 14, 6}); +#else + return __builtin_shuffle (__a, __b, (uint16x8_t) {1, 9, 3, 11, 5, 13, 7, 15}); +#endif +} + +__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) +vtrn2q_s8 (int8x16_t __a, int8x16_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, + (uint8x16_t) {16, 0, 18, 2, 20, 4, 22, 6, 24, 8, 26, 10, 28, 12, 30, 14}); +#else + return __builtin_shuffle (__a, __b, + (uint8x16_t) {1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31}); +#endif +} + +__extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) +vtrn2q_s16 (int16x8_t __a, int16x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x8_t) {8, 0, 10, 2, 12, 4, 14, 6}); +#else + return __builtin_shuffle (__a, __b, (uint16x8_t) {1, 9, 3, 11, 5, 13, 7, 15}); +#endif +} + +__extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) +vtrn2q_s32 (int32x4_t __a, int32x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x4_t) {4, 0, 6, 2}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {1, 5, 3, 7}); +#endif +} + +__extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) +vtrn2q_s64 (int64x2_t __a, int64x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint64x2_t) {2, 0}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {1, 3}); +#endif +} + +__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) +vtrn2q_u8 (uint8x16_t __a, uint8x16_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, + (uint8x16_t) {16, 0, 18, 2, 20, 4, 22, 6, 24, 8, 26, 10, 28, 12, 30, 14}); +#else + return __builtin_shuffle (__a, __b, + (uint8x16_t) {1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31}); +#endif +} + +__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) +vtrn2q_u16 (uint16x8_t __a, uint16x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x8_t) {8, 0, 10, 2, 12, 4, 14, 6}); +#else + return __builtin_shuffle (__a, __b, (uint16x8_t) {1, 9, 3, 11, 5, 13, 7, 15}); +#endif +} + +__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) +vtrn2q_u32 (uint32x4_t __a, uint32x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x4_t) {4, 0, 6, 2}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {1, 5, 3, 7}); +#endif +} + +__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) +vtrn2q_u64 (uint64x2_t __a, uint64x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint64x2_t) {2, 0}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {1, 3}); +#endif +} + __extension__ static __inline float32x2x2_t __attribute__ ((__always_inline__)) vtrn_f32 (float32x2_t a, float32x2_t b) { @@ -25145,73 +24403,73 @@ vtstd_u64 (uint64x1_t __a, uint64x1_t __b) __extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) vuqadd_s8 (int8x8_t __a, uint8x8_t __b) { - return (int8x8_t) __builtin_aarch64_suqaddv8qi (__a, (int8x8_t) __b); + return __builtin_aarch64_suqaddv8qi_ssu (__a, __b); } __extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) vuqadd_s16 (int16x4_t __a, uint16x4_t __b) { - return (int16x4_t) __builtin_aarch64_suqaddv4hi (__a, (int16x4_t) __b); + return __builtin_aarch64_suqaddv4hi_ssu (__a, __b); } __extension__ static __inline int32x2_t __attribute__ ((__always_inline__)) vuqadd_s32 (int32x2_t __a, uint32x2_t __b) { - return (int32x2_t) __builtin_aarch64_suqaddv2si (__a, (int32x2_t) __b); + return __builtin_aarch64_suqaddv2si_ssu (__a, __b); } __extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) vuqadd_s64 (int64x1_t __a, uint64x1_t __b) { - return (int64x1_t) __builtin_aarch64_suqadddi (__a, (int64x1_t) __b); + return __builtin_aarch64_suqadddi_ssu (__a, __b); } __extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) vuqaddq_s8 (int8x16_t __a, uint8x16_t __b) { - return (int8x16_t) __builtin_aarch64_suqaddv16qi (__a, (int8x16_t) __b); + return __builtin_aarch64_suqaddv16qi_ssu (__a, __b); } __extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) vuqaddq_s16 (int16x8_t __a, uint16x8_t __b) { - return (int16x8_t) __builtin_aarch64_suqaddv8hi (__a, (int16x8_t) __b); + return __builtin_aarch64_suqaddv8hi_ssu (__a, __b); } __extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) vuqaddq_s32 (int32x4_t __a, uint32x4_t __b) { - return (int32x4_t) __builtin_aarch64_suqaddv4si (__a, (int32x4_t) __b); + return __builtin_aarch64_suqaddv4si_ssu (__a, __b); } __extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) vuqaddq_s64 (int64x2_t __a, uint64x2_t __b) { - return (int64x2_t) __builtin_aarch64_suqaddv2di (__a, (int64x2_t) __b); + return __builtin_aarch64_suqaddv2di_ssu (__a, __b); } __extension__ static __inline int8x1_t __attribute__ ((__always_inline__)) vuqaddb_s8 (int8x1_t __a, uint8x1_t __b) { - return (int8x1_t) __builtin_aarch64_suqaddqi (__a, (int8x1_t) __b); + return __builtin_aarch64_suqaddqi_ssu (__a, __b); } __extension__ static __inline int16x1_t __attribute__ ((__always_inline__)) vuqaddh_s16 (int16x1_t __a, uint16x1_t __b) { - return (int16x1_t) __builtin_aarch64_suqaddhi (__a, (int16x1_t) __b); + return __builtin_aarch64_suqaddhi_ssu (__a, __b); } __extension__ static __inline int32x1_t __attribute__ ((__always_inline__)) vuqadds_s32 (int32x1_t __a, uint32x1_t __b) { - return (int32x1_t) __builtin_aarch64_suqaddsi (__a, (int32x1_t) __b); + return __builtin_aarch64_suqaddsi_ssu (__a, __b); } __extension__ static __inline int64x1_t __attribute__ ((__always_inline__)) vuqaddd_s64 (int64x1_t __a, uint64x1_t __b) { - return (int64x1_t) __builtin_aarch64_suqadddi (__a, (int64x1_t) __b); + return __builtin_aarch64_suqadddi_ssu (__a, __b); } #define __DEFINTERLEAVE(op, rettype, intype, funcsuffix, Q) \ @@ -25245,6 +24503,438 @@ vuqaddd_s64 (int64x1_t __a, uint64x1_t __b) /* vuzp */ +__extension__ static __inline float32x2_t __attribute__ ((__always_inline__)) +vuzp1_f32 (float32x2_t __a, float32x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x2_t) {3, 1}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {0, 2}); +#endif +} + +__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) +vuzp1_p8 (poly8x8_t __a, poly8x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint8x8_t) {9, 11, 13, 15, 1, 3, 5, 7}); +#else + return __builtin_shuffle (__a, __b, (uint8x8_t) {0, 2, 4, 6, 8, 10, 12, 14}); +#endif +} + +__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__)) +vuzp1_p16 (poly16x4_t __a, poly16x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x4_t) {5, 7, 1, 3}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {0, 2, 4, 6}); +#endif +} + +__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) +vuzp1_s8 (int8x8_t __a, int8x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint8x8_t) {9, 11, 13, 15, 1, 3, 5, 7}); +#else + return __builtin_shuffle (__a, __b, (uint8x8_t) {0, 2, 4, 6, 8, 10, 12, 14}); +#endif +} + +__extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) +vuzp1_s16 (int16x4_t __a, int16x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x4_t) {5, 7, 1, 3}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {0, 2, 4, 6}); +#endif +} + +__extension__ static __inline int32x2_t __attribute__ ((__always_inline__)) +vuzp1_s32 (int32x2_t __a, int32x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x2_t) {3, 1}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {0, 2}); +#endif +} + +__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) +vuzp1_u8 (uint8x8_t __a, uint8x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint8x8_t) {9, 11, 13, 15, 1, 3, 5, 7}); +#else + return __builtin_shuffle (__a, __b, (uint8x8_t) {0, 2, 4, 6, 8, 10, 12, 14}); +#endif +} + +__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) +vuzp1_u16 (uint16x4_t __a, uint16x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x4_t) {5, 7, 1, 3}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {0, 2, 4, 6}); +#endif +} + +__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) +vuzp1_u32 (uint32x2_t __a, uint32x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x2_t) {3, 1}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {0, 2}); +#endif +} + +__extension__ static __inline float32x4_t __attribute__ ((__always_inline__)) +vuzp1q_f32 (float32x4_t __a, float32x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x4_t) {5, 7, 1, 3}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {0, 2, 4, 6}); +#endif +} + +__extension__ static __inline float64x2_t __attribute__ ((__always_inline__)) +vuzp1q_f64 (float64x2_t __a, float64x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint64x2_t) {3, 1}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {0, 2}); +#endif +} + +__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) +vuzp1q_p8 (poly8x16_t __a, poly8x16_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint8x16_t) + {17, 19, 21, 23, 25, 27, 29, 31, 1, 3, 5, 7, 9, 11, 13, 15}); +#else + return __builtin_shuffle (__a, __b, (uint8x16_t) + {0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30}); +#endif +} + +__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__)) +vuzp1q_p16 (poly16x8_t __a, poly16x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x8_t) {9, 11, 13, 15, 1, 3, 5, 7}); +#else + return __builtin_shuffle (__a, __b, (uint16x8_t) {0, 2, 4, 6, 8, 10, 12, 14}); +#endif +} + +__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) +vuzp1q_s8 (int8x16_t __a, int8x16_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, + (uint8x16_t) {17, 19, 21, 23, 25, 27, 29, 31, 1, 3, 5, 7, 9, 11, 13, 15}); +#else + return __builtin_shuffle (__a, __b, + (uint8x16_t) {0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30}); +#endif +} + +__extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) +vuzp1q_s16 (int16x8_t __a, int16x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x8_t) {9, 11, 13, 15, 1, 3, 5, 7}); +#else + return __builtin_shuffle (__a, __b, (uint16x8_t) {0, 2, 4, 6, 8, 10, 12, 14}); +#endif +} + +__extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) +vuzp1q_s32 (int32x4_t __a, int32x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x4_t) {5, 7, 1, 3}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {0, 2, 4, 6}); +#endif +} + +__extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) +vuzp1q_s64 (int64x2_t __a, int64x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint64x2_t) {3, 1}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {0, 2}); +#endif +} + +__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) +vuzp1q_u8 (uint8x16_t __a, uint8x16_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, + (uint8x16_t) {17, 19, 21, 23, 25, 27, 29, 31, 1, 3, 5, 7, 9, 11, 13, 15}); +#else + return __builtin_shuffle (__a, __b, + (uint8x16_t) {0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30}); +#endif +} + +__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) +vuzp1q_u16 (uint16x8_t __a, uint16x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x8_t) {9, 11, 13, 15, 1, 3, 5, 7}); +#else + return __builtin_shuffle (__a, __b, (uint16x8_t) {0, 2, 4, 6, 8, 10, 12, 14}); +#endif +} + +__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) +vuzp1q_u32 (uint32x4_t __a, uint32x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x4_t) {5, 7, 1, 3}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {0, 2, 4, 6}); +#endif +} + +__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) +vuzp1q_u64 (uint64x2_t __a, uint64x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint64x2_t) {3, 1}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {0, 2}); +#endif +} + +__extension__ static __inline float32x2_t __attribute__ ((__always_inline__)) +vuzp2_f32 (float32x2_t __a, float32x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x2_t) {2, 0}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {1, 3}); +#endif +} + +__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__)) +vuzp2_p8 (poly8x8_t __a, poly8x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint8x8_t) {8, 10, 12, 14, 0, 2, 4, 6}); +#else + return __builtin_shuffle (__a, __b, (uint8x8_t) {1, 3, 5, 7, 9, 11, 13, 15}); +#endif +} + +__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__)) +vuzp2_p16 (poly16x4_t __a, poly16x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x4_t) {4, 6, 0, 2}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {1, 3, 5, 7}); +#endif +} + +__extension__ static __inline int8x8_t __attribute__ ((__always_inline__)) +vuzp2_s8 (int8x8_t __a, int8x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint8x8_t) {8, 10, 12, 14, 0, 2, 4, 6}); +#else + return __builtin_shuffle (__a, __b, (uint8x8_t) {1, 3, 5, 7, 9, 11, 13, 15}); +#endif +} + +__extension__ static __inline int16x4_t __attribute__ ((__always_inline__)) +vuzp2_s16 (int16x4_t __a, int16x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x4_t) {4, 6, 0, 2}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {1, 3, 5, 7}); +#endif +} + +__extension__ static __inline int32x2_t __attribute__ ((__always_inline__)) +vuzp2_s32 (int32x2_t __a, int32x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x2_t) {2, 0}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {1, 3}); +#endif +} + +__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__)) +vuzp2_u8 (uint8x8_t __a, uint8x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint8x8_t) {8, 10, 12, 14, 0, 2, 4, 6}); +#else + return __builtin_shuffle (__a, __b, (uint8x8_t) {1, 3, 5, 7, 9, 11, 13, 15}); +#endif +} + +__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__)) +vuzp2_u16 (uint16x4_t __a, uint16x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x4_t) {4, 6, 0, 2}); +#else + return __builtin_shuffle (__a, __b, (uint16x4_t) {1, 3, 5, 7}); +#endif +} + +__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__)) +vuzp2_u32 (uint32x2_t __a, uint32x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x2_t) {2, 0}); +#else + return __builtin_shuffle (__a, __b, (uint32x2_t) {1, 3}); +#endif +} + +__extension__ static __inline float32x4_t __attribute__ ((__always_inline__)) +vuzp2q_f32 (float32x4_t __a, float32x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x4_t) {4, 6, 0, 2}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {1, 3, 5, 7}); +#endif +} + +__extension__ static __inline float64x2_t __attribute__ ((__always_inline__)) +vuzp2q_f64 (float64x2_t __a, float64x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint64x2_t) {2, 0}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {1, 3}); +#endif +} + +__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__)) +vuzp2q_p8 (poly8x16_t __a, poly8x16_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, + (uint8x16_t) {16, 18, 20, 22, 24, 26, 28, 30, 0, 2, 4, 6, 8, 10, 12, 14}); +#else + return __builtin_shuffle (__a, __b, + (uint8x16_t) {1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31}); +#endif +} + +__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__)) +vuzp2q_p16 (poly16x8_t __a, poly16x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x8_t) {8, 10, 12, 14, 0, 2, 4, 6}); +#else + return __builtin_shuffle (__a, __b, (uint16x8_t) {1, 3, 5, 7, 9, 11, 13, 15}); +#endif +} + +__extension__ static __inline int8x16_t __attribute__ ((__always_inline__)) +vuzp2q_s8 (int8x16_t __a, int8x16_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, + (uint8x16_t) {16, 18, 20, 22, 24, 26, 28, 30, 0, 2, 4, 6, 8, 10, 12, 14}); +#else + return __builtin_shuffle (__a, __b, + (uint8x16_t) {1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31}); +#endif +} + +__extension__ static __inline int16x8_t __attribute__ ((__always_inline__)) +vuzp2q_s16 (int16x8_t __a, int16x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x8_t) {8, 10, 12, 14, 0, 2, 4, 6}); +#else + return __builtin_shuffle (__a, __b, (uint16x8_t) {1, 3, 5, 7, 9, 11, 13, 15}); +#endif +} + +__extension__ static __inline int32x4_t __attribute__ ((__always_inline__)) +vuzp2q_s32 (int32x4_t __a, int32x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x4_t) {4, 6, 0, 2}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {1, 3, 5, 7}); +#endif +} + +__extension__ static __inline int64x2_t __attribute__ ((__always_inline__)) +vuzp2q_s64 (int64x2_t __a, int64x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint64x2_t) {2, 0}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {1, 3}); +#endif +} + +__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__)) +vuzp2q_u8 (uint8x16_t __a, uint8x16_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint8x16_t) + {16, 18, 20, 22, 24, 26, 28, 30, 0, 2, 4, 6, 8, 10, 12, 14}); +#else + return __builtin_shuffle (__a, __b, (uint8x16_t) + {1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31}); +#endif +} + +__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__)) +vuzp2q_u16 (uint16x8_t __a, uint16x8_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint16x8_t) {8, 10, 12, 14, 0, 2, 4, 6}); +#else + return __builtin_shuffle (__a, __b, (uint16x8_t) {1, 3, 5, 7, 9, 11, 13, 15}); +#endif +} + +__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__)) +vuzp2q_u32 (uint32x4_t __a, uint32x4_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint32x4_t) {4, 6, 0, 2}); +#else + return __builtin_shuffle (__a, __b, (uint32x4_t) {1, 3, 5, 7}); +#endif +} + +__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__)) +vuzp2q_u64 (uint64x2_t __a, uint64x2_t __b) +{ +#ifdef __AARCH64EB__ + return __builtin_shuffle (__a, __b, (uint64x2_t) {2, 0}); +#else + return __builtin_shuffle (__a, __b, (uint64x2_t) {1, 3}); +#endif +} + __INTERLEAVE_LIST (uzp) /* vzip */ diff --git a/lib/gcc/aarch64-linux-android/4.9.1/libgcc.a b/lib/gcc/aarch64-linux-android/4.9.1/libgcc.a Binary files differindex c56bcd0..a02e0f2 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/libgcc.a +++ b/lib/gcc/aarch64-linux-android/4.9.1/libgcc.a diff --git a/lib/gcc/aarch64-linux-android/4.9.1/libgcov.a b/lib/gcc/aarch64-linux-android/4.9.1/libgcov.a Binary files differindex 1cde3fe..40e11aa 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/libgcov.a +++ b/lib/gcc/aarch64-linux-android/4.9.1/libgcov.a diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/gtype.state b/lib/gcc/aarch64-linux-android/4.9.1/plugin/gtype.state index 3923d7d..2806ccf 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/gtype.state +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/gtype.state @@ -4,8 +4,8 @@ ;;; Don't parse this file wihout knowing GCC gengtype internals. ;;; This file should be parsed by the same gengtype which wrote it. -(!version "4.9.1 20140529 (prerelease)") -(!srcdir "/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc") +(!version "4.9.1 20140710 (prerelease)") +(!srcdir "/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc") (!languages 9 ada c cp fortran go java lto objc objcp) (!fileslist 190 (!srcfile 511 "../libcpp/include/line-map.h") @@ -227,7 +227,7 @@ (!type union 7 (!type already_seen 5) gc_pointed_to "tree_node" - (!srcfileloc "tree-core.h" 1626) + (!srcfileloc "tree-core.h" 1631) (!fields 38 (!pair "base" (!type struct 8 nil gc_used "tree_base" @@ -310,11 +310,11 @@ (!srcfileloc "tree-core.h" 729) nil ) (!pair "u" - (!type union 12 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/tree-core.h:731" + (!type union 12 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/tree-core.h:731" (!srcfileloc "tree-core.h" 765) (!fields 3 (!pair "bits" - (!type struct 13 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/tree-core.h:734" + (!type struct 13 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/tree-core.h:734" (!srcfileloc "tree-core.h" 757) (!fields 16 (!pair "lang_flag_0" @@ -402,7 +402,7 @@ ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1584) + (!srcfileloc "tree-core.h" 1589) (!options (!option tag string "TS_BASE") ) @@ -421,7 +421,7 @@ nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1585) + (!srcfileloc "tree-core.h" 1590) (!options (!option tag string "TS_TYPED") ) @@ -440,7 +440,7 @@ nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1586) + (!srcfileloc "tree-core.h" 1591) (!options (!option tag string "TS_COMMON") ) @@ -459,7 +459,7 @@ nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1587) + (!srcfileloc "tree-core.h" 1592) (!options (!option tag string "TS_INT_CST") ) @@ -516,7 +516,7 @@ nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1588) + (!srcfileloc "tree-core.h" 1593) (!options (!option tag string "TS_REAL_CST") ) @@ -551,7 +551,7 @@ nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1589) + (!srcfileloc "tree-core.h" 1594) (!options (!option tag string "TS_FIXED_CST") ) @@ -575,7 +575,7 @@ ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1590) + (!srcfileloc "tree-core.h" 1595) (!options (!option tag string "TS_VECTOR") ) @@ -608,7 +608,7 @@ nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1591) + (!srcfileloc "tree-core.h" 1596) (!options (!option tag string "TS_STRING") ) @@ -631,7 +631,7 @@ nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1592) + (!srcfileloc "tree-core.h" 1597) (!options (!option tag string "TS_COMPLEX") ) @@ -674,181 +674,181 @@ nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1593) + (!srcfileloc "tree-core.h" 1598) (!options (!option tag string "TS_IDENTIFIER") ) ) (!pair "decl_minimal" (!type struct 37 nil gc_used "tree_decl_minimal" - (!srcfileloc "tree-core.h" 1314) + (!srcfileloc "tree-core.h" 1319) (!fields 5 (!pair "common" (!type already_seen 15) - (!srcfileloc "tree-core.h" 1309) + (!srcfileloc "tree-core.h" 1314) nil ) (!pair "locus" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1310) + (!srcfileloc "tree-core.h" 1315) nil ) (!pair "uid" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1311) + (!srcfileloc "tree-core.h" 1316) nil ) (!pair "name" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1312) + (!srcfileloc "tree-core.h" 1317) nil ) (!pair "context" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1313) + (!srcfileloc "tree-core.h" 1318) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1594) + (!srcfileloc "tree-core.h" 1599) (!options (!option tag string "TS_DECL_MINIMAL") ) ) (!pair "decl_common" (!type struct 38 nil gc_used "tree_decl_common" - (!srcfileloc "tree-core.h" 1380) + (!srcfileloc "tree-core.h" 1385) (!fields 35 (!pair "common" (!type already_seen 37) - (!srcfileloc "tree-core.h" 1317) + (!srcfileloc "tree-core.h" 1322) nil ) (!pair "size" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1318) + (!srcfileloc "tree-core.h" 1323) nil ) (!pair "mode" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1320) + (!srcfileloc "tree-core.h" 1325) nil ) (!pair "nonlocal_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1322) + (!srcfileloc "tree-core.h" 1327) nil ) (!pair "virtual_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1323) + (!srcfileloc "tree-core.h" 1328) nil ) (!pair "ignored_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1324) + (!srcfileloc "tree-core.h" 1329) nil ) (!pair "abstract_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1325) + (!srcfileloc "tree-core.h" 1330) nil ) (!pair "artificial_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1326) + (!srcfileloc "tree-core.h" 1331) nil ) (!pair "preserve_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1327) + (!srcfileloc "tree-core.h" 1332) nil ) (!pair "debug_expr_is_from" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1328) + (!srcfileloc "tree-core.h" 1333) nil ) (!pair "lang_flag_0" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1330) + (!srcfileloc "tree-core.h" 1335) nil ) (!pair "lang_flag_1" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1331) + (!srcfileloc "tree-core.h" 1336) nil ) (!pair "lang_flag_2" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1332) + (!srcfileloc "tree-core.h" 1337) nil ) (!pair "lang_flag_3" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1333) + (!srcfileloc "tree-core.h" 1338) nil ) (!pair "lang_flag_4" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1334) + (!srcfileloc "tree-core.h" 1339) nil ) (!pair "lang_flag_5" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1335) + (!srcfileloc "tree-core.h" 1340) nil ) (!pair "lang_flag_6" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1336) + (!srcfileloc "tree-core.h" 1341) nil ) (!pair "lang_flag_7" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1337) + (!srcfileloc "tree-core.h" 1342) nil ) (!pair "lang_flag_8" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1338) + (!srcfileloc "tree-core.h" 1343) nil ) (!pair "decl_flag_0" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1341) + (!srcfileloc "tree-core.h" 1346) nil ) (!pair "decl_flag_1" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1345) + (!srcfileloc "tree-core.h" 1350) nil ) (!pair "decl_flag_2" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1349) + (!srcfileloc "tree-core.h" 1354) nil ) (!pair "decl_flag_3" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1351) + (!srcfileloc "tree-core.h" 1356) nil ) (!pair "gimple_reg_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1354) + (!srcfileloc "tree-core.h" 1359) nil ) (!pair "decl_by_reference_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1356) + (!srcfileloc "tree-core.h" 1361) nil ) (!pair "decl_read_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1358) + (!srcfileloc "tree-core.h" 1363) nil ) (!pair "decl_nonshareable_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1360) + (!srcfileloc "tree-core.h" 1365) nil ) (!pair "off_align" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1363) + (!srcfileloc "tree-core.h" 1368) nil ) (!pair "align" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1368) + (!srcfileloc "tree-core.h" 1373) nil ) (!pair "pt_uid" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1371) + (!srcfileloc "tree-core.h" 1376) nil ) (!pair "size_unit" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1373) + (!srcfileloc "tree-core.h" 1378) nil ) (!pair "initial" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1374) + (!srcfileloc "tree-core.h" 1379) nil ) (!pair "attributes" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1375) + (!srcfileloc "tree-core.h" 1380) nil ) (!pair "abstract_origin" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1376) + (!srcfileloc "tree-core.h" 1381) nil ) (!pair "lang_specific" (!type pointer 39 nil gc_used @@ -933,15 +933,15 @@ (!type user_struct 46 (!type already_seen 45) gc_pointed_to "vec<tree,va_gc>" - (!srcfileloc "tree-core.h" 1228) + (!srcfileloc "tree-core.h" 1233) (!fields 2 (!pair "va_gc" (!type already_seen 3) - (!srcfileloc "tree-core.h" 1228) + (!srcfileloc "tree-core.h" 1233) nil ) (!pair "tree" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1228) + (!srcfileloc "tree-core.h" 1233) nil ) ) ) @@ -1166,7 +1166,7 @@ ) nil 32 (!type lang_struct 56 nil gc_used "lang_decl_u" - (!srcfileloc "cp/cp-tree.h" 2069) + (!srcfileloc "cp/cp-tree.h" 2071) (!fields 0 ) (!options (!option desc string "%h.base.selector") @@ -1176,223 +1176,223 @@ (!type already_seen 43) (!type union 57 nil gc_used "lang_decl_u" - (!srcfileloc "cp/cp-tree.h" 2069) + (!srcfileloc "cp/cp-tree.h" 2071) (!fields 5 (!pair "base" (!type struct 58 nil gc_used "lang_decl_base" - (!srcfileloc "cp/cp-tree.h" 1948) + (!srcfileloc "cp/cp-tree.h" 1950) (!fields 12 (!pair "selector" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1935) + (!srcfileloc "cp/cp-tree.h" 1937) nil ) (!pair "language" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1936) + (!srcfileloc "cp/cp-tree.h" 1938) nil ) (!pair "use_template" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1937) + (!srcfileloc "cp/cp-tree.h" 1939) nil ) (!pair "not_really_extern" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1938) + (!srcfileloc "cp/cp-tree.h" 1940) nil ) (!pair "initialized_in_class" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1939) + (!srcfileloc "cp/cp-tree.h" 1941) nil ) (!pair "repo_available_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1940) + (!srcfileloc "cp/cp-tree.h" 1942) nil ) (!pair "threadprivate_or_deleted_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1941) + (!srcfileloc "cp/cp-tree.h" 1943) nil ) (!pair "anticipated_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1942) + (!srcfileloc "cp/cp-tree.h" 1944) nil ) (!pair "friend_attr" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1943) + (!srcfileloc "cp/cp-tree.h" 1945) nil ) (!pair "template_conv_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1944) + (!srcfileloc "cp/cp-tree.h" 1946) nil ) (!pair "odr_used" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1945) + (!srcfileloc "cp/cp-tree.h" 1947) nil ) (!pair "u2sel" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1946) + (!srcfileloc "cp/cp-tree.h" 1948) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 2064) + (!srcfileloc "cp/cp-tree.h" 2066) (!options (!option default string "") ) ) (!pair "min" (!type struct 59 nil gc_used "lang_decl_min" - (!srcfileloc "cp/cp-tree.h" 1980) + (!srcfileloc "cp/cp-tree.h" 1982) (!fields 3 (!pair "base" (!type already_seen 58) - (!srcfileloc "cp/cp-tree.h" 1962) + (!srcfileloc "cp/cp-tree.h" 1964) nil ) (!pair "template_info" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1969) + (!srcfileloc "cp/cp-tree.h" 1971) nil ) (!pair "u2" (!type union 60 nil gc_used "lang_decl_u2" - (!srcfileloc "cp/cp-tree.h" 1979) + (!srcfileloc "cp/cp-tree.h" 1981) (!fields 2 (!pair "access" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1975) + (!srcfileloc "cp/cp-tree.h" 1977) (!options (!option tag string "0") ) ) (!pair "discriminator" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1978) + (!srcfileloc "cp/cp-tree.h" 1980) (!options (!option tag string "1") ) ) ) nil 260 nil ) - (!srcfileloc "cp/cp-tree.h" 1979) + (!srcfileloc "cp/cp-tree.h" 1981) (!options (!option desc string "%0.u.base.u2sel") ) ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 2065) + (!srcfileloc "cp/cp-tree.h" 2067) (!options (!option tag string "0") ) ) (!pair "fn" (!type struct 61 nil gc_used "lang_decl_fn" - (!srcfileloc "cp/cp-tree.h" 2041) + (!srcfileloc "cp/cp-tree.h" 2043) (!fields 22 (!pair "min" (!type already_seen 59) - (!srcfileloc "cp/cp-tree.h" 1985) + (!srcfileloc "cp/cp-tree.h" 1987) nil ) (!pair "operator_code" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1989) + (!srcfileloc "cp/cp-tree.h" 1991) nil ) (!pair "global_ctor_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1991) + (!srcfileloc "cp/cp-tree.h" 1993) nil ) (!pair "global_dtor_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1992) + (!srcfileloc "cp/cp-tree.h" 1994) nil ) (!pair "constructor_attr" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1993) + (!srcfileloc "cp/cp-tree.h" 1995) nil ) (!pair "destructor_attr" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1994) + (!srcfileloc "cp/cp-tree.h" 1996) nil ) (!pair "assignment_operator_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1995) + (!srcfileloc "cp/cp-tree.h" 1997) nil ) (!pair "static_function" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1996) + (!srcfileloc "cp/cp-tree.h" 1998) nil ) (!pair "pure_virtual" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1997) + (!srcfileloc "cp/cp-tree.h" 1999) nil ) (!pair "defaulted_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1998) + (!srcfileloc "cp/cp-tree.h" 2000) nil ) (!pair "has_in_charge_parm_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 2000) + (!srcfileloc "cp/cp-tree.h" 2002) nil ) (!pair "has_vtt_parm_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 2001) + (!srcfileloc "cp/cp-tree.h" 2003) nil ) (!pair "pending_inline_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 2002) + (!srcfileloc "cp/cp-tree.h" 2004) nil ) (!pair "nonconverting" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 2003) + (!srcfileloc "cp/cp-tree.h" 2005) nil ) (!pair "thunk_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 2004) + (!srcfileloc "cp/cp-tree.h" 2006) nil ) (!pair "this_thunk_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 2005) + (!srcfileloc "cp/cp-tree.h" 2007) nil ) (!pair "hidden_friend_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 2006) + (!srcfileloc "cp/cp-tree.h" 2008) nil ) (!pair "omp_declare_reduction_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 2007) + (!srcfileloc "cp/cp-tree.h" 2009) nil ) (!pair "befriending_classes" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 2013) + (!srcfileloc "cp/cp-tree.h" 2015) nil ) (!pair "context" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 2021) + (!srcfileloc "cp/cp-tree.h" 2023) nil ) (!pair "u5" (!type union 62 nil gc_used "lang_decl_u5" - (!srcfileloc "cp/cp-tree.h" 2032) + (!srcfileloc "cp/cp-tree.h" 2034) (!fields 2 (!pair "cloned_function" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 2027) + (!srcfileloc "cp/cp-tree.h" 2029) (!options (!option tag string "0") ) ) (!pair "fixed_offset" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 2031) + (!srcfileloc "cp/cp-tree.h" 2033) (!options (!option tag string "1") ) ) ) nil 260 nil ) - (!srcfileloc "cp/cp-tree.h" 2032) + (!srcfileloc "cp/cp-tree.h" 2034) (!options (!option desc string "%1.thunk_p") ) ) (!pair "u" (!type union 63 nil gc_used "lang_decl_u3" - (!srcfileloc "cp/cp-tree.h" 2039) + (!srcfileloc "cp/cp-tree.h" 2041) (!fields 2 (!pair "pending_inline_info" (!type pointer 64 nil gc_used @@ -1468,23 +1468,23 @@ nil ) (!pair "deferred_access_check" (!type struct 73 nil gc_used "deferred_access_check" - (!srcfileloc "cp/cp-tree.h" 5686) + (!srcfileloc "cp/cp-tree.h" 5697) (!fields 4 (!pair "binfo" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 5679) + (!srcfileloc "cp/cp-tree.h" 5690) nil ) (!pair "decl" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 5681) + (!srcfileloc "cp/cp-tree.h" 5692) nil ) (!pair "diag_decl" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 5683) + (!srcfileloc "cp/cp-tree.h" 5694) nil ) (!pair "loc" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 5685) + (!srcfileloc "cp/cp-tree.h" 5696) nil ) ) nil 260 nil nil ) @@ -1539,7 +1539,7 @@ ) nil 260 nil nil ) ) - (!srcfileloc "cp/cp-tree.h" 2036) + (!srcfileloc "cp/cp-tree.h" 2038) (!options (!option tag string "1") ) @@ -1597,7 +1597,7 @@ (!type lang_struct 81 (!type already_seen 80) gc_pointed_to "binding_level" - (!srcfileloc "fortran/f95-lang.c" 294) + (!srcfileloc "fortran/f95-lang.c" 318) (!fields 0 ) nil 40 (!homotypes 2 @@ -1659,19 +1659,19 @@ nil ) (!type struct 85 nil gc_pointed_to "binding_level" - (!srcfileloc "fortran/f95-lang.c" 294) + (!srcfileloc "fortran/f95-lang.c" 318) (!fields 3 (!pair "names" (!type already_seen 5) - (!srcfileloc "fortran/f95-lang.c" 288) + (!srcfileloc "fortran/f95-lang.c" 312) nil ) (!pair "blocks" (!type already_seen 5) - (!srcfileloc "fortran/f95-lang.c" 291) + (!srcfileloc "fortran/f95-lang.c" 315) nil ) (!pair "level_chain" (!type already_seen 80) - (!srcfileloc "fortran/f95-lang.c" 293) + (!srcfileloc "fortran/f95-lang.c" 317) nil ) ) nil 8 @@ -1688,7 +1688,7 @@ nil ) (!type struct 86 nil gc_pointed_to "language_function" - (!srcfileloc "cp/cp-tree.h" 1153) + (!srcfileloc "cp/cp-tree.h" 1155) (!fields 21 (!pair "base" (!type struct 87 nil gc_used "c_language_function" @@ -1720,67 +1720,67 @@ nil ) ) nil 390 nil nil ) - (!srcfileloc "cp/cp-tree.h" 1125) + (!srcfileloc "cp/cp-tree.h" 1127) nil ) (!pair "x_cdtor_label" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1127) + (!srcfileloc "cp/cp-tree.h" 1129) nil ) (!pair "x_current_class_ptr" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1128) + (!srcfileloc "cp/cp-tree.h" 1130) nil ) (!pair "x_current_class_ref" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1129) + (!srcfileloc "cp/cp-tree.h" 1131) nil ) (!pair "x_eh_spec_block" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1130) + (!srcfileloc "cp/cp-tree.h" 1132) nil ) (!pair "x_in_charge_parm" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1131) + (!srcfileloc "cp/cp-tree.h" 1133) nil ) (!pair "x_vtt_parm" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1132) + (!srcfileloc "cp/cp-tree.h" 1134) nil ) (!pair "x_return_value" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1133) + (!srcfileloc "cp/cp-tree.h" 1135) nil ) (!pair "x_auto_return_pattern" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1134) + (!srcfileloc "cp/cp-tree.h" 1136) nil ) (!pair "returns_value" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1136) + (!srcfileloc "cp/cp-tree.h" 1138) nil ) (!pair "returns_null" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1137) + (!srcfileloc "cp/cp-tree.h" 1139) nil ) (!pair "returns_abnormally" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1138) + (!srcfileloc "cp/cp-tree.h" 1140) nil ) (!pair "infinite_loop" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1139) + (!srcfileloc "cp/cp-tree.h" 1141) nil ) (!pair "x_in_function_try_handler" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1140) + (!srcfileloc "cp/cp-tree.h" 1142) nil ) (!pair "x_in_base_initializer" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1141) + (!srcfileloc "cp/cp-tree.h" 1143) nil ) (!pair "can_throw" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1144) + (!srcfileloc "cp/cp-tree.h" 1146) nil ) (!pair "x_named_labels" (!type pointer 90 nil gc_used @@ -2039,7 +2039,7 @@ ) nil nil nil nil nil nil nil nil nil nil ) ) - (!srcfileloc "cp/cp-tree.h" 1146) + (!srcfileloc "cp/cp-tree.h" 1148) (!options (!option param_is type (!type already_seen 93) @@ -2048,15 +2048,15 @@ ) (!pair "bindings" (!type already_seen 94) - (!srcfileloc "cp/cp-tree.h" 1147) + (!srcfileloc "cp/cp-tree.h" 1149) nil ) (!pair "x_local_names" (!type already_seen 45) - (!srcfileloc "cp/cp-tree.h" 1148) + (!srcfileloc "cp/cp-tree.h" 1150) nil ) (!pair "infinite_loops" (!type already_seen 45) - (!srcfileloc "cp/cp-tree.h" 1151) + (!srcfileloc "cp/cp-tree.h" 1153) nil ) (!pair "extern_decl_map" (!type pointer 106 nil gc_used @@ -2069,22 +2069,22 @@ (!type struct 109 (!type already_seen 108) gc_pointed_to "cxx_int_tree_map" - (!srcfileloc "cp/cp-tree.h" 1117) + (!srcfileloc "cp/cp-tree.h" 1119) (!fields 2 (!pair "uid" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1115) + (!srcfileloc "cp/cp-tree.h" 1117) nil ) (!pair "to" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1116) + (!srcfileloc "cp/cp-tree.h" 1118) nil ) ) nil 260 nil nil ) ) nil nil nil nil nil nil nil nil nil nil ) ) - (!srcfileloc "cp/cp-tree.h" 1152) + (!srcfileloc "cp/cp-tree.h" 1154) (!options (!option param_is type (!type already_seen 109) @@ -2314,63 +2314,63 @@ ) ) ) - (!srcfileloc "cp/cp-tree.h" 2038) + (!srcfileloc "cp/cp-tree.h" 2040) (!options (!option tag string "0") ) ) ) nil 260 nil ) - (!srcfileloc "cp/cp-tree.h" 2039) + (!srcfileloc "cp/cp-tree.h" 2041) (!options (!option desc string "%1.pending_inline_p") ) ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 2066) + (!srcfileloc "cp/cp-tree.h" 2068) (!options (!option tag string "1") ) ) (!pair "ns" (!type struct 128 nil gc_used "lang_decl_ns" - (!srcfileloc "cp/cp-tree.h" 2048) + (!srcfileloc "cp/cp-tree.h" 2050) (!fields 2 (!pair "base" (!type already_seen 58) - (!srcfileloc "cp/cp-tree.h" 2046) + (!srcfileloc "cp/cp-tree.h" 2048) nil ) (!pair "level" (!type already_seen 94) - (!srcfileloc "cp/cp-tree.h" 2047) + (!srcfileloc "cp/cp-tree.h" 2049) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 2067) + (!srcfileloc "cp/cp-tree.h" 2069) (!options (!option tag string "2") ) ) (!pair "parm" (!type struct 129 nil gc_used "lang_decl_parm" - (!srcfileloc "cp/cp-tree.h" 2056) + (!srcfileloc "cp/cp-tree.h" 2058) (!fields 3 (!pair "base" (!type already_seen 58) - (!srcfileloc "cp/cp-tree.h" 2053) + (!srcfileloc "cp/cp-tree.h" 2055) nil ) (!pair "level" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 2054) + (!srcfileloc "cp/cp-tree.h" 2056) nil ) (!pair "index" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 2055) + (!srcfileloc "cp/cp-tree.h" 2057) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 2068) + (!srcfileloc "cp/cp-tree.h" 2070) (!options (!option tag string "3") ) @@ -2414,31 +2414,39 @@ nil ) (!type struct 131 nil gc_pointed_to "lang_decl" - (!srcfileloc "fortran/trans.h" 827) - (!fields 6 + (!srcfileloc "fortran/trans.h" 834) + (!fields 8 (!pair "saved_descriptor" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 817) + (!srcfileloc "fortran/trans.h" 822) nil ) (!pair "stringlen" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 822) + (!srcfileloc "fortran/trans.h" 827) nil ) (!pair "addr" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 823) + (!srcfileloc "fortran/trans.h" 828) nil ) (!pair "span" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 824) + (!srcfileloc "fortran/trans.h" 829) nil ) (!pair "token" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 826) + (!srcfileloc "fortran/trans.h" 831) nil ) (!pair "caf_offset" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 826) + (!srcfileloc "fortran/trans.h" 831) + nil ) + (!pair "scalar_allocatable" + (!type already_seen 9) + (!srcfileloc "fortran/trans.h" 832) + nil ) + (!pair "scalar_pointer" + (!type already_seen 9) + (!srcfileloc "fortran/trans.h" 833) nil ) ) (!options @@ -2449,11 +2457,11 @@ nil ) (!type struct 132 nil gc_pointed_to "lang_decl" - (!srcfileloc "cp/cp-tree.h" 2070) + (!srcfileloc "cp/cp-tree.h" 2072) (!fields 1 (!pair "u" (!type already_seen 56) - (!srcfileloc "cp/cp-tree.h" 2069) + (!srcfileloc "cp/cp-tree.h" 2071) nil ) ) (!options @@ -2495,22 +2503,22 @@ ) ) ) - (!srcfileloc "tree-core.h" 1379) + (!srcfileloc "tree-core.h" 1384) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1595) + (!srcfileloc "tree-core.h" 1600) (!options (!option tag string "TS_DECL_COMMON") ) ) (!pair "decl_with_rtl" (!type struct 135 nil gc_used "tree_decl_with_rtl" - (!srcfileloc "tree-core.h" 1385) + (!srcfileloc "tree-core.h" 1390) (!fields 2 (!pair "common" (!type already_seen 38) - (!srcfileloc "tree-core.h" 1383) + (!srcfileloc "tree-core.h" 1388) nil ) (!pair "rtl" (!type pointer 136 @@ -6604,7 +6612,7 @@ ) ) (!pair "x" - (!type struct 321 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/basic-block.h:187" + (!type struct 321 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/basic-block.h:187" (!srcfileloc "basic-block.h" 190) (!fields 2 (!pair "head_" @@ -7399,289 +7407,289 @@ ) 511 nil nil ) ) - (!srcfileloc "tree-core.h" 1384) + (!srcfileloc "tree-core.h" 1389) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1596) + (!srcfileloc "tree-core.h" 1601) (!options (!option tag string "TS_DECL_WRTL") ) ) (!pair "decl_non_common" (!type struct 339 nil gc_used "tree_decl_non_common" - (!srcfileloc "tree-core.h" 1468) + (!srcfileloc "tree-core.h" 1473) (!fields 5 (!pair "common" (!type struct 340 nil gc_used "tree_decl_with_vis" - (!srcfileloc "tree-core.h" 1452) + (!srcfileloc "tree-core.h" 1457) (!fields 22 (!pair "common" (!type already_seen 135) - (!srcfileloc "tree-core.h" 1417) + (!srcfileloc "tree-core.h" 1422) nil ) (!pair "assembler_name" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1418) + (!srcfileloc "tree-core.h" 1423) nil ) (!pair "section_name" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1419) + (!srcfileloc "tree-core.h" 1424) nil ) (!pair "comdat_group" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1420) + (!srcfileloc "tree-core.h" 1425) nil ) (!pair "defer_output" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1423) + (!srcfileloc "tree-core.h" 1428) nil ) (!pair "hard_register" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1424) + (!srcfileloc "tree-core.h" 1429) nil ) (!pair "common_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1425) + (!srcfileloc "tree-core.h" 1430) nil ) (!pair "in_text_section" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1426) + (!srcfileloc "tree-core.h" 1431) nil ) (!pair "in_constant_pool" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1427) + (!srcfileloc "tree-core.h" 1432) nil ) (!pair "dllimport_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1428) + (!srcfileloc "tree-core.h" 1433) nil ) (!pair "weak_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1430) + (!srcfileloc "tree-core.h" 1435) nil ) (!pair "implicit_section_name_p" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1432) + (!srcfileloc "tree-core.h" 1437) nil ) (!pair "seen_in_bind_expr" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1434) + (!srcfileloc "tree-core.h" 1439) nil ) (!pair "comdat_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1435) + (!srcfileloc "tree-core.h" 1440) nil ) (!pair "visibility" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1436) + (!srcfileloc "tree-core.h" 1441) nil ) (!pair "visibility_specified" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1437) + (!srcfileloc "tree-core.h" 1442) nil ) (!pair "tls_model" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1439) + (!srcfileloc "tree-core.h" 1444) nil ) (!pair "init_priority_p" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1442) + (!srcfileloc "tree-core.h" 1447) nil ) (!pair "shadowed_for_var_p" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1444) + (!srcfileloc "tree-core.h" 1449) nil ) (!pair "cxx_constructor" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1446) + (!srcfileloc "tree-core.h" 1451) nil ) (!pair "cxx_destructor" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1448) + (!srcfileloc "tree-core.h" 1453) nil ) (!pair "final" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1450) + (!srcfileloc "tree-core.h" 1455) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1459) + (!srcfileloc "tree-core.h" 1464) nil ) (!pair "saved_tree" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1461) + (!srcfileloc "tree-core.h" 1466) nil ) (!pair "arguments" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1463) + (!srcfileloc "tree-core.h" 1468) nil ) (!pair "result" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1465) + (!srcfileloc "tree-core.h" 1470) nil ) (!pair "vindex" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1467) + (!srcfileloc "tree-core.h" 1472) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1598) + (!srcfileloc "tree-core.h" 1603) (!options (!option tag string "TS_DECL_NON_COMMON") ) ) (!pair "parm_decl" (!type struct 341 nil gc_used "tree_parm_decl" - (!srcfileloc "tree-core.h" 1414) + (!srcfileloc "tree-core.h" 1419) (!fields 2 (!pair "common" (!type already_seen 135) - (!srcfileloc "tree-core.h" 1412) + (!srcfileloc "tree-core.h" 1417) nil ) (!pair "incoming_rtl" (!type already_seen 136) - (!srcfileloc "tree-core.h" 1413) + (!srcfileloc "tree-core.h" 1418) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1599) + (!srcfileloc "tree-core.h" 1604) (!options (!option tag string "TS_PARM_DECL") ) ) (!pair "decl_with_vis" (!type already_seen 340) - (!srcfileloc "tree-core.h" 1600) + (!srcfileloc "tree-core.h" 1605) (!options (!option tag string "TS_DECL_WITH_VIS") ) ) (!pair "var_decl" (!type struct 342 nil gc_used "tree_var_decl" - (!srcfileloc "tree-core.h" 1456) + (!srcfileloc "tree-core.h" 1461) (!fields 1 (!pair "common" (!type already_seen 340) - (!srcfileloc "tree-core.h" 1455) + (!srcfileloc "tree-core.h" 1460) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1601) + (!srcfileloc "tree-core.h" 1606) (!options (!option tag string "TS_VAR_DECL") ) ) (!pair "field_decl" (!type struct 343 nil gc_used "tree_field_decl" - (!srcfileloc "tree-core.h" 1395) + (!srcfileloc "tree-core.h" 1400) (!fields 6 (!pair "common" (!type already_seen 38) - (!srcfileloc "tree-core.h" 1388) + (!srcfileloc "tree-core.h" 1393) nil ) (!pair "offset" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1390) + (!srcfileloc "tree-core.h" 1395) nil ) (!pair "bit_field_type" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1391) + (!srcfileloc "tree-core.h" 1396) nil ) (!pair "qualifier" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1392) + (!srcfileloc "tree-core.h" 1397) nil ) (!pair "bit_offset" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1393) + (!srcfileloc "tree-core.h" 1398) nil ) (!pair "fcontext" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1394) + (!srcfileloc "tree-core.h" 1399) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1602) + (!srcfileloc "tree-core.h" 1607) (!options (!option tag string "TS_FIELD_DECL") ) ) (!pair "label_decl" (!type struct 344 nil gc_used "tree_label_decl" - (!srcfileloc "tree-core.h" 1401) + (!srcfileloc "tree-core.h" 1406) (!fields 3 (!pair "common" (!type already_seen 135) - (!srcfileloc "tree-core.h" 1398) + (!srcfileloc "tree-core.h" 1403) nil ) (!pair "label_decl_uid" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1399) + (!srcfileloc "tree-core.h" 1404) nil ) (!pair "eh_landing_pad_nr" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1400) + (!srcfileloc "tree-core.h" 1405) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1603) + (!srcfileloc "tree-core.h" 1608) (!options (!option tag string "TS_LABEL_DECL") ) ) (!pair "result_decl" (!type struct 345 nil gc_used "tree_result_decl" - (!srcfileloc "tree-core.h" 1405) + (!srcfileloc "tree-core.h" 1410) (!fields 1 (!pair "common" (!type already_seen 135) - (!srcfileloc "tree-core.h" 1404) + (!srcfileloc "tree-core.h" 1409) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1604) + (!srcfileloc "tree-core.h" 1609) (!options (!option tag string "TS_RESULT_DECL") ) ) (!pair "const_decl" (!type struct 346 nil gc_used "tree_const_decl" - (!srcfileloc "tree-core.h" 1409) + (!srcfileloc "tree-core.h" 1414) (!fields 1 (!pair "common" (!type already_seen 38) - (!srcfileloc "tree-core.h" 1408) + (!srcfileloc "tree-core.h" 1413) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1605) + (!srcfileloc "tree-core.h" 1610) (!options (!option tag string "TS_CONST_DECL") ) ) (!pair "type_decl" (!type struct 347 nil gc_used "tree_type_decl" - (!srcfileloc "tree-core.h" 1529) + (!srcfileloc "tree-core.h" 1534) (!fields 1 (!pair "common" (!type already_seen 339) - (!srcfileloc "tree-core.h" 1527) + (!srcfileloc "tree-core.h" 1532) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1606) + (!srcfileloc "tree-core.h" 1611) (!options (!option tag string "TS_TYPE_DECL") ) ) (!pair "function_decl" (!type struct 348 nil gc_used "tree_function_decl" - (!srcfileloc "tree-core.h" 1516) + (!srcfileloc "tree-core.h" 1521) (!fields 26 (!pair "common" (!type already_seen 339) - (!srcfileloc "tree-core.h" 1476) + (!srcfileloc "tree-core.h" 1481) nil ) (!pair "f" (!type pointer 349 nil gc_used @@ -8400,41 +8408,41 @@ (!type struct 404 (!type already_seen 403) gc_maybe_pointed_to "machine_function" - (!srcfileloc "config/aarch64/aarch64.h" 534) + (!srcfileloc "config/aarch64/aarch64.h" 543) (!fields 2 (!pair "frame" (!type struct 405 nil gc_used "aarch64_frame" - (!srcfileloc "config/aarch64/aarch64.h" 524) + (!srcfileloc "config/aarch64/aarch64.h" 533) (!fields 5 (!pair "reg_offset" (!type array 406 nil gc_used "FIRST_PSEUDO_REGISTER" (!type already_seen 9) ) - (!srcfileloc "config/aarch64/aarch64.h" 516) + (!srcfileloc "config/aarch64/aarch64.h" 525) nil ) (!pair "saved_regs_size" (!type already_seen 9) - (!srcfileloc "config/aarch64/aarch64.h" 517) + (!srcfileloc "config/aarch64/aarch64.h" 526) nil ) (!pair "padding0" (!type already_seen 9) - (!srcfileloc "config/aarch64/aarch64.h" 520) + (!srcfileloc "config/aarch64/aarch64.h" 529) nil ) (!pair "hardfp_offset" (!type already_seen 9) - (!srcfileloc "config/aarch64/aarch64.h" 521) + (!srcfileloc "config/aarch64/aarch64.h" 530) nil ) (!pair "laid_out" (!type already_seen 9) - (!srcfileloc "config/aarch64/aarch64.h" 523) + (!srcfileloc "config/aarch64/aarch64.h" 532) nil ) ) nil 511 nil nil ) - (!srcfileloc "config/aarch64/aarch64.h" 528) + (!srcfileloc "config/aarch64/aarch64.h" 537) nil ) (!pair "saved_varargs_size" (!type already_seen 9) - (!srcfileloc "config/aarch64/aarch64.h" 532) + (!srcfileloc "config/aarch64/aarch64.h" 541) nil ) ) nil 511 nil nil ) @@ -8519,7 +8527,7 @@ (!srcfileloc "dwarf2out.h" 55) nil ) (!pair "dw_cfi_oprnd1" - (!type union 413 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/dwarf2out.h:47" + (!type union 413 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/dwarf2out.h:47" (!srcfileloc "dwarf2out.h" 52) (!fields 4 (!pair "dw_cfi_reg_num" @@ -9263,299 +9271,299 @@ ) nil 511 nil nil ) ) - (!srcfileloc "tree-core.h" 1478) + (!srcfileloc "tree-core.h" 1483) nil ) (!pair "personality" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1481) + (!srcfileloc "tree-core.h" 1486) nil ) (!pair "function_specific_target" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1484) + (!srcfileloc "tree-core.h" 1489) nil ) (!pair "function_specific_optimization" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1485) + (!srcfileloc "tree-core.h" 1490) nil ) (!pair "function_code" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1491) + (!srcfileloc "tree-core.h" 1496) nil ) (!pair "built_in_class" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1492) + (!srcfileloc "tree-core.h" 1497) nil ) (!pair "static_ctor_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1494) + (!srcfileloc "tree-core.h" 1499) nil ) (!pair "static_dtor_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1495) + (!srcfileloc "tree-core.h" 1500) nil ) (!pair "uninlinable" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1496) + (!srcfileloc "tree-core.h" 1501) nil ) (!pair "possibly_inlined" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1498) + (!srcfileloc "tree-core.h" 1503) nil ) (!pair "novops_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1499) + (!srcfileloc "tree-core.h" 1504) nil ) (!pair "returns_twice_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1500) + (!srcfileloc "tree-core.h" 1505) nil ) (!pair "malloc_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1501) + (!srcfileloc "tree-core.h" 1506) nil ) (!pair "operator_new_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1502) + (!srcfileloc "tree-core.h" 1507) nil ) (!pair "declared_inline_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1503) + (!srcfileloc "tree-core.h" 1508) nil ) (!pair "regdecl_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1504) + (!srcfileloc "tree-core.h" 1509) nil ) (!pair "no_inline_warning_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1505) + (!srcfileloc "tree-core.h" 1510) nil ) (!pair "no_instrument_function_entry_exit" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1507) + (!srcfileloc "tree-core.h" 1512) nil ) (!pair "no_limit_stack" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1508) + (!srcfileloc "tree-core.h" 1513) nil ) (!pair "disregard_inline_limits" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1509) + (!srcfileloc "tree-core.h" 1514) nil ) (!pair "pure_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1510) + (!srcfileloc "tree-core.h" 1515) nil ) (!pair "looping_const_or_pure_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1511) + (!srcfileloc "tree-core.h" 1516) nil ) (!pair "has_debug_args_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1512) + (!srcfileloc "tree-core.h" 1517) nil ) (!pair "tm_clone_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1513) + (!srcfileloc "tree-core.h" 1518) nil ) (!pair "versioned_function" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1514) + (!srcfileloc "tree-core.h" 1519) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1607) + (!srcfileloc "tree-core.h" 1612) (!options (!option tag string "TS_FUNCTION_DECL") ) ) (!pair "translation_unit_decl" (!type struct 443 nil gc_used "tree_translation_unit_decl" - (!srcfileloc "tree-core.h" 1524) + (!srcfileloc "tree-core.h" 1529) (!fields 2 (!pair "common" (!type already_seen 38) - (!srcfileloc "tree-core.h" 1519) + (!srcfileloc "tree-core.h" 1524) nil ) (!pair "language" (!type already_seen 36) - (!srcfileloc "tree-core.h" 1521) + (!srcfileloc "tree-core.h" 1526) (!options (!option skip string "") ) ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1609) + (!srcfileloc "tree-core.h" 1614) (!options (!option tag string "TS_TRANSLATION_UNIT_DECL") ) ) (!pair "type_common" (!type struct 444 nil gc_used "tree_type_common" - (!srcfileloc "tree-core.h" 1276) + (!srcfileloc "tree-core.h" 1281) (!fields 30 (!pair "common" (!type already_seen 15) - (!srcfileloc "tree-core.h" 1238) + (!srcfileloc "tree-core.h" 1243) nil ) (!pair "size" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1239) + (!srcfileloc "tree-core.h" 1244) nil ) (!pair "size_unit" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1240) + (!srcfileloc "tree-core.h" 1245) nil ) (!pair "attributes" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1241) + (!srcfileloc "tree-core.h" 1246) nil ) (!pair "uid" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1242) + (!srcfileloc "tree-core.h" 1247) nil ) (!pair "precision" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1244) + (!srcfileloc "tree-core.h" 1249) nil ) (!pair "no_force_blk_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1245) + (!srcfileloc "tree-core.h" 1250) nil ) (!pair "needs_constructing_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1246) + (!srcfileloc "tree-core.h" 1251) nil ) (!pair "transparent_aggr_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1247) + (!srcfileloc "tree-core.h" 1252) nil ) (!pair "restrict_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1248) + (!srcfileloc "tree-core.h" 1253) nil ) (!pair "contains_placeholder_bits" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1249) + (!srcfileloc "tree-core.h" 1254) nil ) (!pair "mode" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1251) + (!srcfileloc "tree-core.h" 1256) nil ) (!pair "string_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1253) + (!srcfileloc "tree-core.h" 1258) nil ) (!pair "lang_flag_0" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1254) + (!srcfileloc "tree-core.h" 1259) nil ) (!pair "lang_flag_1" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1255) + (!srcfileloc "tree-core.h" 1260) nil ) (!pair "lang_flag_2" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1256) + (!srcfileloc "tree-core.h" 1261) nil ) (!pair "lang_flag_3" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1257) + (!srcfileloc "tree-core.h" 1262) nil ) (!pair "lang_flag_4" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1258) + (!srcfileloc "tree-core.h" 1263) nil ) (!pair "lang_flag_5" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1259) + (!srcfileloc "tree-core.h" 1264) nil ) (!pair "lang_flag_6" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1260) + (!srcfileloc "tree-core.h" 1265) nil ) (!pair "align" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1262) + (!srcfileloc "tree-core.h" 1267) nil ) (!pair "alias_set" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1263) + (!srcfileloc "tree-core.h" 1268) nil ) (!pair "pointer_to" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1264) + (!srcfileloc "tree-core.h" 1269) nil ) (!pair "reference_to" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1265) + (!srcfileloc "tree-core.h" 1270) nil ) (!pair "symtab" (!type union 445 nil gc_used "tree_type_symtab" - (!srcfileloc "tree-core.h" 1270) + (!srcfileloc "tree-core.h" 1275) (!fields 3 (!pair "address" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1267) + (!srcfileloc "tree-core.h" 1272) (!options (!option tag string "TYPE_SYMTAB_IS_ADDRESS") ) ) (!pair "pointer" (!type already_seen 36) - (!srcfileloc "tree-core.h" 1268) + (!srcfileloc "tree-core.h" 1273) (!options (!option tag string "TYPE_SYMTAB_IS_POINTER") ) ) (!pair "die" (!type already_seen 426) - (!srcfileloc "tree-core.h" 1269) + (!srcfileloc "tree-core.h" 1274) (!options (!option tag string "TYPE_SYMTAB_IS_DIE") ) ) ) nil 511 nil ) - (!srcfileloc "tree-core.h" 1270) + (!srcfileloc "tree-core.h" 1275) (!options (!option desc string "debug_hooks->tree_type_symtab_field") ) ) (!pair "canonical" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1271) + (!srcfileloc "tree-core.h" 1276) nil ) (!pair "next_variant" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1272) + (!srcfileloc "tree-core.h" 1277) nil ) (!pair "main_variant" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1273) + (!srcfileloc "tree-core.h" 1278) nil ) (!pair "context" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1274) + (!srcfileloc "tree-core.h" 1279) nil ) (!pair "name" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1275) + (!srcfileloc "tree-core.h" 1280) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1610) + (!srcfileloc "tree-core.h" 1615) (!options (!option tag string "TS_TYPE_COMMON") ) ) (!pair "type_with_lang_specific" (!type struct 446 nil gc_used "tree_type_with_lang_specific" - (!srcfileloc "tree-core.h" 1282) + (!srcfileloc "tree-core.h" 1287) (!fields 2 (!pair "common" (!type already_seen 444) - (!srcfileloc "tree-core.h" 1279) + (!srcfileloc "tree-core.h" 1284) nil ) (!pair "lang_specific" (!type pointer 447 nil gc_used @@ -10011,75 +10019,75 @@ nil ) (!type struct 473 nil gc_pointed_to "lang_type" - (!srcfileloc "fortran/trans.h" 813) + (!srcfileloc "fortran/trans.h" 818) (!fields 15 (!pair "rank" (!type already_seen 9) - (!srcfileloc "fortran/trans.h" 799) + (!srcfileloc "fortran/trans.h" 804) nil ) (!pair "corank" (!type already_seen 9) - (!srcfileloc "fortran/trans.h" 799) + (!srcfileloc "fortran/trans.h" 804) nil ) (!pair "akind" (!type already_seen 9) - (!srcfileloc "fortran/trans.h" 800) + (!srcfileloc "fortran/trans.h" 805) nil ) (!pair "lbound" (!type array 474 nil gc_used "GFC_MAX_DIMENSIONS" (!type already_seen 5) ) - (!srcfileloc "fortran/trans.h" 801) + (!srcfileloc "fortran/trans.h" 806) nil ) (!pair "ubound" (!type array 475 nil gc_used "GFC_MAX_DIMENSIONS" (!type already_seen 5) ) - (!srcfileloc "fortran/trans.h" 802) + (!srcfileloc "fortran/trans.h" 807) nil ) (!pair "stride" (!type array 476 nil gc_used "GFC_MAX_DIMENSIONS" (!type already_seen 5) ) - (!srcfileloc "fortran/trans.h" 803) + (!srcfileloc "fortran/trans.h" 808) nil ) (!pair "size" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 804) + (!srcfileloc "fortran/trans.h" 809) nil ) (!pair "offset" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 805) + (!srcfileloc "fortran/trans.h" 810) nil ) (!pair "dtype" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 806) + (!srcfileloc "fortran/trans.h" 811) nil ) (!pair "dataptr_type" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 807) + (!srcfileloc "fortran/trans.h" 812) nil ) (!pair "span" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 808) + (!srcfileloc "fortran/trans.h" 813) nil ) (!pair "base_decl" (!type array 477 nil gc_used "2" (!type already_seen 5) ) - (!srcfileloc "fortran/trans.h" 809) + (!srcfileloc "fortran/trans.h" 814) nil ) (!pair "nonrestricted_type" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 810) + (!srcfileloc "fortran/trans.h" 815) nil ) (!pair "caf_token" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 811) + (!srcfileloc "fortran/trans.h" 816) nil ) (!pair "caf_offset" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 812) + (!srcfileloc "fortran/trans.h" 817) nil ) ) (!options @@ -10090,257 +10098,257 @@ nil ) (!type struct 478 nil gc_pointed_to "lang_type" - (!srcfileloc "cp/cp-tree.h" 1491) + (!srcfileloc "cp/cp-tree.h" 1493) (!fields 1 (!pair "u" (!type union 479 nil gc_used "lang_type_u" - (!srcfileloc "cp/cp-tree.h" 1490) + (!srcfileloc "cp/cp-tree.h" 1492) (!fields 3 (!pair "h" (!type struct 480 nil gc_used "lang_type_header" - (!srcfileloc "cp/cp-tree.h" 1375) + (!srcfileloc "cp/cp-tree.h" 1377) (!fields 8 (!pair "is_lang_type_class" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1365) + (!srcfileloc "cp/cp-tree.h" 1367) nil ) (!pair "has_type_conversion" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1367) + (!srcfileloc "cp/cp-tree.h" 1369) nil ) (!pair "has_copy_ctor" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1368) + (!srcfileloc "cp/cp-tree.h" 1370) nil ) (!pair "has_default_ctor" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1369) + (!srcfileloc "cp/cp-tree.h" 1371) nil ) (!pair "const_needs_init" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1370) + (!srcfileloc "cp/cp-tree.h" 1372) nil ) (!pair "ref_needs_init" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1371) + (!srcfileloc "cp/cp-tree.h" 1373) nil ) (!pair "has_const_copy_assign" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1372) + (!srcfileloc "cp/cp-tree.h" 1374) nil ) (!pair "spare" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1374) + (!srcfileloc "cp/cp-tree.h" 1376) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 1487) + (!srcfileloc "cp/cp-tree.h" 1489) (!options (!option skip string "") ) ) (!pair "c" (!type struct 481 nil gc_used "lang_type_class" - (!srcfileloc "cp/cp-tree.h" 1477) + (!srcfileloc "cp/cp-tree.h" 1479) (!fields 63 (!pair "h" (!type already_seen 480) - (!srcfileloc "cp/cp-tree.h" 1391) + (!srcfileloc "cp/cp-tree.h" 1393) nil ) (!pair "align" (!type already_seen 28) - (!srcfileloc "cp/cp-tree.h" 1393) + (!srcfileloc "cp/cp-tree.h" 1395) nil ) (!pair "has_mutable" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1395) + (!srcfileloc "cp/cp-tree.h" 1397) nil ) (!pair "com_interface" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1396) + (!srcfileloc "cp/cp-tree.h" 1398) nil ) (!pair "non_pod_class" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1397) + (!srcfileloc "cp/cp-tree.h" 1399) nil ) (!pair "nearly_empty_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1398) + (!srcfileloc "cp/cp-tree.h" 1400) nil ) (!pair "user_align" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1399) + (!srcfileloc "cp/cp-tree.h" 1401) nil ) (!pair "has_copy_assign" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1400) + (!srcfileloc "cp/cp-tree.h" 1402) nil ) (!pair "has_new" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1401) + (!srcfileloc "cp/cp-tree.h" 1403) nil ) (!pair "has_array_new" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1402) + (!srcfileloc "cp/cp-tree.h" 1404) nil ) (!pair "gets_delete" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1404) + (!srcfileloc "cp/cp-tree.h" 1406) nil ) (!pair "interface_only" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1405) + (!srcfileloc "cp/cp-tree.h" 1407) nil ) (!pair "interface_unknown" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1406) + (!srcfileloc "cp/cp-tree.h" 1408) nil ) (!pair "contains_empty_class_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1407) + (!srcfileloc "cp/cp-tree.h" 1409) nil ) (!pair "anon_aggr" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1408) + (!srcfileloc "cp/cp-tree.h" 1410) nil ) (!pair "non_zero_init" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1409) + (!srcfileloc "cp/cp-tree.h" 1411) nil ) (!pair "empty_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1410) + (!srcfileloc "cp/cp-tree.h" 1412) nil ) (!pair "vec_new_uses_cookie" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1412) + (!srcfileloc "cp/cp-tree.h" 1414) nil ) (!pair "declared_class" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1413) + (!srcfileloc "cp/cp-tree.h" 1415) nil ) (!pair "diamond_shaped" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1414) + (!srcfileloc "cp/cp-tree.h" 1416) nil ) (!pair "repeated_base" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1415) + (!srcfileloc "cp/cp-tree.h" 1417) nil ) (!pair "being_defined" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1416) + (!srcfileloc "cp/cp-tree.h" 1418) nil ) (!pair "java_interface" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1417) + (!srcfileloc "cp/cp-tree.h" 1419) nil ) (!pair "debug_requested" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1418) + (!srcfileloc "cp/cp-tree.h" 1420) nil ) (!pair "fields_readonly" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1419) + (!srcfileloc "cp/cp-tree.h" 1421) nil ) (!pair "use_template" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1421) + (!srcfileloc "cp/cp-tree.h" 1423) nil ) (!pair "ptrmemfunc_flag" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1422) + (!srcfileloc "cp/cp-tree.h" 1424) nil ) (!pair "was_anonymous" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1423) + (!srcfileloc "cp/cp-tree.h" 1425) nil ) (!pair "lazy_default_ctor" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1424) + (!srcfileloc "cp/cp-tree.h" 1426) nil ) (!pair "lazy_copy_ctor" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1425) + (!srcfileloc "cp/cp-tree.h" 1427) nil ) (!pair "lazy_copy_assign" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1426) + (!srcfileloc "cp/cp-tree.h" 1428) nil ) (!pair "lazy_destructor" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1427) + (!srcfileloc "cp/cp-tree.h" 1429) nil ) (!pair "has_const_copy_ctor" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1429) + (!srcfileloc "cp/cp-tree.h" 1431) nil ) (!pair "has_complex_copy_ctor" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1430) + (!srcfileloc "cp/cp-tree.h" 1432) nil ) (!pair "has_complex_copy_assign" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1431) + (!srcfileloc "cp/cp-tree.h" 1433) nil ) (!pair "non_aggregate" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1432) + (!srcfileloc "cp/cp-tree.h" 1434) nil ) (!pair "has_complex_dflt" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1433) + (!srcfileloc "cp/cp-tree.h" 1435) nil ) (!pair "has_list_ctor" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1434) + (!srcfileloc "cp/cp-tree.h" 1436) nil ) (!pair "non_std_layout" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1435) + (!srcfileloc "cp/cp-tree.h" 1437) nil ) (!pair "is_literal" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1436) + (!srcfileloc "cp/cp-tree.h" 1438) nil ) (!pair "lazy_move_ctor" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1438) + (!srcfileloc "cp/cp-tree.h" 1440) nil ) (!pair "lazy_move_assign" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1439) + (!srcfileloc "cp/cp-tree.h" 1441) nil ) (!pair "has_complex_move_ctor" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1440) + (!srcfileloc "cp/cp-tree.h" 1442) nil ) (!pair "has_complex_move_assign" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1441) + (!srcfileloc "cp/cp-tree.h" 1443) nil ) (!pair "has_constexpr_ctor" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1442) + (!srcfileloc "cp/cp-tree.h" 1444) nil ) (!pair "dummy" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1451) + (!srcfileloc "cp/cp-tree.h" 1453) nil ) (!pair "primary_base" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1453) + (!srcfileloc "cp/cp-tree.h" 1455) nil ) (!pair "vcall_indices" (!type pointer 482 nil gc_used (!type user_struct 483 (!type already_seen 482) gc_pointed_to "vec<tree_pair_s,va_gc>" - (!srcfileloc "cp/cp-tree.h" 1454) + (!srcfileloc "cp/cp-tree.h" 1456) (!fields 2 (!pair "va_gc" (!type already_seen 3) - (!srcfileloc "cp/cp-tree.h" 1454) + (!srcfileloc "cp/cp-tree.h" 1456) nil ) (!pair "tree_pair_s" (!type struct 484 @@ -10348,36 +10356,36 @@ (!type already_seen 484) ) gc_used "tree_pair_s" - (!srcfileloc "cp/cp-tree.h" 1358) + (!srcfileloc "cp/cp-tree.h" 1360) (!fields 2 (!pair "purpose" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1356) + (!srcfileloc "cp/cp-tree.h" 1358) nil ) (!pair "value" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1357) + (!srcfileloc "cp/cp-tree.h" 1359) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 1454) + (!srcfileloc "cp/cp-tree.h" 1456) nil ) ) ) ) - (!srcfileloc "cp/cp-tree.h" 1454) + (!srcfileloc "cp/cp-tree.h" 1456) nil ) (!pair "vtables" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1455) + (!srcfileloc "cp/cp-tree.h" 1457) nil ) (!pair "typeinfo_var" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1456) + (!srcfileloc "cp/cp-tree.h" 1458) nil ) (!pair "vbases" (!type already_seen 45) - (!srcfileloc "cp/cp-tree.h" 1457) + (!srcfileloc "cp/cp-tree.h" 1459) nil ) (!pair "nested_udts" (!type pointer 486 nil gc_used @@ -10428,46 +10436,46 @@ ) nil 260 nil nil ) ) - (!srcfileloc "cp/cp-tree.h" 1458) + (!srcfileloc "cp/cp-tree.h" 1460) nil ) (!pair "as_base" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1459) + (!srcfileloc "cp/cp-tree.h" 1461) nil ) (!pair "pure_virtuals" (!type already_seen 45) - (!srcfileloc "cp/cp-tree.h" 1460) + (!srcfileloc "cp/cp-tree.h" 1462) nil ) (!pair "friend_classes" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1461) + (!srcfileloc "cp/cp-tree.h" 1463) nil ) (!pair "methods" (!type already_seen 45) - (!srcfileloc "cp/cp-tree.h" 1462) + (!srcfileloc "cp/cp-tree.h" 1464) (!options (!option reorder string "resort_type_method_vec") ) ) (!pair "key_method" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1463) + (!srcfileloc "cp/cp-tree.h" 1465) nil ) (!pair "decl_list" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1464) + (!srcfileloc "cp/cp-tree.h" 1466) nil ) (!pair "template_info" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1465) + (!srcfileloc "cp/cp-tree.h" 1467) nil ) (!pair "befriending_classes" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1466) + (!srcfileloc "cp/cp-tree.h" 1468) nil ) (!pair "objc_info" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1470) + (!srcfileloc "cp/cp-tree.h" 1472) nil ) (!pair "sorted_fields" (!type pointer 491 nil gc_used @@ -10495,44 +10503,44 @@ ) 390 nil nil ) ) - (!srcfileloc "cp/cp-tree.h" 1474) + (!srcfileloc "cp/cp-tree.h" 1476) (!options (!option reorder string "resort_sorted_fields") ) ) (!pair "lambda_expr" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1476) + (!srcfileloc "cp/cp-tree.h" 1478) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 1488) + (!srcfileloc "cp/cp-tree.h" 1490) (!options (!option tag string "1") ) ) (!pair "ptrmem" (!type struct 494 nil gc_used "lang_type_ptrmem" - (!srcfileloc "cp/cp-tree.h" 1482) + (!srcfileloc "cp/cp-tree.h" 1484) (!fields 2 (!pair "h" (!type already_seen 480) - (!srcfileloc "cp/cp-tree.h" 1480) + (!srcfileloc "cp/cp-tree.h" 1482) nil ) (!pair "record" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1481) + (!srcfileloc "cp/cp-tree.h" 1483) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 1489) + (!srcfileloc "cp/cp-tree.h" 1491) (!options (!option tag string "0") ) ) ) nil 260 nil ) - (!srcfileloc "cp/cp-tree.h" 1490) + (!srcfileloc "cp/cp-tree.h" 1492) (!options (!option desc string "%h.h.is_lang_type_class") ) @@ -10592,42 +10600,42 @@ ) ) ) - (!srcfileloc "tree-core.h" 1281) + (!srcfileloc "tree-core.h" 1286) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1612) + (!srcfileloc "tree-core.h" 1617) (!options (!option tag string "TS_TYPE_WITH_LANG_SPECIFIC") ) ) (!pair "type_non_common" (!type struct 497 nil gc_used "tree_type_non_common" - (!srcfileloc "tree-core.h" 1290) + (!srcfileloc "tree-core.h" 1295) (!fields 5 (!pair "with_lang_specific" (!type already_seen 446) - (!srcfileloc "tree-core.h" 1285) + (!srcfileloc "tree-core.h" 1290) nil ) (!pair "values" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1286) + (!srcfileloc "tree-core.h" 1291) nil ) (!pair "minval" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1287) + (!srcfileloc "tree-core.h" 1292) nil ) (!pair "maxval" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1288) + (!srcfileloc "tree-core.h" 1293) nil ) (!pair "binfo" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1289) + (!srcfileloc "tree-core.h" 1294) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1614) + (!srcfileloc "tree-core.h" 1619) (!options (!option tag string "TS_TYPE_NON_COMMON") ) @@ -10650,7 +10658,7 @@ nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1615) + (!srcfileloc "tree-core.h" 1620) (!options (!option tag string "TS_LIST") ) @@ -10674,26 +10682,26 @@ ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1616) + (!srcfileloc "tree-core.h" 1621) (!options (!option tag string "TS_VEC") ) ) (!pair "exp" (!type struct 501 nil gc_used "tree_exp" - (!srcfileloc "tree-core.h" 1150) + (!srcfileloc "tree-core.h" 1155) (!fields 3 (!pair "typed" (!type already_seen 14) - (!srcfileloc "tree-core.h" 1145) + (!srcfileloc "tree-core.h" 1150) nil ) (!pair "locus" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1146) + (!srcfileloc "tree-core.h" 1151) nil ) (!pair "operands" (!type union 502 nil gc_used "tree_exp_subunion" - (!srcfileloc "tree-core.h" 1149) + (!srcfileloc "tree-core.h" 1154) (!fields 1 (!pair "" (!type array 503 nil gc_used "1" @@ -10711,7 +10719,7 @@ (!option dot string "") ) 511 nil ) - (!srcfileloc "tree-core.h" 1149) + (!srcfileloc "tree-core.h" 1154) (!options (!option desc string "TREE_CODE ((tree) &%0)") (!option special string "tree_exp") @@ -10719,30 +10727,30 @@ ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1617) + (!srcfileloc "tree-core.h" 1622) (!options (!option tag string "TS_EXP") ) ) (!pair "ssa_name" (!type struct 504 nil gc_used "tree_ssa_name" - (!srcfileloc "tree-core.h" 1186) + (!srcfileloc "tree-core.h" 1191) (!fields 5 (!pair "typed" (!type already_seen 14) - (!srcfileloc "tree-core.h" 1167) + (!srcfileloc "tree-core.h" 1172) nil ) (!pair "var" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1170) + (!srcfileloc "tree-core.h" 1175) nil ) (!pair "def_stmt" (!type already_seen 310) - (!srcfileloc "tree-core.h" 1173) + (!srcfileloc "tree-core.h" 1178) nil ) (!pair "info" (!type union 505 nil gc_used "ssa_name_info_type" - (!srcfileloc "tree-core.h" 1181) + (!srcfileloc "tree-core.h" 1186) (!fields 2 (!pair "ptr_info" (!type pointer 506 nil gc_used @@ -10766,7 +10774,7 @@ ) nil 511 nil nil ) ) - (!srcfileloc "tree-core.h" 1178) + (!srcfileloc "tree-core.h" 1183) (!options (!option tag string "0") ) @@ -10793,14 +10801,14 @@ ) nil 511 nil nil ) ) - (!srcfileloc "tree-core.h" 1180) + (!srcfileloc "tree-core.h" 1185) (!options (!option tag string "1") ) ) ) nil 511 nil ) - (!srcfileloc "tree-core.h" 1182) + (!srcfileloc "tree-core.h" 1187) (!options (!option desc string "%1.typed.type ?!POINTER_TYPE_P (TREE_TYPE ((tree)&%1)) : 2") ) @@ -10815,191 +10823,191 @@ (!fields 4 (!pair "prev" (!type already_seen 511) - (!srcfileloc "tree-core.h" 1155) + (!srcfileloc "tree-core.h" 1160) (!options (!option skip string "") ) ) (!pair "next" (!type already_seen 511) - (!srcfileloc "tree-core.h" 1156) + (!srcfileloc "tree-core.h" 1161) (!options (!option skip string "") ) ) (!pair "loc" - (!type union 512 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/tree-core.h:1162" - (!srcfileloc "tree-core.h" 1162) + (!type union 512 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/tree-core.h:1167" + (!srcfileloc "tree-core.h" 1167) (!fields 2 (!pair "stmt" (!type already_seen 310) - (!srcfileloc "tree-core.h" 1162) + (!srcfileloc "tree-core.h" 1167) nil ) (!pair "ssa_name" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1162) + (!srcfileloc "tree-core.h" 1167) nil ) ) nil 511 nil ) - (!srcfileloc "tree-core.h" 1162) + (!srcfileloc "tree-core.h" 1167) (!options (!option skip string "") ) ) (!pair "use" (!type already_seen 6) - (!srcfileloc "tree-core.h" 1163) + (!srcfileloc "tree-core.h" 1168) (!options (!option skip string "") ) ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1185) + (!srcfileloc "tree-core.h" 1190) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1618) + (!srcfileloc "tree-core.h" 1623) (!options (!option tag string "TS_SSA_NAME") ) ) (!pair "block" (!type struct 513 nil gc_used "tree_block" - (!srcfileloc "tree-core.h" 1235) + (!srcfileloc "tree-core.h" 1240) (!fields 12 (!pair "base" (!type already_seen 8) - (!srcfileloc "tree-core.h" 1219) + (!srcfileloc "tree-core.h" 1224) nil ) (!pair "chain" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1220) + (!srcfileloc "tree-core.h" 1225) nil ) (!pair "abstract_flag" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1222) + (!srcfileloc "tree-core.h" 1227) nil ) (!pair "block_num" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1223) + (!srcfileloc "tree-core.h" 1228) nil ) (!pair "locus" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1225) + (!srcfileloc "tree-core.h" 1230) nil ) (!pair "vars" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1227) + (!srcfileloc "tree-core.h" 1232) nil ) (!pair "nonlocalized_vars" (!type already_seen 45) - (!srcfileloc "tree-core.h" 1228) + (!srcfileloc "tree-core.h" 1233) nil ) (!pair "subblocks" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1230) + (!srcfileloc "tree-core.h" 1235) nil ) (!pair "supercontext" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1231) + (!srcfileloc "tree-core.h" 1236) nil ) (!pair "abstract_origin" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1232) + (!srcfileloc "tree-core.h" 1237) nil ) (!pair "fragment_origin" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1233) + (!srcfileloc "tree-core.h" 1238) nil ) (!pair "fragment_chain" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1234) + (!srcfileloc "tree-core.h" 1239) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1619) + (!srcfileloc "tree-core.h" 1624) (!options (!option tag string "TS_BLOCK") ) ) (!pair "binfo" (!type struct 514 nil gc_used "tree_binfo" - (!srcfileloc "tree-core.h" 1306) + (!srcfileloc "tree-core.h" 1311) (!fields 10 (!pair "common" (!type already_seen 15) - (!srcfileloc "tree-core.h" 1293) + (!srcfileloc "tree-core.h" 1298) nil ) (!pair "offset" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1295) + (!srcfileloc "tree-core.h" 1300) nil ) (!pair "vtable" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1296) + (!srcfileloc "tree-core.h" 1301) nil ) (!pair "virtuals" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1297) + (!srcfileloc "tree-core.h" 1302) nil ) (!pair "vptr_field" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1298) + (!srcfileloc "tree-core.h" 1303) nil ) (!pair "base_accesses" (!type already_seen 45) - (!srcfileloc "tree-core.h" 1299) + (!srcfileloc "tree-core.h" 1304) nil ) (!pair "inheritance" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1300) + (!srcfileloc "tree-core.h" 1305) nil ) (!pair "vtt_subvtt" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1302) + (!srcfileloc "tree-core.h" 1307) nil ) (!pair "vtt_vptr" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1303) + (!srcfileloc "tree-core.h" 1308) nil ) (!pair "base_binfos" (!type already_seen 46) - (!srcfileloc "tree-core.h" 1305) + (!srcfileloc "tree-core.h" 1310) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1620) + (!srcfileloc "tree-core.h" 1625) (!options (!option tag string "TS_BINFO") ) ) (!pair "stmt_list" (!type struct 515 nil gc_used "tree_statement_list" - (!srcfileloc "tree-core.h" 1543) + (!srcfileloc "tree-core.h" 1548) (!fields 3 (!pair "typed" (!type already_seen 14) - (!srcfileloc "tree-core.h" 1540) + (!srcfileloc "tree-core.h" 1545) nil ) (!pair "head" (!type pointer 516 nil gc_used (!type struct 517 (!type already_seen 516) gc_pointed_to "tree_statement_list_node" - (!srcfileloc "tree-core.h" 1536) + (!srcfileloc "tree-core.h" 1541) (!fields 3 (!pair "prev" (!type already_seen 516) - (!srcfileloc "tree-core.h" 1533) + (!srcfileloc "tree-core.h" 1538) nil ) (!pair "next" (!type already_seen 516) - (!srcfileloc "tree-core.h" 1534) + (!srcfileloc "tree-core.h" 1539) nil ) (!pair "stmt" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1535) + (!srcfileloc "tree-core.h" 1540) nil ) ) (!options @@ -11008,15 +11016,15 @@ ) 511 nil nil ) ) - (!srcfileloc "tree-core.h" 1541) + (!srcfileloc "tree-core.h" 1546) nil ) (!pair "tail" (!type already_seen 516) - (!srcfileloc "tree-core.h" 1542) + (!srcfileloc "tree-core.h" 1547) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1621) + (!srcfileloc "tree-core.h" 1626) (!options (!option tag string "TS_STATEMENT_LIST") ) @@ -11035,93 +11043,93 @@ nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1622) + (!srcfileloc "tree-core.h" 1627) (!options (!option tag string "TS_CONSTRUCTOR") ) ) (!pair "omp_clause" (!type struct 519 nil gc_used "tree_omp_clause" - (!srcfileloc "tree-core.h" 1216) + (!srcfileloc "tree-core.h" 1221) (!fields 7 (!pair "common" (!type already_seen 15) - (!srcfileloc "tree-core.h" 1197) + (!srcfileloc "tree-core.h" 1202) nil ) (!pair "locus" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1198) + (!srcfileloc "tree-core.h" 1203) nil ) (!pair "code" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1199) + (!srcfileloc "tree-core.h" 1204) nil ) (!pair "subcode" (!type union 520 nil gc_unused "omp_clause_subcode" - (!srcfileloc "tree-core.h" 1207) + (!srcfileloc "tree-core.h" 1212) (!fields 6 (!pair "default_kind" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1201) + (!srcfileloc "tree-core.h" 1206) nil ) (!pair "schedule_kind" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1202) + (!srcfileloc "tree-core.h" 1207) nil ) (!pair "depend_kind" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1203) + (!srcfileloc "tree-core.h" 1208) nil ) (!pair "map_kind" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1204) + (!srcfileloc "tree-core.h" 1209) nil ) (!pair "proc_bind_kind" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1205) + (!srcfileloc "tree-core.h" 1210) nil ) (!pair "reduction_code" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1206) + (!srcfileloc "tree-core.h" 1211) nil ) ) nil 511 nil ) - (!srcfileloc "tree-core.h" 1207) + (!srcfileloc "tree-core.h" 1212) (!options (!option skip string "") ) ) (!pair "gimple_reduction_init" (!type already_seen 310) - (!srcfileloc "tree-core.h" 1211) + (!srcfileloc "tree-core.h" 1216) nil ) (!pair "gimple_reduction_merge" (!type already_seen 310) - (!srcfileloc "tree-core.h" 1212) + (!srcfileloc "tree-core.h" 1217) nil ) (!pair "ops" (!type array 521 nil gc_used "1" (!type already_seen 5) ) - (!srcfileloc "tree-core.h" 1215) + (!srcfileloc "tree-core.h" 1220) (!options (!option length string "omp_clause_num_ops[OMP_CLAUSE_CODE ((tree)&%h)]") ) ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1623) + (!srcfileloc "tree-core.h" 1628) (!options (!option tag string "TS_OMP_CLAUSE") ) ) (!pair "optimization" (!type struct 522 nil gc_used "tree_optimization_option" - (!srcfileloc "tree-core.h" 1561) + (!srcfileloc "tree-core.h" 1566) (!fields 4 (!pair "common" (!type already_seen 15) - (!srcfileloc "tree-core.h" 1549) + (!srcfileloc "tree-core.h" 1554) nil ) (!pair "opts" (!type struct 523 nil gc_used "cl_optimization" @@ -11881,11 +11889,11 @@ nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1552) + (!srcfileloc "tree-core.h" 1557) nil ) (!pair "optabs" (!type already_seen 10) - (!srcfileloc "tree-core.h" 1556) + (!srcfileloc "tree-core.h" 1561) (!options (!option atomic string "") ) @@ -11898,25 +11906,25 @@ (!fields 0 ) nil 0 nil nil ) ) - (!srcfileloc "tree-core.h" 1560) + (!srcfileloc "tree-core.h" 1565) (!options (!option skip string "") ) ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1624) + (!srcfileloc "tree-core.h" 1629) (!options (!option tag string "TS_OPTIMIZATION") ) ) (!pair "target_option" (!type struct 526 nil gc_used "tree_target_option" - (!srcfileloc "tree-core.h" 1577) + (!srcfileloc "tree-core.h" 1582) (!fields 3 (!pair "common" (!type already_seen 15) - (!srcfileloc "tree-core.h" 1570) + (!srcfileloc "tree-core.h" 1575) nil ) (!pair "globals" (!type pointer 527 nil gc_used @@ -12171,7 +12179,7 @@ ) nil 511 nil nil ) ) - (!srcfileloc "tree-core.h" 1573) + (!srcfileloc "tree-core.h" 1578) nil ) (!pair "opts" (!type struct 555 nil gc_used "cl_target_option" @@ -12187,11 +12195,11 @@ nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1576) + (!srcfileloc "tree-core.h" 1581) nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1625) + (!srcfileloc "tree-core.h" 1630) (!options (!option tag string "TS_TARGET_OPTION") ) @@ -12308,7 +12316,7 @@ nil ) (!type struct 564 nil gc_used "lang_identifier" - (!srcfileloc "cp/cp-tree.h" 251) + (!srcfileloc "cp/cp-tree.h" 253) (!fields 5 (!pair "c_common" (!type struct 565 nil gc_used "c_common_identifier" @@ -12610,23 +12618,23 @@ nil ) ) nil 390 nil nil ) - (!srcfileloc "cp/cp-tree.h" 246) + (!srcfileloc "cp/cp-tree.h" 248) nil ) (!pair "namespace_bindings" (!type already_seen 99) - (!srcfileloc "cp/cp-tree.h" 247) + (!srcfileloc "cp/cp-tree.h" 249) nil ) (!pair "bindings" (!type already_seen 99) - (!srcfileloc "cp/cp-tree.h" 248) + (!srcfileloc "cp/cp-tree.h" 250) nil ) (!pair "class_template_info" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 249) + (!srcfileloc "cp/cp-tree.h" 251) nil ) (!pair "label_value" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 250) + (!srcfileloc "cp/cp-tree.h" 252) nil ) ) nil 260 @@ -12648,7 +12656,7 @@ (!srcfileloc "c/c-decl.c" 212) (!fields 11 (!pair "u" - (!type union 585 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/c/c-decl.c:198" + (!type union 585 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/c/c-decl.c:198" (!srcfileloc "c/c-decl.c" 201) (!fields 2 (!pair "type" @@ -12946,11 +12954,11 @@ ) (!type union 597 nil gc_pointed_to "lang_tree_node" - (!srcfileloc "cp/cp-tree.h" 829) + (!srcfileloc "cp/cp-tree.h" 831) (!fields 14 (!pair "generic" (!type already_seen 7) - (!srcfileloc "cp/cp-tree.h" 809) + (!srcfileloc "cp/cp-tree.h" 811) (!options (!option desc string "tree_node_structure (&%h)") (!option tag string "TS_CP_GENERIC") @@ -12958,31 +12966,31 @@ ) (!pair "tpi" (!type struct 598 nil gc_used "template_parm_index_s" - (!srcfileloc "cp/cp-tree.h" 278) + (!srcfileloc "cp/cp-tree.h" 280) (!fields 5 (!pair "common" (!type already_seen 15) - (!srcfileloc "cp/cp-tree.h" 273) + (!srcfileloc "cp/cp-tree.h" 275) nil ) (!pair "index" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 274) + (!srcfileloc "cp/cp-tree.h" 276) nil ) (!pair "level" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 275) + (!srcfileloc "cp/cp-tree.h" 277) nil ) (!pair "orig_level" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 276) + (!srcfileloc "cp/cp-tree.h" 278) nil ) (!pair "decl" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 277) + (!srcfileloc "cp/cp-tree.h" 279) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 810) + (!srcfileloc "cp/cp-tree.h" 812) (!options (!option tag string "TS_CP_TPI") ) @@ -12993,328 +13001,328 @@ (!type already_seen 599) ) gc_used "ptrmem_cst" - (!srcfileloc "cp/cp-tree.h" 284) + (!srcfileloc "cp/cp-tree.h" 286) (!fields 2 (!pair "common" (!type already_seen 15) - (!srcfileloc "cp/cp-tree.h" 282) + (!srcfileloc "cp/cp-tree.h" 284) nil ) (!pair "member" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 283) + (!srcfileloc "cp/cp-tree.h" 285) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 811) + (!srcfileloc "cp/cp-tree.h" 813) (!options (!option tag string "TS_CP_PTRMEM") ) ) (!pair "overload" (!type struct 601 nil gc_used "tree_overload" - (!srcfileloc "cp/cp-tree.h" 359) + (!srcfileloc "cp/cp-tree.h" 361) (!fields 2 (!pair "common" (!type already_seen 15) - (!srcfileloc "cp/cp-tree.h" 357) + (!srcfileloc "cp/cp-tree.h" 359) nil ) (!pair "function" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 358) + (!srcfileloc "cp/cp-tree.h" 360) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 812) + (!srcfileloc "cp/cp-tree.h" 814) (!options (!option tag string "TS_CP_OVERLOAD") ) ) (!pair "baselink" (!type struct 602 nil gc_used "tree_baselink" - (!srcfileloc "cp/cp-tree.h" 393) + (!srcfileloc "cp/cp-tree.h" 395) (!fields 4 (!pair "common" (!type already_seen 15) - (!srcfileloc "cp/cp-tree.h" 389) + (!srcfileloc "cp/cp-tree.h" 391) nil ) (!pair "binfo" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 390) + (!srcfileloc "cp/cp-tree.h" 392) nil ) (!pair "functions" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 391) + (!srcfileloc "cp/cp-tree.h" 393) nil ) (!pair "access_binfo" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 392) + (!srcfileloc "cp/cp-tree.h" 394) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 813) + (!srcfileloc "cp/cp-tree.h" 815) (!options (!option tag string "TS_CP_BASELINK") ) ) (!pair "default_arg" (!type struct 603 nil gc_used "tree_default_arg" - (!srcfileloc "cp/cp-tree.h" 571) + (!srcfileloc "cp/cp-tree.h" 573) (!fields 3 (!pair "common" (!type already_seen 15) - (!srcfileloc "cp/cp-tree.h" 568) + (!srcfileloc "cp/cp-tree.h" 570) nil ) (!pair "tokens" (!type already_seen 64) - (!srcfileloc "cp/cp-tree.h" 569) + (!srcfileloc "cp/cp-tree.h" 571) nil ) (!pair "instantiations" (!type already_seen 45) - (!srcfileloc "cp/cp-tree.h" 570) + (!srcfileloc "cp/cp-tree.h" 572) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 814) + (!srcfileloc "cp/cp-tree.h" 816) (!options (!option tag string "TS_CP_DEFAULT_ARG") ) ) (!pair "deferred_noexcept" (!type struct 604 nil gc_used "tree_deferred_noexcept" - (!srcfileloc "cp/cp-tree.h" 587) + (!srcfileloc "cp/cp-tree.h" 589) (!fields 3 (!pair "base" (!type already_seen 8) - (!srcfileloc "cp/cp-tree.h" 584) + (!srcfileloc "cp/cp-tree.h" 586) nil ) (!pair "pattern" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 585) + (!srcfileloc "cp/cp-tree.h" 587) nil ) (!pair "args" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 586) + (!srcfileloc "cp/cp-tree.h" 588) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 815) + (!srcfileloc "cp/cp-tree.h" 817) (!options (!option tag string "TS_CP_DEFERRED_NOEXCEPT") ) ) (!pair "identifier" (!type already_seen 559) - (!srcfileloc "cp/cp-tree.h" 816) + (!srcfileloc "cp/cp-tree.h" 818) (!options (!option tag string "TS_CP_IDENTIFIER") ) ) (!pair "static_assertion" (!type struct 605 nil gc_used "tree_static_assert" - (!srcfileloc "cp/cp-tree.h" 610) + (!srcfileloc "cp/cp-tree.h" 612) (!fields 4 (!pair "common" (!type already_seen 15) - (!srcfileloc "cp/cp-tree.h" 606) + (!srcfileloc "cp/cp-tree.h" 608) nil ) (!pair "condition" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 607) + (!srcfileloc "cp/cp-tree.h" 609) nil ) (!pair "message" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 608) + (!srcfileloc "cp/cp-tree.h" 610) nil ) (!pair "location" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 609) + (!srcfileloc "cp/cp-tree.h" 611) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 818) + (!srcfileloc "cp/cp-tree.h" 820) (!options (!option tag string "TS_CP_STATIC_ASSERT") ) ) (!pair "argument_pack_select" (!type struct 606 nil gc_used "tree_argument_pack_select" - (!srcfileloc "cp/cp-tree.h" 616) + (!srcfileloc "cp/cp-tree.h" 618) (!fields 3 (!pair "common" (!type already_seen 15) - (!srcfileloc "cp/cp-tree.h" 613) + (!srcfileloc "cp/cp-tree.h" 615) nil ) (!pair "argument_pack" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 614) + (!srcfileloc "cp/cp-tree.h" 616) nil ) (!pair "index" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 615) + (!srcfileloc "cp/cp-tree.h" 617) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 820) + (!srcfileloc "cp/cp-tree.h" 822) (!options (!option tag string "TS_CP_ARGUMENT_PACK_SELECT") ) ) (!pair "trait_expression" (!type struct 607 nil gc_used "tree_trait_expr" - (!srcfileloc "cp/cp-tree.h" 664) + (!srcfileloc "cp/cp-tree.h" 666) (!fields 4 (!pair "common" (!type already_seen 15) - (!srcfileloc "cp/cp-tree.h" 660) + (!srcfileloc "cp/cp-tree.h" 662) nil ) (!pair "type1" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 661) + (!srcfileloc "cp/cp-tree.h" 663) nil ) (!pair "type2" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 662) + (!srcfileloc "cp/cp-tree.h" 664) nil ) (!pair "kind" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 663) + (!srcfileloc "cp/cp-tree.h" 665) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 822) + (!srcfileloc "cp/cp-tree.h" 824) (!options (!option tag string "TS_CP_TRAIT_EXPR") ) ) (!pair "lambda_expression" (!type struct 608 nil gc_used "tree_lambda_expr" - (!srcfileloc "cp/cp-tree.h" 749) + (!srcfileloc "cp/cp-tree.h" 751) (!fields 10 (!pair "typed" (!type already_seen 14) - (!srcfileloc "cp/cp-tree.h" 739) + (!srcfileloc "cp/cp-tree.h" 741) nil ) (!pair "capture_list" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 740) + (!srcfileloc "cp/cp-tree.h" 742) nil ) (!pair "this_capture" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 741) + (!srcfileloc "cp/cp-tree.h" 743) nil ) (!pair "return_type" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 742) + (!srcfileloc "cp/cp-tree.h" 744) nil ) (!pair "extra_scope" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 743) + (!srcfileloc "cp/cp-tree.h" 745) nil ) (!pair "closure" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 744) + (!srcfileloc "cp/cp-tree.h" 746) nil ) (!pair "pending_proxies" (!type already_seen 45) - (!srcfileloc "cp/cp-tree.h" 745) + (!srcfileloc "cp/cp-tree.h" 747) nil ) (!pair "locus" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 746) + (!srcfileloc "cp/cp-tree.h" 748) nil ) (!pair "default_capture_mode" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 747) + (!srcfileloc "cp/cp-tree.h" 749) nil ) (!pair "discriminator" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 748) + (!srcfileloc "cp/cp-tree.h" 750) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 824) + (!srcfileloc "cp/cp-tree.h" 826) (!options (!option tag string "TS_CP_LAMBDA_EXPR") ) ) (!pair "template_info" (!type struct 609 nil gc_used "tree_template_info" - (!srcfileloc "cp/cp-tree.h" 783) + (!srcfileloc "cp/cp-tree.h" 785) (!fields 2 (!pair "common" (!type already_seen 15) - (!srcfileloc "cp/cp-tree.h" 781) + (!srcfileloc "cp/cp-tree.h" 783) nil ) (!pair "typedefs_needing_access_checking" (!type pointer 610 nil gc_used (!type user_struct 611 (!type already_seen 610) gc_pointed_to "vec<qualified_typedef_usage_t,va_gc>" - (!srcfileloc "cp/cp-tree.h" 782) + (!srcfileloc "cp/cp-tree.h" 784) (!fields 2 (!pair "va_gc" (!type already_seen 3) - (!srcfileloc "cp/cp-tree.h" 782) + (!srcfileloc "cp/cp-tree.h" 784) nil ) (!pair "qualified_typedef_usage_t" (!type struct 612 nil gc_used "qualified_typedef_usage_s" - (!srcfileloc "cp/cp-tree.h" 769) + (!srcfileloc "cp/cp-tree.h" 771) (!fields 3 (!pair "typedef_decl" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 766) + (!srcfileloc "cp/cp-tree.h" 768) nil ) (!pair "context" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 767) + (!srcfileloc "cp/cp-tree.h" 769) nil ) (!pair "locus" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 768) + (!srcfileloc "cp/cp-tree.h" 770) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 782) + (!srcfileloc "cp/cp-tree.h" 784) nil ) ) ) ) - (!srcfileloc "cp/cp-tree.h" 782) + (!srcfileloc "cp/cp-tree.h" 784) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 826) + (!srcfileloc "cp/cp-tree.h" 828) (!options (!option tag string "TS_CP_TEMPLATE_INFO") ) ) (!pair "userdef_literal" (!type struct 613 nil gc_used "tree_userdef_literal" - (!srcfileloc "c-family/c-common.h" 1267) + (!srcfileloc "c-family/c-common.h" 1268) (!fields 5 (!pair "base" (!type already_seen 8) - (!srcfileloc "c-family/c-common.h" 1262) + (!srcfileloc "c-family/c-common.h" 1263) nil ) (!pair "suffix_id" (!type already_seen 5) - (!srcfileloc "c-family/c-common.h" 1263) + (!srcfileloc "c-family/c-common.h" 1264) nil ) (!pair "value" (!type already_seen 5) - (!srcfileloc "c-family/c-common.h" 1264) + (!srcfileloc "c-family/c-common.h" 1265) nil ) (!pair "num_string" (!type already_seen 5) - (!srcfileloc "c-family/c-common.h" 1265) + (!srcfileloc "c-family/c-common.h" 1266) nil ) (!pair "overflow" (!type already_seen 9) - (!srcfileloc "c-family/c-common.h" 1266) + (!srcfileloc "c-family/c-common.h" 1267) nil ) ) nil 390 nil nil ) - (!srcfileloc "cp/cp-tree.h" 828) + (!srcfileloc "cp/cp-tree.h" 830) (!options (!option tag string "TS_CP_USERDEF_LITERAL") ) @@ -13667,7 +13675,7 @@ (!fields 0 ) nil 0 nil nil ) - (!type struct 644 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/java/expr.c:2060"nil + (!type struct 644 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/java/expr.c:2060"nil (!fields 0 ) nil 0 nil nil ) @@ -13773,19 +13781,19 @@ nil 0 nil nil ) (!type struct 653 nil gc_used "gfc_powdecl_list" - (!srcfileloc "fortran/trans.h" 723) + (!srcfileloc "fortran/trans.h" 728) (!fields 3 (!pair "integer" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 720) + (!srcfileloc "fortran/trans.h" 725) nil ) (!pair "real" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 721) + (!srcfileloc "fortran/trans.h" 726) nil ) (!pair "cmplx" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 722) + (!srcfileloc "fortran/trans.h" 727) nil ) ) nil 8 nil nil ) @@ -13794,11 +13802,11 @@ (!fields 0 ) nil 0 nil nil ) - (!type struct 655 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/fortran/trans.h:340"nil + (!type struct 655 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/fortran/trans.h:340"nil (!fields 0 ) nil 0 nil nil ) - (!type struct 656 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/fortran/trans.h:328"nil + (!type struct 656 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/fortran/trans.h:328"nil (!fields 0 ) nil 0 nil nil ) @@ -13822,7 +13830,7 @@ (!fields 0 ) nil 0 nil nil ) - (!type struct 662 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/fortran/trans.h:33"nil + (!type struct 662 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/fortran/trans.h:33"nil (!fields 0 ) nil 0 nil nil ) @@ -13965,19 +13973,19 @@ (!type already_seen 668) ) gc_pointed_to "module_htab_entry" - (!srcfileloc "fortran/trans.h" 571) + (!srcfileloc "fortran/trans.h" 574) (!fields 3 (!pair "name" (!type already_seen 36) - (!srcfileloc "fortran/trans.h" 568) + (!srcfileloc "fortran/trans.h" 571) nil ) (!pair "namespace_decl" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 569) + (!srcfileloc "fortran/trans.h" 572) nil ) (!pair "decls" (!type already_seen 53) - (!srcfileloc "fortran/trans.h" 570) + (!srcfileloc "fortran/trans.h" 573) (!options (!option param_is type (!type already_seen 7) @@ -14095,27 +14103,27 @@ (!type already_seen 684) ) gc_pointed_to "vec<tree_int,va_gc>" - (!srcfileloc "cp/parser.c" 8654) + (!srcfileloc "cp/parser.c" 8652) (!fields 2 (!pair "va_gc" (!type already_seen 3) - (!srcfileloc "cp/parser.c" 8654) + (!srcfileloc "cp/parser.c" 8652) nil ) (!pair "tree_int" (!type struct 686 nil gc_used "tree_int" - (!srcfileloc "cp/parser.c" 8653) + (!srcfileloc "cp/parser.c" 8651) (!fields 2 (!pair "t" (!type already_seen 5) - (!srcfileloc "cp/parser.c" 8651) + (!srcfileloc "cp/parser.c" 8649) nil ) (!pair "i" (!type already_seen 9) - (!srcfileloc "cp/parser.c" 8652) + (!srcfileloc "cp/parser.c" 8650) nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/parser.c" 8654) + (!srcfileloc "cp/parser.c" 8652) nil ) ) ) @@ -14479,39 +14487,39 @@ (!type already_seen 709) ) gc_pointed_to "constexpr_call" - (!srcfileloc "cp/semantics.c" 8081) + (!srcfileloc "cp/semantics.c" 8083) (!fields 4 (!pair "fundef" (!type pointer 711 nil gc_used (!type struct 712 (!type already_seen 711) gc_pointed_to "constexpr_fundef" - (!srcfileloc "cp/semantics.c" 7398) + (!srcfileloc "cp/semantics.c" 7400) (!fields 2 (!pair "decl" (!type already_seen 5) - (!srcfileloc "cp/semantics.c" 7396) + (!srcfileloc "cp/semantics.c" 7398) nil ) (!pair "body" (!type already_seen 5) - (!srcfileloc "cp/semantics.c" 7397) + (!srcfileloc "cp/semantics.c" 7399) nil ) ) nil 260 nil nil ) ) - (!srcfileloc "cp/semantics.c" 8065) + (!srcfileloc "cp/semantics.c" 8067) nil ) (!pair "bindings" (!type already_seen 5) - (!srcfileloc "cp/semantics.c" 8072) + (!srcfileloc "cp/semantics.c" 8074) nil ) (!pair "result" (!type already_seen 5) - (!srcfileloc "cp/semantics.c" 8077) + (!srcfileloc "cp/semantics.c" 8079) nil ) (!pair "hash" (!type already_seen 9) - (!srcfileloc "cp/semantics.c" 8080) + (!srcfileloc "cp/semantics.c" 8082) nil ) ) nil 260 nil nil ) @@ -14610,27 +14618,27 @@ (!type struct 726 (!type already_seen 725) gc_pointed_to "tinst_level" - (!srcfileloc "cp/cp-tree.h" 4985) + (!srcfileloc "cp/cp-tree.h" 4995) (!fields 5 (!pair "next" (!type already_seen 725) - (!srcfileloc "cp/cp-tree.h" 4970) + (!srcfileloc "cp/cp-tree.h" 4980) nil ) (!pair "decl" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 4975) + (!srcfileloc "cp/cp-tree.h" 4985) nil ) (!pair "locus" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4978) + (!srcfileloc "cp/cp-tree.h" 4988) nil ) (!pair "errors" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4981) + (!srcfileloc "cp/cp-tree.h" 4991) nil ) (!pair "in_system_header_p" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4984) + (!srcfileloc "cp/cp-tree.h" 4994) nil ) ) (!options @@ -14743,23 +14751,23 @@ nil 0 nil nil ) (!type struct 745 nil gc_used "operator_name_info_t" - (!srcfileloc "cp/cp-tree.h" 4729) + (!srcfileloc "cp/cp-tree.h" 4744) (!fields 4 (!pair "identifier" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 4722) + (!srcfileloc "cp/cp-tree.h" 4737) nil ) (!pair "name" (!type already_seen 36) - (!srcfileloc "cp/cp-tree.h" 4724) + (!srcfileloc "cp/cp-tree.h" 4739) nil ) (!pair "mangled_name" (!type already_seen 36) - (!srcfileloc "cp/cp-tree.h" 4726) + (!srcfileloc "cp/cp-tree.h" 4741) nil ) (!pair "arity" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4728) + (!srcfileloc "cp/cp-tree.h" 4743) nil ) ) nil 260 nil nil ) @@ -14805,11 +14813,11 @@ (!type already_seen 747) ) gc_pointed_to "vec<cxx_saved_binding,va_gc>" - (!srcfileloc "cp/cp-tree.h" 1020) + (!srcfileloc "cp/cp-tree.h" 1022) (!fields 2 (!pair "va_gc" (!type already_seen 3) - (!srcfileloc "cp/cp-tree.h" 1020) + (!srcfileloc "cp/cp-tree.h" 1022) nil ) (!pair "cxx_saved_binding" (!type struct 749 nil gc_used "cxx_saved_binding" @@ -14829,7 +14837,7 @@ nil ) ) nil 260 nil nil ) - (!srcfileloc "cp/cp-tree.h" 1020) + (!srcfileloc "cp/cp-tree.h" 1022) nil ) ) ) @@ -14867,111 +14875,111 @@ (!type already_seen 750) ) gc_pointed_to "saved_scope" - (!srcfileloc "cp/cp-tree.h" 1056) + (!srcfileloc "cp/cp-tree.h" 1058) (!fields 26 (!pair "old_bindings" (!type already_seen 748) - (!srcfileloc "cp/cp-tree.h" 1020) + (!srcfileloc "cp/cp-tree.h" 1022) nil ) (!pair "old_namespace" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1021) + (!srcfileloc "cp/cp-tree.h" 1023) nil ) (!pair "decl_ns_list" (!type already_seen 45) - (!srcfileloc "cp/cp-tree.h" 1022) + (!srcfileloc "cp/cp-tree.h" 1024) nil ) (!pair "class_name" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1023) + (!srcfileloc "cp/cp-tree.h" 1025) nil ) (!pair "class_type" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1024) + (!srcfileloc "cp/cp-tree.h" 1026) nil ) (!pair "access_specifier" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1025) + (!srcfileloc "cp/cp-tree.h" 1027) nil ) (!pair "function_decl" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1026) + (!srcfileloc "cp/cp-tree.h" 1028) nil ) (!pair "lang_base" (!type already_seen 45) - (!srcfileloc "cp/cp-tree.h" 1027) + (!srcfileloc "cp/cp-tree.h" 1029) nil ) (!pair "lang_name" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1028) + (!srcfileloc "cp/cp-tree.h" 1030) nil ) (!pair "template_parms" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1029) + (!srcfileloc "cp/cp-tree.h" 1031) nil ) (!pair "x_previous_class_level" (!type already_seen 94) - (!srcfileloc "cp/cp-tree.h" 1030) + (!srcfileloc "cp/cp-tree.h" 1032) nil ) (!pair "x_saved_tree" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1031) + (!srcfileloc "cp/cp-tree.h" 1033) nil ) (!pair "x_current_class_ptr" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1034) + (!srcfileloc "cp/cp-tree.h" 1036) nil ) (!pair "x_current_class_ref" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 1035) + (!srcfileloc "cp/cp-tree.h" 1037) nil ) (!pair "x_processing_template_decl" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1037) + (!srcfileloc "cp/cp-tree.h" 1039) nil ) (!pair "x_processing_specialization" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1038) + (!srcfileloc "cp/cp-tree.h" 1040) nil ) (!pair "x_processing_explicit_instantiation" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1039) + (!srcfileloc "cp/cp-tree.h" 1041) nil ) (!pair "need_pop_function_context" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1040) + (!srcfileloc "cp/cp-tree.h" 1042) nil ) (!pair "unevaluated_operand" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1042) + (!srcfileloc "cp/cp-tree.h" 1044) nil ) (!pair "inhibit_evaluation_warnings" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1043) + (!srcfileloc "cp/cp-tree.h" 1045) nil ) (!pair "omp_declare_target_attribute" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 1046) + (!srcfileloc "cp/cp-tree.h" 1048) nil ) (!pair "x_stmt_tree" (!type already_seen 88) - (!srcfileloc "cp/cp-tree.h" 1048) + (!srcfileloc "cp/cp-tree.h" 1050) nil ) (!pair "class_bindings" (!type already_seen 94) - (!srcfileloc "cp/cp-tree.h" 1050) + (!srcfileloc "cp/cp-tree.h" 1052) nil ) (!pair "bindings" (!type already_seen 94) - (!srcfileloc "cp/cp-tree.h" 1051) + (!srcfileloc "cp/cp-tree.h" 1053) nil ) (!pair "x_local_specializations" (!type already_seen 385) - (!srcfileloc "cp/cp-tree.h" 1053) + (!srcfileloc "cp/cp-tree.h" 1055) nil ) (!pair "prev" (!type already_seen 751) - (!srcfileloc "cp/cp-tree.h" 1055) + (!srcfileloc "cp/cp-tree.h" 1057) nil ) ) nil 260 nil nil ) @@ -15073,7 +15081,7 @@ (!fields 0 ) nil 0 nil nil ) - (!type struct 759 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/c/c-parser.c:6077"nil + (!type struct 759 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/c/c-parser.c:6071"nil (!fields 0 ) nil 0 nil nil ) @@ -15200,7 +15208,7 @@ (!fields 0 ) nil 0 nil nil ) - (!type struct 768 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/c-family/c-pragma.c:1169"nil + (!type struct 768 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/c-family/c-pragma.c:1169"nil (!fields 0 ) nil 0 nil nil ) @@ -15385,15 +15393,15 @@ (!type already_seen 788) ) gc_pointed_to "vec<tree_gc_vec,va_gc>" - (!srcfileloc "c-family/c-common.c" 11540) + (!srcfileloc "c-family/c-common.c" 11546) (!fields 2 (!pair "va_gc" (!type already_seen 3) - (!srcfileloc "c-family/c-common.c" 11540) + (!srcfileloc "c-family/c-common.c" 11546) nil ) (!pair "tree_gc_vec" (!type already_seen 45) - (!srcfileloc "c-family/c-common.c" 11540) + (!srcfileloc "c-family/c-common.c" 11546) nil ) ) ) @@ -15409,15 +15417,15 @@ (!type already_seen 791) ) gc_pointed_to "vec<const_char_p,va_gc>" - (!srcfileloc "c-family/c-common.c" 9005) + (!srcfileloc "c-family/c-common.c" 9011) (!fields 2 (!pair "va_gc" (!type already_seen 3) - (!srcfileloc "c-family/c-common.c" 9005) + (!srcfileloc "c-family/c-common.c" 9011) nil ) (!pair "const_char_p" (!type already_seen 36) - (!srcfileloc "c-family/c-common.c" 9005) + (!srcfileloc "c-family/c-common.c" 9011) nil ) ) ) @@ -15732,7 +15740,7 @@ (!fields 0 ) nil 0 nil nil ) - (!type struct 828 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/ada/gcc-interface/decl.c:178"nil + (!type struct 828 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/ada/gcc-interface/decl.c:178"nil (!fields 0 ) nil 0 nil nil ) @@ -15919,11 +15927,11 @@ (!fields 2 (!pair "type" (!type struct 858 nil gc_used "tree_map_base" - (!srcfileloc "tree-core.h" 1715) + (!srcfileloc "tree-core.h" 1720) (!fields 1 (!pair "from" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1714) + (!srcfileloc "tree-core.h" 1719) nil ) ) nil 511 nil nil ) @@ -17279,14 +17287,14 @@ nil 0 nil nil ) (!type user_struct 970 nil gc_unused "vec<ipa_parm_adjustment>" - (!srcfileloc "ipa-prop.h" 685) + (!srcfileloc "ipa-prop.h" 686) (!fields 1 (!pair "ipa_parm_adjustment" (!type struct 971 nil gc_unused "ipa_parm_adjustment" - (!srcfileloc "ipa-prop.h" 685) + (!srcfileloc "ipa-prop.h" 686) (!fields 0 ) nil 511 nil nil ) - (!srcfileloc "ipa-prop.h" 685) + (!srcfileloc "ipa-prop.h" 686) nil ) ) ) @@ -17653,7 +17661,7 @@ (!fields 0 ) nil 0 nil nil ) - (!type struct 1014 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/config/aarch64/aarch64.c:6436"nil + (!type struct 1014 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/config/aarch64/aarch64.c:7357"nil (!fields 0 ) nil 0 nil nil ) @@ -17983,7 +17991,7 @@ (!fields 0 ) nil 0 nil nil ) - (!type union 1079 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/tree-eh.c:59"nil + (!type union 1079 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/tree-eh.c:59"nil (!fields 0 ) nil 0 nil ) @@ -18474,19 +18482,19 @@ (!pair "args" (!type array 1112 nil gc_used "1" (!type struct 1113 nil gc_used "phi_arg_d" - (!srcfileloc "tree-core.h" 1194) + (!srcfileloc "tree-core.h" 1199) (!fields 3 (!pair "imm_use" (!type already_seen 510) - (!srcfileloc "tree-core.h" 1191) + (!srcfileloc "tree-core.h" 1196) nil ) (!pair "def" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1192) + (!srcfileloc "tree-core.h" 1197) nil ) (!pair "locus" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1193) + (!srcfileloc "tree-core.h" 1198) nil ) ) nil 511 nil nil ) @@ -18613,7 +18621,7 @@ (!srcfileloc "gimple.h" 277) nil ) (!pair "u" - (!type union 1120 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/gimple.h:280" + (!type union 1120 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/gimple.h:280" (!srcfileloc "gimple.h" 283) (!fields 2 (!pair "fntype" @@ -18718,11 +18726,11 @@ (!fields 0 ) nil 0 nil nil ) - (!type struct 1128 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/varasm.c:4818"nil + (!type struct 1128 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/varasm.c:4818"nil (!fields 0 ) nil 0 nil nil ) - (!type struct 1129 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/varasm.c:4574"nil + (!type struct 1129 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/varasm.c:4574"nil (!fields 0 ) nil 0 nil nil ) @@ -18828,7 +18836,7 @@ (!fields 0 ) nil 0 nil nil ) - (!type union 1148 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/mcf.c:344"nil + (!type union 1148 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/mcf.c:344"nil (!fields 0 ) nil 0 nil ) @@ -19744,7 +19752,7 @@ (!fields 0 ) nil 0 nil nil ) - (!type struct 1268 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/dwarf2cfi.c:218"nil + (!type struct 1268 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/dwarf2cfi.c:218"nil (!fields 0 ) nil 0 nil nil ) @@ -19756,7 +19764,7 @@ (!type pointer 1271 nil gc_unused (!type already_seen 1270) ) - gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/dwarf2cfi.c:101"nil + gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/dwarf2cfi.c:101"nil (!fields 0 ) nil 0 nil nil ) @@ -21059,15 +21067,15 @@ (!type already_seen 1384) ) gc_pointed_to "vec<alias_pair,va_gc>" - (!srcfileloc "tree-core.h" 1793) + (!srcfileloc "tree-core.h" 1798) (!fields 2 (!pair "va_gc" (!type already_seen 3) - (!srcfileloc "tree-core.h" 1793) + (!srcfileloc "tree-core.h" 1798) nil ) (!pair "alias_pair" (!type struct 1386 nil gc_used "alias_pair" - (!srcfileloc "tree-core.h" 1793) + (!srcfileloc "tree-core.h" 1798) (!fields 2 (!pair "decl" (!type already_seen 5) @@ -21079,25 +21087,25 @@ nil ) ) nil 511 nil nil ) - (!srcfileloc "tree-core.h" 1793) + (!srcfileloc "tree-core.h" 1798) nil ) ) ) (!type struct 1387 nil gc_used "builtin_info_type" - (!srcfileloc "tree-core.h" 1828) + (!srcfileloc "tree-core.h" 1833) (!fields 2 (!pair "decl" (!type array 1388 nil gc_used "(int)END_BUILTINS" (!type already_seen 5) ) - (!srcfileloc "tree-core.h" 1771) + (!srcfileloc "tree-core.h" 1776) nil ) (!pair "implicit_p" (!type array 1389 nil gc_used "(int)END_BUILTINS" (!type already_seen 9) ) - (!srcfileloc "tree-core.h" 1772) + (!srcfileloc "tree-core.h" 1777) nil ) ) nil 511 nil nil ) @@ -21115,15 +21123,15 @@ (!type already_seen 1392) ) gc_pointed_to "tree_vec_map" - (!srcfileloc "tree-core.h" 1748) + (!srcfileloc "tree-core.h" 1753) (!fields 2 (!pair "base" (!type already_seen 858) - (!srcfileloc "tree-core.h" 1746) + (!srcfileloc "tree-core.h" 1751) nil ) (!pair "to" (!type already_seen 45) - (!srcfileloc "tree-core.h" 1747) + (!srcfileloc "tree-core.h" 1752) nil ) ) nil 511 nil nil ) @@ -21133,19 +21141,19 @@ (!type already_seen 1394) ) gc_pointed_to "tree_priority_map" - (!srcfileloc "tree-core.h" 1742) + (!srcfileloc "tree-core.h" 1747) (!fields 3 (!pair "base" (!type already_seen 858) - (!srcfileloc "tree-core.h" 1739) + (!srcfileloc "tree-core.h" 1744) nil ) (!pair "init" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1740) + (!srcfileloc "tree-core.h" 1745) nil ) (!pair "fini" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1741) + (!srcfileloc "tree-core.h" 1746) nil ) ) nil 511 nil nil ) @@ -21155,15 +21163,15 @@ (!type already_seen 1396) ) gc_pointed_to "tree_int_map" - (!srcfileloc "tree-core.h" 1735) + (!srcfileloc "tree-core.h" 1740) (!fields 2 (!pair "base" (!type already_seen 858) - (!srcfileloc "tree-core.h" 1733) + (!srcfileloc "tree-core.h" 1738) nil ) (!pair "to" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1734) + (!srcfileloc "tree-core.h" 1739) nil ) ) nil 511 nil nil ) @@ -21173,15 +21181,15 @@ (!type already_seen 1398) ) gc_pointed_to "tree_decl_map" - (!srcfileloc "tree-core.h" 1729) + (!srcfileloc "tree-core.h" 1734) (!fields 2 (!pair "base" (!type already_seen 858) - (!srcfileloc "tree-core.h" 1727) + (!srcfileloc "tree-core.h" 1732) nil ) (!pair "to" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1728) + (!srcfileloc "tree-core.h" 1733) nil ) ) nil 511 nil nil ) @@ -21191,19 +21199,19 @@ (!type already_seen 1400) ) gc_pointed_to "tree_map" - (!srcfileloc "tree-core.h" 1723) + (!srcfileloc "tree-core.h" 1728) (!fields 3 (!pair "base" (!type already_seen 858) - (!srcfileloc "tree-core.h" 1720) + (!srcfileloc "tree-core.h" 1725) nil ) (!pair "hash" (!type already_seen 9) - (!srcfileloc "tree-core.h" 1721) + (!srcfileloc "tree-core.h" 1726) nil ) (!pair "to" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1722) + (!srcfileloc "tree-core.h" 1727) nil ) ) nil 511 nil nil ) @@ -21809,7 +21817,7 @@ (!type already_seen 1318) - (!type struct 1426 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/config/aarch64/aarch64.h:576"nil + (!type struct 1426 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/config/aarch64/aarch64.h:585"nil (!fields 0 ) nil 0 nil nil ) @@ -21897,11 +21905,11 @@ (!fields 0 ) nil 0 nil nil ) - (!type struct 1442 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/../libcpp/include/cpplib.h:999"nil + (!type struct 1442 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/../libcpp/include/cpplib.h:999"nil (!fields 0 ) nil 0 nil nil ) - (!type struct 1443 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/../libcpp/include/cpplib.h:988"nil + (!type struct 1443 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/../libcpp/include/cpplib.h:988"nil (!fields 0 ) nil 0 nil nil ) @@ -21955,7 +21963,7 @@ (!fields 0 ) nil 0 nil nil ) - (!type struct 1452 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/../libcpp/include/line-map.h:627"nil + (!type struct 1452 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/../libcpp/include/line-map.h:627"nil (!fields 0 ) nil 0 nil nil ) @@ -22317,15 +22325,15 @@ nil ) (!pair "gfc_interface_mapping" (!type already_seen 651) - (!srcfileloc "fortran/trans.h" 949) + (!srcfileloc "fortran/trans.h" 963) nil ) (!pair "gfc_interface_sym_mapping" (!type already_seen 652) - (!srcfileloc "fortran/trans.h" 934) + (!srcfileloc "fortran/trans.h" 948) nil ) (!pair "gfc_powdecl_list" (!type already_seen 653) - (!srcfileloc "fortran/trans.h" 724) + (!srcfileloc "fortran/trans.h" 729) nil ) (!pair "gfc_wrapped_block" (!type already_seen 655) @@ -22425,11 +22433,11 @@ nil ) (!pair "vec<tree_int,va_gc>" (!type already_seen 684) - (!srcfileloc "cp/parser.c" 8654) + (!srcfileloc "cp/parser.c" 8652) nil ) (!pair "tree_int" (!type already_seen 686) - (!srcfileloc "cp/parser.c" 8653) + (!srcfileloc "cp/parser.c" 8651) nil ) (!pair "cp_parser_expression_stack" (!type array 1464 nil gc_unused "NUM_PREC_VALUES" @@ -22531,15 +22539,15 @@ nil ) (!pair "cplus_array_info" (!type already_seen 708) - (!srcfileloc "cp/tree.c" 730) + (!srcfileloc "cp/tree.c" 741) nil ) (!pair "constexpr_call" (!type already_seen 709) - (!srcfileloc "cp/semantics.c" 8081) + (!srcfileloc "cp/semantics.c" 8083) nil ) (!pair "constexpr_fundef" (!type already_seen 712) - (!srcfileloc "cp/semantics.c" 7398) + (!srcfileloc "cp/semantics.c" 7400) nil ) (!pair "vec<deferred_access,va_gc>" (!type already_seen 714) @@ -22607,159 +22615,159 @@ nil ) (!pair "deferred_access_check" (!type already_seen 73) - (!srcfileloc "cp/cp-tree.h" 5686) + (!srcfileloc "cp/cp-tree.h" 5697) nil ) (!pair "walk_namespaces_fn" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 5267) + (!srcfileloc "cp/cp-tree.h" 5277) nil ) (!pair "cp_parameter_declarator" (!type already_seen 742) - (!srcfileloc "cp/cp-tree.h" 4880) + (!srcfileloc "cp/cp-tree.h" 4890) nil ) (!pair "cp_declarator" (!type already_seen 743) - (!srcfileloc "cp/cp-tree.h" 4878) + (!srcfileloc "cp/cp-tree.h" 4888) nil ) (!pair "cp_declarator_kind" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4874) + (!srcfileloc "cp/cp-tree.h" 4884) nil ) (!pair "cp_decl_specifier_seq" (!type already_seen 744) - (!srcfileloc "cp/cp-tree.h" 4862) + (!srcfileloc "cp/cp-tree.h" 4872) nil ) (!pair "cp_decl_spec" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4818) + (!srcfileloc "cp/cp-tree.h" 4828) nil ) (!pair "cp_storage_class" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4787) + (!srcfileloc "cp/cp-tree.h" 4797) nil ) (!pair "cp_virt_specifiers" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4757) - nil ) - (!pair "cp_cv_quals" - (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4741) + (!srcfileloc "cp/cp-tree.h" 4767) nil ) (!pair "operator_name_info_t" (!type already_seen 745) - (!srcfileloc "cp/cp-tree.h" 4729) + (!srcfileloc "cp/cp-tree.h" 4744) nil ) (!pair "unification_kind_t" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4370) + (!srcfileloc "cp/cp-tree.h" 4385) + nil ) + (!pair "cp_cv_quals" + (!type already_seen 9) + (!srcfileloc "cp/cp-tree.h" 4362) nil ) (!pair "base_kind" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4331) + (!srcfileloc "cp/cp-tree.h" 4340) nil ) (!pair "deferring_kind" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4318) + (!srcfileloc "cp/cp-tree.h" 4327) nil ) (!pair "base_access" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4311) + (!srcfileloc "cp/cp-tree.h" 4320) nil ) (!pair "tsubst_flags_t" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4297) + (!srcfileloc "cp/cp-tree.h" 4306) nil ) (!pair "duration_kind" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4268) + (!srcfileloc "cp/cp-tree.h" 4277) nil ) (!pair "linkage_kind" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4261) + (!srcfileloc "cp/cp-tree.h" 4270) nil ) (!pair "special_function_kind" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4238) + (!srcfileloc "cp/cp-tree.h" 4247) nil ) (!pair "access_kind" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4217) + (!srcfileloc "cp/cp-tree.h" 4226) nil ) (!pair "tmpl_spec_kind" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4206) + (!srcfileloc "cp/cp-tree.h" 4215) nil ) (!pair "cp_lvalue_kind" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 4188) + (!srcfileloc "cp/cp-tree.h" 4197) nil ) (!pair "aggr_init_expr_arg_iterator" (!type already_seen 746) - (!srcfileloc "cp/cp-tree.h" 3078) + (!srcfileloc "cp/cp-tree.h" 3084) nil ) (!pair "vec<tree_pair_s,va_gc>" (!type already_seen 483) - (!srcfileloc "cp/cp-tree.h" 1454) + (!srcfileloc "cp/cp-tree.h" 1456) nil ) (!pair "tree_pair_p" (!type already_seen 485) - (!srcfileloc "cp/cp-tree.h" 1359) + (!srcfileloc "cp/cp-tree.h" 1361) nil ) (!pair "tree_pair_s" (!type already_seen 484) - (!srcfileloc "cp/cp-tree.h" 1358) + (!srcfileloc "cp/cp-tree.h" 1360) nil ) (!pair "vec<cxx_saved_binding,va_gc>" (!type already_seen 747) - (!srcfileloc "cp/cp-tree.h" 1020) + (!srcfileloc "cp/cp-tree.h" 1022) nil ) (!pair "vec<qualified_typedef_usage_t,va_gc>" (!type already_seen 611) - (!srcfileloc "cp/cp-tree.h" 782) + (!srcfileloc "cp/cp-tree.h" 784) nil ) (!pair "qualified_typedef_usage_t" (!type already_seen 612) - (!srcfileloc "cp/cp-tree.h" 770) + (!srcfileloc "cp/cp-tree.h" 772) nil ) (!pair "cp_trait_kind" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 646) + (!srcfileloc "cp/cp-tree.h" 648) nil ) (!pair "abstract_class_use" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 499) + (!srcfileloc "cp/cp-tree.h" 501) nil ) (!pair "impl_conv_void" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 486) + (!srcfileloc "cp/cp-tree.h" 488) nil ) (!pair "impl_conv_rhs" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 475) + (!srcfileloc "cp/cp-tree.h" 477) nil ) (!pair "expr_list_kind" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 465) + (!srcfileloc "cp/cp-tree.h" 467) nil ) (!pair "composite_pointer_operation" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 458) + (!srcfileloc "cp/cp-tree.h" 460) nil ) (!pair "cpp0x_warn_str" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 446) + (!srcfileloc "cp/cp-tree.h" 448) nil ) (!pair "cp_id_kind" (!type already_seen 9) - (!srcfileloc "cp/cp-tree.h" 409) + (!srcfileloc "cp/cp-tree.h" 411) nil ) (!pair "ptrmem_cst_t" (!type already_seen 600) - (!srcfileloc "cp/cp-tree.h" 285) + (!srcfileloc "cp/cp-tree.h" 287) nil ) (!pair "template_parm_index" (!type already_seen 598) - (!srcfileloc "cp/cp-tree.h" 279) + (!srcfileloc "cp/cp-tree.h" 281) nil ) (!pair "vec<cp_label_binding,va_gc>" (!type already_seen 102) @@ -22903,15 +22911,15 @@ nil ) (!pair "an_loop_parts" (!type already_seen 782) - (!srcfileloc "c-family/c-common.h" 1351) + (!srcfileloc "c-family/c-common.h" 1352) nil ) (!pair "an_parts" (!type already_seen 783) - (!srcfileloc "c-family/c-common.h" 1340) + (!srcfileloc "c-family/c-common.h" 1341) nil ) (!pair "omp_clause_mask" (!type already_seen 9) - (!srcfileloc "c-family/c-common.h" 1050) + (!srcfileloc "c-family/c-common.h" 1051) nil ) (!pair "stmt_tree" (!type already_seen 89) @@ -22927,23 +22935,23 @@ nil ) (!pair "vec<tree_gc_vec,va_gc>" (!type already_seen 788) - (!srcfileloc "c-family/c-common.c" 11540) + (!srcfileloc "c-family/c-common.c" 11546) nil ) (!pair "tree_gc_vec" (!type already_seen 45) - (!srcfileloc "c-family/c-common.c" 11539) + (!srcfileloc "c-family/c-common.c" 11545) nil ) (!pair "vec<const_char_p,va_gc>" (!type already_seen 791) - (!srcfileloc "c-family/c-common.c" 9005) + (!srcfileloc "c-family/c-common.c" 9011) nil ) (!pair "const_char_p" (!type already_seen 29) - (!srcfileloc "c-family/c-common.c" 9004) + (!srcfileloc "c-family/c-common.c" 9010) nil ) (!pair "disabled_builtin" (!type already_seen 793) - (!srcfileloc "c-family/c-common.c" 5755) + (!srcfileloc "c-family/c-common.c" 5761) nil ) (!pair "c_binding_ptr" (!type already_seen 583) @@ -23175,15 +23183,15 @@ nil ) (!pair "ipa_parm_adjustment_vec" (!type already_seen 970) - (!srcfileloc "ipa-prop.h" 685) + (!srcfileloc "ipa-prop.h" 686) nil ) (!pair "ipa_parm_adjustment" (!type already_seen 971) - (!srcfileloc "ipa-prop.h" 685) + (!srcfileloc "ipa-prop.h" 686) nil ) (!pair "vec<ipa_parm_adjustment>" (!type already_seen 970) - (!srcfileloc "ipa-prop.h" 685) + (!srcfileloc "ipa-prop.h" 686) nil ) (!pair "ipa_edge_args" (!type already_seen 974) @@ -23231,11 +23239,11 @@ nil ) (!pair "aarch64_simd_mangle_map_entry" (!type already_seen 1014) - (!srcfileloc "config/aarch64/aarch64.c" 6440) + (!srcfileloc "config/aarch64/aarch64.c" 7361) nil ) (!pair "aarch64_cc" (!type already_seen 9) - (!srcfileloc "config/aarch64/aarch64.c" 324) + (!srcfileloc "config/aarch64/aarch64.c" 376) nil ) (!pair "omp_context" (!type already_seen 1029) @@ -23937,23 +23945,23 @@ nil ) (!pair "builtin_info_type" (!type already_seen 1387) - (!srcfileloc "tree-core.h" 1828) + (!srcfileloc "tree-core.h" 1833) nil ) (!pair "alias_pair" (!type already_seen 1386) - (!srcfileloc "tree-core.h" 1793) + (!srcfileloc "tree-core.h" 1798) nil ) (!pair "vec<alias_pair,va_gc>" (!type already_seen 1384) - (!srcfileloc "tree-core.h" 1793) + (!srcfileloc "tree-core.h" 1798) nil ) (!pair "record_layout_info" (!type already_seen 1404) - (!srcfileloc "tree-core.h" 1705) + (!srcfileloc "tree-core.h" 1710) nil ) (!pair "vec<tree,va_gc>" (!type already_seen 46) - (!srcfileloc "tree-core.h" 1228) + (!srcfileloc "tree-core.h" 1233) nil ) (!pair "constructor_elt" (!type already_seen 467) @@ -24117,7 +24125,7 @@ nil ) (!pair "machine_function" (!type already_seen 404) - (!srcfileloc "config/aarch64/aarch64.h" 534) + (!srcfileloc "config/aarch64/aarch64.h" 543) nil ) (!pair "gt_pointer_operator" (!type already_seen 9) @@ -25188,151 +25196,151 @@ nil ) (!pair "gfor_fndecl_sr_kind" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 773) + (!srcfileloc "fortran/trans.h" 778) nil ) (!pair "gfor_fndecl_si_kind" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 772) + (!srcfileloc "fortran/trans.h" 777) nil ) (!pair "gfor_fndecl_sc_kind" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 771) + (!srcfileloc "fortran/trans.h" 776) nil ) (!pair "gfor_fndecl_iargc" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 768) + (!srcfileloc "fortran/trans.h" 773) nil ) (!pair "gfor_fndecl_size1" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 767) + (!srcfileloc "fortran/trans.h" 772) nil ) (!pair "gfor_fndecl_size0" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 766) + (!srcfileloc "fortran/trans.h" 771) nil ) (!pair "gfor_fndecl_convert_char4_to_char1" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 763) + (!srcfileloc "fortran/trans.h" 768) nil ) (!pair "gfor_fndecl_convert_char1_to_char4" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 762) + (!srcfileloc "fortran/trans.h" 767) nil ) (!pair "gfor_fndecl_select_string_char4" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 759) + (!srcfileloc "fortran/trans.h" 764) nil ) (!pair "gfor_fndecl_adjustr_char4" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 758) + (!srcfileloc "fortran/trans.h" 763) nil ) (!pair "gfor_fndecl_adjustl_char4" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 757) + (!srcfileloc "fortran/trans.h" 762) nil ) (!pair "gfor_fndecl_string_minmax_char4" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 756) + (!srcfileloc "fortran/trans.h" 761) nil ) (!pair "gfor_fndecl_string_trim_char4" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 755) + (!srcfileloc "fortran/trans.h" 760) nil ) (!pair "gfor_fndecl_string_verify_char4" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 754) + (!srcfileloc "fortran/trans.h" 759) nil ) (!pair "gfor_fndecl_string_scan_char4" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 753) + (!srcfileloc "fortran/trans.h" 758) nil ) (!pair "gfor_fndecl_string_index_char4" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 752) + (!srcfileloc "fortran/trans.h" 757) nil ) (!pair "gfor_fndecl_string_len_trim_char4" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 751) + (!srcfileloc "fortran/trans.h" 756) nil ) (!pair "gfor_fndecl_concat_string_char4" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 750) + (!srcfileloc "fortran/trans.h" 755) nil ) (!pair "gfor_fndecl_compare_string_char4" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 749) + (!srcfileloc "fortran/trans.h" 754) nil ) (!pair "gfor_fndecl_select_string" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 748) + (!srcfileloc "fortran/trans.h" 753) nil ) (!pair "gfor_fndecl_adjustr" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 747) + (!srcfileloc "fortran/trans.h" 752) nil ) (!pair "gfor_fndecl_adjustl" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 746) + (!srcfileloc "fortran/trans.h" 751) nil ) (!pair "gfor_fndecl_string_minmax" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 745) + (!srcfileloc "fortran/trans.h" 750) nil ) (!pair "gfor_fndecl_string_trim" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 744) + (!srcfileloc "fortran/trans.h" 749) nil ) (!pair "gfor_fndecl_string_verify" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 743) + (!srcfileloc "fortran/trans.h" 748) nil ) (!pair "gfor_fndecl_string_scan" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 742) + (!srcfileloc "fortran/trans.h" 747) nil ) (!pair "gfor_fndecl_string_index" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 741) + (!srcfileloc "fortran/trans.h" 746) nil ) (!pair "gfor_fndecl_string_len_trim" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 740) + (!srcfileloc "fortran/trans.h" 745) nil ) (!pair "gfor_fndecl_concat_string" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 739) + (!srcfileloc "fortran/trans.h" 744) nil ) (!pair "gfor_fndecl_compare_string" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 738) + (!srcfileloc "fortran/trans.h" 743) nil ) (!pair "gfor_fndecl_zgemm" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 735) + (!srcfileloc "fortran/trans.h" 740) nil ) (!pair "gfor_fndecl_cgemm" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 734) + (!srcfileloc "fortran/trans.h" 739) nil ) (!pair "gfor_fndecl_dgemm" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 733) + (!srcfileloc "fortran/trans.h" 738) nil ) (!pair "gfor_fndecl_sgemm" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 732) + (!srcfileloc "fortran/trans.h" 737) nil ) (!pair "gfor_fndecl_math_ishftc16" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 729) + (!srcfileloc "fortran/trans.h" 734) nil ) (!pair "gfor_fndecl_math_ishftc8" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 728) + (!srcfileloc "fortran/trans.h" 733) nil ) (!pair "gfor_fndecl_math_ishftc4" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 727) + (!srcfileloc "fortran/trans.h" 732) nil ) (!pair "gfor_fndecl_math_powi" (!type array 1547 nil gc_used "4" @@ -25340,139 +25348,139 @@ (!type already_seen 653) ) ) - (!srcfileloc "fortran/trans.h" 726) + (!srcfileloc "fortran/trans.h" 731) nil ) (!pair "gfort_gvar_caf_this_image" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 713) + (!srcfileloc "fortran/trans.h" 718) nil ) (!pair "gfort_gvar_caf_num_images" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 712) + (!srcfileloc "fortran/trans.h" 717) nil ) (!pair "gfor_fndecl_caf_error_stop_str" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 709) + (!srcfileloc "fortran/trans.h" 714) nil ) (!pair "gfor_fndecl_caf_error_stop" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 708) + (!srcfileloc "fortran/trans.h" 713) nil ) (!pair "gfor_fndecl_caf_sync_images" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 707) + (!srcfileloc "fortran/trans.h" 712) nil ) (!pair "gfor_fndecl_caf_sync_all" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 706) + (!srcfileloc "fortran/trans.h" 711) nil ) (!pair "gfor_fndecl_caf_end_critical" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 705) + (!srcfileloc "fortran/trans.h" 710) nil ) (!pair "gfor_fndecl_caf_critical" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 704) + (!srcfileloc "fortran/trans.h" 709) nil ) (!pair "gfor_fndecl_caf_deregister" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 703) + (!srcfileloc "fortran/trans.h" 708) nil ) (!pair "gfor_fndecl_caf_register" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 702) + (!srcfileloc "fortran/trans.h" 707) nil ) (!pair "gfor_fndecl_caf_finalize" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 701) + (!srcfileloc "fortran/trans.h" 706) nil ) (!pair "gfor_fndecl_caf_init" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 700) + (!srcfileloc "fortran/trans.h" 705) nil ) (!pair "gfor_fndecl_associated" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 696) + (!srcfileloc "fortran/trans.h" 701) nil ) (!pair "gfor_fndecl_in_unpack" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 695) + (!srcfileloc "fortran/trans.h" 700) nil ) (!pair "gfor_fndecl_in_pack" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 694) + (!srcfileloc "fortran/trans.h" 699) nil ) (!pair "gfor_fndecl_fdate" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 693) + (!srcfileloc "fortran/trans.h" 698) nil ) (!pair "gfor_fndecl_ctime" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 692) + (!srcfileloc "fortran/trans.h" 697) nil ) (!pair "gfor_fndecl_ttynam" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 691) + (!srcfileloc "fortran/trans.h" 696) nil ) (!pair "gfor_fndecl_set_options" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 690) + (!srcfileloc "fortran/trans.h" 695) nil ) (!pair "gfor_fndecl_set_fpe" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 689) + (!srcfileloc "fortran/trans.h" 694) nil ) (!pair "gfor_fndecl_generate_error" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 688) + (!srcfileloc "fortran/trans.h" 693) nil ) (!pair "gfor_fndecl_os_error" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 687) + (!srcfileloc "fortran/trans.h" 692) nil ) (!pair "gfor_fndecl_runtime_warning_at" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 686) + (!srcfileloc "fortran/trans.h" 691) nil ) (!pair "gfor_fndecl_runtime_error_at" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 685) + (!srcfileloc "fortran/trans.h" 690) nil ) (!pair "gfor_fndecl_runtime_error" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 684) + (!srcfileloc "fortran/trans.h" 689) nil ) (!pair "gfor_fndecl_error_stop_string" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 683) + (!srcfileloc "fortran/trans.h" 688) nil ) (!pair "gfor_fndecl_error_stop_numeric" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 682) + (!srcfileloc "fortran/trans.h" 687) nil ) (!pair "gfor_fndecl_stop_string" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 681) + (!srcfileloc "fortran/trans.h" 686) nil ) (!pair "gfor_fndecl_stop_numeric_f08" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 680) + (!srcfileloc "fortran/trans.h" 685) nil ) (!pair "gfor_fndecl_stop_numeric" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 679) + (!srcfileloc "fortran/trans.h" 684) nil ) (!pair "gfor_fndecl_pause_string" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 678) + (!srcfileloc "fortran/trans.h" 683) nil ) (!pair "gfor_fndecl_pause_numeric" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 677) + (!srcfileloc "fortran/trans.h" 682) nil ) (!pair "gfc_static_ctors" (!type already_seen 5) - (!srcfileloc "fortran/trans.h" 582) + (!srcfileloc "fortran/trans.h" 585) nil ) (!pair "gfc_charlen_type_node" (!type already_seen 5) @@ -25574,7 +25582,7 @@ (!type array 1557 nil gc_used "2" (!type already_seen 5) ) - (!srcfileloc "fortran/trans-stmt.c" 2127) + (!srcfileloc "fortran/trans-stmt.c" 2130) nil ) (!pair "dt_parm" (!type already_seen 5) @@ -25606,7 +25614,7 @@ nil ) (!pair "module_htab" (!type already_seen 1472) - (!srcfileloc "fortran/trans-decl.c" 4069) + (!srcfileloc "fortran/trans-decl.c" 4112) (!options (!option param_is type (!type already_seen 668) @@ -25639,19 +25647,19 @@ nil ) (!pair "global_binding_level" (!type already_seen 80) - (!srcfileloc "fortran/f95-lang.c" 301) + (!srcfileloc "fortran/f95-lang.c" 325) nil ) (!pair "current_binding_level" (!type already_seen 80) - (!srcfileloc "fortran/f95-lang.c" 297) + (!srcfileloc "fortran/f95-lang.c" 321) nil ) (!pair "current_translation_unit" (!type already_seen 5) - (!srcfileloc "fortran/f95-lang.c" 162) + (!srcfileloc "fortran/f95-lang.c" 186) nil ) (!pair "free_binding_level" (!type already_seen 80) - (!srcfileloc "fortran/f95-lang.c" 156) + (!srcfileloc "fortran/f95-lang.c" 180) nil ) (!pair "vlt_register_set_fndecl" (!type already_seen 5) @@ -25706,23 +25714,23 @@ nil ) (!pair "generic_parm_count" (!type already_seen 9) - (!srcfileloc "cp/parser.c" 31972) + (!srcfileloc "cp/parser.c" 31989) nil ) (!pair "the_parser" (!type already_seen 690) - (!srcfileloc "cp/parser.c" 31395) + (!srcfileloc "cp/parser.c" 31412) nil ) (!pair "lambda_scope_stack" (!type already_seen 685) - (!srcfileloc "cp/parser.c" 8654) + (!srcfileloc "cp/parser.c" 8652) nil ) (!pair "lambda_count" (!type already_seen 9) - (!srcfileloc "cp/parser.c" 8648) + (!srcfileloc "cp/parser.c" 8646) nil ) (!pair "lambda_scope" (!type already_seen 5) - (!srcfileloc "cp/parser.c" 8647) + (!srcfileloc "cp/parser.c" 8645) nil ) (!pair "cp_parser_context_free_list" (!type already_seen 696) @@ -25733,7 +25741,7 @@ ) (!pair "list_hash_table" (!type already_seen 53) - (!srcfileloc "cp/tree.c" 1598) + (!srcfileloc "cp/tree.c" 1609) (!options (!option param_is type (!type already_seen 7) @@ -25742,7 +25750,7 @@ ) (!pair "cplus_array_htab" (!type already_seen 53) - (!srcfileloc "cp/tree.c" 746) + (!srcfileloc "cp/tree.c" 757) (!options (!option param_is type (!type already_seen 7) @@ -25751,7 +25759,7 @@ ) (!pair "constexpr_call_table" (!type already_seen 1476) - (!srcfileloc "cp/semantics.c" 8086) + (!srcfileloc "cp/semantics.c" 8088) (!options (!option param_is type (!type already_seen 709) @@ -25760,7 +25768,7 @@ ) (!pair "constexpr_fundef_table" (!type already_seen 1478) - (!srcfileloc "cp/semantics.c" 7403) + (!srcfileloc "cp/semantics.c" 7405) (!options (!option param_is type (!type already_seen 712) @@ -25781,7 +25789,7 @@ nil ) (!pair "last_error_tinst_level" (!type already_seen 725) - (!srcfileloc "cp/pt.c" 8230) + (!srcfileloc "cp/pt.c" 8235) nil ) (!pair "canonical_template_parms" (!type already_seen 45) @@ -25855,7 +25863,7 @@ nil ) (!pair "start_cleanup_cnt" (!type already_seen 9) - (!srcfileloc "cp/decl.c" 6739) + (!srcfileloc "cp/decl.c" 6746) nil ) (!pair "typename_htab" (!type already_seen 53) @@ -25872,57 +25880,57 @@ nil ) (!pair "java_iface_lookup_fn" (!type already_seen 5) - (!srcfileloc "cp/call.c" 7318) + (!srcfileloc "cp/call.c" 7341) nil ) (!pair "default_arg_context" (!type already_seen 45) - (!srcfileloc "cp/call.c" 6491) + (!srcfileloc "cp/call.c" 6505) nil ) (!pair "unemitted_tinfo_decls" (!type already_seen 45) - (!srcfileloc "cp/cp-tree.h" 5620) + (!srcfileloc "cp/cp-tree.h" 5631) nil ) (!pair "deferred_mark_used_calls" (!type already_seen 45) - (!srcfileloc "cp/cp-tree.h" 5291) + (!srcfileloc "cp/cp-tree.h" 5301) nil ) (!pair "assignment_operator_name_info" (!type array 1562 nil gc_used "(int) MAX_TREE_CODES" (!type already_seen 745) ) - (!srcfileloc "cp/cp-tree.h" 4736) + (!srcfileloc "cp/cp-tree.h" 4751) nil ) (!pair "operator_name_info" (!type array 1563 nil gc_used "(int) MAX_TREE_CODES" (!type already_seen 745) ) - (!srcfileloc "cp/cp-tree.h" 4733) + (!srcfileloc "cp/cp-tree.h" 4748) nil ) (!pair "tls_aggregates" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 4478) + (!srcfileloc "cp/cp-tree.h" 4493) nil ) (!pair "static_aggregates" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 4476) + (!srcfileloc "cp/cp-tree.h" 4491) nil ) (!pair "local_classes" (!type already_seen 45) - (!srcfileloc "cp/cp-tree.h" 4378) + (!srcfileloc "cp/cp-tree.h" 4393) nil ) (!pair "integer_two_node" (!type already_seen 5) - (!srcfileloc "cp/cp-tree.h" 4339) + (!srcfileloc "cp/cp-tree.h" 4348) nil ) (!pair "scope_chain" (!type already_seen 751) - (!srcfileloc "cp/cp-tree.h" 1112) + (!srcfileloc "cp/cp-tree.h" 1114) nil ) (!pair "cp_global_trees" (!type array 1564 nil gc_used "CPTI_MAX" (!type already_seen 5) ) - (!srcfileloc "cp/cp-tree.h" 902) + (!srcfileloc "cp/cp-tree.h" 904) nil ) (!pair "lambda_cnt" (!type already_seen 9) @@ -25971,11 +25979,11 @@ nil ) (!pair "temp_count" (!type already_seen 9) - (!srcfileloc "cp/mangle.c" 3858) + (!srcfileloc "cp/mangle.c" 3862) nil ) (!pair "conv_type_names" (!type already_seen 53) - (!srcfileloc "cp/mangle.c" 3730) + (!srcfileloc "cp/mangle.c" 3734) (!options (!option param_is type (!type already_seen 7) @@ -26036,7 +26044,7 @@ nil ) (!pair "pending_lang_change" (!type already_seen 9) - (!srcfileloc "c-family/c-common.h" 943) + (!srcfileloc "c-family/c-common.h" 944) nil ) (!pair "c_global_trees" (!type array 1567 nil gc_used "CTI_MAX" @@ -26053,24 +26061,24 @@ ) (!pair "tree_vector_cache" (!type already_seen 789) - (!srcfileloc "c-family/c-common.c" 11540) + (!srcfileloc "c-family/c-common.c" 11546) (!options (!option deletable string "") ) ) (!pair "optimize_args" (!type already_seen 792) - (!srcfileloc "c-family/c-common.c" 9005) + (!srcfileloc "c-family/c-common.c" 9011) nil ) (!pair "compound_literal_number" (!type already_seen 9) - (!srcfileloc "c-family/c-common.c" 5726) + (!srcfileloc "c-family/c-common.c" 5732) nil ) (!pair "built_in_attributes" (!type array 1568 nil gc_used "(int) ATTR_LAST" (!type already_seen 5) ) - (!srcfileloc "c-family/c-common.c" 5101) + (!srcfileloc "c-family/c-common.c" 5107) nil ) (!pair "type_hash_table" (!type already_seen 53) @@ -26387,7 +26395,7 @@ nil ) (!pair "clone_fn_id_num" (!type already_seen 9) - (!srcfileloc "cgraphclones.c" 479) + (!srcfileloc "cgraphclones.c" 491) nil ) (!pair "vtable_entry_type" (!type already_seen 5) @@ -26406,15 +26414,15 @@ ) (!pair "va_list_type" (!type already_seen 5) - (!srcfileloc "config/aarch64/aarch64.c" 5553) + (!srcfileloc "config/aarch64/aarch64.c" 6473) nil ) (!pair "tls_get_addr_libfunc" (!type already_seen 136) - (!srcfileloc "config/aarch64/aarch64.c" 492) + (!srcfileloc "config/aarch64/aarch64.c" 562) nil ) (!pair "gty_dummy" (!type already_seen 9) - (!srcfileloc "config/aarch64/aarch64.c" 310) + (!srcfileloc "config/aarch64/aarch64.c" 362) nil ) (!pair "stack_chk_fail_decl" (!type already_seen 5) @@ -26426,7 +26434,7 @@ nil ) (!pair "critical_name_mutexes" (!type already_seen 1488) - (!srcfileloc "omp-low.c" 8832) + (!srcfileloc "omp-low.c" 8883) (!options (!option param2_is type (!type already_seen 5) @@ -26607,7 +26615,7 @@ nil ) (!pair "gcc_eh_personality_decl" (!type already_seen 5) - (!srcfileloc "tree.c" 11746) + (!srcfileloc "tree.c" 11751) nil ) (!pair "nonstandard_integer_type_cache" (!type array 1579 nil gc_used "2 * MAX_INT_CACHED_PREC + 2" @@ -26795,7 +26803,7 @@ nil ) (!pair "next_block_index" (!type already_seen 9) - (!srcfileloc "function.c" 4327) + (!srcfileloc "function.c" 4332) nil ) (!pair "temp_slot_address_table" (!type already_seen 1508) @@ -27551,41 +27559,41 @@ nil ) (!pair "current_function_func_begin_label" (!type already_seen 36) - (!srcfileloc "tree-core.h" 1841) + (!srcfileloc "tree-core.h" 1846) nil ) (!pair "current_function_decl" (!type already_seen 5) - (!srcfileloc "tree-core.h" 1838) + (!srcfileloc "tree-core.h" 1843) nil ) (!pair "builtin_info" (!type already_seen 1387) - (!srcfileloc "tree-core.h" 1828) + (!srcfileloc "tree-core.h" 1833) nil ) (!pair "sizetype_tab" (!type array 1582 nil gc_used "(int) stk_type_kind_last" (!type already_seen 5) ) - (!srcfileloc "tree-core.h" 1816) + (!srcfileloc "tree-core.h" 1821) nil ) (!pair "integer_types" (!type array 1583 nil gc_used "itk_none" (!type already_seen 5) ) - (!srcfileloc "tree-core.h" 1813) + (!srcfileloc "tree-core.h" 1818) nil ) (!pair "global_trees" (!type array 1584 nil gc_used "TI_MAX" (!type already_seen 5) ) - (!srcfileloc "tree-core.h" 1809) + (!srcfileloc "tree-core.h" 1814) nil ) (!pair "all_translation_units" (!type already_seen 45) - (!srcfileloc "tree-core.h" 1806) + (!srcfileloc "tree-core.h" 1811) nil ) (!pair "alias_pairs" (!type already_seen 1385) - (!srcfileloc "tree-core.h" 1793) + (!srcfileloc "tree-core.h" 1798) nil ) (!pair "stack_limit_rtx" (!type already_seen 136) diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/auto-host.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/auto-host.h index b14aea0..c2d78a8 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/auto-host.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/auto-host.h @@ -1739,7 +1739,7 @@ /* Define to 1 if HOST_WIDE_INT must be 64 bits wide (see hwint.h). */ #ifndef USED_FOR_TARGET -#define NEED_64BIT_HOST_WIDE_INT 1 +/* #undef NEED_64BIT_HOST_WIDE_INT */ #endif diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/b-header-vars b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/b-header-vars index 458b1d5..6183b6f 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/b-header-vars +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/b-header-vars @@ -1,4 +1,4 @@ -USER_H=arm_neon.h +USER_H=arm_acle.h HASHTAB_H=hashtab.h OBSTACK_H=obstack.h SPLAY_TREE_H=splay-tree.h diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-family/c-common.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-family/c-common.h index 24959d8..06dcb7c 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-family/c-common.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-family/c-common.h @@ -758,6 +758,7 @@ extern tree c_wrap_maybe_const (tree, bool); extern tree c_save_expr (tree); extern tree c_common_truthvalue_conversion (location_t, tree); extern void c_apply_type_quals_to_decl (int, tree); +extern unsigned int min_align_of_type (tree); extern tree c_sizeof_or_alignof_type (location_t, tree, bool, bool, int); extern tree c_alignof_expr (location_t, tree); /* Print an error message for invalid operands to arith operation CODE. diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-tree.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-tree.h index 85df885..53768d6 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-tree.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-tree.h @@ -612,7 +612,8 @@ extern void push_init_level (int, struct obstack *); extern struct c_expr pop_init_level (int, struct obstack *); extern void set_init_index (tree, tree, struct obstack *); extern void set_init_label (tree, struct obstack *); -extern void process_init_element (struct c_expr, bool, struct obstack *); +extern void process_init_element (location_t, struct c_expr, bool, + struct obstack *); extern tree build_compound_literal (location_t, tree, tree, bool); extern void check_compound_literal_type (location_t, struct c_type_name *); extern tree c_start_case (location_t, location_t, tree); diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-linux.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-linux.h index dae41b0..f40a03e 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-linux.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-linux.h @@ -21,7 +21,7 @@ #ifndef GCC_AARCH64_LINUX_H #define GCC_AARCH64_LINUX_H -#define GLIBC_DYNAMIC_LINKER "/lib/ld-linux-aarch64%{mbig-endian:_be}.so.1" +#define GLIBC_DYNAMIC_LINKER "/lib/ld-linux-aarch64%{mbig-endian:_be}%{mabi=ilp32:_ilp32}.so.1" #undef BIONIC_DYNAMIC_LINKER #define BIONIC_DYNAMIC_LINKER "/system/bin/linker64" @@ -35,7 +35,7 @@ -dynamic-linker " GNU_USER_DYNAMIC_LINKER " \ -X \ %{mbig-endian:-EB} %{mlittle-endian:-EL} \ - -maarch64linux%{mbig-endian:b}" + -maarch64linux%{mabi=ilp32:32}%{mbig-endian:b}" #define TARGET_OS_CPP_BUILTINS() \ do \ diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-protos.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-protos.h index 04cbc78..c4f75b3 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-protos.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-protos.h @@ -108,9 +108,22 @@ enum aarch64_symbol_type cost models and vectors for address cost calculations, register move costs and memory move costs. */ +/* Scaled addressing modes can vary cost depending on the mode of the + value to be loaded/stored. QImode values cannot use scaled + addressing modes. */ + +struct scale_addr_mode_cost +{ + const int hi; + const int si; + const int di; + const int ti; +}; + /* Additional cost for addresses. */ struct cpu_addrcost_table { + const struct scale_addr_mode_cost addr_scale_costs; const int pre_modify; const int post_modify; const int register_offset; @@ -167,6 +180,7 @@ bool aarch64_cannot_change_mode_class (enum machine_mode, enum aarch64_symbol_type aarch64_classify_symbolic_expression (rtx, enum aarch64_symbol_context); bool aarch64_constant_address_p (rtx); +bool aarch64_expand_movmem (rtx *); bool aarch64_float_const_zero_rtx_p (rtx); bool aarch64_function_arg_regno_p (unsigned); bool aarch64_gen_movmemqi (rtx *); @@ -202,6 +216,8 @@ enum aarch64_symbol_type aarch64_classify_symbol (rtx, enum aarch64_symbol_type aarch64_classify_tls_symbol (rtx); enum reg_class aarch64_regno_regclass (unsigned); int aarch64_asm_preferred_eh_data_format (int, int); +enum machine_mode aarch64_hard_regno_caller_save_mode (unsigned, unsigned, + enum machine_mode); int aarch64_hard_regno_mode_ok (unsigned, enum machine_mode); int aarch64_hard_regno_nregs (unsigned, enum machine_mode); int aarch64_simd_attr_length_move (rtx); @@ -291,4 +307,5 @@ extern void aarch64_split_combinev16qi (rtx operands[3]); extern void aarch64_expand_vec_perm (rtx target, rtx op0, rtx op1, rtx sel); extern bool aarch64_expand_vec_perm_const (rtx target, rtx op0, rtx op1, rtx sel); +void aarch64_atomic_assign_expand_fenv (tree *, tree *, tree *); #endif /* GCC_AARCH64_PROTOS_H */ diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64.h index c9b30d0..e0384b3 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64.h @@ -35,6 +35,9 @@ if (TARGET_SIMD) \ builtin_define ("__ARM_NEON"); \ \ + if (TARGET_CRC32) \ + builtin_define ("__ARM_FEATURE_CRC32"); \ + \ switch (aarch64_cmodel) \ { \ case AARCH64_CMODEL_TINY: \ @@ -188,6 +191,9 @@ extern unsigned long aarch64_tune_flags; /* Crypto is an optional extension to AdvSIMD. */ #define TARGET_CRYPTO (TARGET_SIMD && AARCH64_ISA_CRYPTO) +/* CRC instructions that can be enabled through +crc arch extension. */ +#define TARGET_CRC32 (AARCH64_ISA_CRC) + /* Standard register usage. */ /* 31 64-bit general purpose registers R0-R30: @@ -408,6 +414,7 @@ extern unsigned long aarch64_tune_flags; enum reg_class { NO_REGS, + CALLER_SAVE_REGS, CORE_REGS, GENERAL_REGS, STACK_REG, @@ -423,6 +430,7 @@ enum reg_class #define REG_CLASS_NAMES \ { \ "NO_REGS", \ + "CALLER_SAVE_REGS", \ "CORE_REGS", \ "GENERAL_REGS", \ "STACK_REG", \ @@ -435,6 +443,7 @@ enum reg_class #define REG_CLASS_CONTENTS \ { \ { 0x00000000, 0x00000000, 0x00000000 }, /* NO_REGS */ \ + { 0x0007ffff, 0x00000000, 0x00000000 }, /* CALLER_SAVE_REGS */ \ { 0x7fffffff, 0x00000000, 0x00000003 }, /* CORE_REGS */ \ { 0x7fffffff, 0x00000000, 0x00000003 }, /* GENERAL_REGS */ \ { 0x80000000, 0x00000000, 0x00000000 }, /* STACK_REG */ \ @@ -659,12 +668,14 @@ do { \ /* The base cost overhead of a memcpy call, for MOVE_RATIO and friends. */ #define AARCH64_CALL_RATIO 8 -/* When optimizing for size, give a better estimate of the length of a memcpy - call, but use the default otherwise. But move_by_pieces_ninsns() counts - memory-to-memory moves, and we'll have to generate a load & store for each, - so halve the value to take that into account. */ +/* MOVE_RATIO dictates when we will use the move_by_pieces infrastructure. + move_by_pieces will continually copy the largest safe chunks. So a + 7-byte copy is a 4-byte + 2-byte + byte copy. This proves inefficient + for both size and speed of copy, so we will instead use the "movmem" + standard name to implement the copy. This logic does not apply when + targeting -mstrict-align, so keep a sensible default in that case. */ #define MOVE_RATIO(speed) \ - (((speed) ? 15 : AARCH64_CALL_RATIO) / 2) + (!STRICT_ALIGNMENT ? 2 : (((speed) ? 15 : AARCH64_CALL_RATIO) / 2)) /* For CLEAR_RATIO, when optimizing for size, give a better estimate of the length of a memset call, but use the default otherwise. */ @@ -824,6 +835,11 @@ do { \ #define SHIFT_COUNT_TRUNCATED !TARGET_SIMD +/* Choose appropriate mode for caller saves, so we do the minimum + required size of load/store. */ +#define HARD_REGNO_CALLER_SAVE_MODE(REGNO, NREGS, MODE) \ + aarch64_hard_regno_caller_save_mode ((REGNO), (NREGS), (MODE)) + /* Callee only saves lower 64-bits of a 128-bit register. Tell the compiler the callee clobbers the top 64-bits when restoring the bottom 64-bits. */ diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/configargs.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/configargs.h index dbe4a51..1b088a2 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/configargs.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/configargs.h @@ -1,5 +1,5 @@ /* Generated automatically. */ -static const char configuration_arguments[] = "/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/configure --prefix=/tmp/android-toolchain-eabi --host=x86_64-linux-gnu --build=x86_64-linux-gnu --with-gnu-as --with-gnu-ld --enable-languages=c,c++ --with-gmp=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/temp-install --with-mpfr=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/temp-install --with-mpc=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/temp-install --disable-libssp --disable-nls --disable-libmudflap --disable-shared --disable-sjlj-exceptions --disable-libquadmath --disable-libitm --disable-libsanitizer --prefix=/tmp/android-toolchain-eabi --disable-docs --disable-nls --with-gcc-version=linaro-4.9-2014.06 --with-binutils-version=current --with-gmp-version=current --with-mpfr-version=current --with-mpc-version=current --disable-graphite --target=aarch64-linux-android --enable-threads --enable-tls --enable-libgomp --disable-libitm --disable-libsanitizer --enable-libatomic --disable-libstdc__-v3 : (reconfigured) /mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/configure --prefix=/tmp/android-toolchain-eabi --host=x86_64-linux-gnu --build=x86_64-linux-gnu --with-gnu-as --with-gnu-ld --enable-languages=c,c++ --with-gmp=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/temp-install --with-mpfr=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/temp-install --with-mpc=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/temp-install --disable-libssp --disable-nls --disable-libmudflap --disable-shared --disable-sjlj-exceptions --disable-libquadmath --disable-libitm --disable-libsanitizer --prefix=/tmp/android-toolchain-eabi --disable-docs --disable-nls --with-gcc-version=linaro-4.9-2014.06 --with-binutils-version=current --with-gmp-version=current --with-mpfr-version=current --with-mpc-version=current --disable-graphite --target=aarch64-linux-android --enable-threads --enable-tls --enable-libgomp --disable-libitm --disable-libsanitizer --enable-libatomic --disable-libstdc__-v3"; +static const char configuration_arguments[] = "/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/configure --prefix=/tmp/android-toolchain-eabi --host=x86_64-linux-gnu --build=x86_64-linux-gnu --with-gnu-as --with-gnu-ld --enable-languages=c,c++ --with-gmp=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/temp-install --with-mpfr=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/temp-install --with-mpc=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/temp-install --disable-libssp --disable-nls --disable-libmudflap --disable-shared --disable-sjlj-exceptions --disable-libquadmath --disable-libitm --disable-libsanitizer --prefix=/tmp/android-toolchain-eabi --disable-docs --disable-nls --with-gcc-version=linaro-4.9-2014.07 --with-binutils-version=current --with-gmp-version=current --with-mpfr-version=current --with-mpc-version=current --disable-graphite --target=aarch64-linux-android --enable-threads --enable-tls --disable-libgomp --disable-libstdc__-v3 --disable-libitm --disable-libsanitizer --disable-libatomic"; static const char thread_model[] = "posix"; static const struct { diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/cp/cp-tree.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/cp/cp-tree.h index e9fe86e..26a63d0 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/cp/cp-tree.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/cp/cp-tree.h @@ -101,12 +101,14 @@ c-common.h, not after. FNDECL_USED_AUTO (in FUNCTION_DECL) DECLTYPE_FOR_LAMBDA_PROXY (in DECLTYPE_TYPE) REF_PARENTHESIZED_P (in COMPONENT_REF, SCOPE_REF) + AGGR_INIT_ZERO_FIRST (in AGGR_INIT_EXPR) 3: (TREE_REFERENCE_EXPR) (in NON_LVALUE_EXPR) (commented-out). ICS_BAD_FLAG (in _CONV) FN_TRY_BLOCK_P (in TRY_BLOCK) IDENTIFIER_CTOR_OR_DTOR_P (in IDENTIFIER_NODE) BIND_EXPR_BODY_BLOCK (in BIND_EXPR) DECL_NON_TRIVIALLY_INITIALIZED_P (in VAR_DECL) + CALL_EXPR_LIST_INIT_P (in CALL_EXPR, AGGR_INIT_EXPR) 4: TREE_HAS_CONSTRUCTOR (in INDIRECT_REF, SAVE_EXPR, CONSTRUCTOR, or FIELD_DECL). IDENTIFIER_TYPENAME_P (in IDENTIFIER_NODE) @@ -3026,6 +3028,10 @@ extern void decl_shadowed_for_var_insert (tree, tree); should be performed at instantiation time. */ #define KOENIG_LOOKUP_P(NODE) TREE_LANG_FLAG_0 (CALL_EXPR_CHECK (NODE)) +/* True if CALL_EXPR expresses list-initialization of an object. */ +#define CALL_EXPR_LIST_INIT_P(NODE) \ + TREE_LANG_FLAG_3 (TREE_CHECK2 ((NODE),CALL_EXPR,AGGR_INIT_EXPR)) + /* Indicates whether a string literal has been parenthesized. Such usages are disallowed in certain circumstances. */ @@ -3430,6 +3436,9 @@ more_aggr_init_expr_args_p (const aggr_init_expr_arg_iterator *iter) B b{1,2}, not B b({1,2}) or B b = {1,2}. */ #define CONSTRUCTOR_IS_DIRECT_INIT(NODE) (TREE_LANG_FLAG_0 (CONSTRUCTOR_CHECK (NODE))) +#define DIRECT_LIST_INIT_P(NODE) \ + (BRACE_ENCLOSED_INITIALIZER_P (NODE) && CONSTRUCTOR_IS_DIRECT_INIT (NODE)) + /* True if NODE represents a conversion for direct-initialization in a template. Set by perform_implicit_conversion_flags. */ #define IMPLICIT_CONV_EXPR_DIRECT_INIT(NODE) \ @@ -4347,6 +4356,11 @@ extern int function_depth; PARM_DECLs in cp_tree_equal. */ extern int comparing_specializations; +/* A type-qualifier, or bitmask therefore, using the TYPE_QUAL + constants. */ + +typedef int cp_cv_quals; + /* In parser.c. */ /* Nonzero if we are parsing an unevaluated operand: an operand to @@ -4356,6 +4370,7 @@ extern int comparing_specializations; extern int cp_unevaluated_operand; extern tree cp_convert_range_for (tree, tree, tree, bool); extern bool parsing_nsdmi (void); +extern void inject_this_parameter (tree, cp_cv_quals); /* in pt.c */ @@ -4735,11 +4750,6 @@ extern GTY(()) operator_name_info_t operator_name_info extern GTY(()) operator_name_info_t assignment_operator_name_info [(int) MAX_TREE_CODES]; -/* A type-qualifier, or bitmask therefore, using the TYPE_QUAL - constants. */ - -typedef int cp_cv_quals; - /* Non-static member functions have an optional virt-specifier-seq. There is a VIRT_SPEC value for each virt-specifier. They can be combined by bitwise-or to form the complete set of @@ -5415,6 +5425,7 @@ extern tree get_type_value (tree); extern tree build_zero_init (tree, tree, bool); extern tree build_value_init (tree, tsubst_flags_t); extern tree build_value_init_noctor (tree, tsubst_flags_t); +extern tree get_nsdmi (tree, bool); extern tree build_offset_ref (tree, tree, bool, tsubst_flags_t); extern tree throw_bad_array_new_length (void); @@ -6151,6 +6162,7 @@ extern tree store_init_value (tree, tree, vec<tree, va_gc>**, int); extern void check_narrowing (tree, tree); extern tree digest_init (tree, tree, tsubst_flags_t); extern tree digest_init_flags (tree, tree, int); +extern tree digest_nsdmi_init (tree, tree); extern tree build_scoped_ref (tree, tree, tree *); extern tree build_x_arrow (location_t, tree, tsubst_flags_t); @@ -6201,7 +6213,7 @@ extern tree cxx_omp_clause_default_ctor (tree, tree, tree); extern tree cxx_omp_clause_copy_ctor (tree, tree, tree); extern tree cxx_omp_clause_assign_op (tree, tree, tree); extern tree cxx_omp_clause_dtor (tree, tree); -extern void cxx_omp_finish_clause (tree); +extern void cxx_omp_finish_clause (tree, gimple_seq *); extern bool cxx_omp_privatize_by_reference (const_tree); /* in name-lookup.c */ diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/gimplify.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/gimplify.h index 47e7213..5085ccf 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/gimplify.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/gimplify.h @@ -77,6 +77,7 @@ extern enum gimplify_status gimplify_expr (tree *, gimple_seq *, gimple_seq *, extern void gimplify_type_sizes (tree, gimple_seq *); extern void gimplify_one_sizepos (tree *, gimple_seq *); extern gimple gimplify_body (tree, bool); +extern enum gimplify_status gimplify_arg (tree *, gimple_seq *, location_t); extern void gimplify_function_tree (tree); extern enum gimplify_status gimplify_va_arg_expr (tree *, gimple_seq *, gimple_seq *); diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-codes.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-codes.h index 3385748..9eca9bb 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-codes.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-codes.h @@ -64,2224 +64,2340 @@ enum insn_code { CODE_FOR_cstorehi_neg = 359, CODE_FOR_cstoresi_neg = 360, CODE_FOR_cstoredi_neg = 361, - CODE_FOR_csinc3si_insn = 372, - CODE_FOR_csinc3di_insn = 373, - CODE_FOR_andsi3 = 378, - CODE_FOR_iorsi3 = 379, - CODE_FOR_xorsi3 = 380, - CODE_FOR_anddi3 = 381, - CODE_FOR_iordi3 = 382, - CODE_FOR_xordi3 = 383, - CODE_FOR_one_cmplsi2 = 447, - CODE_FOR_one_cmpldi2 = 448, - CODE_FOR_clzsi2 = 502, - CODE_FOR_clzdi2 = 503, - CODE_FOR_clrsbsi2 = 504, - CODE_FOR_clrsbdi2 = 505, - CODE_FOR_rbitsi2 = 506, - CODE_FOR_rbitdi2 = 507, - CODE_FOR_bswapsi2 = 587, - CODE_FOR_bswapdi2 = 588, - CODE_FOR_bswaphi2 = 589, - CODE_FOR_rev16si2 = 590, - CODE_FOR_rev16di2 = 591, - CODE_FOR_rev16si2_alt = 592, - CODE_FOR_rev16di2_alt = 593, - CODE_FOR_btruncsf2 = 595, - CODE_FOR_ceilsf2 = 596, - CODE_FOR_floorsf2 = 597, - CODE_FOR_frintnsf2 = 598, - CODE_FOR_nearbyintsf2 = 599, - CODE_FOR_rintsf2 = 600, - CODE_FOR_roundsf2 = 601, - CODE_FOR_btruncdf2 = 602, - CODE_FOR_ceildf2 = 603, - CODE_FOR_floordf2 = 604, - CODE_FOR_frintndf2 = 605, - CODE_FOR_nearbyintdf2 = 606, - CODE_FOR_rintdf2 = 607, - CODE_FOR_rounddf2 = 608, - CODE_FOR_lbtruncsfsi2 = 609, - CODE_FOR_lceilsfsi2 = 610, - CODE_FOR_lfloorsfsi2 = 611, - CODE_FOR_lroundsfsi2 = 612, - CODE_FOR_lfrintnsfsi2 = 613, - CODE_FOR_lbtruncusfsi2 = 614, - CODE_FOR_lceilusfsi2 = 615, - CODE_FOR_lfloorusfsi2 = 616, - CODE_FOR_lroundusfsi2 = 617, - CODE_FOR_lfrintnusfsi2 = 618, - CODE_FOR_lbtruncdfsi2 = 619, - CODE_FOR_lceildfsi2 = 620, - CODE_FOR_lfloordfsi2 = 621, - CODE_FOR_lrounddfsi2 = 622, - CODE_FOR_lfrintndfsi2 = 623, - CODE_FOR_lbtruncudfsi2 = 624, - CODE_FOR_lceiludfsi2 = 625, - CODE_FOR_lfloorudfsi2 = 626, - CODE_FOR_lroundudfsi2 = 627, - CODE_FOR_lfrintnudfsi2 = 628, - CODE_FOR_lbtruncsfdi2 = 629, - CODE_FOR_lceilsfdi2 = 630, - CODE_FOR_lfloorsfdi2 = 631, - CODE_FOR_lroundsfdi2 = 632, - CODE_FOR_lfrintnsfdi2 = 633, - CODE_FOR_lbtruncusfdi2 = 634, - CODE_FOR_lceilusfdi2 = 635, - CODE_FOR_lfloorusfdi2 = 636, - CODE_FOR_lroundusfdi2 = 637, - CODE_FOR_lfrintnusfdi2 = 638, - CODE_FOR_lbtruncdfdi2 = 639, - CODE_FOR_lceildfdi2 = 640, - CODE_FOR_lfloordfdi2 = 641, - CODE_FOR_lrounddfdi2 = 642, - CODE_FOR_lfrintndfdi2 = 643, - CODE_FOR_lbtruncudfdi2 = 644, - CODE_FOR_lceiludfdi2 = 645, - CODE_FOR_lfloorudfdi2 = 646, - CODE_FOR_lroundudfdi2 = 647, - CODE_FOR_lfrintnudfdi2 = 648, - CODE_FOR_fmasf4 = 649, - CODE_FOR_fmadf4 = 650, - CODE_FOR_fnmasf4 = 651, - CODE_FOR_fnmadf4 = 652, - CODE_FOR_fmssf4 = 653, - CODE_FOR_fmsdf4 = 654, - CODE_FOR_fnmssf4 = 655, - CODE_FOR_fnmsdf4 = 656, - CODE_FOR_extendsfdf2 = 659, - CODE_FOR_truncdfsf2 = 660, - CODE_FOR_fix_truncsfsi2 = 661, - CODE_FOR_fix_truncdfsi2 = 662, - CODE_FOR_fix_truncsfdi2 = 663, - CODE_FOR_fix_truncdfdi2 = 664, - CODE_FOR_fixuns_truncsfsi2 = 665, - CODE_FOR_fixuns_truncdfsi2 = 666, - CODE_FOR_fixuns_truncsfdi2 = 667, - CODE_FOR_fixuns_truncdfdi2 = 668, - CODE_FOR_floatsisf2 = 669, - CODE_FOR_floatsidf2 = 670, - CODE_FOR_floatdisf2 = 671, - CODE_FOR_floatdidf2 = 672, - CODE_FOR_floatunssisf2 = 673, - CODE_FOR_floatunssidf2 = 674, - CODE_FOR_floatunsdisf2 = 675, - CODE_FOR_floatunsdidf2 = 676, - CODE_FOR_addsf3 = 677, - CODE_FOR_adddf3 = 678, - CODE_FOR_subsf3 = 679, - CODE_FOR_subdf3 = 680, - CODE_FOR_mulsf3 = 681, - CODE_FOR_muldf3 = 682, - CODE_FOR_divsf3 = 685, - CODE_FOR_divdf3 = 686, - CODE_FOR_negsf2 = 687, - CODE_FOR_negdf2 = 688, - CODE_FOR_sqrtsf2 = 689, - CODE_FOR_sqrtdf2 = 690, - CODE_FOR_abssf2 = 691, - CODE_FOR_absdf2 = 692, - CODE_FOR_smaxsf3 = 693, - CODE_FOR_smaxdf3 = 694, - CODE_FOR_sminsf3 = 695, - CODE_FOR_smindf3 = 696, - CODE_FOR_aarch64_movdi_tilow = 697, - CODE_FOR_aarch64_movdi_tflow = 698, - CODE_FOR_aarch64_movdi_tihigh = 699, - CODE_FOR_aarch64_movdi_tfhigh = 700, - CODE_FOR_aarch64_movtihigh_di = 701, - CODE_FOR_aarch64_movtfhigh_di = 702, - CODE_FOR_aarch64_movtilow_di = 703, - CODE_FOR_aarch64_movtflow_di = 704, - CODE_FOR_aarch64_movtilow_tilow = 705, - CODE_FOR_add_losym_si = 706, - CODE_FOR_add_losym_di = 707, - CODE_FOR_ldr_got_small_si = 708, - CODE_FOR_ldr_got_small_di = 709, - CODE_FOR_ldr_got_small_sidi = 710, - CODE_FOR_ldr_got_tiny = 711, - CODE_FOR_aarch64_load_tp_hard = 712, - CODE_FOR_tlsie_small_si = 714, - CODE_FOR_tlsie_small_di = 715, - CODE_FOR_tlsie_small_sidi = 716, - CODE_FOR_tlsle_small_si = 717, - CODE_FOR_tlsle_small_di = 718, - CODE_FOR_tlsdesc_small_si = 719, - CODE_FOR_tlsdesc_small_di = 720, - CODE_FOR_stack_tie = 721, - CODE_FOR_stack_protect_set_si = 722, - CODE_FOR_stack_protect_set_di = 723, - CODE_FOR_stack_protect_test_si = 724, - CODE_FOR_stack_protect_test_di = 725, - CODE_FOR_aarch64_simd_dupv8qi = 726, - CODE_FOR_aarch64_simd_dupv16qi = 727, - CODE_FOR_aarch64_simd_dupv4hi = 728, - CODE_FOR_aarch64_simd_dupv8hi = 729, - CODE_FOR_aarch64_simd_dupv2si = 730, - CODE_FOR_aarch64_simd_dupv4si = 731, - CODE_FOR_aarch64_simd_dupv2di = 732, - CODE_FOR_aarch64_simd_dupv2sf = 733, - CODE_FOR_aarch64_simd_dupv4sf = 734, - CODE_FOR_aarch64_simd_dupv2df = 735, - CODE_FOR_aarch64_dup_lanev8qi = 736, - CODE_FOR_aarch64_dup_lanev16qi = 737, - CODE_FOR_aarch64_dup_lanev4hi = 738, - CODE_FOR_aarch64_dup_lanev8hi = 739, - CODE_FOR_aarch64_dup_lanev2si = 740, - CODE_FOR_aarch64_dup_lanev4si = 741, - CODE_FOR_aarch64_dup_lanev2di = 742, - CODE_FOR_aarch64_dup_lanev2sf = 743, - CODE_FOR_aarch64_dup_lanev4sf = 744, - CODE_FOR_aarch64_dup_lanev2df = 745, - CODE_FOR_aarch64_dup_lane_to_128v8qi = 746, - CODE_FOR_aarch64_dup_lane_to_64v16qi = 747, - CODE_FOR_aarch64_dup_lane_to_128v4hi = 748, - CODE_FOR_aarch64_dup_lane_to_64v8hi = 749, - CODE_FOR_aarch64_dup_lane_to_128v2si = 750, - CODE_FOR_aarch64_dup_lane_to_64v4si = 751, - CODE_FOR_aarch64_dup_lane_to_64v2di = 752, - CODE_FOR_aarch64_dup_lane_to_128v2sf = 753, - CODE_FOR_aarch64_dup_lane_to_64v4sf = 754, - CODE_FOR_aarch64_dup_lane_to_64v2df = 755, - CODE_FOR_aarch64_simd_mov_from_v16qilow = 766, - CODE_FOR_aarch64_simd_mov_from_v8hilow = 767, - CODE_FOR_aarch64_simd_mov_from_v4silow = 768, - CODE_FOR_aarch64_simd_mov_from_v2dilow = 769, - CODE_FOR_aarch64_simd_mov_from_v4sflow = 770, - CODE_FOR_aarch64_simd_mov_from_v2dflow = 771, - CODE_FOR_aarch64_simd_mov_from_v16qihigh = 772, - CODE_FOR_aarch64_simd_mov_from_v8hihigh = 773, - CODE_FOR_aarch64_simd_mov_from_v4sihigh = 774, - CODE_FOR_aarch64_simd_mov_from_v2dihigh = 775, - CODE_FOR_aarch64_simd_mov_from_v4sfhigh = 776, - CODE_FOR_aarch64_simd_mov_from_v2dfhigh = 777, - CODE_FOR_ornv8qi3 = 778, - CODE_FOR_ornv16qi3 = 779, - CODE_FOR_ornv4hi3 = 780, - CODE_FOR_ornv8hi3 = 781, - CODE_FOR_ornv2si3 = 782, - CODE_FOR_ornv4si3 = 783, - CODE_FOR_ornv2di3 = 784, - CODE_FOR_bicv8qi3 = 785, - CODE_FOR_bicv16qi3 = 786, - CODE_FOR_bicv4hi3 = 787, - CODE_FOR_bicv8hi3 = 788, - CODE_FOR_bicv2si3 = 789, - CODE_FOR_bicv4si3 = 790, - CODE_FOR_bicv2di3 = 791, - CODE_FOR_addv8qi3 = 792, - CODE_FOR_addv16qi3 = 793, - CODE_FOR_addv4hi3 = 794, - CODE_FOR_addv8hi3 = 795, - CODE_FOR_addv2si3 = 796, - CODE_FOR_addv4si3 = 797, - CODE_FOR_addv2di3 = 798, - CODE_FOR_subv8qi3 = 799, - CODE_FOR_subv16qi3 = 800, - CODE_FOR_subv4hi3 = 801, - CODE_FOR_subv8hi3 = 802, - CODE_FOR_subv2si3 = 803, - CODE_FOR_subv4si3 = 804, - CODE_FOR_subv2di3 = 805, - CODE_FOR_mulv8qi3 = 806, - CODE_FOR_mulv16qi3 = 807, - CODE_FOR_mulv4hi3 = 808, - CODE_FOR_mulv8hi3 = 809, - CODE_FOR_mulv2si3 = 810, - CODE_FOR_mulv4si3 = 811, - CODE_FOR_bswapv4hi = 812, - CODE_FOR_bswapv8hi = 813, - CODE_FOR_bswapv2si = 814, - CODE_FOR_bswapv4si = 815, - CODE_FOR_bswapv2di = 816, - CODE_FOR_negv8qi2 = 832, - CODE_FOR_negv16qi2 = 833, - CODE_FOR_negv4hi2 = 834, - CODE_FOR_negv8hi2 = 835, - CODE_FOR_negv2si2 = 836, - CODE_FOR_negv4si2 = 837, - CODE_FOR_negv2di2 = 838, - CODE_FOR_absv8qi2 = 839, - CODE_FOR_absv16qi2 = 840, - CODE_FOR_absv4hi2 = 841, - CODE_FOR_absv8hi2 = 842, - CODE_FOR_absv2si2 = 843, - CODE_FOR_absv4si2 = 844, - CODE_FOR_absv2di2 = 845, - CODE_FOR_abdv8qi_3 = 846, - CODE_FOR_abdv16qi_3 = 847, - CODE_FOR_abdv4hi_3 = 848, - CODE_FOR_abdv8hi_3 = 849, - CODE_FOR_abdv2si_3 = 850, - CODE_FOR_abdv4si_3 = 851, - CODE_FOR_abav8qi_3 = 852, - CODE_FOR_abav16qi_3 = 853, - CODE_FOR_abav4hi_3 = 854, - CODE_FOR_abav8hi_3 = 855, - CODE_FOR_abav2si_3 = 856, - CODE_FOR_abav4si_3 = 857, - CODE_FOR_fabdv2sf_3 = 858, - CODE_FOR_fabdv4sf_3 = 859, - CODE_FOR_fabdv2df_3 = 860, - CODE_FOR_andv8qi3 = 863, - CODE_FOR_andv16qi3 = 864, - CODE_FOR_andv4hi3 = 865, - CODE_FOR_andv8hi3 = 866, - CODE_FOR_andv2si3 = 867, - CODE_FOR_andv4si3 = 868, - CODE_FOR_andv2di3 = 869, - CODE_FOR_iorv8qi3 = 870, - CODE_FOR_iorv16qi3 = 871, - CODE_FOR_iorv4hi3 = 872, - CODE_FOR_iorv8hi3 = 873, - CODE_FOR_iorv2si3 = 874, - CODE_FOR_iorv4si3 = 875, - CODE_FOR_iorv2di3 = 876, - CODE_FOR_xorv8qi3 = 877, - CODE_FOR_xorv16qi3 = 878, - CODE_FOR_xorv4hi3 = 879, - CODE_FOR_xorv8hi3 = 880, - CODE_FOR_xorv2si3 = 881, - CODE_FOR_xorv4si3 = 882, - CODE_FOR_xorv2di3 = 883, - CODE_FOR_one_cmplv8qi2 = 884, - CODE_FOR_one_cmplv16qi2 = 885, - CODE_FOR_one_cmplv4hi2 = 886, - CODE_FOR_one_cmplv8hi2 = 887, - CODE_FOR_one_cmplv2si2 = 888, - CODE_FOR_one_cmplv4si2 = 889, - CODE_FOR_one_cmplv2di2 = 890, - CODE_FOR_aarch64_simd_vec_setv8qi = 891, - CODE_FOR_aarch64_simd_vec_setv16qi = 892, - CODE_FOR_aarch64_simd_vec_setv4hi = 893, - CODE_FOR_aarch64_simd_vec_setv8hi = 894, - CODE_FOR_aarch64_simd_vec_setv2si = 895, - CODE_FOR_aarch64_simd_vec_setv4si = 896, - CODE_FOR_aarch64_simd_lshrv8qi = 897, - CODE_FOR_aarch64_simd_lshrv16qi = 898, - CODE_FOR_aarch64_simd_lshrv4hi = 899, - CODE_FOR_aarch64_simd_lshrv8hi = 900, - CODE_FOR_aarch64_simd_lshrv2si = 901, - CODE_FOR_aarch64_simd_lshrv4si = 902, - CODE_FOR_aarch64_simd_lshrv2di = 903, - CODE_FOR_aarch64_simd_ashrv8qi = 904, - CODE_FOR_aarch64_simd_ashrv16qi = 905, - CODE_FOR_aarch64_simd_ashrv4hi = 906, - CODE_FOR_aarch64_simd_ashrv8hi = 907, - CODE_FOR_aarch64_simd_ashrv2si = 908, - CODE_FOR_aarch64_simd_ashrv4si = 909, - CODE_FOR_aarch64_simd_ashrv2di = 910, - CODE_FOR_aarch64_simd_imm_shlv8qi = 911, - CODE_FOR_aarch64_simd_imm_shlv16qi = 912, - CODE_FOR_aarch64_simd_imm_shlv4hi = 913, - CODE_FOR_aarch64_simd_imm_shlv8hi = 914, - CODE_FOR_aarch64_simd_imm_shlv2si = 915, - CODE_FOR_aarch64_simd_imm_shlv4si = 916, - CODE_FOR_aarch64_simd_imm_shlv2di = 917, - CODE_FOR_aarch64_simd_reg_sshlv8qi = 918, - CODE_FOR_aarch64_simd_reg_sshlv16qi = 919, - CODE_FOR_aarch64_simd_reg_sshlv4hi = 920, - CODE_FOR_aarch64_simd_reg_sshlv8hi = 921, - CODE_FOR_aarch64_simd_reg_sshlv2si = 922, - CODE_FOR_aarch64_simd_reg_sshlv4si = 923, - CODE_FOR_aarch64_simd_reg_sshlv2di = 924, - CODE_FOR_aarch64_simd_reg_shlv8qi_unsigned = 925, - CODE_FOR_aarch64_simd_reg_shlv16qi_unsigned = 926, - CODE_FOR_aarch64_simd_reg_shlv4hi_unsigned = 927, - CODE_FOR_aarch64_simd_reg_shlv8hi_unsigned = 928, - CODE_FOR_aarch64_simd_reg_shlv2si_unsigned = 929, - CODE_FOR_aarch64_simd_reg_shlv4si_unsigned = 930, - CODE_FOR_aarch64_simd_reg_shlv2di_unsigned = 931, - CODE_FOR_aarch64_simd_reg_shlv8qi_signed = 932, - CODE_FOR_aarch64_simd_reg_shlv16qi_signed = 933, - CODE_FOR_aarch64_simd_reg_shlv4hi_signed = 934, - CODE_FOR_aarch64_simd_reg_shlv8hi_signed = 935, - CODE_FOR_aarch64_simd_reg_shlv2si_signed = 936, - CODE_FOR_aarch64_simd_reg_shlv4si_signed = 937, - CODE_FOR_aarch64_simd_reg_shlv2di_signed = 938, - CODE_FOR_aarch64_sshr_simddi = 939, - CODE_FOR_aarch64_ushr_simddi = 940, - CODE_FOR_aarch64_simd_vec_setv2di = 941, - CODE_FOR_aarch64_simd_vec_setv2sf = 942, - CODE_FOR_aarch64_simd_vec_setv4sf = 943, - CODE_FOR_aarch64_simd_vec_setv2df = 944, - CODE_FOR_aarch64_mlav8qi = 945, - CODE_FOR_aarch64_mlav16qi = 946, - CODE_FOR_aarch64_mlav4hi = 947, - CODE_FOR_aarch64_mlav8hi = 948, - CODE_FOR_aarch64_mlav2si = 949, - CODE_FOR_aarch64_mlav4si = 950, - CODE_FOR_aarch64_mlsv8qi = 959, - CODE_FOR_aarch64_mlsv16qi = 960, - CODE_FOR_aarch64_mlsv4hi = 961, - CODE_FOR_aarch64_mlsv8hi = 962, - CODE_FOR_aarch64_mlsv2si = 963, - CODE_FOR_aarch64_mlsv4si = 964, - CODE_FOR_smaxv8qi3 = 973, - CODE_FOR_sminv8qi3 = 974, - CODE_FOR_umaxv8qi3 = 975, - CODE_FOR_uminv8qi3 = 976, - CODE_FOR_smaxv16qi3 = 977, - CODE_FOR_sminv16qi3 = 978, - CODE_FOR_umaxv16qi3 = 979, - CODE_FOR_uminv16qi3 = 980, - CODE_FOR_smaxv4hi3 = 981, - CODE_FOR_sminv4hi3 = 982, - CODE_FOR_umaxv4hi3 = 983, - CODE_FOR_uminv4hi3 = 984, - CODE_FOR_smaxv8hi3 = 985, - CODE_FOR_sminv8hi3 = 986, - CODE_FOR_umaxv8hi3 = 987, - CODE_FOR_uminv8hi3 = 988, - CODE_FOR_smaxv2si3 = 989, - CODE_FOR_sminv2si3 = 990, - CODE_FOR_umaxv2si3 = 991, - CODE_FOR_uminv2si3 = 992, - CODE_FOR_smaxv4si3 = 993, - CODE_FOR_sminv4si3 = 994, - CODE_FOR_umaxv4si3 = 995, - CODE_FOR_uminv4si3 = 996, - CODE_FOR_move_lo_quad_v16qi = 997, - CODE_FOR_move_lo_quad_v8hi = 998, - CODE_FOR_move_lo_quad_v4si = 999, - CODE_FOR_move_lo_quad_v2di = 1000, - CODE_FOR_move_lo_quad_v4sf = 1001, - CODE_FOR_move_lo_quad_v2df = 1002, - CODE_FOR_aarch64_simd_move_hi_quad_v16qi = 1003, - CODE_FOR_aarch64_simd_move_hi_quad_v8hi = 1004, - CODE_FOR_aarch64_simd_move_hi_quad_v4si = 1005, - CODE_FOR_aarch64_simd_move_hi_quad_v2di = 1006, - CODE_FOR_aarch64_simd_move_hi_quad_v4sf = 1007, - CODE_FOR_aarch64_simd_move_hi_quad_v2df = 1008, - CODE_FOR_aarch64_simd_vec_pack_trunc_v8hi = 1009, - CODE_FOR_aarch64_simd_vec_pack_trunc_v4si = 1010, - CODE_FOR_aarch64_simd_vec_pack_trunc_v2di = 1011, - CODE_FOR_vec_pack_trunc_v8hi = 1012, - CODE_FOR_vec_pack_trunc_v4si = 1013, - CODE_FOR_vec_pack_trunc_v2di = 1014, - CODE_FOR_aarch64_simd_vec_unpacks_lo_v16qi = 1015, - CODE_FOR_aarch64_simd_vec_unpacku_lo_v16qi = 1016, - CODE_FOR_aarch64_simd_vec_unpacks_lo_v8hi = 1017, - CODE_FOR_aarch64_simd_vec_unpacku_lo_v8hi = 1018, - CODE_FOR_aarch64_simd_vec_unpacks_lo_v4si = 1019, - CODE_FOR_aarch64_simd_vec_unpacku_lo_v4si = 1020, - CODE_FOR_aarch64_simd_vec_unpacks_hi_v16qi = 1021, - CODE_FOR_aarch64_simd_vec_unpacku_hi_v16qi = 1022, - CODE_FOR_aarch64_simd_vec_unpacks_hi_v8hi = 1023, - CODE_FOR_aarch64_simd_vec_unpacku_hi_v8hi = 1024, - CODE_FOR_aarch64_simd_vec_unpacks_hi_v4si = 1025, - CODE_FOR_aarch64_simd_vec_unpacku_hi_v4si = 1026, - CODE_FOR_aarch64_simd_vec_smult_lo_v16qi = 1063, - CODE_FOR_aarch64_simd_vec_umult_lo_v16qi = 1064, - CODE_FOR_aarch64_simd_vec_smult_lo_v8hi = 1065, - CODE_FOR_aarch64_simd_vec_umult_lo_v8hi = 1066, - CODE_FOR_aarch64_simd_vec_smult_lo_v4si = 1067, - CODE_FOR_aarch64_simd_vec_umult_lo_v4si = 1068, - CODE_FOR_aarch64_simd_vec_smult_hi_v16qi = 1069, - CODE_FOR_aarch64_simd_vec_umult_hi_v16qi = 1070, - CODE_FOR_aarch64_simd_vec_smult_hi_v8hi = 1071, - CODE_FOR_aarch64_simd_vec_umult_hi_v8hi = 1072, - CODE_FOR_aarch64_simd_vec_smult_hi_v4si = 1073, - CODE_FOR_aarch64_simd_vec_umult_hi_v4si = 1074, - CODE_FOR_addv2sf3 = 1075, - CODE_FOR_addv4sf3 = 1076, - CODE_FOR_addv2df3 = 1077, - CODE_FOR_subv2sf3 = 1078, - CODE_FOR_subv4sf3 = 1079, - CODE_FOR_subv2df3 = 1080, - CODE_FOR_mulv2sf3 = 1081, - CODE_FOR_mulv4sf3 = 1082, - CODE_FOR_mulv2df3 = 1083, - CODE_FOR_divv2sf3 = 1084, - CODE_FOR_divv4sf3 = 1085, - CODE_FOR_divv2df3 = 1086, - CODE_FOR_negv2sf2 = 1087, - CODE_FOR_negv4sf2 = 1088, - CODE_FOR_negv2df2 = 1089, - CODE_FOR_absv2sf2 = 1090, - CODE_FOR_absv4sf2 = 1091, - CODE_FOR_absv2df2 = 1092, - CODE_FOR_fmav2sf4 = 1093, - CODE_FOR_fmav4sf4 = 1094, - CODE_FOR_fmav2df4 = 1095, - CODE_FOR_fnmav2sf4 = 1103, - CODE_FOR_fnmav4sf4 = 1104, - CODE_FOR_fnmav2df4 = 1105, - CODE_FOR_btruncv2sf2 = 1113, - CODE_FOR_ceilv2sf2 = 1114, - CODE_FOR_floorv2sf2 = 1115, - CODE_FOR_frintnv2sf2 = 1116, - CODE_FOR_nearbyintv2sf2 = 1117, - CODE_FOR_rintv2sf2 = 1118, - CODE_FOR_roundv2sf2 = 1119, - CODE_FOR_btruncv4sf2 = 1120, - CODE_FOR_ceilv4sf2 = 1121, - CODE_FOR_floorv4sf2 = 1122, - CODE_FOR_frintnv4sf2 = 1123, - CODE_FOR_nearbyintv4sf2 = 1124, - CODE_FOR_rintv4sf2 = 1125, - CODE_FOR_roundv4sf2 = 1126, - CODE_FOR_btruncv2df2 = 1127, - CODE_FOR_ceilv2df2 = 1128, - CODE_FOR_floorv2df2 = 1129, - CODE_FOR_frintnv2df2 = 1130, - CODE_FOR_nearbyintv2df2 = 1131, - CODE_FOR_rintv2df2 = 1132, - CODE_FOR_roundv2df2 = 1133, - CODE_FOR_lbtruncv2sfv2si2 = 1134, - CODE_FOR_lceilv2sfv2si2 = 1135, - CODE_FOR_lfloorv2sfv2si2 = 1136, - CODE_FOR_lroundv2sfv2si2 = 1137, - CODE_FOR_lfrintnv2sfv2si2 = 1138, - CODE_FOR_lbtruncuv2sfv2si2 = 1139, - CODE_FOR_lceiluv2sfv2si2 = 1140, - CODE_FOR_lflooruv2sfv2si2 = 1141, - CODE_FOR_lrounduv2sfv2si2 = 1142, - CODE_FOR_lfrintnuv2sfv2si2 = 1143, - CODE_FOR_lbtruncv4sfv4si2 = 1144, - CODE_FOR_lceilv4sfv4si2 = 1145, - CODE_FOR_lfloorv4sfv4si2 = 1146, - CODE_FOR_lroundv4sfv4si2 = 1147, - CODE_FOR_lfrintnv4sfv4si2 = 1148, - CODE_FOR_lbtruncuv4sfv4si2 = 1149, - CODE_FOR_lceiluv4sfv4si2 = 1150, - CODE_FOR_lflooruv4sfv4si2 = 1151, - CODE_FOR_lrounduv4sfv4si2 = 1152, - CODE_FOR_lfrintnuv4sfv4si2 = 1153, - CODE_FOR_lbtruncv2dfv2di2 = 1154, - CODE_FOR_lceilv2dfv2di2 = 1155, - CODE_FOR_lfloorv2dfv2di2 = 1156, - CODE_FOR_lroundv2dfv2di2 = 1157, - CODE_FOR_lfrintnv2dfv2di2 = 1158, - CODE_FOR_lbtruncuv2dfv2di2 = 1159, - CODE_FOR_lceiluv2dfv2di2 = 1160, - CODE_FOR_lflooruv2dfv2di2 = 1161, - CODE_FOR_lrounduv2dfv2di2 = 1162, - CODE_FOR_lfrintnuv2dfv2di2 = 1163, - CODE_FOR_floatv2siv2sf2 = 1164, - CODE_FOR_floatunsv2siv2sf2 = 1165, - CODE_FOR_floatv4siv4sf2 = 1166, - CODE_FOR_floatunsv4siv4sf2 = 1167, - CODE_FOR_floatv2div2df2 = 1168, - CODE_FOR_floatunsv2div2df2 = 1169, - CODE_FOR_vec_unpacks_lo_v4sf = 1170, - CODE_FOR_aarch64_float_extend_lo_v2df = 1171, - CODE_FOR_vec_unpacks_hi_v4sf = 1172, - CODE_FOR_aarch64_float_truncate_lo_v2sf = 1173, - CODE_FOR_aarch64_float_truncate_hi_v4sf = 1174, - CODE_FOR_aarch64_vmlsv2sf = 1175, - CODE_FOR_aarch64_vmlsv4sf = 1176, - CODE_FOR_aarch64_vmlsv2df = 1177, - CODE_FOR_smaxv2sf3 = 1178, - CODE_FOR_sminv2sf3 = 1179, - CODE_FOR_smaxv4sf3 = 1180, - CODE_FOR_sminv4sf3 = 1181, - CODE_FOR_smaxv2df3 = 1182, - CODE_FOR_sminv2df3 = 1183, - CODE_FOR_smax_nanv2sf3 = 1184, - CODE_FOR_smin_nanv2sf3 = 1185, - CODE_FOR_smax_nanv4sf3 = 1186, - CODE_FOR_smin_nanv4sf3 = 1187, - CODE_FOR_smax_nanv2df3 = 1188, - CODE_FOR_smin_nanv2df3 = 1189, - CODE_FOR_reduc_splus_v8qi = 1190, - CODE_FOR_reduc_uplus_v8qi = 1191, - CODE_FOR_reduc_splus_v16qi = 1192, - CODE_FOR_reduc_uplus_v16qi = 1193, - CODE_FOR_reduc_splus_v4hi = 1194, - CODE_FOR_reduc_uplus_v4hi = 1195, - CODE_FOR_reduc_splus_v8hi = 1196, - CODE_FOR_reduc_uplus_v8hi = 1197, - CODE_FOR_reduc_splus_v4si = 1198, - CODE_FOR_reduc_uplus_v4si = 1199, - CODE_FOR_reduc_splus_v2di = 1200, - CODE_FOR_reduc_uplus_v2di = 1201, - CODE_FOR_reduc_splus_v2si = 1202, - CODE_FOR_reduc_uplus_v2si = 1203, - CODE_FOR_reduc_splus_v2sf = 1204, - CODE_FOR_reduc_splus_v2df = 1205, - CODE_FOR_aarch64_addpv4sf = 1206, - CODE_FOR_clzv8qi2 = 1207, - CODE_FOR_clzv16qi2 = 1208, - CODE_FOR_clzv4hi2 = 1209, - CODE_FOR_clzv8hi2 = 1210, - CODE_FOR_clzv2si2 = 1211, - CODE_FOR_clzv4si2 = 1212, - CODE_FOR_reduc_umax_v8qi = 1213, - CODE_FOR_reduc_umin_v8qi = 1214, - CODE_FOR_reduc_smax_v8qi = 1215, - CODE_FOR_reduc_smin_v8qi = 1216, - CODE_FOR_reduc_umax_v16qi = 1217, - CODE_FOR_reduc_umin_v16qi = 1218, - CODE_FOR_reduc_smax_v16qi = 1219, - CODE_FOR_reduc_smin_v16qi = 1220, - CODE_FOR_reduc_umax_v4hi = 1221, - CODE_FOR_reduc_umin_v4hi = 1222, - CODE_FOR_reduc_smax_v4hi = 1223, - CODE_FOR_reduc_smin_v4hi = 1224, - CODE_FOR_reduc_umax_v8hi = 1225, - CODE_FOR_reduc_umin_v8hi = 1226, - CODE_FOR_reduc_smax_v8hi = 1227, - CODE_FOR_reduc_smin_v8hi = 1228, - CODE_FOR_reduc_umax_v4si = 1229, - CODE_FOR_reduc_umin_v4si = 1230, - CODE_FOR_reduc_smax_v4si = 1231, - CODE_FOR_reduc_smin_v4si = 1232, - CODE_FOR_reduc_umax_v2si = 1233, - CODE_FOR_reduc_umin_v2si = 1234, - CODE_FOR_reduc_smax_v2si = 1235, - CODE_FOR_reduc_smin_v2si = 1236, - CODE_FOR_reduc_smax_nan_v2sf = 1237, - CODE_FOR_reduc_smin_nan_v2sf = 1238, - CODE_FOR_reduc_smax_v2sf = 1239, - CODE_FOR_reduc_smin_v2sf = 1240, - CODE_FOR_reduc_smax_nan_v2df = 1241, - CODE_FOR_reduc_smin_nan_v2df = 1242, - CODE_FOR_reduc_smax_v2df = 1243, - CODE_FOR_reduc_smin_v2df = 1244, - CODE_FOR_reduc_smax_nan_v4sf = 1245, - CODE_FOR_reduc_smin_nan_v4sf = 1246, - CODE_FOR_reduc_smax_v4sf = 1247, - CODE_FOR_reduc_smin_v4sf = 1248, - CODE_FOR_aarch64_simd_bslv8qi_internal = 1249, - CODE_FOR_aarch64_simd_bslv16qi_internal = 1250, - CODE_FOR_aarch64_simd_bslv4hi_internal = 1251, - CODE_FOR_aarch64_simd_bslv8hi_internal = 1252, - CODE_FOR_aarch64_simd_bslv2si_internal = 1253, - CODE_FOR_aarch64_simd_bslv4si_internal = 1254, - CODE_FOR_aarch64_simd_bslv2di_internal = 1255, - CODE_FOR_aarch64_simd_bslv2sf_internal = 1256, - CODE_FOR_aarch64_simd_bslv4sf_internal = 1257, - CODE_FOR_aarch64_simd_bslv2df_internal = 1258, - CODE_FOR_aarch64_simd_bsldi_internal = 1259, - CODE_FOR_aarch64_simd_bsldf_internal = 1260, - CODE_FOR_aarch64_get_lanev8qi = 1273, - CODE_FOR_aarch64_get_lanev16qi = 1274, - CODE_FOR_aarch64_get_lanev4hi = 1275, - CODE_FOR_aarch64_get_lanev8hi = 1276, - CODE_FOR_aarch64_get_lanev2si = 1277, - CODE_FOR_aarch64_get_lanev4si = 1278, - CODE_FOR_aarch64_get_lanev2di = 1279, - CODE_FOR_aarch64_get_lanev2sf = 1280, - CODE_FOR_aarch64_get_lanev4sf = 1281, - CODE_FOR_aarch64_get_lanev2df = 1282, - CODE_FOR_aarch64_combinev8qi = 1286, - CODE_FOR_aarch64_combinev4hi = 1287, - CODE_FOR_aarch64_combinev2si = 1288, - CODE_FOR_aarch64_combinev2sf = 1289, - CODE_FOR_aarch64_combinedi = 1290, - CODE_FOR_aarch64_combinedf = 1291, - CODE_FOR_aarch64_saddlv16qi_hi_internal = 1292, - CODE_FOR_aarch64_ssublv16qi_hi_internal = 1293, - CODE_FOR_aarch64_uaddlv16qi_hi_internal = 1294, - CODE_FOR_aarch64_usublv16qi_hi_internal = 1295, - CODE_FOR_aarch64_saddlv8hi_hi_internal = 1296, - CODE_FOR_aarch64_ssublv8hi_hi_internal = 1297, - CODE_FOR_aarch64_uaddlv8hi_hi_internal = 1298, - CODE_FOR_aarch64_usublv8hi_hi_internal = 1299, - CODE_FOR_aarch64_saddlv4si_hi_internal = 1300, - CODE_FOR_aarch64_ssublv4si_hi_internal = 1301, - CODE_FOR_aarch64_uaddlv4si_hi_internal = 1302, - CODE_FOR_aarch64_usublv4si_hi_internal = 1303, - CODE_FOR_aarch64_saddlv16qi_lo_internal = 1304, - CODE_FOR_aarch64_ssublv16qi_lo_internal = 1305, - CODE_FOR_aarch64_uaddlv16qi_lo_internal = 1306, - CODE_FOR_aarch64_usublv16qi_lo_internal = 1307, - CODE_FOR_aarch64_saddlv8hi_lo_internal = 1308, - CODE_FOR_aarch64_ssublv8hi_lo_internal = 1309, - CODE_FOR_aarch64_uaddlv8hi_lo_internal = 1310, - CODE_FOR_aarch64_usublv8hi_lo_internal = 1311, - CODE_FOR_aarch64_saddlv4si_lo_internal = 1312, - CODE_FOR_aarch64_ssublv4si_lo_internal = 1313, - CODE_FOR_aarch64_uaddlv4si_lo_internal = 1314, - CODE_FOR_aarch64_usublv4si_lo_internal = 1315, - CODE_FOR_aarch64_saddlv8qi = 1316, - CODE_FOR_aarch64_ssublv8qi = 1317, - CODE_FOR_aarch64_uaddlv8qi = 1318, - CODE_FOR_aarch64_usublv8qi = 1319, - CODE_FOR_aarch64_saddlv4hi = 1320, - CODE_FOR_aarch64_ssublv4hi = 1321, - CODE_FOR_aarch64_uaddlv4hi = 1322, - CODE_FOR_aarch64_usublv4hi = 1323, - CODE_FOR_aarch64_saddlv2si = 1324, - CODE_FOR_aarch64_ssublv2si = 1325, - CODE_FOR_aarch64_uaddlv2si = 1326, - CODE_FOR_aarch64_usublv2si = 1327, - CODE_FOR_aarch64_saddwv8qi = 1328, - CODE_FOR_aarch64_ssubwv8qi = 1329, - CODE_FOR_aarch64_uaddwv8qi = 1330, - CODE_FOR_aarch64_usubwv8qi = 1331, - CODE_FOR_aarch64_saddwv4hi = 1332, - CODE_FOR_aarch64_ssubwv4hi = 1333, - CODE_FOR_aarch64_uaddwv4hi = 1334, - CODE_FOR_aarch64_usubwv4hi = 1335, - CODE_FOR_aarch64_saddwv2si = 1336, - CODE_FOR_aarch64_ssubwv2si = 1337, - CODE_FOR_aarch64_uaddwv2si = 1338, - CODE_FOR_aarch64_usubwv2si = 1339, - CODE_FOR_aarch64_saddw2v16qi_internal = 1340, - CODE_FOR_aarch64_ssubw2v16qi_internal = 1341, - CODE_FOR_aarch64_uaddw2v16qi_internal = 1342, - CODE_FOR_aarch64_usubw2v16qi_internal = 1343, - CODE_FOR_aarch64_saddw2v8hi_internal = 1344, - CODE_FOR_aarch64_ssubw2v8hi_internal = 1345, - CODE_FOR_aarch64_uaddw2v8hi_internal = 1346, - CODE_FOR_aarch64_usubw2v8hi_internal = 1347, - CODE_FOR_aarch64_saddw2v4si_internal = 1348, - CODE_FOR_aarch64_ssubw2v4si_internal = 1349, - CODE_FOR_aarch64_uaddw2v4si_internal = 1350, - CODE_FOR_aarch64_usubw2v4si_internal = 1351, - CODE_FOR_aarch64_shaddv8qi = 1352, - CODE_FOR_aarch64_uhaddv8qi = 1353, - CODE_FOR_aarch64_srhaddv8qi = 1354, - CODE_FOR_aarch64_urhaddv8qi = 1355, - CODE_FOR_aarch64_shsubv8qi = 1356, - CODE_FOR_aarch64_uhsubv8qi = 1357, - CODE_FOR_aarch64_srhsubv8qi = 1358, - CODE_FOR_aarch64_urhsubv8qi = 1359, - CODE_FOR_aarch64_shaddv16qi = 1360, - CODE_FOR_aarch64_uhaddv16qi = 1361, - CODE_FOR_aarch64_srhaddv16qi = 1362, - CODE_FOR_aarch64_urhaddv16qi = 1363, - CODE_FOR_aarch64_shsubv16qi = 1364, - CODE_FOR_aarch64_uhsubv16qi = 1365, - CODE_FOR_aarch64_srhsubv16qi = 1366, - CODE_FOR_aarch64_urhsubv16qi = 1367, - CODE_FOR_aarch64_shaddv4hi = 1368, - CODE_FOR_aarch64_uhaddv4hi = 1369, - CODE_FOR_aarch64_srhaddv4hi = 1370, - CODE_FOR_aarch64_urhaddv4hi = 1371, - CODE_FOR_aarch64_shsubv4hi = 1372, - CODE_FOR_aarch64_uhsubv4hi = 1373, - CODE_FOR_aarch64_srhsubv4hi = 1374, - CODE_FOR_aarch64_urhsubv4hi = 1375, - CODE_FOR_aarch64_shaddv8hi = 1376, - CODE_FOR_aarch64_uhaddv8hi = 1377, - CODE_FOR_aarch64_srhaddv8hi = 1378, - CODE_FOR_aarch64_urhaddv8hi = 1379, - CODE_FOR_aarch64_shsubv8hi = 1380, - CODE_FOR_aarch64_uhsubv8hi = 1381, - CODE_FOR_aarch64_srhsubv8hi = 1382, - CODE_FOR_aarch64_urhsubv8hi = 1383, - CODE_FOR_aarch64_shaddv2si = 1384, - CODE_FOR_aarch64_uhaddv2si = 1385, - CODE_FOR_aarch64_srhaddv2si = 1386, - CODE_FOR_aarch64_urhaddv2si = 1387, - CODE_FOR_aarch64_shsubv2si = 1388, - CODE_FOR_aarch64_uhsubv2si = 1389, - CODE_FOR_aarch64_srhsubv2si = 1390, - CODE_FOR_aarch64_urhsubv2si = 1391, - CODE_FOR_aarch64_shaddv4si = 1392, - CODE_FOR_aarch64_uhaddv4si = 1393, - CODE_FOR_aarch64_srhaddv4si = 1394, - CODE_FOR_aarch64_urhaddv4si = 1395, - CODE_FOR_aarch64_shsubv4si = 1396, - CODE_FOR_aarch64_uhsubv4si = 1397, - CODE_FOR_aarch64_srhsubv4si = 1398, - CODE_FOR_aarch64_urhsubv4si = 1399, - CODE_FOR_aarch64_addhnv8hi = 1400, - CODE_FOR_aarch64_raddhnv8hi = 1401, - CODE_FOR_aarch64_subhnv8hi = 1402, - CODE_FOR_aarch64_rsubhnv8hi = 1403, - CODE_FOR_aarch64_addhnv4si = 1404, - CODE_FOR_aarch64_raddhnv4si = 1405, - CODE_FOR_aarch64_subhnv4si = 1406, - CODE_FOR_aarch64_rsubhnv4si = 1407, - CODE_FOR_aarch64_addhnv2di = 1408, - CODE_FOR_aarch64_raddhnv2di = 1409, - CODE_FOR_aarch64_subhnv2di = 1410, - CODE_FOR_aarch64_rsubhnv2di = 1411, - CODE_FOR_aarch64_addhn2v8hi = 1412, - CODE_FOR_aarch64_raddhn2v8hi = 1413, - CODE_FOR_aarch64_subhn2v8hi = 1414, - CODE_FOR_aarch64_rsubhn2v8hi = 1415, - CODE_FOR_aarch64_addhn2v4si = 1416, - CODE_FOR_aarch64_raddhn2v4si = 1417, - CODE_FOR_aarch64_subhn2v4si = 1418, - CODE_FOR_aarch64_rsubhn2v4si = 1419, - CODE_FOR_aarch64_addhn2v2di = 1420, - CODE_FOR_aarch64_raddhn2v2di = 1421, - CODE_FOR_aarch64_subhn2v2di = 1422, - CODE_FOR_aarch64_rsubhn2v2di = 1423, - CODE_FOR_aarch64_pmulv8qi = 1424, - CODE_FOR_aarch64_pmulv16qi = 1425, - CODE_FOR_aarch64_sqaddv8qi = 1426, - CODE_FOR_aarch64_uqaddv8qi = 1427, - CODE_FOR_aarch64_sqsubv8qi = 1428, - CODE_FOR_aarch64_uqsubv8qi = 1429, - CODE_FOR_aarch64_sqaddv16qi = 1430, - CODE_FOR_aarch64_uqaddv16qi = 1431, - CODE_FOR_aarch64_sqsubv16qi = 1432, - CODE_FOR_aarch64_uqsubv16qi = 1433, - CODE_FOR_aarch64_sqaddv4hi = 1434, - CODE_FOR_aarch64_uqaddv4hi = 1435, - CODE_FOR_aarch64_sqsubv4hi = 1436, - CODE_FOR_aarch64_uqsubv4hi = 1437, - CODE_FOR_aarch64_sqaddv8hi = 1438, - CODE_FOR_aarch64_uqaddv8hi = 1439, - CODE_FOR_aarch64_sqsubv8hi = 1440, - CODE_FOR_aarch64_uqsubv8hi = 1441, - CODE_FOR_aarch64_sqaddv2si = 1442, - CODE_FOR_aarch64_uqaddv2si = 1443, - CODE_FOR_aarch64_sqsubv2si = 1444, - CODE_FOR_aarch64_uqsubv2si = 1445, - CODE_FOR_aarch64_sqaddv4si = 1446, - CODE_FOR_aarch64_uqaddv4si = 1447, - CODE_FOR_aarch64_sqsubv4si = 1448, - CODE_FOR_aarch64_uqsubv4si = 1449, - CODE_FOR_aarch64_sqaddv2di = 1450, - CODE_FOR_aarch64_uqaddv2di = 1451, - CODE_FOR_aarch64_sqsubv2di = 1452, - CODE_FOR_aarch64_uqsubv2di = 1453, - CODE_FOR_aarch64_sqaddqi = 1454, - CODE_FOR_aarch64_uqaddqi = 1455, - CODE_FOR_aarch64_sqsubqi = 1456, - CODE_FOR_aarch64_uqsubqi = 1457, - CODE_FOR_aarch64_sqaddhi = 1458, - CODE_FOR_aarch64_uqaddhi = 1459, - CODE_FOR_aarch64_sqsubhi = 1460, - CODE_FOR_aarch64_uqsubhi = 1461, - CODE_FOR_aarch64_sqaddsi = 1462, - CODE_FOR_aarch64_uqaddsi = 1463, - CODE_FOR_aarch64_sqsubsi = 1464, - CODE_FOR_aarch64_uqsubsi = 1465, - CODE_FOR_aarch64_sqadddi = 1466, - CODE_FOR_aarch64_uqadddi = 1467, - CODE_FOR_aarch64_sqsubdi = 1468, - CODE_FOR_aarch64_uqsubdi = 1469, - CODE_FOR_aarch64_suqaddv8qi = 1470, - CODE_FOR_aarch64_usqaddv8qi = 1471, - CODE_FOR_aarch64_suqaddv16qi = 1472, - CODE_FOR_aarch64_usqaddv16qi = 1473, - CODE_FOR_aarch64_suqaddv4hi = 1474, - CODE_FOR_aarch64_usqaddv4hi = 1475, - CODE_FOR_aarch64_suqaddv8hi = 1476, - CODE_FOR_aarch64_usqaddv8hi = 1477, - CODE_FOR_aarch64_suqaddv2si = 1478, - CODE_FOR_aarch64_usqaddv2si = 1479, - CODE_FOR_aarch64_suqaddv4si = 1480, - CODE_FOR_aarch64_usqaddv4si = 1481, - CODE_FOR_aarch64_suqaddv2di = 1482, - CODE_FOR_aarch64_usqaddv2di = 1483, - CODE_FOR_aarch64_suqaddqi = 1484, - CODE_FOR_aarch64_usqaddqi = 1485, - CODE_FOR_aarch64_suqaddhi = 1486, - CODE_FOR_aarch64_usqaddhi = 1487, - CODE_FOR_aarch64_suqaddsi = 1488, - CODE_FOR_aarch64_usqaddsi = 1489, - CODE_FOR_aarch64_suqadddi = 1490, - CODE_FOR_aarch64_usqadddi = 1491, - CODE_FOR_aarch64_sqmovunv8hi = 1492, - CODE_FOR_aarch64_sqmovunv4si = 1493, - CODE_FOR_aarch64_sqmovunv2di = 1494, - CODE_FOR_aarch64_sqmovunhi = 1495, - CODE_FOR_aarch64_sqmovunsi = 1496, - CODE_FOR_aarch64_sqmovundi = 1497, - CODE_FOR_aarch64_sqmovnv8hi = 1498, - CODE_FOR_aarch64_uqmovnv8hi = 1499, - CODE_FOR_aarch64_sqmovnv4si = 1500, - CODE_FOR_aarch64_uqmovnv4si = 1501, - CODE_FOR_aarch64_sqmovnv2di = 1502, - CODE_FOR_aarch64_uqmovnv2di = 1503, - CODE_FOR_aarch64_sqmovnhi = 1504, - CODE_FOR_aarch64_uqmovnhi = 1505, - CODE_FOR_aarch64_sqmovnsi = 1506, - CODE_FOR_aarch64_uqmovnsi = 1507, - CODE_FOR_aarch64_sqmovndi = 1508, - CODE_FOR_aarch64_uqmovndi = 1509, - CODE_FOR_aarch64_sqnegv8qi = 1510, - CODE_FOR_aarch64_sqabsv8qi = 1511, - CODE_FOR_aarch64_sqnegv16qi = 1512, - CODE_FOR_aarch64_sqabsv16qi = 1513, - CODE_FOR_aarch64_sqnegv4hi = 1514, - CODE_FOR_aarch64_sqabsv4hi = 1515, - CODE_FOR_aarch64_sqnegv8hi = 1516, - CODE_FOR_aarch64_sqabsv8hi = 1517, - CODE_FOR_aarch64_sqnegv2si = 1518, - CODE_FOR_aarch64_sqabsv2si = 1519, - CODE_FOR_aarch64_sqnegv4si = 1520, - CODE_FOR_aarch64_sqabsv4si = 1521, - CODE_FOR_aarch64_sqnegv2di = 1522, - CODE_FOR_aarch64_sqabsv2di = 1523, - CODE_FOR_aarch64_sqnegqi = 1524, - CODE_FOR_aarch64_sqabsqi = 1525, - CODE_FOR_aarch64_sqneghi = 1526, - CODE_FOR_aarch64_sqabshi = 1527, - CODE_FOR_aarch64_sqnegsi = 1528, - CODE_FOR_aarch64_sqabssi = 1529, - CODE_FOR_aarch64_sqnegdi = 1530, - CODE_FOR_aarch64_sqabsdi = 1531, - CODE_FOR_aarch64_sqdmulhv4hi = 1532, - CODE_FOR_aarch64_sqrdmulhv4hi = 1533, - CODE_FOR_aarch64_sqdmulhv8hi = 1534, - CODE_FOR_aarch64_sqrdmulhv8hi = 1535, - CODE_FOR_aarch64_sqdmulhv2si = 1536, - CODE_FOR_aarch64_sqrdmulhv2si = 1537, - CODE_FOR_aarch64_sqdmulhv4si = 1538, - CODE_FOR_aarch64_sqrdmulhv4si = 1539, - CODE_FOR_aarch64_sqdmulhhi = 1540, - CODE_FOR_aarch64_sqrdmulhhi = 1541, - CODE_FOR_aarch64_sqdmulhsi = 1542, - CODE_FOR_aarch64_sqrdmulhsi = 1543, - CODE_FOR_aarch64_sqdmulh_lanev4hi = 1544, - CODE_FOR_aarch64_sqrdmulh_lanev4hi = 1545, - CODE_FOR_aarch64_sqdmulh_lanev8hi = 1546, - CODE_FOR_aarch64_sqrdmulh_lanev8hi = 1547, - CODE_FOR_aarch64_sqdmulh_lanev2si = 1548, - CODE_FOR_aarch64_sqrdmulh_lanev2si = 1549, - CODE_FOR_aarch64_sqdmulh_lanev4si = 1550, - CODE_FOR_aarch64_sqrdmulh_lanev4si = 1551, - CODE_FOR_aarch64_sqdmulh_laneqv4hi = 1552, - CODE_FOR_aarch64_sqrdmulh_laneqv4hi = 1553, - CODE_FOR_aarch64_sqdmulh_laneqv8hi = 1554, - CODE_FOR_aarch64_sqrdmulh_laneqv8hi = 1555, - CODE_FOR_aarch64_sqdmulh_laneqv2si = 1556, - CODE_FOR_aarch64_sqrdmulh_laneqv2si = 1557, - CODE_FOR_aarch64_sqdmulh_laneqv4si = 1558, - CODE_FOR_aarch64_sqrdmulh_laneqv4si = 1559, - CODE_FOR_aarch64_sqdmulh_lanehi = 1560, - CODE_FOR_aarch64_sqrdmulh_lanehi = 1561, - CODE_FOR_aarch64_sqdmulh_lanesi = 1562, - CODE_FOR_aarch64_sqrdmulh_lanesi = 1563, - CODE_FOR_aarch64_sqdmlalv4hi = 1564, - CODE_FOR_aarch64_sqdmlslv4hi = 1565, - CODE_FOR_aarch64_sqdmlalv2si = 1566, - CODE_FOR_aarch64_sqdmlslv2si = 1567, - CODE_FOR_aarch64_sqdmlalhi = 1568, - CODE_FOR_aarch64_sqdmlslhi = 1569, - CODE_FOR_aarch64_sqdmlalsi = 1570, - CODE_FOR_aarch64_sqdmlslsi = 1571, - CODE_FOR_aarch64_sqdmlal_lanev4hi_internal = 1572, - CODE_FOR_aarch64_sqdmlsl_lanev4hi_internal = 1573, - CODE_FOR_aarch64_sqdmlal_lanev2si_internal = 1574, - CODE_FOR_aarch64_sqdmlsl_lanev2si_internal = 1575, - CODE_FOR_aarch64_sqdmlal_lanehi_internal = 1576, - CODE_FOR_aarch64_sqdmlsl_lanehi_internal = 1577, - CODE_FOR_aarch64_sqdmlal_lanesi_internal = 1578, - CODE_FOR_aarch64_sqdmlsl_lanesi_internal = 1579, - CODE_FOR_aarch64_sqdmlal_nv4hi = 1580, - CODE_FOR_aarch64_sqdmlsl_nv4hi = 1581, - CODE_FOR_aarch64_sqdmlal_nv2si = 1582, - CODE_FOR_aarch64_sqdmlsl_nv2si = 1583, - CODE_FOR_aarch64_sqdmlal2v8hi_internal = 1584, - CODE_FOR_aarch64_sqdmlsl2v8hi_internal = 1585, - CODE_FOR_aarch64_sqdmlal2v4si_internal = 1586, - CODE_FOR_aarch64_sqdmlsl2v4si_internal = 1587, - CODE_FOR_aarch64_sqdmlal2_lanev8hi_internal = 1588, - CODE_FOR_aarch64_sqdmlsl2_lanev8hi_internal = 1589, - CODE_FOR_aarch64_sqdmlal2_lanev4si_internal = 1590, - CODE_FOR_aarch64_sqdmlsl2_lanev4si_internal = 1591, - CODE_FOR_aarch64_sqdmlal2_nv8hi_internal = 1592, - CODE_FOR_aarch64_sqdmlsl2_nv8hi_internal = 1593, - CODE_FOR_aarch64_sqdmlal2_nv4si_internal = 1594, - CODE_FOR_aarch64_sqdmlsl2_nv4si_internal = 1595, - CODE_FOR_aarch64_sqdmullv4hi = 1596, - CODE_FOR_aarch64_sqdmullv2si = 1597, - CODE_FOR_aarch64_sqdmullhi = 1598, - CODE_FOR_aarch64_sqdmullsi = 1599, - CODE_FOR_aarch64_sqdmull_lanev4hi_internal = 1600, - CODE_FOR_aarch64_sqdmull_lanev2si_internal = 1601, - CODE_FOR_aarch64_sqdmull_lanehi_internal = 1602, - CODE_FOR_aarch64_sqdmull_lanesi_internal = 1603, - CODE_FOR_aarch64_sqdmull_nv4hi = 1604, - CODE_FOR_aarch64_sqdmull_nv2si = 1605, - CODE_FOR_aarch64_sqdmull2v8hi_internal = 1606, - CODE_FOR_aarch64_sqdmull2v4si_internal = 1607, - CODE_FOR_aarch64_sqdmull2_lanev8hi_internal = 1608, - CODE_FOR_aarch64_sqdmull2_lanev4si_internal = 1609, - CODE_FOR_aarch64_sqdmull2_nv8hi_internal = 1610, - CODE_FOR_aarch64_sqdmull2_nv4si_internal = 1611, - CODE_FOR_aarch64_sshlv8qi = 1612, - CODE_FOR_aarch64_ushlv8qi = 1613, - CODE_FOR_aarch64_srshlv8qi = 1614, - CODE_FOR_aarch64_urshlv8qi = 1615, - CODE_FOR_aarch64_sshlv16qi = 1616, - CODE_FOR_aarch64_ushlv16qi = 1617, - CODE_FOR_aarch64_srshlv16qi = 1618, - CODE_FOR_aarch64_urshlv16qi = 1619, - CODE_FOR_aarch64_sshlv4hi = 1620, - CODE_FOR_aarch64_ushlv4hi = 1621, - CODE_FOR_aarch64_srshlv4hi = 1622, - CODE_FOR_aarch64_urshlv4hi = 1623, - CODE_FOR_aarch64_sshlv8hi = 1624, - CODE_FOR_aarch64_ushlv8hi = 1625, - CODE_FOR_aarch64_srshlv8hi = 1626, - CODE_FOR_aarch64_urshlv8hi = 1627, - CODE_FOR_aarch64_sshlv2si = 1628, - CODE_FOR_aarch64_ushlv2si = 1629, - CODE_FOR_aarch64_srshlv2si = 1630, - CODE_FOR_aarch64_urshlv2si = 1631, - CODE_FOR_aarch64_sshlv4si = 1632, - CODE_FOR_aarch64_ushlv4si = 1633, - CODE_FOR_aarch64_srshlv4si = 1634, - CODE_FOR_aarch64_urshlv4si = 1635, - CODE_FOR_aarch64_sshlv2di = 1636, - CODE_FOR_aarch64_ushlv2di = 1637, - CODE_FOR_aarch64_srshlv2di = 1638, - CODE_FOR_aarch64_urshlv2di = 1639, - CODE_FOR_aarch64_sshldi = 1640, - CODE_FOR_aarch64_ushldi = 1641, - CODE_FOR_aarch64_srshldi = 1642, - CODE_FOR_aarch64_urshldi = 1643, - CODE_FOR_aarch64_sqshlv8qi = 1644, - CODE_FOR_aarch64_uqshlv8qi = 1645, - CODE_FOR_aarch64_sqrshlv8qi = 1646, - CODE_FOR_aarch64_uqrshlv8qi = 1647, - CODE_FOR_aarch64_sqshlv16qi = 1648, - CODE_FOR_aarch64_uqshlv16qi = 1649, - CODE_FOR_aarch64_sqrshlv16qi = 1650, - CODE_FOR_aarch64_uqrshlv16qi = 1651, - CODE_FOR_aarch64_sqshlv4hi = 1652, - CODE_FOR_aarch64_uqshlv4hi = 1653, - CODE_FOR_aarch64_sqrshlv4hi = 1654, - CODE_FOR_aarch64_uqrshlv4hi = 1655, - CODE_FOR_aarch64_sqshlv8hi = 1656, - CODE_FOR_aarch64_uqshlv8hi = 1657, - CODE_FOR_aarch64_sqrshlv8hi = 1658, - CODE_FOR_aarch64_uqrshlv8hi = 1659, - CODE_FOR_aarch64_sqshlv2si = 1660, - CODE_FOR_aarch64_uqshlv2si = 1661, - CODE_FOR_aarch64_sqrshlv2si = 1662, - CODE_FOR_aarch64_uqrshlv2si = 1663, - CODE_FOR_aarch64_sqshlv4si = 1664, - CODE_FOR_aarch64_uqshlv4si = 1665, - CODE_FOR_aarch64_sqrshlv4si = 1666, - CODE_FOR_aarch64_uqrshlv4si = 1667, - CODE_FOR_aarch64_sqshlv2di = 1668, - CODE_FOR_aarch64_uqshlv2di = 1669, - CODE_FOR_aarch64_sqrshlv2di = 1670, - CODE_FOR_aarch64_uqrshlv2di = 1671, - CODE_FOR_aarch64_sqshlqi = 1672, - CODE_FOR_aarch64_uqshlqi = 1673, - CODE_FOR_aarch64_sqrshlqi = 1674, - CODE_FOR_aarch64_uqrshlqi = 1675, - CODE_FOR_aarch64_sqshlhi = 1676, - CODE_FOR_aarch64_uqshlhi = 1677, - CODE_FOR_aarch64_sqrshlhi = 1678, - CODE_FOR_aarch64_uqrshlhi = 1679, - CODE_FOR_aarch64_sqshlsi = 1680, - CODE_FOR_aarch64_uqshlsi = 1681, - CODE_FOR_aarch64_sqrshlsi = 1682, - CODE_FOR_aarch64_uqrshlsi = 1683, - CODE_FOR_aarch64_sqshldi = 1684, - CODE_FOR_aarch64_uqshldi = 1685, - CODE_FOR_aarch64_sqrshldi = 1686, - CODE_FOR_aarch64_uqrshldi = 1687, - CODE_FOR_aarch64_sshll_nv8qi = 1688, - CODE_FOR_aarch64_ushll_nv8qi = 1689, - CODE_FOR_aarch64_sshll_nv4hi = 1690, - CODE_FOR_aarch64_ushll_nv4hi = 1691, - CODE_FOR_aarch64_sshll_nv2si = 1692, - CODE_FOR_aarch64_ushll_nv2si = 1693, - CODE_FOR_aarch64_sshll2_nv16qi = 1694, - CODE_FOR_aarch64_ushll2_nv16qi = 1695, - CODE_FOR_aarch64_sshll2_nv8hi = 1696, - CODE_FOR_aarch64_ushll2_nv8hi = 1697, - CODE_FOR_aarch64_sshll2_nv4si = 1698, - CODE_FOR_aarch64_ushll2_nv4si = 1699, - CODE_FOR_aarch64_srshr_nv8qi = 1700, - CODE_FOR_aarch64_urshr_nv8qi = 1701, - CODE_FOR_aarch64_srshr_nv16qi = 1702, - CODE_FOR_aarch64_urshr_nv16qi = 1703, - CODE_FOR_aarch64_srshr_nv4hi = 1704, - CODE_FOR_aarch64_urshr_nv4hi = 1705, - CODE_FOR_aarch64_srshr_nv8hi = 1706, - CODE_FOR_aarch64_urshr_nv8hi = 1707, - CODE_FOR_aarch64_srshr_nv2si = 1708, - CODE_FOR_aarch64_urshr_nv2si = 1709, - CODE_FOR_aarch64_srshr_nv4si = 1710, - CODE_FOR_aarch64_urshr_nv4si = 1711, - CODE_FOR_aarch64_srshr_nv2di = 1712, - CODE_FOR_aarch64_urshr_nv2di = 1713, - CODE_FOR_aarch64_srshr_ndi = 1714, - CODE_FOR_aarch64_urshr_ndi = 1715, - CODE_FOR_aarch64_ssra_nv8qi = 1716, - CODE_FOR_aarch64_usra_nv8qi = 1717, - CODE_FOR_aarch64_srsra_nv8qi = 1718, - CODE_FOR_aarch64_ursra_nv8qi = 1719, - CODE_FOR_aarch64_ssra_nv16qi = 1720, - CODE_FOR_aarch64_usra_nv16qi = 1721, - CODE_FOR_aarch64_srsra_nv16qi = 1722, - CODE_FOR_aarch64_ursra_nv16qi = 1723, - CODE_FOR_aarch64_ssra_nv4hi = 1724, - CODE_FOR_aarch64_usra_nv4hi = 1725, - CODE_FOR_aarch64_srsra_nv4hi = 1726, - CODE_FOR_aarch64_ursra_nv4hi = 1727, - CODE_FOR_aarch64_ssra_nv8hi = 1728, - CODE_FOR_aarch64_usra_nv8hi = 1729, - CODE_FOR_aarch64_srsra_nv8hi = 1730, - CODE_FOR_aarch64_ursra_nv8hi = 1731, - CODE_FOR_aarch64_ssra_nv2si = 1732, - CODE_FOR_aarch64_usra_nv2si = 1733, - CODE_FOR_aarch64_srsra_nv2si = 1734, - CODE_FOR_aarch64_ursra_nv2si = 1735, - CODE_FOR_aarch64_ssra_nv4si = 1736, - CODE_FOR_aarch64_usra_nv4si = 1737, - CODE_FOR_aarch64_srsra_nv4si = 1738, - CODE_FOR_aarch64_ursra_nv4si = 1739, - CODE_FOR_aarch64_ssra_nv2di = 1740, - CODE_FOR_aarch64_usra_nv2di = 1741, - CODE_FOR_aarch64_srsra_nv2di = 1742, - CODE_FOR_aarch64_ursra_nv2di = 1743, - CODE_FOR_aarch64_ssra_ndi = 1744, - CODE_FOR_aarch64_usra_ndi = 1745, - CODE_FOR_aarch64_srsra_ndi = 1746, - CODE_FOR_aarch64_ursra_ndi = 1747, - CODE_FOR_aarch64_ssli_nv8qi = 1748, - CODE_FOR_aarch64_usli_nv8qi = 1749, - CODE_FOR_aarch64_ssri_nv8qi = 1750, - CODE_FOR_aarch64_usri_nv8qi = 1751, - CODE_FOR_aarch64_ssli_nv16qi = 1752, - CODE_FOR_aarch64_usli_nv16qi = 1753, - CODE_FOR_aarch64_ssri_nv16qi = 1754, - CODE_FOR_aarch64_usri_nv16qi = 1755, - CODE_FOR_aarch64_ssli_nv4hi = 1756, - CODE_FOR_aarch64_usli_nv4hi = 1757, - CODE_FOR_aarch64_ssri_nv4hi = 1758, - CODE_FOR_aarch64_usri_nv4hi = 1759, - CODE_FOR_aarch64_ssli_nv8hi = 1760, - CODE_FOR_aarch64_usli_nv8hi = 1761, - CODE_FOR_aarch64_ssri_nv8hi = 1762, - CODE_FOR_aarch64_usri_nv8hi = 1763, - CODE_FOR_aarch64_ssli_nv2si = 1764, - CODE_FOR_aarch64_usli_nv2si = 1765, - CODE_FOR_aarch64_ssri_nv2si = 1766, - CODE_FOR_aarch64_usri_nv2si = 1767, - CODE_FOR_aarch64_ssli_nv4si = 1768, - CODE_FOR_aarch64_usli_nv4si = 1769, - CODE_FOR_aarch64_ssri_nv4si = 1770, - CODE_FOR_aarch64_usri_nv4si = 1771, - CODE_FOR_aarch64_ssli_nv2di = 1772, - CODE_FOR_aarch64_usli_nv2di = 1773, - CODE_FOR_aarch64_ssri_nv2di = 1774, - CODE_FOR_aarch64_usri_nv2di = 1775, - CODE_FOR_aarch64_ssli_ndi = 1776, - CODE_FOR_aarch64_usli_ndi = 1777, - CODE_FOR_aarch64_ssri_ndi = 1778, - CODE_FOR_aarch64_usri_ndi = 1779, - CODE_FOR_aarch64_sqshlu_nv8qi = 1780, - CODE_FOR_aarch64_sqshl_nv8qi = 1781, - CODE_FOR_aarch64_uqshl_nv8qi = 1782, - CODE_FOR_aarch64_sqshlu_nv16qi = 1783, - CODE_FOR_aarch64_sqshl_nv16qi = 1784, - CODE_FOR_aarch64_uqshl_nv16qi = 1785, - CODE_FOR_aarch64_sqshlu_nv4hi = 1786, - CODE_FOR_aarch64_sqshl_nv4hi = 1787, - CODE_FOR_aarch64_uqshl_nv4hi = 1788, - CODE_FOR_aarch64_sqshlu_nv8hi = 1789, - CODE_FOR_aarch64_sqshl_nv8hi = 1790, - CODE_FOR_aarch64_uqshl_nv8hi = 1791, - CODE_FOR_aarch64_sqshlu_nv2si = 1792, - CODE_FOR_aarch64_sqshl_nv2si = 1793, - CODE_FOR_aarch64_uqshl_nv2si = 1794, - CODE_FOR_aarch64_sqshlu_nv4si = 1795, - CODE_FOR_aarch64_sqshl_nv4si = 1796, - CODE_FOR_aarch64_uqshl_nv4si = 1797, - CODE_FOR_aarch64_sqshlu_nv2di = 1798, - CODE_FOR_aarch64_sqshl_nv2di = 1799, - CODE_FOR_aarch64_uqshl_nv2di = 1800, - CODE_FOR_aarch64_sqshlu_nqi = 1801, - CODE_FOR_aarch64_sqshl_nqi = 1802, - CODE_FOR_aarch64_uqshl_nqi = 1803, - CODE_FOR_aarch64_sqshlu_nhi = 1804, - CODE_FOR_aarch64_sqshl_nhi = 1805, - CODE_FOR_aarch64_uqshl_nhi = 1806, - CODE_FOR_aarch64_sqshlu_nsi = 1807, - CODE_FOR_aarch64_sqshl_nsi = 1808, - CODE_FOR_aarch64_uqshl_nsi = 1809, - CODE_FOR_aarch64_sqshlu_ndi = 1810, - CODE_FOR_aarch64_sqshl_ndi = 1811, - CODE_FOR_aarch64_uqshl_ndi = 1812, - CODE_FOR_aarch64_sqshrun_nv8hi = 1813, - CODE_FOR_aarch64_sqrshrun_nv8hi = 1814, - CODE_FOR_aarch64_sqshrn_nv8hi = 1815, - CODE_FOR_aarch64_uqshrn_nv8hi = 1816, - CODE_FOR_aarch64_sqrshrn_nv8hi = 1817, - CODE_FOR_aarch64_uqrshrn_nv8hi = 1818, - CODE_FOR_aarch64_sqshrun_nv4si = 1819, - CODE_FOR_aarch64_sqrshrun_nv4si = 1820, - CODE_FOR_aarch64_sqshrn_nv4si = 1821, - CODE_FOR_aarch64_uqshrn_nv4si = 1822, - CODE_FOR_aarch64_sqrshrn_nv4si = 1823, - CODE_FOR_aarch64_uqrshrn_nv4si = 1824, - CODE_FOR_aarch64_sqshrun_nv2di = 1825, - CODE_FOR_aarch64_sqrshrun_nv2di = 1826, - CODE_FOR_aarch64_sqshrn_nv2di = 1827, - CODE_FOR_aarch64_uqshrn_nv2di = 1828, - CODE_FOR_aarch64_sqrshrn_nv2di = 1829, - CODE_FOR_aarch64_uqrshrn_nv2di = 1830, - CODE_FOR_aarch64_sqshrun_nhi = 1831, - CODE_FOR_aarch64_sqrshrun_nhi = 1832, - CODE_FOR_aarch64_sqshrn_nhi = 1833, - CODE_FOR_aarch64_uqshrn_nhi = 1834, - CODE_FOR_aarch64_sqrshrn_nhi = 1835, - CODE_FOR_aarch64_uqrshrn_nhi = 1836, - CODE_FOR_aarch64_sqshrun_nsi = 1837, - CODE_FOR_aarch64_sqrshrun_nsi = 1838, - CODE_FOR_aarch64_sqshrn_nsi = 1839, - CODE_FOR_aarch64_uqshrn_nsi = 1840, - CODE_FOR_aarch64_sqrshrn_nsi = 1841, - CODE_FOR_aarch64_uqrshrn_nsi = 1842, - CODE_FOR_aarch64_sqshrun_ndi = 1843, - CODE_FOR_aarch64_sqrshrun_ndi = 1844, - CODE_FOR_aarch64_sqshrn_ndi = 1845, - CODE_FOR_aarch64_uqshrn_ndi = 1846, - CODE_FOR_aarch64_sqrshrn_ndi = 1847, - CODE_FOR_aarch64_uqrshrn_ndi = 1848, - CODE_FOR_aarch64_cmltv8qi = 1849, - CODE_FOR_aarch64_cmlev8qi = 1850, - CODE_FOR_aarch64_cmeqv8qi = 1851, - CODE_FOR_aarch64_cmgev8qi = 1852, - CODE_FOR_aarch64_cmgtv8qi = 1853, - CODE_FOR_aarch64_cmltv16qi = 1854, - CODE_FOR_aarch64_cmlev16qi = 1855, - CODE_FOR_aarch64_cmeqv16qi = 1856, - CODE_FOR_aarch64_cmgev16qi = 1857, - CODE_FOR_aarch64_cmgtv16qi = 1858, - CODE_FOR_aarch64_cmltv4hi = 1859, - CODE_FOR_aarch64_cmlev4hi = 1860, - CODE_FOR_aarch64_cmeqv4hi = 1861, - CODE_FOR_aarch64_cmgev4hi = 1862, - CODE_FOR_aarch64_cmgtv4hi = 1863, - CODE_FOR_aarch64_cmltv8hi = 1864, - CODE_FOR_aarch64_cmlev8hi = 1865, - CODE_FOR_aarch64_cmeqv8hi = 1866, - CODE_FOR_aarch64_cmgev8hi = 1867, - CODE_FOR_aarch64_cmgtv8hi = 1868, - CODE_FOR_aarch64_cmltv2si = 1869, - CODE_FOR_aarch64_cmlev2si = 1870, - CODE_FOR_aarch64_cmeqv2si = 1871, - CODE_FOR_aarch64_cmgev2si = 1872, - CODE_FOR_aarch64_cmgtv2si = 1873, - CODE_FOR_aarch64_cmltv4si = 1874, - CODE_FOR_aarch64_cmlev4si = 1875, - CODE_FOR_aarch64_cmeqv4si = 1876, - CODE_FOR_aarch64_cmgev4si = 1877, - CODE_FOR_aarch64_cmgtv4si = 1878, - CODE_FOR_aarch64_cmltv2di = 1879, - CODE_FOR_aarch64_cmlev2di = 1880, - CODE_FOR_aarch64_cmeqv2di = 1881, - CODE_FOR_aarch64_cmgev2di = 1882, - CODE_FOR_aarch64_cmgtv2di = 1883, - CODE_FOR_aarch64_cmltdi = 1884, - CODE_FOR_aarch64_cmledi = 1885, - CODE_FOR_aarch64_cmeqdi = 1886, - CODE_FOR_aarch64_cmgedi = 1887, - CODE_FOR_aarch64_cmgtdi = 1888, - CODE_FOR_aarch64_cmltuv8qi = 1894, - CODE_FOR_aarch64_cmleuv8qi = 1895, - CODE_FOR_aarch64_cmgeuv8qi = 1896, - CODE_FOR_aarch64_cmgtuv8qi = 1897, - CODE_FOR_aarch64_cmltuv16qi = 1898, - CODE_FOR_aarch64_cmleuv16qi = 1899, - CODE_FOR_aarch64_cmgeuv16qi = 1900, - CODE_FOR_aarch64_cmgtuv16qi = 1901, - CODE_FOR_aarch64_cmltuv4hi = 1902, - CODE_FOR_aarch64_cmleuv4hi = 1903, - CODE_FOR_aarch64_cmgeuv4hi = 1904, - CODE_FOR_aarch64_cmgtuv4hi = 1905, - CODE_FOR_aarch64_cmltuv8hi = 1906, - CODE_FOR_aarch64_cmleuv8hi = 1907, - CODE_FOR_aarch64_cmgeuv8hi = 1908, - CODE_FOR_aarch64_cmgtuv8hi = 1909, - CODE_FOR_aarch64_cmltuv2si = 1910, - CODE_FOR_aarch64_cmleuv2si = 1911, - CODE_FOR_aarch64_cmgeuv2si = 1912, - CODE_FOR_aarch64_cmgtuv2si = 1913, - CODE_FOR_aarch64_cmltuv4si = 1914, - CODE_FOR_aarch64_cmleuv4si = 1915, - CODE_FOR_aarch64_cmgeuv4si = 1916, - CODE_FOR_aarch64_cmgtuv4si = 1917, - CODE_FOR_aarch64_cmltuv2di = 1918, - CODE_FOR_aarch64_cmleuv2di = 1919, - CODE_FOR_aarch64_cmgeuv2di = 1920, - CODE_FOR_aarch64_cmgtuv2di = 1921, - CODE_FOR_aarch64_cmltudi = 1922, - CODE_FOR_aarch64_cmleudi = 1923, - CODE_FOR_aarch64_cmgeudi = 1924, - CODE_FOR_aarch64_cmgtudi = 1925, - CODE_FOR_aarch64_cmtstv8qi = 1930, - CODE_FOR_aarch64_cmtstv16qi = 1931, - CODE_FOR_aarch64_cmtstv4hi = 1932, - CODE_FOR_aarch64_cmtstv8hi = 1933, - CODE_FOR_aarch64_cmtstv2si = 1934, - CODE_FOR_aarch64_cmtstv4si = 1935, - CODE_FOR_aarch64_cmtstv2di = 1936, - CODE_FOR_aarch64_cmtstdi = 1937, - CODE_FOR_aarch64_cmltv2sf = 1939, - CODE_FOR_aarch64_cmlev2sf = 1940, - CODE_FOR_aarch64_cmeqv2sf = 1941, - CODE_FOR_aarch64_cmgev2sf = 1942, - CODE_FOR_aarch64_cmgtv2sf = 1943, - CODE_FOR_aarch64_cmltv4sf = 1944, - CODE_FOR_aarch64_cmlev4sf = 1945, - CODE_FOR_aarch64_cmeqv4sf = 1946, - CODE_FOR_aarch64_cmgev4sf = 1947, - CODE_FOR_aarch64_cmgtv4sf = 1948, - CODE_FOR_aarch64_cmltv2df = 1949, - CODE_FOR_aarch64_cmlev2df = 1950, - CODE_FOR_aarch64_cmeqv2df = 1951, - CODE_FOR_aarch64_cmgev2df = 1952, - CODE_FOR_aarch64_cmgtv2df = 1953, - CODE_FOR_aarch64_cmltsf = 1954, - CODE_FOR_aarch64_cmlesf = 1955, - CODE_FOR_aarch64_cmeqsf = 1956, - CODE_FOR_aarch64_cmgesf = 1957, - CODE_FOR_aarch64_cmgtsf = 1958, - CODE_FOR_aarch64_cmltdf = 1959, - CODE_FOR_aarch64_cmledf = 1960, - CODE_FOR_aarch64_cmeqdf = 1961, - CODE_FOR_aarch64_cmgedf = 1962, - CODE_FOR_aarch64_cmgtdf = 1963, - CODE_FOR_aarch64_addpv8qi = 1984, - CODE_FOR_aarch64_addpv4hi = 1985, - CODE_FOR_aarch64_addpv2si = 1986, - CODE_FOR_aarch64_addpdi = 1987, - CODE_FOR_sqrtv2sf2 = 1988, - CODE_FOR_sqrtv4sf2 = 1989, - CODE_FOR_sqrtv2df2 = 1990, - CODE_FOR_vec_load_lanesoiv16qi = 1991, - CODE_FOR_vec_load_lanesoiv8hi = 1992, - CODE_FOR_vec_load_lanesoiv4si = 1993, - CODE_FOR_vec_load_lanesoiv2di = 1994, - CODE_FOR_vec_load_lanesoiv4sf = 1995, - CODE_FOR_vec_load_lanesoiv2df = 1996, - CODE_FOR_vec_store_lanesoiv16qi = 1997, - CODE_FOR_vec_store_lanesoiv8hi = 1998, - CODE_FOR_vec_store_lanesoiv4si = 1999, - CODE_FOR_vec_store_lanesoiv2di = 2000, - CODE_FOR_vec_store_lanesoiv4sf = 2001, - CODE_FOR_vec_store_lanesoiv2df = 2002, - CODE_FOR_vec_store_lanesoi_lanev16qi = 2003, - CODE_FOR_vec_store_lanesoi_lanev8hi = 2004, - CODE_FOR_vec_store_lanesoi_lanev4si = 2005, - CODE_FOR_vec_store_lanesoi_lanev2di = 2006, - CODE_FOR_vec_store_lanesoi_lanev4sf = 2007, - CODE_FOR_vec_store_lanesoi_lanev2df = 2008, - CODE_FOR_vec_load_lanesciv16qi = 2009, - CODE_FOR_vec_load_lanesciv8hi = 2010, - CODE_FOR_vec_load_lanesciv4si = 2011, - CODE_FOR_vec_load_lanesciv2di = 2012, - CODE_FOR_vec_load_lanesciv4sf = 2013, - CODE_FOR_vec_load_lanesciv2df = 2014, - CODE_FOR_vec_store_lanesciv16qi = 2015, - CODE_FOR_vec_store_lanesciv8hi = 2016, - CODE_FOR_vec_store_lanesciv4si = 2017, - CODE_FOR_vec_store_lanesciv2di = 2018, - CODE_FOR_vec_store_lanesciv4sf = 2019, - CODE_FOR_vec_store_lanesciv2df = 2020, - CODE_FOR_vec_store_lanesci_lanev16qi = 2021, - CODE_FOR_vec_store_lanesci_lanev8hi = 2022, - CODE_FOR_vec_store_lanesci_lanev4si = 2023, - CODE_FOR_vec_store_lanesci_lanev2di = 2024, - CODE_FOR_vec_store_lanesci_lanev4sf = 2025, - CODE_FOR_vec_store_lanesci_lanev2df = 2026, - CODE_FOR_vec_load_lanesxiv16qi = 2027, - CODE_FOR_vec_load_lanesxiv8hi = 2028, - CODE_FOR_vec_load_lanesxiv4si = 2029, - CODE_FOR_vec_load_lanesxiv2di = 2030, - CODE_FOR_vec_load_lanesxiv4sf = 2031, - CODE_FOR_vec_load_lanesxiv2df = 2032, - CODE_FOR_vec_store_lanesxiv16qi = 2033, - CODE_FOR_vec_store_lanesxiv8hi = 2034, - CODE_FOR_vec_store_lanesxiv4si = 2035, - CODE_FOR_vec_store_lanesxiv2di = 2036, - CODE_FOR_vec_store_lanesxiv4sf = 2037, - CODE_FOR_vec_store_lanesxiv2df = 2038, - CODE_FOR_vec_store_lanesxi_lanev16qi = 2039, - CODE_FOR_vec_store_lanesxi_lanev8hi = 2040, - CODE_FOR_vec_store_lanesxi_lanev4si = 2041, - CODE_FOR_vec_store_lanesxi_lanev2di = 2042, - CODE_FOR_vec_store_lanesxi_lanev4sf = 2043, - CODE_FOR_vec_store_lanesxi_lanev2df = 2044, - CODE_FOR_aarch64_be_ld1v8qi = 2048, - CODE_FOR_aarch64_be_ld1v16qi = 2049, - CODE_FOR_aarch64_be_ld1v4hi = 2050, - CODE_FOR_aarch64_be_ld1v8hi = 2051, - CODE_FOR_aarch64_be_ld1v2si = 2052, - CODE_FOR_aarch64_be_ld1v4si = 2053, - CODE_FOR_aarch64_be_ld1v2di = 2054, - CODE_FOR_aarch64_be_ld1v2sf = 2055, - CODE_FOR_aarch64_be_ld1v4sf = 2056, - CODE_FOR_aarch64_be_ld1v2df = 2057, - CODE_FOR_aarch64_be_ld1di = 2058, - CODE_FOR_aarch64_be_st1v8qi = 2059, - CODE_FOR_aarch64_be_st1v16qi = 2060, - CODE_FOR_aarch64_be_st1v4hi = 2061, - CODE_FOR_aarch64_be_st1v8hi = 2062, - CODE_FOR_aarch64_be_st1v2si = 2063, - CODE_FOR_aarch64_be_st1v4si = 2064, - CODE_FOR_aarch64_be_st1v2di = 2065, - CODE_FOR_aarch64_be_st1v2sf = 2066, - CODE_FOR_aarch64_be_st1v4sf = 2067, - CODE_FOR_aarch64_be_st1v2df = 2068, - CODE_FOR_aarch64_be_st1di = 2069, - CODE_FOR_aarch64_ld2v8qi_dreg = 2070, - CODE_FOR_aarch64_ld2v4hi_dreg = 2071, - CODE_FOR_aarch64_ld2v2si_dreg = 2072, - CODE_FOR_aarch64_ld2v2sf_dreg = 2073, - CODE_FOR_aarch64_ld2di_dreg = 2074, - CODE_FOR_aarch64_ld2df_dreg = 2075, - CODE_FOR_aarch64_ld3v8qi_dreg = 2076, - CODE_FOR_aarch64_ld3v4hi_dreg = 2077, - CODE_FOR_aarch64_ld3v2si_dreg = 2078, - CODE_FOR_aarch64_ld3v2sf_dreg = 2079, - CODE_FOR_aarch64_ld3di_dreg = 2080, - CODE_FOR_aarch64_ld3df_dreg = 2081, - CODE_FOR_aarch64_ld4v8qi_dreg = 2082, - CODE_FOR_aarch64_ld4v4hi_dreg = 2083, - CODE_FOR_aarch64_ld4v2si_dreg = 2084, - CODE_FOR_aarch64_ld4v2sf_dreg = 2085, - CODE_FOR_aarch64_ld4di_dreg = 2086, - CODE_FOR_aarch64_ld4df_dreg = 2087, - CODE_FOR_aarch64_tbl1v8qi = 2088, - CODE_FOR_aarch64_tbl1v16qi = 2089, - CODE_FOR_aarch64_tbl2v16qi = 2090, - CODE_FOR_aarch64_combinev16qi = 2091, - CODE_FOR_aarch64_zip1v8qi = 2092, - CODE_FOR_aarch64_zip2v8qi = 2093, - CODE_FOR_aarch64_trn1v8qi = 2094, - CODE_FOR_aarch64_trn2v8qi = 2095, - CODE_FOR_aarch64_uzp1v8qi = 2096, - CODE_FOR_aarch64_uzp2v8qi = 2097, - CODE_FOR_aarch64_zip1v16qi = 2098, - CODE_FOR_aarch64_zip2v16qi = 2099, - CODE_FOR_aarch64_trn1v16qi = 2100, - CODE_FOR_aarch64_trn2v16qi = 2101, - CODE_FOR_aarch64_uzp1v16qi = 2102, - CODE_FOR_aarch64_uzp2v16qi = 2103, - CODE_FOR_aarch64_zip1v4hi = 2104, - CODE_FOR_aarch64_zip2v4hi = 2105, - CODE_FOR_aarch64_trn1v4hi = 2106, - CODE_FOR_aarch64_trn2v4hi = 2107, - CODE_FOR_aarch64_uzp1v4hi = 2108, - CODE_FOR_aarch64_uzp2v4hi = 2109, - CODE_FOR_aarch64_zip1v8hi = 2110, - CODE_FOR_aarch64_zip2v8hi = 2111, - CODE_FOR_aarch64_trn1v8hi = 2112, - CODE_FOR_aarch64_trn2v8hi = 2113, - CODE_FOR_aarch64_uzp1v8hi = 2114, - CODE_FOR_aarch64_uzp2v8hi = 2115, - CODE_FOR_aarch64_zip1v2si = 2116, - CODE_FOR_aarch64_zip2v2si = 2117, - CODE_FOR_aarch64_trn1v2si = 2118, - CODE_FOR_aarch64_trn2v2si = 2119, - CODE_FOR_aarch64_uzp1v2si = 2120, - CODE_FOR_aarch64_uzp2v2si = 2121, - CODE_FOR_aarch64_zip1v4si = 2122, - CODE_FOR_aarch64_zip2v4si = 2123, - CODE_FOR_aarch64_trn1v4si = 2124, - CODE_FOR_aarch64_trn2v4si = 2125, - CODE_FOR_aarch64_uzp1v4si = 2126, - CODE_FOR_aarch64_uzp2v4si = 2127, - CODE_FOR_aarch64_zip1v2di = 2128, - CODE_FOR_aarch64_zip2v2di = 2129, - CODE_FOR_aarch64_trn1v2di = 2130, - CODE_FOR_aarch64_trn2v2di = 2131, - CODE_FOR_aarch64_uzp1v2di = 2132, - CODE_FOR_aarch64_uzp2v2di = 2133, - CODE_FOR_aarch64_zip1v2sf = 2134, - CODE_FOR_aarch64_zip2v2sf = 2135, - CODE_FOR_aarch64_trn1v2sf = 2136, - CODE_FOR_aarch64_trn2v2sf = 2137, - CODE_FOR_aarch64_uzp1v2sf = 2138, - CODE_FOR_aarch64_uzp2v2sf = 2139, - CODE_FOR_aarch64_zip1v4sf = 2140, - CODE_FOR_aarch64_zip2v4sf = 2141, - CODE_FOR_aarch64_trn1v4sf = 2142, - CODE_FOR_aarch64_trn2v4sf = 2143, - CODE_FOR_aarch64_uzp1v4sf = 2144, - CODE_FOR_aarch64_uzp2v4sf = 2145, - CODE_FOR_aarch64_zip1v2df = 2146, - CODE_FOR_aarch64_zip2v2df = 2147, - CODE_FOR_aarch64_trn1v2df = 2148, - CODE_FOR_aarch64_trn2v2df = 2149, - CODE_FOR_aarch64_uzp1v2df = 2150, - CODE_FOR_aarch64_uzp2v2df = 2151, - CODE_FOR_aarch64_st2v8qi_dreg = 2152, - CODE_FOR_aarch64_st2v4hi_dreg = 2153, - CODE_FOR_aarch64_st2v2si_dreg = 2154, - CODE_FOR_aarch64_st2v2sf_dreg = 2155, - CODE_FOR_aarch64_st2di_dreg = 2156, - CODE_FOR_aarch64_st2df_dreg = 2157, - CODE_FOR_aarch64_st3v8qi_dreg = 2158, - CODE_FOR_aarch64_st3v4hi_dreg = 2159, - CODE_FOR_aarch64_st3v2si_dreg = 2160, - CODE_FOR_aarch64_st3v2sf_dreg = 2161, - CODE_FOR_aarch64_st3di_dreg = 2162, - CODE_FOR_aarch64_st3df_dreg = 2163, - CODE_FOR_aarch64_st4v8qi_dreg = 2164, - CODE_FOR_aarch64_st4v4hi_dreg = 2165, - CODE_FOR_aarch64_st4v2si_dreg = 2166, - CODE_FOR_aarch64_st4v2sf_dreg = 2167, - CODE_FOR_aarch64_st4di_dreg = 2168, - CODE_FOR_aarch64_st4df_dreg = 2169, - CODE_FOR_aarch64_frecpev2sf = 2181, - CODE_FOR_aarch64_frecpev4sf = 2182, - CODE_FOR_aarch64_frecpev2df = 2183, - CODE_FOR_aarch64_frecpesf = 2184, - CODE_FOR_aarch64_frecpxsf = 2185, - CODE_FOR_aarch64_frecpedf = 2186, - CODE_FOR_aarch64_frecpxdf = 2187, - CODE_FOR_aarch64_frecpsv2sf = 2188, - CODE_FOR_aarch64_frecpsv4sf = 2189, - CODE_FOR_aarch64_frecpsv2df = 2190, - CODE_FOR_aarch64_frecpssf = 2191, - CODE_FOR_aarch64_frecpsdf = 2192, - CODE_FOR_aarch64_crypto_aesev16qi = 2193, - CODE_FOR_aarch64_crypto_aesdv16qi = 2194, - CODE_FOR_aarch64_crypto_aesmcv16qi = 2195, - CODE_FOR_aarch64_crypto_aesimcv16qi = 2196, - CODE_FOR_aarch64_crypto_sha1hsi = 2197, - CODE_FOR_aarch64_crypto_sha1su1v4si = 2198, - CODE_FOR_aarch64_crypto_sha1cv4si = 2199, - CODE_FOR_aarch64_crypto_sha1mv4si = 2200, - CODE_FOR_aarch64_crypto_sha1pv4si = 2201, - CODE_FOR_aarch64_crypto_sha1su0v4si = 2202, - CODE_FOR_aarch64_crypto_sha256hv4si = 2203, - CODE_FOR_aarch64_crypto_sha256h2v4si = 2204, - CODE_FOR_aarch64_crypto_sha256su0v4si = 2205, - CODE_FOR_aarch64_crypto_sha256su1v4si = 2206, - CODE_FOR_aarch64_crypto_pmulldi = 2207, - CODE_FOR_aarch64_crypto_pmullv2di = 2208, - CODE_FOR_atomic_compare_and_swapqi_1 = 2209, - CODE_FOR_atomic_compare_and_swaphi_1 = 2210, - CODE_FOR_atomic_compare_and_swapsi_1 = 2211, - CODE_FOR_atomic_compare_and_swapdi_1 = 2212, - CODE_FOR_atomic_exchangeqi = 2213, - CODE_FOR_atomic_exchangehi = 2214, - CODE_FOR_atomic_exchangesi = 2215, - CODE_FOR_atomic_exchangedi = 2216, - CODE_FOR_atomic_addqi = 2217, - CODE_FOR_atomic_subqi = 2218, - CODE_FOR_atomic_orqi = 2219, - CODE_FOR_atomic_xorqi = 2220, - CODE_FOR_atomic_andqi = 2221, - CODE_FOR_atomic_addhi = 2222, - CODE_FOR_atomic_subhi = 2223, - CODE_FOR_atomic_orhi = 2224, - CODE_FOR_atomic_xorhi = 2225, - CODE_FOR_atomic_andhi = 2226, - CODE_FOR_atomic_addsi = 2227, - CODE_FOR_atomic_subsi = 2228, - CODE_FOR_atomic_orsi = 2229, - CODE_FOR_atomic_xorsi = 2230, - CODE_FOR_atomic_andsi = 2231, - CODE_FOR_atomic_adddi = 2232, - CODE_FOR_atomic_subdi = 2233, - CODE_FOR_atomic_ordi = 2234, - CODE_FOR_atomic_xordi = 2235, - CODE_FOR_atomic_anddi = 2236, - CODE_FOR_atomic_nandqi = 2237, - CODE_FOR_atomic_nandhi = 2238, - CODE_FOR_atomic_nandsi = 2239, - CODE_FOR_atomic_nanddi = 2240, - CODE_FOR_atomic_fetch_addqi = 2241, - CODE_FOR_atomic_fetch_subqi = 2242, - CODE_FOR_atomic_fetch_orqi = 2243, - CODE_FOR_atomic_fetch_xorqi = 2244, - CODE_FOR_atomic_fetch_andqi = 2245, - CODE_FOR_atomic_fetch_addhi = 2246, - CODE_FOR_atomic_fetch_subhi = 2247, - CODE_FOR_atomic_fetch_orhi = 2248, - CODE_FOR_atomic_fetch_xorhi = 2249, - CODE_FOR_atomic_fetch_andhi = 2250, - CODE_FOR_atomic_fetch_addsi = 2251, - CODE_FOR_atomic_fetch_subsi = 2252, - CODE_FOR_atomic_fetch_orsi = 2253, - CODE_FOR_atomic_fetch_xorsi = 2254, - CODE_FOR_atomic_fetch_andsi = 2255, - CODE_FOR_atomic_fetch_adddi = 2256, - CODE_FOR_atomic_fetch_subdi = 2257, - CODE_FOR_atomic_fetch_ordi = 2258, - CODE_FOR_atomic_fetch_xordi = 2259, - CODE_FOR_atomic_fetch_anddi = 2260, - CODE_FOR_atomic_fetch_nandqi = 2261, - CODE_FOR_atomic_fetch_nandhi = 2262, - CODE_FOR_atomic_fetch_nandsi = 2263, - CODE_FOR_atomic_fetch_nanddi = 2264, - CODE_FOR_atomic_add_fetchqi = 2265, - CODE_FOR_atomic_sub_fetchqi = 2266, - CODE_FOR_atomic_or_fetchqi = 2267, - CODE_FOR_atomic_xor_fetchqi = 2268, - CODE_FOR_atomic_and_fetchqi = 2269, - CODE_FOR_atomic_add_fetchhi = 2270, - CODE_FOR_atomic_sub_fetchhi = 2271, - CODE_FOR_atomic_or_fetchhi = 2272, - CODE_FOR_atomic_xor_fetchhi = 2273, - CODE_FOR_atomic_and_fetchhi = 2274, - CODE_FOR_atomic_add_fetchsi = 2275, - CODE_FOR_atomic_sub_fetchsi = 2276, - CODE_FOR_atomic_or_fetchsi = 2277, - CODE_FOR_atomic_xor_fetchsi = 2278, - CODE_FOR_atomic_and_fetchsi = 2279, - CODE_FOR_atomic_add_fetchdi = 2280, - CODE_FOR_atomic_sub_fetchdi = 2281, - CODE_FOR_atomic_or_fetchdi = 2282, - CODE_FOR_atomic_xor_fetchdi = 2283, - CODE_FOR_atomic_and_fetchdi = 2284, - CODE_FOR_atomic_nand_fetchqi = 2285, - CODE_FOR_atomic_nand_fetchhi = 2286, - CODE_FOR_atomic_nand_fetchsi = 2287, - CODE_FOR_atomic_nand_fetchdi = 2288, - CODE_FOR_atomic_loadqi = 2289, - CODE_FOR_atomic_loadhi = 2290, - CODE_FOR_atomic_loadsi = 2291, - CODE_FOR_atomic_loaddi = 2292, - CODE_FOR_atomic_storeqi = 2293, - CODE_FOR_atomic_storehi = 2294, - CODE_FOR_atomic_storesi = 2295, - CODE_FOR_atomic_storedi = 2296, - CODE_FOR_aarch64_load_exclusiveqi = 2297, - CODE_FOR_aarch64_load_exclusivehi = 2298, - CODE_FOR_aarch64_load_exclusivesi = 2299, - CODE_FOR_aarch64_load_exclusivedi = 2300, - CODE_FOR_aarch64_store_exclusiveqi = 2301, - CODE_FOR_aarch64_store_exclusivehi = 2302, - CODE_FOR_aarch64_store_exclusivesi = 2303, - CODE_FOR_aarch64_store_exclusivedi = 2304, - CODE_FOR_cbranchsi4 = 2306, - CODE_FOR_cbranchdi4 = 2307, - CODE_FOR_cbranchsf4 = 2308, - CODE_FOR_cbranchdf4 = 2309, - CODE_FOR_casesi = 2310, - CODE_FOR_prologue = 2311, - CODE_FOR_epilogue = 2312, - CODE_FOR_sibcall_epilogue = 2313, - CODE_FOR_call = 2315, - CODE_FOR_call_value = 2316, - CODE_FOR_sibcall = 2317, - CODE_FOR_sibcall_value = 2318, - CODE_FOR_untyped_call = 2319, - CODE_FOR_movqi = 2320, - CODE_FOR_movhi = 2321, - CODE_FOR_movsi = 2322, - CODE_FOR_movdi = 2323, - CODE_FOR_movti = 2324, - CODE_FOR_movsf = 2326, - CODE_FOR_movdf = 2327, - CODE_FOR_movtf = 2328, - CODE_FOR_extendsidi2 = 2330, - CODE_FOR_zero_extendsidi2 = 2331, - CODE_FOR_extendqisi2 = 2332, - CODE_FOR_zero_extendqisi2 = 2333, - CODE_FOR_extendqidi2 = 2334, - CODE_FOR_zero_extendqidi2 = 2335, - CODE_FOR_extendhisi2 = 2336, - CODE_FOR_zero_extendhisi2 = 2337, - CODE_FOR_extendhidi2 = 2338, - CODE_FOR_zero_extendhidi2 = 2339, - CODE_FOR_extendqihi2 = 2340, - CODE_FOR_zero_extendqihi2 = 2341, - CODE_FOR_addsi3 = 2342, - CODE_FOR_adddi3 = 2343, - CODE_FOR_addti3 = 2344, - CODE_FOR_subti3 = 2345, - CODE_FOR_mulditi3 = 2347, - CODE_FOR_umulditi3 = 2348, - CODE_FOR_multi3 = 2349, - CODE_FOR_cstoresi4 = 2350, - CODE_FOR_cstoredi4 = 2351, - CODE_FOR_cstoresf4 = 2352, - CODE_FOR_cstoredf4 = 2353, - CODE_FOR_cmovsi6 = 2354, - CODE_FOR_cmovdi6 = 2355, - CODE_FOR_cmovsf6 = 2356, - CODE_FOR_cmovdf6 = 2357, - CODE_FOR_movqicc = 2358, - CODE_FOR_movhicc = 2359, - CODE_FOR_movsicc = 2360, - CODE_FOR_movdicc = 2361, - CODE_FOR_movsfsicc = 2362, - CODE_FOR_movdfsicc = 2363, - CODE_FOR_movsfdicc = 2364, - CODE_FOR_movdfdicc = 2365, - CODE_FOR_movsfcc = 2366, - CODE_FOR_movdfcc = 2367, - CODE_FOR_ffssi2 = 2368, - CODE_FOR_ffsdi2 = 2369, - CODE_FOR_ctzsi2 = 2370, - CODE_FOR_ctzdi2 = 2371, - CODE_FOR_ashlsi3 = 2372, - CODE_FOR_ashrsi3 = 2373, - CODE_FOR_lshrsi3 = 2374, - CODE_FOR_ashldi3 = 2375, - CODE_FOR_ashrdi3 = 2376, - CODE_FOR_lshrdi3 = 2377, - CODE_FOR_ashlqi3 = 2378, - CODE_FOR_ashlhi3 = 2379, - CODE_FOR_rotrsi3 = 2380, - CODE_FOR_rotrdi3 = 2381, - CODE_FOR_rotlsi3 = 2382, - CODE_FOR_rotldi3 = 2383, - CODE_FOR_extv = 2388, - CODE_FOR_extzv = 2389, - CODE_FOR_insvsi = 2390, - CODE_FOR_insvdi = 2391, - CODE_FOR_aarch64_reload_movti = 2392, - CODE_FOR_aarch64_reload_movtf = 2393, - CODE_FOR_add_losym = 2394, - CODE_FOR_tlsgd_small = 2395, - CODE_FOR_tlsle_small = 2396, - CODE_FOR_get_thread_pointerdi = 2397, - CODE_FOR_stack_protect_set = 2398, - CODE_FOR_stack_protect_test = 2399, - CODE_FOR_movv8qi = 2400, - CODE_FOR_movv16qi = 2401, - CODE_FOR_movv4hi = 2402, - CODE_FOR_movv8hi = 2403, - CODE_FOR_movv2si = 2404, - CODE_FOR_movv4si = 2405, - CODE_FOR_movv2di = 2406, - CODE_FOR_movv2sf = 2407, - CODE_FOR_movv4sf = 2408, - CODE_FOR_movv2df = 2409, - CODE_FOR_movmisalignv8qi = 2410, - CODE_FOR_movmisalignv16qi = 2411, - CODE_FOR_movmisalignv4hi = 2412, - CODE_FOR_movmisalignv8hi = 2413, - CODE_FOR_movmisalignv2si = 2414, - CODE_FOR_movmisalignv4si = 2415, - CODE_FOR_movmisalignv2di = 2416, - CODE_FOR_movmisalignv2sf = 2417, - CODE_FOR_movmisalignv4sf = 2418, - CODE_FOR_movmisalignv2df = 2419, - CODE_FOR_aarch64_split_simd_movv16qi = 2432, - CODE_FOR_aarch64_split_simd_movv8hi = 2433, - CODE_FOR_aarch64_split_simd_movv4si = 2434, - CODE_FOR_aarch64_split_simd_movv2di = 2435, - CODE_FOR_aarch64_split_simd_movv4sf = 2436, - CODE_FOR_aarch64_split_simd_movv2df = 2437, - CODE_FOR_ashlv8qi3 = 2438, - CODE_FOR_ashlv16qi3 = 2439, - CODE_FOR_ashlv4hi3 = 2440, - CODE_FOR_ashlv8hi3 = 2441, - CODE_FOR_ashlv2si3 = 2442, - CODE_FOR_ashlv4si3 = 2443, - CODE_FOR_ashlv2di3 = 2444, - CODE_FOR_lshrv8qi3 = 2445, - CODE_FOR_lshrv16qi3 = 2446, - CODE_FOR_lshrv4hi3 = 2447, - CODE_FOR_lshrv8hi3 = 2448, - CODE_FOR_lshrv2si3 = 2449, - CODE_FOR_lshrv4si3 = 2450, - CODE_FOR_lshrv2di3 = 2451, - CODE_FOR_ashrv8qi3 = 2452, - CODE_FOR_ashrv16qi3 = 2453, - CODE_FOR_ashrv4hi3 = 2454, - CODE_FOR_ashrv8hi3 = 2455, - CODE_FOR_ashrv2si3 = 2456, - CODE_FOR_ashrv4si3 = 2457, - CODE_FOR_ashrv2di3 = 2458, - CODE_FOR_vashlv8qi3 = 2459, - CODE_FOR_vashlv16qi3 = 2460, - CODE_FOR_vashlv4hi3 = 2461, - CODE_FOR_vashlv8hi3 = 2462, - CODE_FOR_vashlv2si3 = 2463, - CODE_FOR_vashlv4si3 = 2464, - CODE_FOR_vashlv2di3 = 2465, - CODE_FOR_vashrv8qi3 = 2466, - CODE_FOR_vashrv16qi3 = 2467, - CODE_FOR_vashrv4hi3 = 2468, - CODE_FOR_vashrv8hi3 = 2469, - CODE_FOR_vashrv2si3 = 2470, - CODE_FOR_vashrv4si3 = 2471, - CODE_FOR_aarch64_ashr_simddi = 2472, - CODE_FOR_vlshrv8qi3 = 2473, - CODE_FOR_vlshrv16qi3 = 2474, - CODE_FOR_vlshrv4hi3 = 2475, - CODE_FOR_vlshrv8hi3 = 2476, - CODE_FOR_vlshrv2si3 = 2477, - CODE_FOR_vlshrv4si3 = 2478, - CODE_FOR_aarch64_lshr_simddi = 2479, - CODE_FOR_vec_setv8qi = 2480, - CODE_FOR_vec_setv16qi = 2481, - CODE_FOR_vec_setv4hi = 2482, - CODE_FOR_vec_setv8hi = 2483, - CODE_FOR_vec_setv2si = 2484, - CODE_FOR_vec_setv4si = 2485, - CODE_FOR_vec_setv2di = 2486, - CODE_FOR_vec_setv2sf = 2487, - CODE_FOR_vec_setv4sf = 2488, - CODE_FOR_vec_setv2df = 2489, - CODE_FOR_move_hi_quad_v16qi = 2490, - CODE_FOR_move_hi_quad_v8hi = 2491, - CODE_FOR_move_hi_quad_v4si = 2492, - CODE_FOR_move_hi_quad_v2di = 2493, - CODE_FOR_move_hi_quad_v4sf = 2494, - CODE_FOR_move_hi_quad_v2df = 2495, - CODE_FOR_vec_pack_trunc_v4hi = 2496, - CODE_FOR_vec_pack_trunc_v2si = 2497, - CODE_FOR_vec_pack_trunc_di = 2498, - CODE_FOR_vec_unpacks_hi_v16qi = 2499, - CODE_FOR_vec_unpacku_hi_v16qi = 2500, - CODE_FOR_vec_unpacks_hi_v8hi = 2501, - CODE_FOR_vec_unpacku_hi_v8hi = 2502, - CODE_FOR_vec_unpacks_hi_v4si = 2503, - CODE_FOR_vec_unpacku_hi_v4si = 2504, - CODE_FOR_vec_unpacks_lo_v16qi = 2505, - CODE_FOR_vec_unpacku_lo_v16qi = 2506, - CODE_FOR_vec_unpacks_lo_v8hi = 2507, - CODE_FOR_vec_unpacku_lo_v8hi = 2508, - CODE_FOR_vec_unpacks_lo_v4si = 2509, - CODE_FOR_vec_unpacku_lo_v4si = 2510, - CODE_FOR_vec_widen_smult_lo_v16qi = 2511, - CODE_FOR_vec_widen_umult_lo_v16qi = 2512, - CODE_FOR_vec_widen_smult_lo_v8hi = 2513, - CODE_FOR_vec_widen_umult_lo_v8hi = 2514, - CODE_FOR_vec_widen_smult_lo_v4si = 2515, - CODE_FOR_vec_widen_umult_lo_v4si = 2516, - CODE_FOR_vec_widen_smult_hi_v16qi = 2517, - CODE_FOR_vec_widen_umult_hi_v16qi = 2518, - CODE_FOR_vec_widen_smult_hi_v8hi = 2519, - CODE_FOR_vec_widen_umult_hi_v8hi = 2520, - CODE_FOR_vec_widen_smult_hi_v4si = 2521, - CODE_FOR_vec_widen_umult_hi_v4si = 2522, - CODE_FOR_fixv2sfv2si2 = 2523, - CODE_FOR_fixunsv2sfv2si2 = 2524, - CODE_FOR_fixv4sfv4si2 = 2525, - CODE_FOR_fixunsv4sfv4si2 = 2526, - CODE_FOR_fixv2dfv2di2 = 2527, - CODE_FOR_fixunsv2dfv2di2 = 2528, - CODE_FOR_fix_truncv2sfv2si2 = 2529, - CODE_FOR_fixuns_truncv2sfv2si2 = 2530, - CODE_FOR_fix_truncv4sfv4si2 = 2531, - CODE_FOR_fixuns_truncv4sfv4si2 = 2532, - CODE_FOR_fix_truncv2dfv2di2 = 2533, - CODE_FOR_fixuns_truncv2dfv2di2 = 2534, - CODE_FOR_ftruncv2sf2 = 2535, - CODE_FOR_ftruncv4sf2 = 2536, - CODE_FOR_ftruncv2df2 = 2537, - CODE_FOR_vec_pack_trunc_v2df = 2538, - CODE_FOR_vec_pack_trunc_df = 2539, - CODE_FOR_reduc_splus_v4sf = 2540, - CODE_FOR_aarch64_simd_bslv8qi = 2541, - CODE_FOR_aarch64_simd_bslv16qi = 2542, - CODE_FOR_aarch64_simd_bslv4hi = 2543, - CODE_FOR_aarch64_simd_bslv8hi = 2544, - CODE_FOR_aarch64_simd_bslv2si = 2545, - CODE_FOR_aarch64_simd_bslv4si = 2546, - CODE_FOR_aarch64_simd_bslv2di = 2547, - CODE_FOR_aarch64_simd_bslv2sf = 2548, - CODE_FOR_aarch64_simd_bslv4sf = 2549, - CODE_FOR_aarch64_simd_bslv2df = 2550, - CODE_FOR_aarch64_simd_bsldi = 2551, - CODE_FOR_aarch64_simd_bsldf = 2552, - CODE_FOR_aarch64_vcond_internalv8qiv8qi = 2553, - CODE_FOR_aarch64_vcond_internalv16qiv16qi = 2554, - CODE_FOR_aarch64_vcond_internalv4hiv4hi = 2555, - CODE_FOR_aarch64_vcond_internalv8hiv8hi = 2556, - CODE_FOR_aarch64_vcond_internalv2siv2si = 2557, - CODE_FOR_aarch64_vcond_internalv4siv4si = 2558, - CODE_FOR_aarch64_vcond_internalv2div2di = 2559, - CODE_FOR_aarch64_vcond_internalv2sfv2sf = 2560, - CODE_FOR_aarch64_vcond_internalv2siv2sf = 2561, - CODE_FOR_aarch64_vcond_internalv4sfv2sf = 2562, - CODE_FOR_aarch64_vcond_internalv4siv2sf = 2563, - CODE_FOR_aarch64_vcond_internalv2dfv2sf = 2564, - CODE_FOR_aarch64_vcond_internalv2div2sf = 2565, - CODE_FOR_aarch64_vcond_internalv2sfv4sf = 2566, - CODE_FOR_aarch64_vcond_internalv2siv4sf = 2567, - CODE_FOR_aarch64_vcond_internalv4sfv4sf = 2568, - CODE_FOR_aarch64_vcond_internalv4siv4sf = 2569, - CODE_FOR_aarch64_vcond_internalv2dfv4sf = 2570, - CODE_FOR_aarch64_vcond_internalv2div4sf = 2571, - CODE_FOR_aarch64_vcond_internalv2sfv2df = 2572, - CODE_FOR_aarch64_vcond_internalv2siv2df = 2573, - CODE_FOR_aarch64_vcond_internalv4sfv2df = 2574, - CODE_FOR_aarch64_vcond_internalv4siv2df = 2575, - CODE_FOR_aarch64_vcond_internalv2dfv2df = 2576, - CODE_FOR_aarch64_vcond_internalv2div2df = 2577, - CODE_FOR_vcondv8qiv8qi = 2578, - CODE_FOR_vcondv16qiv16qi = 2579, - CODE_FOR_vcondv4hiv4hi = 2580, - CODE_FOR_vcondv8hiv8hi = 2581, - CODE_FOR_vcondv2siv2si = 2582, - CODE_FOR_vcondv4siv4si = 2583, - CODE_FOR_vcondv2div2di = 2584, - CODE_FOR_vcondv2sfv2sf = 2585, - CODE_FOR_vcondv4sfv4sf = 2586, - CODE_FOR_vcondv2dfv2df = 2587, - CODE_FOR_vcondv2siv2sf = 2588, - CODE_FOR_vcondv4siv4sf = 2589, - CODE_FOR_vcondv2div2df = 2590, - CODE_FOR_vconduv8qiv8qi = 2591, - CODE_FOR_vconduv16qiv16qi = 2592, - CODE_FOR_vconduv4hiv4hi = 2593, - CODE_FOR_vconduv8hiv8hi = 2594, - CODE_FOR_vconduv2siv2si = 2595, - CODE_FOR_vconduv4siv4si = 2596, - CODE_FOR_vconduv2div2di = 2597, - CODE_FOR_aarch64_createv8qi = 2598, - CODE_FOR_aarch64_createv4hi = 2599, - CODE_FOR_aarch64_createv2si = 2600, - CODE_FOR_aarch64_createdi = 2601, - CODE_FOR_aarch64_createdf = 2602, - CODE_FOR_aarch64_createv2sf = 2603, - CODE_FOR_aarch64_be_checked_get_lanev8qi = 2604, - CODE_FOR_aarch64_be_checked_get_lanev16qi = 2605, - CODE_FOR_aarch64_be_checked_get_lanev4hi = 2606, - CODE_FOR_aarch64_be_checked_get_lanev8hi = 2607, - CODE_FOR_aarch64_be_checked_get_lanev2si = 2608, - CODE_FOR_aarch64_be_checked_get_lanev4si = 2609, - CODE_FOR_aarch64_be_checked_get_lanev2di = 2610, - CODE_FOR_aarch64_be_checked_get_lanev2sf = 2611, - CODE_FOR_aarch64_be_checked_get_lanev4sf = 2612, - CODE_FOR_aarch64_be_checked_get_lanev2df = 2613, - CODE_FOR_aarch64_get_lanedi = 2614, - CODE_FOR_aarch64_reinterpretv8qiv8qi = 2615, - CODE_FOR_aarch64_reinterpretv8qiv4hi = 2616, - CODE_FOR_aarch64_reinterpretv8qiv2si = 2617, - CODE_FOR_aarch64_reinterpretv8qiv2sf = 2618, - CODE_FOR_aarch64_reinterpretv8qidi = 2619, - CODE_FOR_aarch64_reinterpretv8qidf = 2620, - CODE_FOR_aarch64_reinterpretv4hiv8qi = 2621, - CODE_FOR_aarch64_reinterpretv4hiv4hi = 2622, - CODE_FOR_aarch64_reinterpretv4hiv2si = 2623, - CODE_FOR_aarch64_reinterpretv4hiv2sf = 2624, - CODE_FOR_aarch64_reinterpretv4hidi = 2625, - CODE_FOR_aarch64_reinterpretv4hidf = 2626, - CODE_FOR_aarch64_reinterpretv2siv8qi = 2627, - CODE_FOR_aarch64_reinterpretv2siv4hi = 2628, - CODE_FOR_aarch64_reinterpretv2siv2si = 2629, - CODE_FOR_aarch64_reinterpretv2siv2sf = 2630, - CODE_FOR_aarch64_reinterpretv2sidi = 2631, - CODE_FOR_aarch64_reinterpretv2sidf = 2632, - CODE_FOR_aarch64_reinterpretv2sfv8qi = 2633, - CODE_FOR_aarch64_reinterpretv2sfv4hi = 2634, - CODE_FOR_aarch64_reinterpretv2sfv2si = 2635, - CODE_FOR_aarch64_reinterpretv2sfv2sf = 2636, - CODE_FOR_aarch64_reinterpretv2sfdi = 2637, - CODE_FOR_aarch64_reinterpretv2sfdf = 2638, - CODE_FOR_aarch64_reinterpretdiv8qi = 2639, - CODE_FOR_aarch64_reinterpretdiv4hi = 2640, - CODE_FOR_aarch64_reinterpretdiv2si = 2641, - CODE_FOR_aarch64_reinterpretdidi = 2642, - CODE_FOR_aarch64_reinterpretdidf = 2643, - CODE_FOR_aarch64_reinterpretdiv2sf = 2644, - CODE_FOR_aarch64_reinterpretdfv8qi = 2645, - CODE_FOR_aarch64_reinterpretdfv4hi = 2646, - CODE_FOR_aarch64_reinterpretdfv2si = 2647, - CODE_FOR_aarch64_reinterpretdfdi = 2648, - CODE_FOR_aarch64_reinterpretdfdf = 2649, - CODE_FOR_aarch64_reinterpretdfv2sf = 2650, - CODE_FOR_aarch64_reinterpretv16qiv16qi = 2651, - CODE_FOR_aarch64_reinterpretv16qiv8hi = 2652, - CODE_FOR_aarch64_reinterpretv16qiv4si = 2653, - CODE_FOR_aarch64_reinterpretv16qiv2di = 2654, - CODE_FOR_aarch64_reinterpretv16qiv4sf = 2655, - CODE_FOR_aarch64_reinterpretv16qiv2df = 2656, - CODE_FOR_aarch64_reinterpretv8hiv16qi = 2657, - CODE_FOR_aarch64_reinterpretv8hiv8hi = 2658, - CODE_FOR_aarch64_reinterpretv8hiv4si = 2659, - CODE_FOR_aarch64_reinterpretv8hiv2di = 2660, - CODE_FOR_aarch64_reinterpretv8hiv4sf = 2661, - CODE_FOR_aarch64_reinterpretv8hiv2df = 2662, - CODE_FOR_aarch64_reinterpretv4siv16qi = 2663, - CODE_FOR_aarch64_reinterpretv4siv8hi = 2664, - CODE_FOR_aarch64_reinterpretv4siv4si = 2665, - CODE_FOR_aarch64_reinterpretv4siv2di = 2666, - CODE_FOR_aarch64_reinterpretv4siv4sf = 2667, - CODE_FOR_aarch64_reinterpretv4siv2df = 2668, - CODE_FOR_aarch64_reinterpretv4sfv16qi = 2669, - CODE_FOR_aarch64_reinterpretv4sfv8hi = 2670, - CODE_FOR_aarch64_reinterpretv4sfv4si = 2671, - CODE_FOR_aarch64_reinterpretv4sfv2di = 2672, - CODE_FOR_aarch64_reinterpretv4sfv4sf = 2673, - CODE_FOR_aarch64_reinterpretv4sfv2df = 2674, - CODE_FOR_aarch64_reinterpretv2div16qi = 2675, - CODE_FOR_aarch64_reinterpretv2div8hi = 2676, - CODE_FOR_aarch64_reinterpretv2div4si = 2677, - CODE_FOR_aarch64_reinterpretv2div2di = 2678, - CODE_FOR_aarch64_reinterpretv2div4sf = 2679, - CODE_FOR_aarch64_reinterpretv2div2df = 2680, - CODE_FOR_aarch64_reinterpretv2dfv16qi = 2681, - CODE_FOR_aarch64_reinterpretv2dfv8hi = 2682, - CODE_FOR_aarch64_reinterpretv2dfv4si = 2683, - CODE_FOR_aarch64_reinterpretv2dfv2di = 2684, - CODE_FOR_aarch64_reinterpretv2dfv4sf = 2685, - CODE_FOR_aarch64_reinterpretv2dfv2df = 2686, - CODE_FOR_aarch64_simd_combinev8qi = 2693, - CODE_FOR_aarch64_simd_combinev4hi = 2694, - CODE_FOR_aarch64_simd_combinev2si = 2695, - CODE_FOR_aarch64_simd_combinev2sf = 2696, - CODE_FOR_aarch64_simd_combinedi = 2697, - CODE_FOR_aarch64_simd_combinedf = 2698, - CODE_FOR_aarch64_saddl2v16qi = 2699, - CODE_FOR_aarch64_saddl2v8hi = 2700, - CODE_FOR_aarch64_saddl2v4si = 2701, - CODE_FOR_aarch64_uaddl2v16qi = 2702, - CODE_FOR_aarch64_uaddl2v8hi = 2703, - CODE_FOR_aarch64_uaddl2v4si = 2704, - CODE_FOR_aarch64_ssubl2v16qi = 2705, - CODE_FOR_aarch64_ssubl2v8hi = 2706, - CODE_FOR_aarch64_ssubl2v4si = 2707, - CODE_FOR_aarch64_usubl2v16qi = 2708, - CODE_FOR_aarch64_usubl2v8hi = 2709, - CODE_FOR_aarch64_usubl2v4si = 2710, - CODE_FOR_aarch64_saddw2v16qi = 2711, - CODE_FOR_aarch64_saddw2v8hi = 2712, - CODE_FOR_aarch64_saddw2v4si = 2713, - CODE_FOR_aarch64_uaddw2v16qi = 2714, - CODE_FOR_aarch64_uaddw2v8hi = 2715, - CODE_FOR_aarch64_uaddw2v4si = 2716, - CODE_FOR_aarch64_ssubw2v16qi = 2717, - CODE_FOR_aarch64_ssubw2v8hi = 2718, - CODE_FOR_aarch64_ssubw2v4si = 2719, - CODE_FOR_aarch64_usubw2v16qi = 2720, - CODE_FOR_aarch64_usubw2v8hi = 2721, - CODE_FOR_aarch64_usubw2v4si = 2722, - CODE_FOR_aarch64_sqdmlal_lanev4hi = 2723, - CODE_FOR_aarch64_sqdmlal_lanev2si = 2724, - CODE_FOR_aarch64_sqdmlal_lanehi = 2725, - CODE_FOR_aarch64_sqdmlal_lanesi = 2726, - CODE_FOR_aarch64_sqdmlal_laneqv4hi = 2727, - CODE_FOR_aarch64_sqdmlal_laneqv2si = 2728, - CODE_FOR_aarch64_sqdmlal_laneqhi = 2729, - CODE_FOR_aarch64_sqdmlal_laneqsi = 2730, - CODE_FOR_aarch64_sqdmlsl_lanev4hi = 2731, - CODE_FOR_aarch64_sqdmlsl_lanev2si = 2732, - CODE_FOR_aarch64_sqdmlsl_lanehi = 2733, - CODE_FOR_aarch64_sqdmlsl_lanesi = 2734, - CODE_FOR_aarch64_sqdmlsl_laneqv4hi = 2735, - CODE_FOR_aarch64_sqdmlsl_laneqv2si = 2736, - CODE_FOR_aarch64_sqdmlsl_laneqhi = 2737, - CODE_FOR_aarch64_sqdmlsl_laneqsi = 2738, - CODE_FOR_aarch64_sqdmlal2v8hi = 2739, - CODE_FOR_aarch64_sqdmlal2v4si = 2740, - CODE_FOR_aarch64_sqdmlsl2v8hi = 2741, - CODE_FOR_aarch64_sqdmlsl2v4si = 2742, - CODE_FOR_aarch64_sqdmlal2_lanev8hi = 2743, - CODE_FOR_aarch64_sqdmlal2_lanev4si = 2744, - CODE_FOR_aarch64_sqdmlal2_laneqv8hi = 2745, - CODE_FOR_aarch64_sqdmlal2_laneqv4si = 2746, - CODE_FOR_aarch64_sqdmlsl2_lanev8hi = 2747, - CODE_FOR_aarch64_sqdmlsl2_lanev4si = 2748, - CODE_FOR_aarch64_sqdmlsl2_laneqv8hi = 2749, - CODE_FOR_aarch64_sqdmlsl2_laneqv4si = 2750, - CODE_FOR_aarch64_sqdmlal2_nv8hi = 2751, - CODE_FOR_aarch64_sqdmlal2_nv4si = 2752, - CODE_FOR_aarch64_sqdmlsl2_nv8hi = 2753, - CODE_FOR_aarch64_sqdmlsl2_nv4si = 2754, - CODE_FOR_aarch64_sqdmull_lanev4hi = 2755, - CODE_FOR_aarch64_sqdmull_lanev2si = 2756, - CODE_FOR_aarch64_sqdmull_lanehi = 2757, - CODE_FOR_aarch64_sqdmull_lanesi = 2758, - CODE_FOR_aarch64_sqdmull_laneqv4hi = 2759, - CODE_FOR_aarch64_sqdmull_laneqv2si = 2760, - CODE_FOR_aarch64_sqdmull2v8hi = 2761, - CODE_FOR_aarch64_sqdmull2v4si = 2762, - CODE_FOR_aarch64_sqdmull2_lanev8hi = 2763, - CODE_FOR_aarch64_sqdmull2_lanev4si = 2764, - CODE_FOR_aarch64_sqdmull2_laneqv8hi = 2765, - CODE_FOR_aarch64_sqdmull2_laneqv4si = 2766, - CODE_FOR_aarch64_sqdmull2_nv8hi = 2767, - CODE_FOR_aarch64_sqdmull2_nv4si = 2768, - CODE_FOR_movoi = 2779, - CODE_FOR_movci = 2780, - CODE_FOR_movxi = 2781, - CODE_FOR_aarch64_ld2v8qi = 2785, - CODE_FOR_aarch64_ld2v4hi = 2786, - CODE_FOR_aarch64_ld2v2si = 2787, - CODE_FOR_aarch64_ld2v2sf = 2788, - CODE_FOR_aarch64_ld2di = 2789, - CODE_FOR_aarch64_ld2df = 2790, - CODE_FOR_aarch64_ld3v8qi = 2791, - CODE_FOR_aarch64_ld3v4hi = 2792, - CODE_FOR_aarch64_ld3v2si = 2793, - CODE_FOR_aarch64_ld3v2sf = 2794, - CODE_FOR_aarch64_ld3di = 2795, - CODE_FOR_aarch64_ld3df = 2796, - CODE_FOR_aarch64_ld4v8qi = 2797, - CODE_FOR_aarch64_ld4v4hi = 2798, - CODE_FOR_aarch64_ld4v2si = 2799, - CODE_FOR_aarch64_ld4v2sf = 2800, - CODE_FOR_aarch64_ld4di = 2801, - CODE_FOR_aarch64_ld4df = 2802, - CODE_FOR_aarch64_ld1v8qi = 2803, - CODE_FOR_aarch64_ld1v16qi = 2804, - CODE_FOR_aarch64_ld1v4hi = 2805, - CODE_FOR_aarch64_ld1v8hi = 2806, - CODE_FOR_aarch64_ld1v2si = 2807, - CODE_FOR_aarch64_ld1v4si = 2808, - CODE_FOR_aarch64_ld1v2di = 2809, - CODE_FOR_aarch64_ld1v2sf = 2810, - CODE_FOR_aarch64_ld1v4sf = 2811, - CODE_FOR_aarch64_ld1v2df = 2812, - CODE_FOR_aarch64_ld2v16qi = 2813, - CODE_FOR_aarch64_ld3v16qi = 2814, - CODE_FOR_aarch64_ld4v16qi = 2815, - CODE_FOR_aarch64_ld2v8hi = 2816, - CODE_FOR_aarch64_ld3v8hi = 2817, - CODE_FOR_aarch64_ld4v8hi = 2818, - CODE_FOR_aarch64_ld2v4si = 2819, - CODE_FOR_aarch64_ld3v4si = 2820, - CODE_FOR_aarch64_ld4v4si = 2821, - CODE_FOR_aarch64_ld2v2di = 2822, - CODE_FOR_aarch64_ld3v2di = 2823, - CODE_FOR_aarch64_ld4v2di = 2824, - CODE_FOR_aarch64_ld2v4sf = 2825, - CODE_FOR_aarch64_ld3v4sf = 2826, - CODE_FOR_aarch64_ld4v4sf = 2827, - CODE_FOR_aarch64_ld2v2df = 2828, - CODE_FOR_aarch64_ld3v2df = 2829, - CODE_FOR_aarch64_ld4v2df = 2830, - CODE_FOR_aarch64_get_dregoiv8qi = 2831, - CODE_FOR_aarch64_get_dregoiv4hi = 2832, - CODE_FOR_aarch64_get_dregoiv2si = 2833, - CODE_FOR_aarch64_get_dregoiv2sf = 2834, - CODE_FOR_aarch64_get_dregoidi = 2835, - CODE_FOR_aarch64_get_dregoidf = 2836, - CODE_FOR_aarch64_get_dregciv8qi = 2837, - CODE_FOR_aarch64_get_dregciv4hi = 2838, - CODE_FOR_aarch64_get_dregciv2si = 2839, - CODE_FOR_aarch64_get_dregciv2sf = 2840, - CODE_FOR_aarch64_get_dregcidi = 2841, - CODE_FOR_aarch64_get_dregcidf = 2842, - CODE_FOR_aarch64_get_dregxiv8qi = 2843, - CODE_FOR_aarch64_get_dregxiv4hi = 2844, - CODE_FOR_aarch64_get_dregxiv2si = 2845, - CODE_FOR_aarch64_get_dregxiv2sf = 2846, - CODE_FOR_aarch64_get_dregxidi = 2847, - CODE_FOR_aarch64_get_dregxidf = 2848, - CODE_FOR_aarch64_get_qregoiv16qi = 2849, - CODE_FOR_aarch64_get_qregciv16qi = 2850, - CODE_FOR_aarch64_get_qregxiv16qi = 2851, - CODE_FOR_aarch64_get_qregoiv8hi = 2852, - CODE_FOR_aarch64_get_qregciv8hi = 2853, - CODE_FOR_aarch64_get_qregxiv8hi = 2854, - CODE_FOR_aarch64_get_qregoiv4si = 2855, - CODE_FOR_aarch64_get_qregciv4si = 2856, - CODE_FOR_aarch64_get_qregxiv4si = 2857, - CODE_FOR_aarch64_get_qregoiv2di = 2858, - CODE_FOR_aarch64_get_qregciv2di = 2859, - CODE_FOR_aarch64_get_qregxiv2di = 2860, - CODE_FOR_aarch64_get_qregoiv4sf = 2861, - CODE_FOR_aarch64_get_qregciv4sf = 2862, - CODE_FOR_aarch64_get_qregxiv4sf = 2863, - CODE_FOR_aarch64_get_qregoiv2df = 2864, - CODE_FOR_aarch64_get_qregciv2df = 2865, - CODE_FOR_aarch64_get_qregxiv2df = 2866, - CODE_FOR_vec_perm_constv8qi = 2867, - CODE_FOR_vec_perm_constv16qi = 2868, - CODE_FOR_vec_perm_constv4hi = 2869, - CODE_FOR_vec_perm_constv8hi = 2870, - CODE_FOR_vec_perm_constv2si = 2871, - CODE_FOR_vec_perm_constv4si = 2872, - CODE_FOR_vec_perm_constv2di = 2873, - CODE_FOR_vec_perm_constv2sf = 2874, - CODE_FOR_vec_perm_constv4sf = 2875, - CODE_FOR_vec_perm_constv2df = 2876, - CODE_FOR_vec_permv8qi = 2877, - CODE_FOR_vec_permv16qi = 2878, - CODE_FOR_aarch64_st2v8qi = 2880, - CODE_FOR_aarch64_st2v4hi = 2881, - CODE_FOR_aarch64_st2v2si = 2882, - CODE_FOR_aarch64_st2v2sf = 2883, - CODE_FOR_aarch64_st2di = 2884, - CODE_FOR_aarch64_st2df = 2885, - CODE_FOR_aarch64_st3v8qi = 2886, - CODE_FOR_aarch64_st3v4hi = 2887, - CODE_FOR_aarch64_st3v2si = 2888, - CODE_FOR_aarch64_st3v2sf = 2889, - CODE_FOR_aarch64_st3di = 2890, - CODE_FOR_aarch64_st3df = 2891, - CODE_FOR_aarch64_st4v8qi = 2892, - CODE_FOR_aarch64_st4v4hi = 2893, - CODE_FOR_aarch64_st4v2si = 2894, - CODE_FOR_aarch64_st4v2sf = 2895, - CODE_FOR_aarch64_st4di = 2896, - CODE_FOR_aarch64_st4df = 2897, - CODE_FOR_aarch64_st2v16qi = 2898, - CODE_FOR_aarch64_st3v16qi = 2899, - CODE_FOR_aarch64_st4v16qi = 2900, - CODE_FOR_aarch64_st2v8hi = 2901, - CODE_FOR_aarch64_st3v8hi = 2902, - CODE_FOR_aarch64_st4v8hi = 2903, - CODE_FOR_aarch64_st2v4si = 2904, - CODE_FOR_aarch64_st3v4si = 2905, - CODE_FOR_aarch64_st4v4si = 2906, - CODE_FOR_aarch64_st2v2di = 2907, - CODE_FOR_aarch64_st3v2di = 2908, - CODE_FOR_aarch64_st4v2di = 2909, - CODE_FOR_aarch64_st2v4sf = 2910, - CODE_FOR_aarch64_st3v4sf = 2911, - CODE_FOR_aarch64_st4v4sf = 2912, - CODE_FOR_aarch64_st2v2df = 2913, - CODE_FOR_aarch64_st3v2df = 2914, - CODE_FOR_aarch64_st4v2df = 2915, - CODE_FOR_aarch64_st2_lanev16qi = 2916, - CODE_FOR_aarch64_st2_lanev8hi = 2917, - CODE_FOR_aarch64_st2_lanev4si = 2918, - CODE_FOR_aarch64_st2_lanev2di = 2919, - CODE_FOR_aarch64_st2_lanev4sf = 2920, - CODE_FOR_aarch64_st2_lanev2df = 2921, - CODE_FOR_aarch64_st3_lanev16qi = 2922, - CODE_FOR_aarch64_st3_lanev8hi = 2923, - CODE_FOR_aarch64_st3_lanev4si = 2924, - CODE_FOR_aarch64_st3_lanev2di = 2925, - CODE_FOR_aarch64_st3_lanev4sf = 2926, - CODE_FOR_aarch64_st3_lanev2df = 2927, - CODE_FOR_aarch64_st4_lanev16qi = 2928, - CODE_FOR_aarch64_st4_lanev8hi = 2929, - CODE_FOR_aarch64_st4_lanev4si = 2930, - CODE_FOR_aarch64_st4_lanev2di = 2931, - CODE_FOR_aarch64_st4_lanev4sf = 2932, - CODE_FOR_aarch64_st4_lanev2df = 2933, - CODE_FOR_aarch64_st1v8qi = 2934, - CODE_FOR_aarch64_st1v16qi = 2935, - CODE_FOR_aarch64_st1v4hi = 2936, - CODE_FOR_aarch64_st1v8hi = 2937, - CODE_FOR_aarch64_st1v2si = 2938, - CODE_FOR_aarch64_st1v4si = 2939, - CODE_FOR_aarch64_st1v2di = 2940, - CODE_FOR_aarch64_st1v2sf = 2941, - CODE_FOR_aarch64_st1v4sf = 2942, - CODE_FOR_aarch64_st1v2df = 2943, - CODE_FOR_aarch64_set_qregoiv16qi = 2944, - CODE_FOR_aarch64_set_qregciv16qi = 2945, - CODE_FOR_aarch64_set_qregxiv16qi = 2946, - CODE_FOR_aarch64_set_qregoiv8hi = 2947, - CODE_FOR_aarch64_set_qregciv8hi = 2948, - CODE_FOR_aarch64_set_qregxiv8hi = 2949, - CODE_FOR_aarch64_set_qregoiv4si = 2950, - CODE_FOR_aarch64_set_qregciv4si = 2951, - CODE_FOR_aarch64_set_qregxiv4si = 2952, - CODE_FOR_aarch64_set_qregoiv2di = 2953, - CODE_FOR_aarch64_set_qregciv2di = 2954, - CODE_FOR_aarch64_set_qregxiv2di = 2955, - CODE_FOR_aarch64_set_qregoiv4sf = 2956, - CODE_FOR_aarch64_set_qregciv4sf = 2957, - CODE_FOR_aarch64_set_qregxiv4sf = 2958, - CODE_FOR_aarch64_set_qregoiv2df = 2959, - CODE_FOR_aarch64_set_qregciv2df = 2960, - CODE_FOR_aarch64_set_qregxiv2df = 2961, - CODE_FOR_vec_initv8qi = 2962, - CODE_FOR_vec_initv16qi = 2963, - CODE_FOR_vec_initv4hi = 2964, - CODE_FOR_vec_initv8hi = 2965, - CODE_FOR_vec_initv2si = 2966, - CODE_FOR_vec_initv4si = 2967, - CODE_FOR_vec_initv2di = 2968, - CODE_FOR_vec_initv2sf = 2969, - CODE_FOR_vec_initv4sf = 2970, - CODE_FOR_vec_initv2df = 2971, - CODE_FOR_vec_extractv8qi = 2972, - CODE_FOR_vec_extractv16qi = 2973, - CODE_FOR_vec_extractv4hi = 2974, - CODE_FOR_vec_extractv8hi = 2975, - CODE_FOR_vec_extractv2si = 2976, - CODE_FOR_vec_extractv4si = 2977, - CODE_FOR_vec_extractv2di = 2978, - CODE_FOR_vec_extractv2sf = 2979, - CODE_FOR_vec_extractv4sf = 2980, - CODE_FOR_vec_extractv2df = 2981, - CODE_FOR_atomic_compare_and_swapqi = 2982, - CODE_FOR_atomic_compare_and_swaphi = 2983, - CODE_FOR_atomic_compare_and_swapsi = 2984, - CODE_FOR_atomic_compare_and_swapdi = 2985, - CODE_FOR_mem_thread_fence = 3066, - CODE_FOR_dmb = 3067, + CODE_FOR_aarch64_crc32b = 370, + CODE_FOR_aarch64_crc32h = 371, + CODE_FOR_aarch64_crc32w = 372, + CODE_FOR_aarch64_crc32x = 373, + CODE_FOR_aarch64_crc32cb = 374, + CODE_FOR_aarch64_crc32ch = 375, + CODE_FOR_aarch64_crc32cw = 376, + CODE_FOR_aarch64_crc32cx = 377, + CODE_FOR_csinc3si_insn = 380, + CODE_FOR_csinc3di_insn = 381, + CODE_FOR_andsi3 = 386, + CODE_FOR_iorsi3 = 387, + CODE_FOR_xorsi3 = 388, + CODE_FOR_anddi3 = 389, + CODE_FOR_iordi3 = 390, + CODE_FOR_xordi3 = 391, + CODE_FOR_one_cmplsi2 = 455, + CODE_FOR_one_cmpldi2 = 456, + CODE_FOR_clzsi2 = 510, + CODE_FOR_clzdi2 = 511, + CODE_FOR_clrsbsi2 = 512, + CODE_FOR_clrsbdi2 = 513, + CODE_FOR_rbitsi2 = 514, + CODE_FOR_rbitdi2 = 515, + CODE_FOR_bswapsi2 = 595, + CODE_FOR_bswapdi2 = 596, + CODE_FOR_bswaphi2 = 597, + CODE_FOR_rev16si2 = 598, + CODE_FOR_rev16di2 = 599, + CODE_FOR_rev16si2_alt = 600, + CODE_FOR_rev16di2_alt = 601, + CODE_FOR_btruncsf2 = 603, + CODE_FOR_ceilsf2 = 604, + CODE_FOR_floorsf2 = 605, + CODE_FOR_frintnsf2 = 606, + CODE_FOR_nearbyintsf2 = 607, + CODE_FOR_rintsf2 = 608, + CODE_FOR_roundsf2 = 609, + CODE_FOR_btruncdf2 = 610, + CODE_FOR_ceildf2 = 611, + CODE_FOR_floordf2 = 612, + CODE_FOR_frintndf2 = 613, + CODE_FOR_nearbyintdf2 = 614, + CODE_FOR_rintdf2 = 615, + CODE_FOR_rounddf2 = 616, + CODE_FOR_lbtruncsfsi2 = 617, + CODE_FOR_lceilsfsi2 = 618, + CODE_FOR_lfloorsfsi2 = 619, + CODE_FOR_lroundsfsi2 = 620, + CODE_FOR_lfrintnsfsi2 = 621, + CODE_FOR_lbtruncusfsi2 = 622, + CODE_FOR_lceilusfsi2 = 623, + CODE_FOR_lfloorusfsi2 = 624, + CODE_FOR_lroundusfsi2 = 625, + CODE_FOR_lfrintnusfsi2 = 626, + CODE_FOR_lbtruncdfsi2 = 627, + CODE_FOR_lceildfsi2 = 628, + CODE_FOR_lfloordfsi2 = 629, + CODE_FOR_lrounddfsi2 = 630, + CODE_FOR_lfrintndfsi2 = 631, + CODE_FOR_lbtruncudfsi2 = 632, + CODE_FOR_lceiludfsi2 = 633, + CODE_FOR_lfloorudfsi2 = 634, + CODE_FOR_lroundudfsi2 = 635, + CODE_FOR_lfrintnudfsi2 = 636, + CODE_FOR_lbtruncsfdi2 = 637, + CODE_FOR_lceilsfdi2 = 638, + CODE_FOR_lfloorsfdi2 = 639, + CODE_FOR_lroundsfdi2 = 640, + CODE_FOR_lfrintnsfdi2 = 641, + CODE_FOR_lbtruncusfdi2 = 642, + CODE_FOR_lceilusfdi2 = 643, + CODE_FOR_lfloorusfdi2 = 644, + CODE_FOR_lroundusfdi2 = 645, + CODE_FOR_lfrintnusfdi2 = 646, + CODE_FOR_lbtruncdfdi2 = 647, + CODE_FOR_lceildfdi2 = 648, + CODE_FOR_lfloordfdi2 = 649, + CODE_FOR_lrounddfdi2 = 650, + CODE_FOR_lfrintndfdi2 = 651, + CODE_FOR_lbtruncudfdi2 = 652, + CODE_FOR_lceiludfdi2 = 653, + CODE_FOR_lfloorudfdi2 = 654, + CODE_FOR_lroundudfdi2 = 655, + CODE_FOR_lfrintnudfdi2 = 656, + CODE_FOR_fmasf4 = 657, + CODE_FOR_fmadf4 = 658, + CODE_FOR_fnmasf4 = 659, + CODE_FOR_fnmadf4 = 660, + CODE_FOR_fmssf4 = 661, + CODE_FOR_fmsdf4 = 662, + CODE_FOR_fnmssf4 = 663, + CODE_FOR_fnmsdf4 = 664, + CODE_FOR_extendsfdf2 = 667, + CODE_FOR_truncdfsf2 = 668, + CODE_FOR_fix_truncsfsi2 = 669, + CODE_FOR_fix_truncdfsi2 = 670, + CODE_FOR_fix_truncsfdi2 = 671, + CODE_FOR_fix_truncdfdi2 = 672, + CODE_FOR_fixuns_truncsfsi2 = 673, + CODE_FOR_fixuns_truncdfsi2 = 674, + CODE_FOR_fixuns_truncsfdi2 = 675, + CODE_FOR_fixuns_truncdfdi2 = 676, + CODE_FOR_floatsisf2 = 677, + CODE_FOR_floatsidf2 = 678, + CODE_FOR_floatdisf2 = 679, + CODE_FOR_floatdidf2 = 680, + CODE_FOR_floatunssisf2 = 681, + CODE_FOR_floatunssidf2 = 682, + CODE_FOR_floatunsdisf2 = 683, + CODE_FOR_floatunsdidf2 = 684, + CODE_FOR_addsf3 = 685, + CODE_FOR_adddf3 = 686, + CODE_FOR_subsf3 = 687, + CODE_FOR_subdf3 = 688, + CODE_FOR_mulsf3 = 689, + CODE_FOR_muldf3 = 690, + CODE_FOR_divsf3 = 693, + CODE_FOR_divdf3 = 694, + CODE_FOR_negsf2 = 695, + CODE_FOR_negdf2 = 696, + CODE_FOR_sqrtsf2 = 697, + CODE_FOR_sqrtdf2 = 698, + CODE_FOR_abssf2 = 699, + CODE_FOR_absdf2 = 700, + CODE_FOR_smaxsf3 = 701, + CODE_FOR_smaxdf3 = 702, + CODE_FOR_sminsf3 = 703, + CODE_FOR_smindf3 = 704, + CODE_FOR_aarch64_movdi_tilow = 705, + CODE_FOR_aarch64_movdi_tflow = 706, + CODE_FOR_aarch64_movdi_tihigh = 707, + CODE_FOR_aarch64_movdi_tfhigh = 708, + CODE_FOR_aarch64_movtihigh_di = 709, + CODE_FOR_aarch64_movtfhigh_di = 710, + CODE_FOR_aarch64_movtilow_di = 711, + CODE_FOR_aarch64_movtflow_di = 712, + CODE_FOR_aarch64_movtilow_tilow = 713, + CODE_FOR_add_losym_si = 714, + CODE_FOR_add_losym_di = 715, + CODE_FOR_ldr_got_small_si = 716, + CODE_FOR_ldr_got_small_di = 717, + CODE_FOR_ldr_got_small_sidi = 718, + CODE_FOR_ldr_got_tiny = 719, + CODE_FOR_aarch64_load_tp_hard = 720, + CODE_FOR_tlsie_small_si = 722, + CODE_FOR_tlsie_small_di = 723, + CODE_FOR_tlsie_small_sidi = 724, + CODE_FOR_tlsle_small_si = 725, + CODE_FOR_tlsle_small_di = 726, + CODE_FOR_tlsdesc_small_si = 727, + CODE_FOR_tlsdesc_small_di = 728, + CODE_FOR_stack_tie = 729, + CODE_FOR_stack_protect_set_si = 730, + CODE_FOR_stack_protect_set_di = 731, + CODE_FOR_stack_protect_test_si = 732, + CODE_FOR_stack_protect_test_di = 733, + CODE_FOR_set_fpcr = 734, + CODE_FOR_get_fpcr = 735, + CODE_FOR_set_fpsr = 736, + CODE_FOR_get_fpsr = 737, + CODE_FOR_aarch64_simd_dupv8qi = 738, + CODE_FOR_aarch64_simd_dupv16qi = 739, + CODE_FOR_aarch64_simd_dupv4hi = 740, + CODE_FOR_aarch64_simd_dupv8hi = 741, + CODE_FOR_aarch64_simd_dupv2si = 742, + CODE_FOR_aarch64_simd_dupv4si = 743, + CODE_FOR_aarch64_simd_dupv2di = 744, + CODE_FOR_aarch64_simd_dupv2sf = 745, + CODE_FOR_aarch64_simd_dupv4sf = 746, + CODE_FOR_aarch64_simd_dupv2df = 747, + CODE_FOR_aarch64_dup_lanev8qi = 748, + CODE_FOR_aarch64_dup_lanev16qi = 749, + CODE_FOR_aarch64_dup_lanev4hi = 750, + CODE_FOR_aarch64_dup_lanev8hi = 751, + CODE_FOR_aarch64_dup_lanev2si = 752, + CODE_FOR_aarch64_dup_lanev4si = 753, + CODE_FOR_aarch64_dup_lanev2di = 754, + CODE_FOR_aarch64_dup_lanev2sf = 755, + CODE_FOR_aarch64_dup_lanev4sf = 756, + CODE_FOR_aarch64_dup_lanev2df = 757, + CODE_FOR_aarch64_dup_lane_to_128v8qi = 758, + CODE_FOR_aarch64_dup_lane_to_64v16qi = 759, + CODE_FOR_aarch64_dup_lane_to_128v4hi = 760, + CODE_FOR_aarch64_dup_lane_to_64v8hi = 761, + CODE_FOR_aarch64_dup_lane_to_128v2si = 762, + CODE_FOR_aarch64_dup_lane_to_64v4si = 763, + CODE_FOR_aarch64_dup_lane_to_64v2di = 764, + CODE_FOR_aarch64_dup_lane_to_128v2sf = 765, + CODE_FOR_aarch64_dup_lane_to_64v4sf = 766, + CODE_FOR_aarch64_dup_lane_to_64v2df = 767, + CODE_FOR_aarch64_simd_mov_from_v16qilow = 778, + CODE_FOR_aarch64_simd_mov_from_v8hilow = 779, + CODE_FOR_aarch64_simd_mov_from_v4silow = 780, + CODE_FOR_aarch64_simd_mov_from_v2dilow = 781, + CODE_FOR_aarch64_simd_mov_from_v4sflow = 782, + CODE_FOR_aarch64_simd_mov_from_v2dflow = 783, + CODE_FOR_aarch64_simd_mov_from_v16qihigh = 784, + CODE_FOR_aarch64_simd_mov_from_v8hihigh = 785, + CODE_FOR_aarch64_simd_mov_from_v4sihigh = 786, + CODE_FOR_aarch64_simd_mov_from_v2dihigh = 787, + CODE_FOR_aarch64_simd_mov_from_v4sfhigh = 788, + CODE_FOR_aarch64_simd_mov_from_v2dfhigh = 789, + CODE_FOR_ornv8qi3 = 790, + CODE_FOR_ornv16qi3 = 791, + CODE_FOR_ornv4hi3 = 792, + CODE_FOR_ornv8hi3 = 793, + CODE_FOR_ornv2si3 = 794, + CODE_FOR_ornv4si3 = 795, + CODE_FOR_ornv2di3 = 796, + CODE_FOR_bicv8qi3 = 797, + CODE_FOR_bicv16qi3 = 798, + CODE_FOR_bicv4hi3 = 799, + CODE_FOR_bicv8hi3 = 800, + CODE_FOR_bicv2si3 = 801, + CODE_FOR_bicv4si3 = 802, + CODE_FOR_bicv2di3 = 803, + CODE_FOR_addv8qi3 = 804, + CODE_FOR_addv16qi3 = 805, + CODE_FOR_addv4hi3 = 806, + CODE_FOR_addv8hi3 = 807, + CODE_FOR_addv2si3 = 808, + CODE_FOR_addv4si3 = 809, + CODE_FOR_addv2di3 = 810, + CODE_FOR_subv8qi3 = 811, + CODE_FOR_subv16qi3 = 812, + CODE_FOR_subv4hi3 = 813, + CODE_FOR_subv8hi3 = 814, + CODE_FOR_subv2si3 = 815, + CODE_FOR_subv4si3 = 816, + CODE_FOR_subv2di3 = 817, + CODE_FOR_mulv8qi3 = 818, + CODE_FOR_mulv16qi3 = 819, + CODE_FOR_mulv4hi3 = 820, + CODE_FOR_mulv8hi3 = 821, + CODE_FOR_mulv2si3 = 822, + CODE_FOR_mulv4si3 = 823, + CODE_FOR_bswapv4hi = 824, + CODE_FOR_bswapv8hi = 825, + CODE_FOR_bswapv2si = 826, + CODE_FOR_bswapv4si = 827, + CODE_FOR_bswapv2di = 828, + CODE_FOR_negv8qi2 = 844, + CODE_FOR_negv16qi2 = 845, + CODE_FOR_negv4hi2 = 846, + CODE_FOR_negv8hi2 = 847, + CODE_FOR_negv2si2 = 848, + CODE_FOR_negv4si2 = 849, + CODE_FOR_negv2di2 = 850, + CODE_FOR_absv8qi2 = 851, + CODE_FOR_absv16qi2 = 852, + CODE_FOR_absv4hi2 = 853, + CODE_FOR_absv8hi2 = 854, + CODE_FOR_absv2si2 = 855, + CODE_FOR_absv4si2 = 856, + CODE_FOR_absv2di2 = 857, + CODE_FOR_abdv8qi_3 = 858, + CODE_FOR_abdv16qi_3 = 859, + CODE_FOR_abdv4hi_3 = 860, + CODE_FOR_abdv8hi_3 = 861, + CODE_FOR_abdv2si_3 = 862, + CODE_FOR_abdv4si_3 = 863, + CODE_FOR_abav8qi_3 = 864, + CODE_FOR_abav16qi_3 = 865, + CODE_FOR_abav4hi_3 = 866, + CODE_FOR_abav8hi_3 = 867, + CODE_FOR_abav2si_3 = 868, + CODE_FOR_abav4si_3 = 869, + CODE_FOR_fabdv2sf_3 = 870, + CODE_FOR_fabdv4sf_3 = 871, + CODE_FOR_fabdv2df_3 = 872, + CODE_FOR_andv8qi3 = 875, + CODE_FOR_andv16qi3 = 876, + CODE_FOR_andv4hi3 = 877, + CODE_FOR_andv8hi3 = 878, + CODE_FOR_andv2si3 = 879, + CODE_FOR_andv4si3 = 880, + CODE_FOR_andv2di3 = 881, + CODE_FOR_iorv8qi3 = 882, + CODE_FOR_iorv16qi3 = 883, + CODE_FOR_iorv4hi3 = 884, + CODE_FOR_iorv8hi3 = 885, + CODE_FOR_iorv2si3 = 886, + CODE_FOR_iorv4si3 = 887, + CODE_FOR_iorv2di3 = 888, + CODE_FOR_xorv8qi3 = 889, + CODE_FOR_xorv16qi3 = 890, + CODE_FOR_xorv4hi3 = 891, + CODE_FOR_xorv8hi3 = 892, + CODE_FOR_xorv2si3 = 893, + CODE_FOR_xorv4si3 = 894, + CODE_FOR_xorv2di3 = 895, + CODE_FOR_one_cmplv8qi2 = 896, + CODE_FOR_one_cmplv16qi2 = 897, + CODE_FOR_one_cmplv4hi2 = 898, + CODE_FOR_one_cmplv8hi2 = 899, + CODE_FOR_one_cmplv2si2 = 900, + CODE_FOR_one_cmplv4si2 = 901, + CODE_FOR_one_cmplv2di2 = 902, + CODE_FOR_aarch64_simd_vec_setv8qi = 903, + CODE_FOR_aarch64_simd_vec_setv16qi = 904, + CODE_FOR_aarch64_simd_vec_setv4hi = 905, + CODE_FOR_aarch64_simd_vec_setv8hi = 906, + CODE_FOR_aarch64_simd_vec_setv2si = 907, + CODE_FOR_aarch64_simd_vec_setv4si = 908, + CODE_FOR_aarch64_simd_lshrv8qi = 909, + CODE_FOR_aarch64_simd_lshrv16qi = 910, + CODE_FOR_aarch64_simd_lshrv4hi = 911, + CODE_FOR_aarch64_simd_lshrv8hi = 912, + CODE_FOR_aarch64_simd_lshrv2si = 913, + CODE_FOR_aarch64_simd_lshrv4si = 914, + CODE_FOR_aarch64_simd_lshrv2di = 915, + CODE_FOR_aarch64_simd_ashrv8qi = 916, + CODE_FOR_aarch64_simd_ashrv16qi = 917, + CODE_FOR_aarch64_simd_ashrv4hi = 918, + CODE_FOR_aarch64_simd_ashrv8hi = 919, + CODE_FOR_aarch64_simd_ashrv2si = 920, + CODE_FOR_aarch64_simd_ashrv4si = 921, + CODE_FOR_aarch64_simd_ashrv2di = 922, + CODE_FOR_aarch64_simd_imm_shlv8qi = 923, + CODE_FOR_aarch64_simd_imm_shlv16qi = 924, + CODE_FOR_aarch64_simd_imm_shlv4hi = 925, + CODE_FOR_aarch64_simd_imm_shlv8hi = 926, + CODE_FOR_aarch64_simd_imm_shlv2si = 927, + CODE_FOR_aarch64_simd_imm_shlv4si = 928, + CODE_FOR_aarch64_simd_imm_shlv2di = 929, + CODE_FOR_aarch64_simd_reg_sshlv8qi = 930, + CODE_FOR_aarch64_simd_reg_sshlv16qi = 931, + CODE_FOR_aarch64_simd_reg_sshlv4hi = 932, + CODE_FOR_aarch64_simd_reg_sshlv8hi = 933, + CODE_FOR_aarch64_simd_reg_sshlv2si = 934, + CODE_FOR_aarch64_simd_reg_sshlv4si = 935, + CODE_FOR_aarch64_simd_reg_sshlv2di = 936, + CODE_FOR_aarch64_simd_reg_shlv8qi_unsigned = 937, + CODE_FOR_aarch64_simd_reg_shlv16qi_unsigned = 938, + CODE_FOR_aarch64_simd_reg_shlv4hi_unsigned = 939, + CODE_FOR_aarch64_simd_reg_shlv8hi_unsigned = 940, + CODE_FOR_aarch64_simd_reg_shlv2si_unsigned = 941, + CODE_FOR_aarch64_simd_reg_shlv4si_unsigned = 942, + CODE_FOR_aarch64_simd_reg_shlv2di_unsigned = 943, + CODE_FOR_aarch64_simd_reg_shlv8qi_signed = 944, + CODE_FOR_aarch64_simd_reg_shlv16qi_signed = 945, + CODE_FOR_aarch64_simd_reg_shlv4hi_signed = 946, + CODE_FOR_aarch64_simd_reg_shlv8hi_signed = 947, + CODE_FOR_aarch64_simd_reg_shlv2si_signed = 948, + CODE_FOR_aarch64_simd_reg_shlv4si_signed = 949, + CODE_FOR_aarch64_simd_reg_shlv2di_signed = 950, + CODE_FOR_aarch64_sshr_simddi = 951, + CODE_FOR_aarch64_ushr_simddi = 952, + CODE_FOR_aarch64_simd_vec_setv2di = 953, + CODE_FOR_aarch64_simd_vec_setv2sf = 954, + CODE_FOR_aarch64_simd_vec_setv4sf = 955, + CODE_FOR_aarch64_simd_vec_setv2df = 956, + CODE_FOR_aarch64_mlav8qi = 957, + CODE_FOR_aarch64_mlav16qi = 958, + CODE_FOR_aarch64_mlav4hi = 959, + CODE_FOR_aarch64_mlav8hi = 960, + CODE_FOR_aarch64_mlav2si = 961, + CODE_FOR_aarch64_mlav4si = 962, + CODE_FOR_aarch64_mlsv8qi = 971, + CODE_FOR_aarch64_mlsv16qi = 972, + CODE_FOR_aarch64_mlsv4hi = 973, + CODE_FOR_aarch64_mlsv8hi = 974, + CODE_FOR_aarch64_mlsv2si = 975, + CODE_FOR_aarch64_mlsv4si = 976, + CODE_FOR_smaxv8qi3 = 985, + CODE_FOR_sminv8qi3 = 986, + CODE_FOR_umaxv8qi3 = 987, + CODE_FOR_uminv8qi3 = 988, + CODE_FOR_smaxv16qi3 = 989, + CODE_FOR_sminv16qi3 = 990, + CODE_FOR_umaxv16qi3 = 991, + CODE_FOR_uminv16qi3 = 992, + CODE_FOR_smaxv4hi3 = 993, + CODE_FOR_sminv4hi3 = 994, + CODE_FOR_umaxv4hi3 = 995, + CODE_FOR_uminv4hi3 = 996, + CODE_FOR_smaxv8hi3 = 997, + CODE_FOR_sminv8hi3 = 998, + CODE_FOR_umaxv8hi3 = 999, + CODE_FOR_uminv8hi3 = 1000, + CODE_FOR_smaxv2si3 = 1001, + CODE_FOR_sminv2si3 = 1002, + CODE_FOR_umaxv2si3 = 1003, + CODE_FOR_uminv2si3 = 1004, + CODE_FOR_smaxv4si3 = 1005, + CODE_FOR_sminv4si3 = 1006, + CODE_FOR_umaxv4si3 = 1007, + CODE_FOR_uminv4si3 = 1008, + CODE_FOR_move_lo_quad_internal_v16qi = 1009, + CODE_FOR_move_lo_quad_internal_v8hi = 1010, + CODE_FOR_move_lo_quad_internal_v4si = 1011, + CODE_FOR_move_lo_quad_internal_v2di = 1012, + CODE_FOR_move_lo_quad_internal_v4sf = 1013, + CODE_FOR_move_lo_quad_internal_v2df = 1014, + CODE_FOR_move_lo_quad_internal_be_v16qi = 1015, + CODE_FOR_move_lo_quad_internal_be_v8hi = 1016, + CODE_FOR_move_lo_quad_internal_be_v4si = 1017, + CODE_FOR_move_lo_quad_internal_be_v2di = 1018, + CODE_FOR_move_lo_quad_internal_be_v4sf = 1019, + CODE_FOR_move_lo_quad_internal_be_v2df = 1020, + CODE_FOR_aarch64_simd_move_hi_quad_v16qi = 1021, + CODE_FOR_aarch64_simd_move_hi_quad_v8hi = 1022, + CODE_FOR_aarch64_simd_move_hi_quad_v4si = 1023, + CODE_FOR_aarch64_simd_move_hi_quad_v2di = 1024, + CODE_FOR_aarch64_simd_move_hi_quad_v4sf = 1025, + CODE_FOR_aarch64_simd_move_hi_quad_v2df = 1026, + CODE_FOR_aarch64_simd_move_hi_quad_be_v16qi = 1027, + CODE_FOR_aarch64_simd_move_hi_quad_be_v8hi = 1028, + CODE_FOR_aarch64_simd_move_hi_quad_be_v4si = 1029, + CODE_FOR_aarch64_simd_move_hi_quad_be_v2di = 1030, + CODE_FOR_aarch64_simd_move_hi_quad_be_v4sf = 1031, + CODE_FOR_aarch64_simd_move_hi_quad_be_v2df = 1032, + CODE_FOR_aarch64_simd_vec_pack_trunc_v8hi = 1033, + CODE_FOR_aarch64_simd_vec_pack_trunc_v4si = 1034, + CODE_FOR_aarch64_simd_vec_pack_trunc_v2di = 1035, + CODE_FOR_vec_pack_trunc_v8hi = 1036, + CODE_FOR_vec_pack_trunc_v4si = 1037, + CODE_FOR_vec_pack_trunc_v2di = 1038, + CODE_FOR_aarch64_simd_vec_unpacks_lo_v16qi = 1039, + CODE_FOR_aarch64_simd_vec_unpacku_lo_v16qi = 1040, + CODE_FOR_aarch64_simd_vec_unpacks_lo_v8hi = 1041, + CODE_FOR_aarch64_simd_vec_unpacku_lo_v8hi = 1042, + CODE_FOR_aarch64_simd_vec_unpacks_lo_v4si = 1043, + CODE_FOR_aarch64_simd_vec_unpacku_lo_v4si = 1044, + CODE_FOR_aarch64_simd_vec_unpacks_hi_v16qi = 1045, + CODE_FOR_aarch64_simd_vec_unpacku_hi_v16qi = 1046, + CODE_FOR_aarch64_simd_vec_unpacks_hi_v8hi = 1047, + CODE_FOR_aarch64_simd_vec_unpacku_hi_v8hi = 1048, + CODE_FOR_aarch64_simd_vec_unpacks_hi_v4si = 1049, + CODE_FOR_aarch64_simd_vec_unpacku_hi_v4si = 1050, + CODE_FOR_aarch64_simd_vec_smult_lo_v16qi = 1087, + CODE_FOR_aarch64_simd_vec_umult_lo_v16qi = 1088, + CODE_FOR_aarch64_simd_vec_smult_lo_v8hi = 1089, + CODE_FOR_aarch64_simd_vec_umult_lo_v8hi = 1090, + CODE_FOR_aarch64_simd_vec_smult_lo_v4si = 1091, + CODE_FOR_aarch64_simd_vec_umult_lo_v4si = 1092, + CODE_FOR_aarch64_simd_vec_smult_hi_v16qi = 1093, + CODE_FOR_aarch64_simd_vec_umult_hi_v16qi = 1094, + CODE_FOR_aarch64_simd_vec_smult_hi_v8hi = 1095, + CODE_FOR_aarch64_simd_vec_umult_hi_v8hi = 1096, + CODE_FOR_aarch64_simd_vec_smult_hi_v4si = 1097, + CODE_FOR_aarch64_simd_vec_umult_hi_v4si = 1098, + CODE_FOR_addv2sf3 = 1099, + CODE_FOR_addv4sf3 = 1100, + CODE_FOR_addv2df3 = 1101, + CODE_FOR_subv2sf3 = 1102, + CODE_FOR_subv4sf3 = 1103, + CODE_FOR_subv2df3 = 1104, + CODE_FOR_mulv2sf3 = 1105, + CODE_FOR_mulv4sf3 = 1106, + CODE_FOR_mulv2df3 = 1107, + CODE_FOR_divv2sf3 = 1108, + CODE_FOR_divv4sf3 = 1109, + CODE_FOR_divv2df3 = 1110, + CODE_FOR_negv2sf2 = 1111, + CODE_FOR_negv4sf2 = 1112, + CODE_FOR_negv2df2 = 1113, + CODE_FOR_absv2sf2 = 1114, + CODE_FOR_absv4sf2 = 1115, + CODE_FOR_absv2df2 = 1116, + CODE_FOR_fmav2sf4 = 1117, + CODE_FOR_fmav4sf4 = 1118, + CODE_FOR_fmav2df4 = 1119, + CODE_FOR_fnmav2sf4 = 1127, + CODE_FOR_fnmav4sf4 = 1128, + CODE_FOR_fnmav2df4 = 1129, + CODE_FOR_btruncv2sf2 = 1137, + CODE_FOR_ceilv2sf2 = 1138, + CODE_FOR_floorv2sf2 = 1139, + CODE_FOR_frintnv2sf2 = 1140, + CODE_FOR_nearbyintv2sf2 = 1141, + CODE_FOR_rintv2sf2 = 1142, + CODE_FOR_roundv2sf2 = 1143, + CODE_FOR_btruncv4sf2 = 1144, + CODE_FOR_ceilv4sf2 = 1145, + CODE_FOR_floorv4sf2 = 1146, + CODE_FOR_frintnv4sf2 = 1147, + CODE_FOR_nearbyintv4sf2 = 1148, + CODE_FOR_rintv4sf2 = 1149, + CODE_FOR_roundv4sf2 = 1150, + CODE_FOR_btruncv2df2 = 1151, + CODE_FOR_ceilv2df2 = 1152, + CODE_FOR_floorv2df2 = 1153, + CODE_FOR_frintnv2df2 = 1154, + CODE_FOR_nearbyintv2df2 = 1155, + CODE_FOR_rintv2df2 = 1156, + CODE_FOR_roundv2df2 = 1157, + CODE_FOR_lbtruncv2sfv2si2 = 1158, + CODE_FOR_lceilv2sfv2si2 = 1159, + CODE_FOR_lfloorv2sfv2si2 = 1160, + CODE_FOR_lroundv2sfv2si2 = 1161, + CODE_FOR_lfrintnv2sfv2si2 = 1162, + CODE_FOR_lbtruncuv2sfv2si2 = 1163, + CODE_FOR_lceiluv2sfv2si2 = 1164, + CODE_FOR_lflooruv2sfv2si2 = 1165, + CODE_FOR_lrounduv2sfv2si2 = 1166, + CODE_FOR_lfrintnuv2sfv2si2 = 1167, + CODE_FOR_lbtruncv4sfv4si2 = 1168, + CODE_FOR_lceilv4sfv4si2 = 1169, + CODE_FOR_lfloorv4sfv4si2 = 1170, + CODE_FOR_lroundv4sfv4si2 = 1171, + CODE_FOR_lfrintnv4sfv4si2 = 1172, + CODE_FOR_lbtruncuv4sfv4si2 = 1173, + CODE_FOR_lceiluv4sfv4si2 = 1174, + CODE_FOR_lflooruv4sfv4si2 = 1175, + CODE_FOR_lrounduv4sfv4si2 = 1176, + CODE_FOR_lfrintnuv4sfv4si2 = 1177, + CODE_FOR_lbtruncv2dfv2di2 = 1178, + CODE_FOR_lceilv2dfv2di2 = 1179, + CODE_FOR_lfloorv2dfv2di2 = 1180, + CODE_FOR_lroundv2dfv2di2 = 1181, + CODE_FOR_lfrintnv2dfv2di2 = 1182, + CODE_FOR_lbtruncuv2dfv2di2 = 1183, + CODE_FOR_lceiluv2dfv2di2 = 1184, + CODE_FOR_lflooruv2dfv2di2 = 1185, + CODE_FOR_lrounduv2dfv2di2 = 1186, + CODE_FOR_lfrintnuv2dfv2di2 = 1187, + CODE_FOR_floatv2siv2sf2 = 1188, + CODE_FOR_floatunsv2siv2sf2 = 1189, + CODE_FOR_floatv4siv4sf2 = 1190, + CODE_FOR_floatunsv4siv4sf2 = 1191, + CODE_FOR_floatv2div2df2 = 1192, + CODE_FOR_floatunsv2div2df2 = 1193, + CODE_FOR_vec_unpacks_lo_v4sf = 1194, + CODE_FOR_aarch64_float_extend_lo_v2df = 1195, + CODE_FOR_vec_unpacks_hi_v4sf = 1196, + CODE_FOR_aarch64_float_truncate_lo_v2sf = 1197, + CODE_FOR_aarch64_float_truncate_hi_v4sf = 1198, + CODE_FOR_aarch64_vmlsv2sf = 1199, + CODE_FOR_aarch64_vmlsv4sf = 1200, + CODE_FOR_aarch64_vmlsv2df = 1201, + CODE_FOR_smaxv2sf3 = 1202, + CODE_FOR_sminv2sf3 = 1203, + CODE_FOR_smaxv4sf3 = 1204, + CODE_FOR_sminv4sf3 = 1205, + CODE_FOR_smaxv2df3 = 1206, + CODE_FOR_sminv2df3 = 1207, + CODE_FOR_smax_nanv2sf3 = 1208, + CODE_FOR_smin_nanv2sf3 = 1209, + CODE_FOR_smax_nanv4sf3 = 1210, + CODE_FOR_smin_nanv4sf3 = 1211, + CODE_FOR_smax_nanv2df3 = 1212, + CODE_FOR_smin_nanv2df3 = 1213, + CODE_FOR_reduc_splus_v8qi = 1214, + CODE_FOR_reduc_uplus_v8qi = 1215, + CODE_FOR_reduc_splus_v16qi = 1216, + CODE_FOR_reduc_uplus_v16qi = 1217, + CODE_FOR_reduc_splus_v4hi = 1218, + CODE_FOR_reduc_uplus_v4hi = 1219, + CODE_FOR_reduc_splus_v8hi = 1220, + CODE_FOR_reduc_uplus_v8hi = 1221, + CODE_FOR_reduc_splus_v4si = 1222, + CODE_FOR_reduc_uplus_v4si = 1223, + CODE_FOR_reduc_splus_v2di = 1224, + CODE_FOR_reduc_uplus_v2di = 1225, + CODE_FOR_reduc_splus_v2si = 1226, + CODE_FOR_reduc_uplus_v2si = 1227, + CODE_FOR_reduc_splus_v2sf = 1228, + CODE_FOR_reduc_splus_v2df = 1229, + CODE_FOR_aarch64_addpv4sf = 1230, + CODE_FOR_clzv8qi2 = 1231, + CODE_FOR_clzv16qi2 = 1232, + CODE_FOR_clzv4hi2 = 1233, + CODE_FOR_clzv8hi2 = 1234, + CODE_FOR_clzv2si2 = 1235, + CODE_FOR_clzv4si2 = 1236, + CODE_FOR_reduc_umax_v8qi = 1237, + CODE_FOR_reduc_umin_v8qi = 1238, + CODE_FOR_reduc_smax_v8qi = 1239, + CODE_FOR_reduc_smin_v8qi = 1240, + CODE_FOR_reduc_umax_v16qi = 1241, + CODE_FOR_reduc_umin_v16qi = 1242, + CODE_FOR_reduc_smax_v16qi = 1243, + CODE_FOR_reduc_smin_v16qi = 1244, + CODE_FOR_reduc_umax_v4hi = 1245, + CODE_FOR_reduc_umin_v4hi = 1246, + CODE_FOR_reduc_smax_v4hi = 1247, + CODE_FOR_reduc_smin_v4hi = 1248, + CODE_FOR_reduc_umax_v8hi = 1249, + CODE_FOR_reduc_umin_v8hi = 1250, + CODE_FOR_reduc_smax_v8hi = 1251, + CODE_FOR_reduc_smin_v8hi = 1252, + CODE_FOR_reduc_umax_v4si = 1253, + CODE_FOR_reduc_umin_v4si = 1254, + CODE_FOR_reduc_smax_v4si = 1255, + CODE_FOR_reduc_smin_v4si = 1256, + CODE_FOR_reduc_umax_v2si = 1257, + CODE_FOR_reduc_umin_v2si = 1258, + CODE_FOR_reduc_smax_v2si = 1259, + CODE_FOR_reduc_smin_v2si = 1260, + CODE_FOR_reduc_smax_nan_v2sf = 1261, + CODE_FOR_reduc_smin_nan_v2sf = 1262, + CODE_FOR_reduc_smax_v2sf = 1263, + CODE_FOR_reduc_smin_v2sf = 1264, + CODE_FOR_reduc_smax_nan_v2df = 1265, + CODE_FOR_reduc_smin_nan_v2df = 1266, + CODE_FOR_reduc_smax_v2df = 1267, + CODE_FOR_reduc_smin_v2df = 1268, + CODE_FOR_reduc_smax_nan_v4sf = 1269, + CODE_FOR_reduc_smin_nan_v4sf = 1270, + CODE_FOR_reduc_smax_v4sf = 1271, + CODE_FOR_reduc_smin_v4sf = 1272, + CODE_FOR_aarch64_simd_bslv8qi_internal = 1273, + CODE_FOR_aarch64_simd_bslv16qi_internal = 1274, + CODE_FOR_aarch64_simd_bslv4hi_internal = 1275, + CODE_FOR_aarch64_simd_bslv8hi_internal = 1276, + CODE_FOR_aarch64_simd_bslv2si_internal = 1277, + CODE_FOR_aarch64_simd_bslv4si_internal = 1278, + CODE_FOR_aarch64_simd_bslv2di_internal = 1279, + CODE_FOR_aarch64_simd_bslv2sf_internal = 1280, + CODE_FOR_aarch64_simd_bslv4sf_internal = 1281, + CODE_FOR_aarch64_simd_bslv2df_internal = 1282, + CODE_FOR_aarch64_simd_bsldi_internal = 1283, + CODE_FOR_aarch64_simd_bsldf_internal = 1284, + CODE_FOR_aarch64_get_lanev8qi = 1297, + CODE_FOR_aarch64_get_lanev16qi = 1298, + CODE_FOR_aarch64_get_lanev4hi = 1299, + CODE_FOR_aarch64_get_lanev8hi = 1300, + CODE_FOR_aarch64_get_lanev2si = 1301, + CODE_FOR_aarch64_get_lanev4si = 1302, + CODE_FOR_aarch64_get_lanev2di = 1303, + CODE_FOR_aarch64_get_lanev2sf = 1304, + CODE_FOR_aarch64_get_lanev4sf = 1305, + CODE_FOR_aarch64_get_lanev2df = 1306, + CODE_FOR_aarch64_combine_internalv8qi = 1313, + CODE_FOR_aarch64_combine_internalv4hi = 1314, + CODE_FOR_aarch64_combine_internalv2si = 1315, + CODE_FOR_aarch64_combine_internalv2sf = 1316, + CODE_FOR_aarch64_combine_internaldi = 1317, + CODE_FOR_aarch64_combine_internaldf = 1318, + CODE_FOR_aarch64_saddlv16qi_hi_internal = 1319, + CODE_FOR_aarch64_ssublv16qi_hi_internal = 1320, + CODE_FOR_aarch64_uaddlv16qi_hi_internal = 1321, + CODE_FOR_aarch64_usublv16qi_hi_internal = 1322, + CODE_FOR_aarch64_saddlv8hi_hi_internal = 1323, + CODE_FOR_aarch64_ssublv8hi_hi_internal = 1324, + CODE_FOR_aarch64_uaddlv8hi_hi_internal = 1325, + CODE_FOR_aarch64_usublv8hi_hi_internal = 1326, + CODE_FOR_aarch64_saddlv4si_hi_internal = 1327, + CODE_FOR_aarch64_ssublv4si_hi_internal = 1328, + CODE_FOR_aarch64_uaddlv4si_hi_internal = 1329, + CODE_FOR_aarch64_usublv4si_hi_internal = 1330, + CODE_FOR_aarch64_saddlv16qi_lo_internal = 1331, + CODE_FOR_aarch64_ssublv16qi_lo_internal = 1332, + CODE_FOR_aarch64_uaddlv16qi_lo_internal = 1333, + CODE_FOR_aarch64_usublv16qi_lo_internal = 1334, + CODE_FOR_aarch64_saddlv8hi_lo_internal = 1335, + CODE_FOR_aarch64_ssublv8hi_lo_internal = 1336, + CODE_FOR_aarch64_uaddlv8hi_lo_internal = 1337, + CODE_FOR_aarch64_usublv8hi_lo_internal = 1338, + CODE_FOR_aarch64_saddlv4si_lo_internal = 1339, + CODE_FOR_aarch64_ssublv4si_lo_internal = 1340, + CODE_FOR_aarch64_uaddlv4si_lo_internal = 1341, + CODE_FOR_aarch64_usublv4si_lo_internal = 1342, + CODE_FOR_aarch64_saddlv8qi = 1343, + CODE_FOR_aarch64_ssublv8qi = 1344, + CODE_FOR_aarch64_uaddlv8qi = 1345, + CODE_FOR_aarch64_usublv8qi = 1346, + CODE_FOR_aarch64_saddlv4hi = 1347, + CODE_FOR_aarch64_ssublv4hi = 1348, + CODE_FOR_aarch64_uaddlv4hi = 1349, + CODE_FOR_aarch64_usublv4hi = 1350, + CODE_FOR_aarch64_saddlv2si = 1351, + CODE_FOR_aarch64_ssublv2si = 1352, + CODE_FOR_aarch64_uaddlv2si = 1353, + CODE_FOR_aarch64_usublv2si = 1354, + CODE_FOR_aarch64_saddwv8qi = 1355, + CODE_FOR_aarch64_ssubwv8qi = 1356, + CODE_FOR_aarch64_uaddwv8qi = 1357, + CODE_FOR_aarch64_usubwv8qi = 1358, + CODE_FOR_aarch64_saddwv4hi = 1359, + CODE_FOR_aarch64_ssubwv4hi = 1360, + CODE_FOR_aarch64_uaddwv4hi = 1361, + CODE_FOR_aarch64_usubwv4hi = 1362, + CODE_FOR_aarch64_saddwv2si = 1363, + CODE_FOR_aarch64_ssubwv2si = 1364, + CODE_FOR_aarch64_uaddwv2si = 1365, + CODE_FOR_aarch64_usubwv2si = 1366, + CODE_FOR_aarch64_saddw2v16qi_internal = 1367, + CODE_FOR_aarch64_ssubw2v16qi_internal = 1368, + CODE_FOR_aarch64_uaddw2v16qi_internal = 1369, + CODE_FOR_aarch64_usubw2v16qi_internal = 1370, + CODE_FOR_aarch64_saddw2v8hi_internal = 1371, + CODE_FOR_aarch64_ssubw2v8hi_internal = 1372, + CODE_FOR_aarch64_uaddw2v8hi_internal = 1373, + CODE_FOR_aarch64_usubw2v8hi_internal = 1374, + CODE_FOR_aarch64_saddw2v4si_internal = 1375, + CODE_FOR_aarch64_ssubw2v4si_internal = 1376, + CODE_FOR_aarch64_uaddw2v4si_internal = 1377, + CODE_FOR_aarch64_usubw2v4si_internal = 1378, + CODE_FOR_aarch64_shaddv8qi = 1379, + CODE_FOR_aarch64_uhaddv8qi = 1380, + CODE_FOR_aarch64_srhaddv8qi = 1381, + CODE_FOR_aarch64_urhaddv8qi = 1382, + CODE_FOR_aarch64_shsubv8qi = 1383, + CODE_FOR_aarch64_uhsubv8qi = 1384, + CODE_FOR_aarch64_srhsubv8qi = 1385, + CODE_FOR_aarch64_urhsubv8qi = 1386, + CODE_FOR_aarch64_shaddv16qi = 1387, + CODE_FOR_aarch64_uhaddv16qi = 1388, + CODE_FOR_aarch64_srhaddv16qi = 1389, + CODE_FOR_aarch64_urhaddv16qi = 1390, + CODE_FOR_aarch64_shsubv16qi = 1391, + CODE_FOR_aarch64_uhsubv16qi = 1392, + CODE_FOR_aarch64_srhsubv16qi = 1393, + CODE_FOR_aarch64_urhsubv16qi = 1394, + CODE_FOR_aarch64_shaddv4hi = 1395, + CODE_FOR_aarch64_uhaddv4hi = 1396, + CODE_FOR_aarch64_srhaddv4hi = 1397, + CODE_FOR_aarch64_urhaddv4hi = 1398, + CODE_FOR_aarch64_shsubv4hi = 1399, + CODE_FOR_aarch64_uhsubv4hi = 1400, + CODE_FOR_aarch64_srhsubv4hi = 1401, + CODE_FOR_aarch64_urhsubv4hi = 1402, + CODE_FOR_aarch64_shaddv8hi = 1403, + CODE_FOR_aarch64_uhaddv8hi = 1404, + CODE_FOR_aarch64_srhaddv8hi = 1405, + CODE_FOR_aarch64_urhaddv8hi = 1406, + CODE_FOR_aarch64_shsubv8hi = 1407, + CODE_FOR_aarch64_uhsubv8hi = 1408, + CODE_FOR_aarch64_srhsubv8hi = 1409, + CODE_FOR_aarch64_urhsubv8hi = 1410, + CODE_FOR_aarch64_shaddv2si = 1411, + CODE_FOR_aarch64_uhaddv2si = 1412, + CODE_FOR_aarch64_srhaddv2si = 1413, + CODE_FOR_aarch64_urhaddv2si = 1414, + CODE_FOR_aarch64_shsubv2si = 1415, + CODE_FOR_aarch64_uhsubv2si = 1416, + CODE_FOR_aarch64_srhsubv2si = 1417, + CODE_FOR_aarch64_urhsubv2si = 1418, + CODE_FOR_aarch64_shaddv4si = 1419, + CODE_FOR_aarch64_uhaddv4si = 1420, + CODE_FOR_aarch64_srhaddv4si = 1421, + CODE_FOR_aarch64_urhaddv4si = 1422, + CODE_FOR_aarch64_shsubv4si = 1423, + CODE_FOR_aarch64_uhsubv4si = 1424, + CODE_FOR_aarch64_srhsubv4si = 1425, + CODE_FOR_aarch64_urhsubv4si = 1426, + CODE_FOR_aarch64_addhnv8hi = 1427, + CODE_FOR_aarch64_raddhnv8hi = 1428, + CODE_FOR_aarch64_subhnv8hi = 1429, + CODE_FOR_aarch64_rsubhnv8hi = 1430, + CODE_FOR_aarch64_addhnv4si = 1431, + CODE_FOR_aarch64_raddhnv4si = 1432, + CODE_FOR_aarch64_subhnv4si = 1433, + CODE_FOR_aarch64_rsubhnv4si = 1434, + CODE_FOR_aarch64_addhnv2di = 1435, + CODE_FOR_aarch64_raddhnv2di = 1436, + CODE_FOR_aarch64_subhnv2di = 1437, + CODE_FOR_aarch64_rsubhnv2di = 1438, + CODE_FOR_aarch64_addhn2v8hi = 1439, + CODE_FOR_aarch64_raddhn2v8hi = 1440, + CODE_FOR_aarch64_subhn2v8hi = 1441, + CODE_FOR_aarch64_rsubhn2v8hi = 1442, + CODE_FOR_aarch64_addhn2v4si = 1443, + CODE_FOR_aarch64_raddhn2v4si = 1444, + CODE_FOR_aarch64_subhn2v4si = 1445, + CODE_FOR_aarch64_rsubhn2v4si = 1446, + CODE_FOR_aarch64_addhn2v2di = 1447, + CODE_FOR_aarch64_raddhn2v2di = 1448, + CODE_FOR_aarch64_subhn2v2di = 1449, + CODE_FOR_aarch64_rsubhn2v2di = 1450, + CODE_FOR_aarch64_pmulv8qi = 1451, + CODE_FOR_aarch64_pmulv16qi = 1452, + CODE_FOR_aarch64_sqaddv8qi = 1453, + CODE_FOR_aarch64_uqaddv8qi = 1454, + CODE_FOR_aarch64_sqsubv8qi = 1455, + CODE_FOR_aarch64_uqsubv8qi = 1456, + CODE_FOR_aarch64_sqaddv16qi = 1457, + CODE_FOR_aarch64_uqaddv16qi = 1458, + CODE_FOR_aarch64_sqsubv16qi = 1459, + CODE_FOR_aarch64_uqsubv16qi = 1460, + CODE_FOR_aarch64_sqaddv4hi = 1461, + CODE_FOR_aarch64_uqaddv4hi = 1462, + CODE_FOR_aarch64_sqsubv4hi = 1463, + CODE_FOR_aarch64_uqsubv4hi = 1464, + CODE_FOR_aarch64_sqaddv8hi = 1465, + CODE_FOR_aarch64_uqaddv8hi = 1466, + CODE_FOR_aarch64_sqsubv8hi = 1467, + CODE_FOR_aarch64_uqsubv8hi = 1468, + CODE_FOR_aarch64_sqaddv2si = 1469, + CODE_FOR_aarch64_uqaddv2si = 1470, + CODE_FOR_aarch64_sqsubv2si = 1471, + CODE_FOR_aarch64_uqsubv2si = 1472, + CODE_FOR_aarch64_sqaddv4si = 1473, + CODE_FOR_aarch64_uqaddv4si = 1474, + CODE_FOR_aarch64_sqsubv4si = 1475, + CODE_FOR_aarch64_uqsubv4si = 1476, + CODE_FOR_aarch64_sqaddv2di = 1477, + CODE_FOR_aarch64_uqaddv2di = 1478, + CODE_FOR_aarch64_sqsubv2di = 1479, + CODE_FOR_aarch64_uqsubv2di = 1480, + CODE_FOR_aarch64_sqaddqi = 1481, + CODE_FOR_aarch64_uqaddqi = 1482, + CODE_FOR_aarch64_sqsubqi = 1483, + CODE_FOR_aarch64_uqsubqi = 1484, + CODE_FOR_aarch64_sqaddhi = 1485, + CODE_FOR_aarch64_uqaddhi = 1486, + CODE_FOR_aarch64_sqsubhi = 1487, + CODE_FOR_aarch64_uqsubhi = 1488, + CODE_FOR_aarch64_sqaddsi = 1489, + CODE_FOR_aarch64_uqaddsi = 1490, + CODE_FOR_aarch64_sqsubsi = 1491, + CODE_FOR_aarch64_uqsubsi = 1492, + CODE_FOR_aarch64_sqadddi = 1493, + CODE_FOR_aarch64_uqadddi = 1494, + CODE_FOR_aarch64_sqsubdi = 1495, + CODE_FOR_aarch64_uqsubdi = 1496, + CODE_FOR_aarch64_suqaddv8qi = 1497, + CODE_FOR_aarch64_usqaddv8qi = 1498, + CODE_FOR_aarch64_suqaddv16qi = 1499, + CODE_FOR_aarch64_usqaddv16qi = 1500, + CODE_FOR_aarch64_suqaddv4hi = 1501, + CODE_FOR_aarch64_usqaddv4hi = 1502, + CODE_FOR_aarch64_suqaddv8hi = 1503, + CODE_FOR_aarch64_usqaddv8hi = 1504, + CODE_FOR_aarch64_suqaddv2si = 1505, + CODE_FOR_aarch64_usqaddv2si = 1506, + CODE_FOR_aarch64_suqaddv4si = 1507, + CODE_FOR_aarch64_usqaddv4si = 1508, + CODE_FOR_aarch64_suqaddv2di = 1509, + CODE_FOR_aarch64_usqaddv2di = 1510, + CODE_FOR_aarch64_suqaddqi = 1511, + CODE_FOR_aarch64_usqaddqi = 1512, + CODE_FOR_aarch64_suqaddhi = 1513, + CODE_FOR_aarch64_usqaddhi = 1514, + CODE_FOR_aarch64_suqaddsi = 1515, + CODE_FOR_aarch64_usqaddsi = 1516, + CODE_FOR_aarch64_suqadddi = 1517, + CODE_FOR_aarch64_usqadddi = 1518, + CODE_FOR_aarch64_sqmovunv8hi = 1519, + CODE_FOR_aarch64_sqmovunv4si = 1520, + CODE_FOR_aarch64_sqmovunv2di = 1521, + CODE_FOR_aarch64_sqmovunhi = 1522, + CODE_FOR_aarch64_sqmovunsi = 1523, + CODE_FOR_aarch64_sqmovundi = 1524, + CODE_FOR_aarch64_sqmovnv8hi = 1525, + CODE_FOR_aarch64_uqmovnv8hi = 1526, + CODE_FOR_aarch64_sqmovnv4si = 1527, + CODE_FOR_aarch64_uqmovnv4si = 1528, + CODE_FOR_aarch64_sqmovnv2di = 1529, + CODE_FOR_aarch64_uqmovnv2di = 1530, + CODE_FOR_aarch64_sqmovnhi = 1531, + CODE_FOR_aarch64_uqmovnhi = 1532, + CODE_FOR_aarch64_sqmovnsi = 1533, + CODE_FOR_aarch64_uqmovnsi = 1534, + CODE_FOR_aarch64_sqmovndi = 1535, + CODE_FOR_aarch64_uqmovndi = 1536, + CODE_FOR_aarch64_sqnegv8qi = 1537, + CODE_FOR_aarch64_sqabsv8qi = 1538, + CODE_FOR_aarch64_sqnegv16qi = 1539, + CODE_FOR_aarch64_sqabsv16qi = 1540, + CODE_FOR_aarch64_sqnegv4hi = 1541, + CODE_FOR_aarch64_sqabsv4hi = 1542, + CODE_FOR_aarch64_sqnegv8hi = 1543, + CODE_FOR_aarch64_sqabsv8hi = 1544, + CODE_FOR_aarch64_sqnegv2si = 1545, + CODE_FOR_aarch64_sqabsv2si = 1546, + CODE_FOR_aarch64_sqnegv4si = 1547, + CODE_FOR_aarch64_sqabsv4si = 1548, + CODE_FOR_aarch64_sqnegv2di = 1549, + CODE_FOR_aarch64_sqabsv2di = 1550, + CODE_FOR_aarch64_sqnegqi = 1551, + CODE_FOR_aarch64_sqabsqi = 1552, + CODE_FOR_aarch64_sqneghi = 1553, + CODE_FOR_aarch64_sqabshi = 1554, + CODE_FOR_aarch64_sqnegsi = 1555, + CODE_FOR_aarch64_sqabssi = 1556, + CODE_FOR_aarch64_sqnegdi = 1557, + CODE_FOR_aarch64_sqabsdi = 1558, + CODE_FOR_aarch64_sqdmulhv4hi = 1559, + CODE_FOR_aarch64_sqrdmulhv4hi = 1560, + CODE_FOR_aarch64_sqdmulhv8hi = 1561, + CODE_FOR_aarch64_sqrdmulhv8hi = 1562, + CODE_FOR_aarch64_sqdmulhv2si = 1563, + CODE_FOR_aarch64_sqrdmulhv2si = 1564, + CODE_FOR_aarch64_sqdmulhv4si = 1565, + CODE_FOR_aarch64_sqrdmulhv4si = 1566, + CODE_FOR_aarch64_sqdmulhhi = 1567, + CODE_FOR_aarch64_sqrdmulhhi = 1568, + CODE_FOR_aarch64_sqdmulhsi = 1569, + CODE_FOR_aarch64_sqrdmulhsi = 1570, + CODE_FOR_aarch64_sqdmulh_lanev4hi_internal = 1571, + CODE_FOR_aarch64_sqrdmulh_lanev4hi_internal = 1572, + CODE_FOR_aarch64_sqdmulh_lanev8hi_internal = 1573, + CODE_FOR_aarch64_sqrdmulh_lanev8hi_internal = 1574, + CODE_FOR_aarch64_sqdmulh_lanev2si_internal = 1575, + CODE_FOR_aarch64_sqrdmulh_lanev2si_internal = 1576, + CODE_FOR_aarch64_sqdmulh_lanev4si_internal = 1577, + CODE_FOR_aarch64_sqrdmulh_lanev4si_internal = 1578, + CODE_FOR_aarch64_sqdmulh_laneqv4hi_internal = 1579, + CODE_FOR_aarch64_sqrdmulh_laneqv4hi_internal = 1580, + CODE_FOR_aarch64_sqdmulh_laneqv8hi_internal = 1581, + CODE_FOR_aarch64_sqrdmulh_laneqv8hi_internal = 1582, + CODE_FOR_aarch64_sqdmulh_laneqv2si_internal = 1583, + CODE_FOR_aarch64_sqrdmulh_laneqv2si_internal = 1584, + CODE_FOR_aarch64_sqdmulh_laneqv4si_internal = 1585, + CODE_FOR_aarch64_sqrdmulh_laneqv4si_internal = 1586, + CODE_FOR_aarch64_sqdmulh_lanehi_internal = 1587, + CODE_FOR_aarch64_sqrdmulh_lanehi_internal = 1588, + CODE_FOR_aarch64_sqdmulh_lanesi_internal = 1589, + CODE_FOR_aarch64_sqrdmulh_lanesi_internal = 1590, + CODE_FOR_aarch64_sqdmlalv4hi = 1591, + CODE_FOR_aarch64_sqdmlslv4hi = 1592, + CODE_FOR_aarch64_sqdmlalv2si = 1593, + CODE_FOR_aarch64_sqdmlslv2si = 1594, + CODE_FOR_aarch64_sqdmlalhi = 1595, + CODE_FOR_aarch64_sqdmlslhi = 1596, + CODE_FOR_aarch64_sqdmlalsi = 1597, + CODE_FOR_aarch64_sqdmlslsi = 1598, + CODE_FOR_aarch64_sqdmlal_lanev4hi_internal = 1599, + CODE_FOR_aarch64_sqdmlsl_lanev4hi_internal = 1600, + CODE_FOR_aarch64_sqdmlal_lanev2si_internal = 1601, + CODE_FOR_aarch64_sqdmlsl_lanev2si_internal = 1602, + CODE_FOR_aarch64_sqdmlal_laneqv4hi_internal = 1603, + CODE_FOR_aarch64_sqdmlsl_laneqv4hi_internal = 1604, + CODE_FOR_aarch64_sqdmlal_laneqv2si_internal = 1605, + CODE_FOR_aarch64_sqdmlsl_laneqv2si_internal = 1606, + CODE_FOR_aarch64_sqdmlal_lanehi_internal = 1607, + CODE_FOR_aarch64_sqdmlsl_lanehi_internal = 1608, + CODE_FOR_aarch64_sqdmlal_lanesi_internal = 1609, + CODE_FOR_aarch64_sqdmlsl_lanesi_internal = 1610, + CODE_FOR_aarch64_sqdmlal_laneqhi_internal = 1611, + CODE_FOR_aarch64_sqdmlsl_laneqhi_internal = 1612, + CODE_FOR_aarch64_sqdmlal_laneqsi_internal = 1613, + CODE_FOR_aarch64_sqdmlsl_laneqsi_internal = 1614, + CODE_FOR_aarch64_sqdmlal_nv4hi = 1615, + CODE_FOR_aarch64_sqdmlsl_nv4hi = 1616, + CODE_FOR_aarch64_sqdmlal_nv2si = 1617, + CODE_FOR_aarch64_sqdmlsl_nv2si = 1618, + CODE_FOR_aarch64_sqdmlal2v8hi_internal = 1619, + CODE_FOR_aarch64_sqdmlsl2v8hi_internal = 1620, + CODE_FOR_aarch64_sqdmlal2v4si_internal = 1621, + CODE_FOR_aarch64_sqdmlsl2v4si_internal = 1622, + CODE_FOR_aarch64_sqdmlal2_lanev8hi_internal = 1623, + CODE_FOR_aarch64_sqdmlsl2_lanev8hi_internal = 1624, + CODE_FOR_aarch64_sqdmlal2_lanev4si_internal = 1625, + CODE_FOR_aarch64_sqdmlsl2_lanev4si_internal = 1626, + CODE_FOR_aarch64_sqdmlal2_laneqv8hi_internal = 1627, + CODE_FOR_aarch64_sqdmlsl2_laneqv8hi_internal = 1628, + CODE_FOR_aarch64_sqdmlal2_laneqv4si_internal = 1629, + CODE_FOR_aarch64_sqdmlsl2_laneqv4si_internal = 1630, + CODE_FOR_aarch64_sqdmlal2_nv8hi_internal = 1631, + CODE_FOR_aarch64_sqdmlsl2_nv8hi_internal = 1632, + CODE_FOR_aarch64_sqdmlal2_nv4si_internal = 1633, + CODE_FOR_aarch64_sqdmlsl2_nv4si_internal = 1634, + CODE_FOR_aarch64_sqdmullv4hi = 1635, + CODE_FOR_aarch64_sqdmullv2si = 1636, + CODE_FOR_aarch64_sqdmullhi = 1637, + CODE_FOR_aarch64_sqdmullsi = 1638, + CODE_FOR_aarch64_sqdmull_lanev4hi_internal = 1639, + CODE_FOR_aarch64_sqdmull_lanev2si_internal = 1640, + CODE_FOR_aarch64_sqdmull_laneqv4hi_internal = 1641, + CODE_FOR_aarch64_sqdmull_laneqv2si_internal = 1642, + CODE_FOR_aarch64_sqdmull_lanehi_internal = 1643, + CODE_FOR_aarch64_sqdmull_lanesi_internal = 1644, + CODE_FOR_aarch64_sqdmull_laneqhi_internal = 1645, + CODE_FOR_aarch64_sqdmull_laneqsi_internal = 1646, + CODE_FOR_aarch64_sqdmull_nv4hi = 1647, + CODE_FOR_aarch64_sqdmull_nv2si = 1648, + CODE_FOR_aarch64_sqdmull2v8hi_internal = 1649, + CODE_FOR_aarch64_sqdmull2v4si_internal = 1650, + CODE_FOR_aarch64_sqdmull2_lanev8hi_internal = 1651, + CODE_FOR_aarch64_sqdmull2_lanev4si_internal = 1652, + CODE_FOR_aarch64_sqdmull2_laneqv8hi_internal = 1653, + CODE_FOR_aarch64_sqdmull2_laneqv4si_internal = 1654, + CODE_FOR_aarch64_sqdmull2_nv8hi_internal = 1655, + CODE_FOR_aarch64_sqdmull2_nv4si_internal = 1656, + CODE_FOR_aarch64_sshlv8qi = 1657, + CODE_FOR_aarch64_ushlv8qi = 1658, + CODE_FOR_aarch64_srshlv8qi = 1659, + CODE_FOR_aarch64_urshlv8qi = 1660, + CODE_FOR_aarch64_sshlv16qi = 1661, + CODE_FOR_aarch64_ushlv16qi = 1662, + CODE_FOR_aarch64_srshlv16qi = 1663, + CODE_FOR_aarch64_urshlv16qi = 1664, + CODE_FOR_aarch64_sshlv4hi = 1665, + CODE_FOR_aarch64_ushlv4hi = 1666, + CODE_FOR_aarch64_srshlv4hi = 1667, + CODE_FOR_aarch64_urshlv4hi = 1668, + CODE_FOR_aarch64_sshlv8hi = 1669, + CODE_FOR_aarch64_ushlv8hi = 1670, + CODE_FOR_aarch64_srshlv8hi = 1671, + CODE_FOR_aarch64_urshlv8hi = 1672, + CODE_FOR_aarch64_sshlv2si = 1673, + CODE_FOR_aarch64_ushlv2si = 1674, + CODE_FOR_aarch64_srshlv2si = 1675, + CODE_FOR_aarch64_urshlv2si = 1676, + CODE_FOR_aarch64_sshlv4si = 1677, + CODE_FOR_aarch64_ushlv4si = 1678, + CODE_FOR_aarch64_srshlv4si = 1679, + CODE_FOR_aarch64_urshlv4si = 1680, + CODE_FOR_aarch64_sshlv2di = 1681, + CODE_FOR_aarch64_ushlv2di = 1682, + CODE_FOR_aarch64_srshlv2di = 1683, + CODE_FOR_aarch64_urshlv2di = 1684, + CODE_FOR_aarch64_sshldi = 1685, + CODE_FOR_aarch64_ushldi = 1686, + CODE_FOR_aarch64_srshldi = 1687, + CODE_FOR_aarch64_urshldi = 1688, + CODE_FOR_aarch64_sqshlv8qi = 1689, + CODE_FOR_aarch64_uqshlv8qi = 1690, + CODE_FOR_aarch64_sqrshlv8qi = 1691, + CODE_FOR_aarch64_uqrshlv8qi = 1692, + CODE_FOR_aarch64_sqshlv16qi = 1693, + CODE_FOR_aarch64_uqshlv16qi = 1694, + CODE_FOR_aarch64_sqrshlv16qi = 1695, + CODE_FOR_aarch64_uqrshlv16qi = 1696, + CODE_FOR_aarch64_sqshlv4hi = 1697, + CODE_FOR_aarch64_uqshlv4hi = 1698, + CODE_FOR_aarch64_sqrshlv4hi = 1699, + CODE_FOR_aarch64_uqrshlv4hi = 1700, + CODE_FOR_aarch64_sqshlv8hi = 1701, + CODE_FOR_aarch64_uqshlv8hi = 1702, + CODE_FOR_aarch64_sqrshlv8hi = 1703, + CODE_FOR_aarch64_uqrshlv8hi = 1704, + CODE_FOR_aarch64_sqshlv2si = 1705, + CODE_FOR_aarch64_uqshlv2si = 1706, + CODE_FOR_aarch64_sqrshlv2si = 1707, + CODE_FOR_aarch64_uqrshlv2si = 1708, + CODE_FOR_aarch64_sqshlv4si = 1709, + CODE_FOR_aarch64_uqshlv4si = 1710, + CODE_FOR_aarch64_sqrshlv4si = 1711, + CODE_FOR_aarch64_uqrshlv4si = 1712, + CODE_FOR_aarch64_sqshlv2di = 1713, + CODE_FOR_aarch64_uqshlv2di = 1714, + CODE_FOR_aarch64_sqrshlv2di = 1715, + CODE_FOR_aarch64_uqrshlv2di = 1716, + CODE_FOR_aarch64_sqshlqi = 1717, + CODE_FOR_aarch64_uqshlqi = 1718, + CODE_FOR_aarch64_sqrshlqi = 1719, + CODE_FOR_aarch64_uqrshlqi = 1720, + CODE_FOR_aarch64_sqshlhi = 1721, + CODE_FOR_aarch64_uqshlhi = 1722, + CODE_FOR_aarch64_sqrshlhi = 1723, + CODE_FOR_aarch64_uqrshlhi = 1724, + CODE_FOR_aarch64_sqshlsi = 1725, + CODE_FOR_aarch64_uqshlsi = 1726, + CODE_FOR_aarch64_sqrshlsi = 1727, + CODE_FOR_aarch64_uqrshlsi = 1728, + CODE_FOR_aarch64_sqshldi = 1729, + CODE_FOR_aarch64_uqshldi = 1730, + CODE_FOR_aarch64_sqrshldi = 1731, + CODE_FOR_aarch64_uqrshldi = 1732, + CODE_FOR_aarch64_sshll_nv8qi = 1733, + CODE_FOR_aarch64_ushll_nv8qi = 1734, + CODE_FOR_aarch64_sshll_nv4hi = 1735, + CODE_FOR_aarch64_ushll_nv4hi = 1736, + CODE_FOR_aarch64_sshll_nv2si = 1737, + CODE_FOR_aarch64_ushll_nv2si = 1738, + CODE_FOR_aarch64_sshll2_nv16qi = 1739, + CODE_FOR_aarch64_ushll2_nv16qi = 1740, + CODE_FOR_aarch64_sshll2_nv8hi = 1741, + CODE_FOR_aarch64_ushll2_nv8hi = 1742, + CODE_FOR_aarch64_sshll2_nv4si = 1743, + CODE_FOR_aarch64_ushll2_nv4si = 1744, + CODE_FOR_aarch64_srshr_nv8qi = 1745, + CODE_FOR_aarch64_urshr_nv8qi = 1746, + CODE_FOR_aarch64_srshr_nv16qi = 1747, + CODE_FOR_aarch64_urshr_nv16qi = 1748, + CODE_FOR_aarch64_srshr_nv4hi = 1749, + CODE_FOR_aarch64_urshr_nv4hi = 1750, + CODE_FOR_aarch64_srshr_nv8hi = 1751, + CODE_FOR_aarch64_urshr_nv8hi = 1752, + CODE_FOR_aarch64_srshr_nv2si = 1753, + CODE_FOR_aarch64_urshr_nv2si = 1754, + CODE_FOR_aarch64_srshr_nv4si = 1755, + CODE_FOR_aarch64_urshr_nv4si = 1756, + CODE_FOR_aarch64_srshr_nv2di = 1757, + CODE_FOR_aarch64_urshr_nv2di = 1758, + CODE_FOR_aarch64_srshr_ndi = 1759, + CODE_FOR_aarch64_urshr_ndi = 1760, + CODE_FOR_aarch64_ssra_nv8qi = 1761, + CODE_FOR_aarch64_usra_nv8qi = 1762, + CODE_FOR_aarch64_srsra_nv8qi = 1763, + CODE_FOR_aarch64_ursra_nv8qi = 1764, + CODE_FOR_aarch64_ssra_nv16qi = 1765, + CODE_FOR_aarch64_usra_nv16qi = 1766, + CODE_FOR_aarch64_srsra_nv16qi = 1767, + CODE_FOR_aarch64_ursra_nv16qi = 1768, + CODE_FOR_aarch64_ssra_nv4hi = 1769, + CODE_FOR_aarch64_usra_nv4hi = 1770, + CODE_FOR_aarch64_srsra_nv4hi = 1771, + CODE_FOR_aarch64_ursra_nv4hi = 1772, + CODE_FOR_aarch64_ssra_nv8hi = 1773, + CODE_FOR_aarch64_usra_nv8hi = 1774, + CODE_FOR_aarch64_srsra_nv8hi = 1775, + CODE_FOR_aarch64_ursra_nv8hi = 1776, + CODE_FOR_aarch64_ssra_nv2si = 1777, + CODE_FOR_aarch64_usra_nv2si = 1778, + CODE_FOR_aarch64_srsra_nv2si = 1779, + CODE_FOR_aarch64_ursra_nv2si = 1780, + CODE_FOR_aarch64_ssra_nv4si = 1781, + CODE_FOR_aarch64_usra_nv4si = 1782, + CODE_FOR_aarch64_srsra_nv4si = 1783, + CODE_FOR_aarch64_ursra_nv4si = 1784, + CODE_FOR_aarch64_ssra_nv2di = 1785, + CODE_FOR_aarch64_usra_nv2di = 1786, + CODE_FOR_aarch64_srsra_nv2di = 1787, + CODE_FOR_aarch64_ursra_nv2di = 1788, + CODE_FOR_aarch64_ssra_ndi = 1789, + CODE_FOR_aarch64_usra_ndi = 1790, + CODE_FOR_aarch64_srsra_ndi = 1791, + CODE_FOR_aarch64_ursra_ndi = 1792, + CODE_FOR_aarch64_ssli_nv8qi = 1793, + CODE_FOR_aarch64_usli_nv8qi = 1794, + CODE_FOR_aarch64_ssri_nv8qi = 1795, + CODE_FOR_aarch64_usri_nv8qi = 1796, + CODE_FOR_aarch64_ssli_nv16qi = 1797, + CODE_FOR_aarch64_usli_nv16qi = 1798, + CODE_FOR_aarch64_ssri_nv16qi = 1799, + CODE_FOR_aarch64_usri_nv16qi = 1800, + CODE_FOR_aarch64_ssli_nv4hi = 1801, + CODE_FOR_aarch64_usli_nv4hi = 1802, + CODE_FOR_aarch64_ssri_nv4hi = 1803, + CODE_FOR_aarch64_usri_nv4hi = 1804, + CODE_FOR_aarch64_ssli_nv8hi = 1805, + CODE_FOR_aarch64_usli_nv8hi = 1806, + CODE_FOR_aarch64_ssri_nv8hi = 1807, + CODE_FOR_aarch64_usri_nv8hi = 1808, + CODE_FOR_aarch64_ssli_nv2si = 1809, + CODE_FOR_aarch64_usli_nv2si = 1810, + CODE_FOR_aarch64_ssri_nv2si = 1811, + CODE_FOR_aarch64_usri_nv2si = 1812, + CODE_FOR_aarch64_ssli_nv4si = 1813, + CODE_FOR_aarch64_usli_nv4si = 1814, + CODE_FOR_aarch64_ssri_nv4si = 1815, + CODE_FOR_aarch64_usri_nv4si = 1816, + CODE_FOR_aarch64_ssli_nv2di = 1817, + CODE_FOR_aarch64_usli_nv2di = 1818, + CODE_FOR_aarch64_ssri_nv2di = 1819, + CODE_FOR_aarch64_usri_nv2di = 1820, + CODE_FOR_aarch64_ssli_ndi = 1821, + CODE_FOR_aarch64_usli_ndi = 1822, + CODE_FOR_aarch64_ssri_ndi = 1823, + CODE_FOR_aarch64_usri_ndi = 1824, + CODE_FOR_aarch64_sqshlu_nv8qi = 1825, + CODE_FOR_aarch64_sqshl_nv8qi = 1826, + CODE_FOR_aarch64_uqshl_nv8qi = 1827, + CODE_FOR_aarch64_sqshlu_nv16qi = 1828, + CODE_FOR_aarch64_sqshl_nv16qi = 1829, + CODE_FOR_aarch64_uqshl_nv16qi = 1830, + CODE_FOR_aarch64_sqshlu_nv4hi = 1831, + CODE_FOR_aarch64_sqshl_nv4hi = 1832, + CODE_FOR_aarch64_uqshl_nv4hi = 1833, + CODE_FOR_aarch64_sqshlu_nv8hi = 1834, + CODE_FOR_aarch64_sqshl_nv8hi = 1835, + CODE_FOR_aarch64_uqshl_nv8hi = 1836, + CODE_FOR_aarch64_sqshlu_nv2si = 1837, + CODE_FOR_aarch64_sqshl_nv2si = 1838, + CODE_FOR_aarch64_uqshl_nv2si = 1839, + CODE_FOR_aarch64_sqshlu_nv4si = 1840, + CODE_FOR_aarch64_sqshl_nv4si = 1841, + CODE_FOR_aarch64_uqshl_nv4si = 1842, + CODE_FOR_aarch64_sqshlu_nv2di = 1843, + CODE_FOR_aarch64_sqshl_nv2di = 1844, + CODE_FOR_aarch64_uqshl_nv2di = 1845, + CODE_FOR_aarch64_sqshlu_nqi = 1846, + CODE_FOR_aarch64_sqshl_nqi = 1847, + CODE_FOR_aarch64_uqshl_nqi = 1848, + CODE_FOR_aarch64_sqshlu_nhi = 1849, + CODE_FOR_aarch64_sqshl_nhi = 1850, + CODE_FOR_aarch64_uqshl_nhi = 1851, + CODE_FOR_aarch64_sqshlu_nsi = 1852, + CODE_FOR_aarch64_sqshl_nsi = 1853, + CODE_FOR_aarch64_uqshl_nsi = 1854, + CODE_FOR_aarch64_sqshlu_ndi = 1855, + CODE_FOR_aarch64_sqshl_ndi = 1856, + CODE_FOR_aarch64_uqshl_ndi = 1857, + CODE_FOR_aarch64_sqshrun_nv8hi = 1858, + CODE_FOR_aarch64_sqrshrun_nv8hi = 1859, + CODE_FOR_aarch64_sqshrn_nv8hi = 1860, + CODE_FOR_aarch64_uqshrn_nv8hi = 1861, + CODE_FOR_aarch64_sqrshrn_nv8hi = 1862, + CODE_FOR_aarch64_uqrshrn_nv8hi = 1863, + CODE_FOR_aarch64_sqshrun_nv4si = 1864, + CODE_FOR_aarch64_sqrshrun_nv4si = 1865, + CODE_FOR_aarch64_sqshrn_nv4si = 1866, + CODE_FOR_aarch64_uqshrn_nv4si = 1867, + CODE_FOR_aarch64_sqrshrn_nv4si = 1868, + CODE_FOR_aarch64_uqrshrn_nv4si = 1869, + CODE_FOR_aarch64_sqshrun_nv2di = 1870, + CODE_FOR_aarch64_sqrshrun_nv2di = 1871, + CODE_FOR_aarch64_sqshrn_nv2di = 1872, + CODE_FOR_aarch64_uqshrn_nv2di = 1873, + CODE_FOR_aarch64_sqrshrn_nv2di = 1874, + CODE_FOR_aarch64_uqrshrn_nv2di = 1875, + CODE_FOR_aarch64_sqshrun_nhi = 1876, + CODE_FOR_aarch64_sqrshrun_nhi = 1877, + CODE_FOR_aarch64_sqshrn_nhi = 1878, + CODE_FOR_aarch64_uqshrn_nhi = 1879, + CODE_FOR_aarch64_sqrshrn_nhi = 1880, + CODE_FOR_aarch64_uqrshrn_nhi = 1881, + CODE_FOR_aarch64_sqshrun_nsi = 1882, + CODE_FOR_aarch64_sqrshrun_nsi = 1883, + CODE_FOR_aarch64_sqshrn_nsi = 1884, + CODE_FOR_aarch64_uqshrn_nsi = 1885, + CODE_FOR_aarch64_sqrshrn_nsi = 1886, + CODE_FOR_aarch64_uqrshrn_nsi = 1887, + CODE_FOR_aarch64_sqshrun_ndi = 1888, + CODE_FOR_aarch64_sqrshrun_ndi = 1889, + CODE_FOR_aarch64_sqshrn_ndi = 1890, + CODE_FOR_aarch64_uqshrn_ndi = 1891, + CODE_FOR_aarch64_sqrshrn_ndi = 1892, + CODE_FOR_aarch64_uqrshrn_ndi = 1893, + CODE_FOR_aarch64_cmltv8qi = 1894, + CODE_FOR_aarch64_cmlev8qi = 1895, + CODE_FOR_aarch64_cmeqv8qi = 1896, + CODE_FOR_aarch64_cmgev8qi = 1897, + CODE_FOR_aarch64_cmgtv8qi = 1898, + CODE_FOR_aarch64_cmltv16qi = 1899, + CODE_FOR_aarch64_cmlev16qi = 1900, + CODE_FOR_aarch64_cmeqv16qi = 1901, + CODE_FOR_aarch64_cmgev16qi = 1902, + CODE_FOR_aarch64_cmgtv16qi = 1903, + CODE_FOR_aarch64_cmltv4hi = 1904, + CODE_FOR_aarch64_cmlev4hi = 1905, + CODE_FOR_aarch64_cmeqv4hi = 1906, + CODE_FOR_aarch64_cmgev4hi = 1907, + CODE_FOR_aarch64_cmgtv4hi = 1908, + CODE_FOR_aarch64_cmltv8hi = 1909, + CODE_FOR_aarch64_cmlev8hi = 1910, + CODE_FOR_aarch64_cmeqv8hi = 1911, + CODE_FOR_aarch64_cmgev8hi = 1912, + CODE_FOR_aarch64_cmgtv8hi = 1913, + CODE_FOR_aarch64_cmltv2si = 1914, + CODE_FOR_aarch64_cmlev2si = 1915, + CODE_FOR_aarch64_cmeqv2si = 1916, + CODE_FOR_aarch64_cmgev2si = 1917, + CODE_FOR_aarch64_cmgtv2si = 1918, + CODE_FOR_aarch64_cmltv4si = 1919, + CODE_FOR_aarch64_cmlev4si = 1920, + CODE_FOR_aarch64_cmeqv4si = 1921, + CODE_FOR_aarch64_cmgev4si = 1922, + CODE_FOR_aarch64_cmgtv4si = 1923, + CODE_FOR_aarch64_cmltv2di = 1924, + CODE_FOR_aarch64_cmlev2di = 1925, + CODE_FOR_aarch64_cmeqv2di = 1926, + CODE_FOR_aarch64_cmgev2di = 1927, + CODE_FOR_aarch64_cmgtv2di = 1928, + CODE_FOR_aarch64_cmltdi = 1929, + CODE_FOR_aarch64_cmledi = 1930, + CODE_FOR_aarch64_cmeqdi = 1931, + CODE_FOR_aarch64_cmgedi = 1932, + CODE_FOR_aarch64_cmgtdi = 1933, + CODE_FOR_aarch64_cmltuv8qi = 1939, + CODE_FOR_aarch64_cmleuv8qi = 1940, + CODE_FOR_aarch64_cmgeuv8qi = 1941, + CODE_FOR_aarch64_cmgtuv8qi = 1942, + CODE_FOR_aarch64_cmltuv16qi = 1943, + CODE_FOR_aarch64_cmleuv16qi = 1944, + CODE_FOR_aarch64_cmgeuv16qi = 1945, + CODE_FOR_aarch64_cmgtuv16qi = 1946, + CODE_FOR_aarch64_cmltuv4hi = 1947, + CODE_FOR_aarch64_cmleuv4hi = 1948, + CODE_FOR_aarch64_cmgeuv4hi = 1949, + CODE_FOR_aarch64_cmgtuv4hi = 1950, + CODE_FOR_aarch64_cmltuv8hi = 1951, + CODE_FOR_aarch64_cmleuv8hi = 1952, + CODE_FOR_aarch64_cmgeuv8hi = 1953, + CODE_FOR_aarch64_cmgtuv8hi = 1954, + CODE_FOR_aarch64_cmltuv2si = 1955, + CODE_FOR_aarch64_cmleuv2si = 1956, + CODE_FOR_aarch64_cmgeuv2si = 1957, + CODE_FOR_aarch64_cmgtuv2si = 1958, + CODE_FOR_aarch64_cmltuv4si = 1959, + CODE_FOR_aarch64_cmleuv4si = 1960, + CODE_FOR_aarch64_cmgeuv4si = 1961, + CODE_FOR_aarch64_cmgtuv4si = 1962, + CODE_FOR_aarch64_cmltuv2di = 1963, + CODE_FOR_aarch64_cmleuv2di = 1964, + CODE_FOR_aarch64_cmgeuv2di = 1965, + CODE_FOR_aarch64_cmgtuv2di = 1966, + CODE_FOR_aarch64_cmltudi = 1967, + CODE_FOR_aarch64_cmleudi = 1968, + CODE_FOR_aarch64_cmgeudi = 1969, + CODE_FOR_aarch64_cmgtudi = 1970, + CODE_FOR_aarch64_cmtstv8qi = 1975, + CODE_FOR_aarch64_cmtstv16qi = 1976, + CODE_FOR_aarch64_cmtstv4hi = 1977, + CODE_FOR_aarch64_cmtstv8hi = 1978, + CODE_FOR_aarch64_cmtstv2si = 1979, + CODE_FOR_aarch64_cmtstv4si = 1980, + CODE_FOR_aarch64_cmtstv2di = 1981, + CODE_FOR_aarch64_cmtstdi = 1982, + CODE_FOR_aarch64_cmltv2sf = 1984, + CODE_FOR_aarch64_cmlev2sf = 1985, + CODE_FOR_aarch64_cmeqv2sf = 1986, + CODE_FOR_aarch64_cmgev2sf = 1987, + CODE_FOR_aarch64_cmgtv2sf = 1988, + CODE_FOR_aarch64_cmltv4sf = 1989, + CODE_FOR_aarch64_cmlev4sf = 1990, + CODE_FOR_aarch64_cmeqv4sf = 1991, + CODE_FOR_aarch64_cmgev4sf = 1992, + CODE_FOR_aarch64_cmgtv4sf = 1993, + CODE_FOR_aarch64_cmltv2df = 1994, + CODE_FOR_aarch64_cmlev2df = 1995, + CODE_FOR_aarch64_cmeqv2df = 1996, + CODE_FOR_aarch64_cmgev2df = 1997, + CODE_FOR_aarch64_cmgtv2df = 1998, + CODE_FOR_aarch64_cmltsf = 1999, + CODE_FOR_aarch64_cmlesf = 2000, + CODE_FOR_aarch64_cmeqsf = 2001, + CODE_FOR_aarch64_cmgesf = 2002, + CODE_FOR_aarch64_cmgtsf = 2003, + CODE_FOR_aarch64_cmltdf = 2004, + CODE_FOR_aarch64_cmledf = 2005, + CODE_FOR_aarch64_cmeqdf = 2006, + CODE_FOR_aarch64_cmgedf = 2007, + CODE_FOR_aarch64_cmgtdf = 2008, + CODE_FOR_aarch64_addpv8qi = 2029, + CODE_FOR_aarch64_addpv4hi = 2030, + CODE_FOR_aarch64_addpv2si = 2031, + CODE_FOR_aarch64_addpdi = 2032, + CODE_FOR_sqrtv2sf2 = 2033, + CODE_FOR_sqrtv4sf2 = 2034, + CODE_FOR_sqrtv2df2 = 2035, + CODE_FOR_vec_load_lanesoiv16qi = 2036, + CODE_FOR_vec_load_lanesoiv8hi = 2037, + CODE_FOR_vec_load_lanesoiv4si = 2038, + CODE_FOR_vec_load_lanesoiv2di = 2039, + CODE_FOR_vec_load_lanesoiv4sf = 2040, + CODE_FOR_vec_load_lanesoiv2df = 2041, + CODE_FOR_vec_store_lanesoiv16qi = 2042, + CODE_FOR_vec_store_lanesoiv8hi = 2043, + CODE_FOR_vec_store_lanesoiv4si = 2044, + CODE_FOR_vec_store_lanesoiv2di = 2045, + CODE_FOR_vec_store_lanesoiv4sf = 2046, + CODE_FOR_vec_store_lanesoiv2df = 2047, + CODE_FOR_vec_store_lanesoi_lanev16qi = 2048, + CODE_FOR_vec_store_lanesoi_lanev8hi = 2049, + CODE_FOR_vec_store_lanesoi_lanev4si = 2050, + CODE_FOR_vec_store_lanesoi_lanev2di = 2051, + CODE_FOR_vec_store_lanesoi_lanev4sf = 2052, + CODE_FOR_vec_store_lanesoi_lanev2df = 2053, + CODE_FOR_vec_load_lanesciv16qi = 2054, + CODE_FOR_vec_load_lanesciv8hi = 2055, + CODE_FOR_vec_load_lanesciv4si = 2056, + CODE_FOR_vec_load_lanesciv2di = 2057, + CODE_FOR_vec_load_lanesciv4sf = 2058, + CODE_FOR_vec_load_lanesciv2df = 2059, + CODE_FOR_vec_store_lanesciv16qi = 2060, + CODE_FOR_vec_store_lanesciv8hi = 2061, + CODE_FOR_vec_store_lanesciv4si = 2062, + CODE_FOR_vec_store_lanesciv2di = 2063, + CODE_FOR_vec_store_lanesciv4sf = 2064, + CODE_FOR_vec_store_lanesciv2df = 2065, + CODE_FOR_vec_store_lanesci_lanev16qi = 2066, + CODE_FOR_vec_store_lanesci_lanev8hi = 2067, + CODE_FOR_vec_store_lanesci_lanev4si = 2068, + CODE_FOR_vec_store_lanesci_lanev2di = 2069, + CODE_FOR_vec_store_lanesci_lanev4sf = 2070, + CODE_FOR_vec_store_lanesci_lanev2df = 2071, + CODE_FOR_vec_load_lanesxiv16qi = 2072, + CODE_FOR_vec_load_lanesxiv8hi = 2073, + CODE_FOR_vec_load_lanesxiv4si = 2074, + CODE_FOR_vec_load_lanesxiv2di = 2075, + CODE_FOR_vec_load_lanesxiv4sf = 2076, + CODE_FOR_vec_load_lanesxiv2df = 2077, + CODE_FOR_vec_store_lanesxiv16qi = 2078, + CODE_FOR_vec_store_lanesxiv8hi = 2079, + CODE_FOR_vec_store_lanesxiv4si = 2080, + CODE_FOR_vec_store_lanesxiv2di = 2081, + CODE_FOR_vec_store_lanesxiv4sf = 2082, + CODE_FOR_vec_store_lanesxiv2df = 2083, + CODE_FOR_vec_store_lanesxi_lanev16qi = 2084, + CODE_FOR_vec_store_lanesxi_lanev8hi = 2085, + CODE_FOR_vec_store_lanesxi_lanev4si = 2086, + CODE_FOR_vec_store_lanesxi_lanev2di = 2087, + CODE_FOR_vec_store_lanesxi_lanev4sf = 2088, + CODE_FOR_vec_store_lanesxi_lanev2df = 2089, + CODE_FOR_aarch64_be_ld1v8qi = 2093, + CODE_FOR_aarch64_be_ld1v16qi = 2094, + CODE_FOR_aarch64_be_ld1v4hi = 2095, + CODE_FOR_aarch64_be_ld1v8hi = 2096, + CODE_FOR_aarch64_be_ld1v2si = 2097, + CODE_FOR_aarch64_be_ld1v4si = 2098, + CODE_FOR_aarch64_be_ld1v2di = 2099, + CODE_FOR_aarch64_be_ld1v2sf = 2100, + CODE_FOR_aarch64_be_ld1v4sf = 2101, + CODE_FOR_aarch64_be_ld1v2df = 2102, + CODE_FOR_aarch64_be_ld1di = 2103, + CODE_FOR_aarch64_be_st1v8qi = 2104, + CODE_FOR_aarch64_be_st1v16qi = 2105, + CODE_FOR_aarch64_be_st1v4hi = 2106, + CODE_FOR_aarch64_be_st1v8hi = 2107, + CODE_FOR_aarch64_be_st1v2si = 2108, + CODE_FOR_aarch64_be_st1v4si = 2109, + CODE_FOR_aarch64_be_st1v2di = 2110, + CODE_FOR_aarch64_be_st1v2sf = 2111, + CODE_FOR_aarch64_be_st1v4sf = 2112, + CODE_FOR_aarch64_be_st1v2df = 2113, + CODE_FOR_aarch64_be_st1di = 2114, + CODE_FOR_aarch64_ld2v8qi_dreg = 2115, + CODE_FOR_aarch64_ld2v4hi_dreg = 2116, + CODE_FOR_aarch64_ld2v2si_dreg = 2117, + CODE_FOR_aarch64_ld2v2sf_dreg = 2118, + CODE_FOR_aarch64_ld2di_dreg = 2119, + CODE_FOR_aarch64_ld2df_dreg = 2120, + CODE_FOR_aarch64_ld3v8qi_dreg = 2121, + CODE_FOR_aarch64_ld3v4hi_dreg = 2122, + CODE_FOR_aarch64_ld3v2si_dreg = 2123, + CODE_FOR_aarch64_ld3v2sf_dreg = 2124, + CODE_FOR_aarch64_ld3di_dreg = 2125, + CODE_FOR_aarch64_ld3df_dreg = 2126, + CODE_FOR_aarch64_ld4v8qi_dreg = 2127, + CODE_FOR_aarch64_ld4v4hi_dreg = 2128, + CODE_FOR_aarch64_ld4v2si_dreg = 2129, + CODE_FOR_aarch64_ld4v2sf_dreg = 2130, + CODE_FOR_aarch64_ld4di_dreg = 2131, + CODE_FOR_aarch64_ld4df_dreg = 2132, + CODE_FOR_aarch64_tbl1v8qi = 2133, + CODE_FOR_aarch64_tbl1v16qi = 2134, + CODE_FOR_aarch64_tbl2v16qi = 2135, + CODE_FOR_aarch64_combinev16qi = 2136, + CODE_FOR_aarch64_zip1v8qi = 2137, + CODE_FOR_aarch64_zip2v8qi = 2138, + CODE_FOR_aarch64_trn1v8qi = 2139, + CODE_FOR_aarch64_trn2v8qi = 2140, + CODE_FOR_aarch64_uzp1v8qi = 2141, + CODE_FOR_aarch64_uzp2v8qi = 2142, + CODE_FOR_aarch64_zip1v16qi = 2143, + CODE_FOR_aarch64_zip2v16qi = 2144, + CODE_FOR_aarch64_trn1v16qi = 2145, + CODE_FOR_aarch64_trn2v16qi = 2146, + CODE_FOR_aarch64_uzp1v16qi = 2147, + CODE_FOR_aarch64_uzp2v16qi = 2148, + CODE_FOR_aarch64_zip1v4hi = 2149, + CODE_FOR_aarch64_zip2v4hi = 2150, + CODE_FOR_aarch64_trn1v4hi = 2151, + CODE_FOR_aarch64_trn2v4hi = 2152, + CODE_FOR_aarch64_uzp1v4hi = 2153, + CODE_FOR_aarch64_uzp2v4hi = 2154, + CODE_FOR_aarch64_zip1v8hi = 2155, + CODE_FOR_aarch64_zip2v8hi = 2156, + CODE_FOR_aarch64_trn1v8hi = 2157, + CODE_FOR_aarch64_trn2v8hi = 2158, + CODE_FOR_aarch64_uzp1v8hi = 2159, + CODE_FOR_aarch64_uzp2v8hi = 2160, + CODE_FOR_aarch64_zip1v2si = 2161, + CODE_FOR_aarch64_zip2v2si = 2162, + CODE_FOR_aarch64_trn1v2si = 2163, + CODE_FOR_aarch64_trn2v2si = 2164, + CODE_FOR_aarch64_uzp1v2si = 2165, + CODE_FOR_aarch64_uzp2v2si = 2166, + CODE_FOR_aarch64_zip1v4si = 2167, + CODE_FOR_aarch64_zip2v4si = 2168, + CODE_FOR_aarch64_trn1v4si = 2169, + CODE_FOR_aarch64_trn2v4si = 2170, + CODE_FOR_aarch64_uzp1v4si = 2171, + CODE_FOR_aarch64_uzp2v4si = 2172, + CODE_FOR_aarch64_zip1v2di = 2173, + CODE_FOR_aarch64_zip2v2di = 2174, + CODE_FOR_aarch64_trn1v2di = 2175, + CODE_FOR_aarch64_trn2v2di = 2176, + CODE_FOR_aarch64_uzp1v2di = 2177, + CODE_FOR_aarch64_uzp2v2di = 2178, + CODE_FOR_aarch64_zip1v2sf = 2179, + CODE_FOR_aarch64_zip2v2sf = 2180, + CODE_FOR_aarch64_trn1v2sf = 2181, + CODE_FOR_aarch64_trn2v2sf = 2182, + CODE_FOR_aarch64_uzp1v2sf = 2183, + CODE_FOR_aarch64_uzp2v2sf = 2184, + CODE_FOR_aarch64_zip1v4sf = 2185, + CODE_FOR_aarch64_zip2v4sf = 2186, + CODE_FOR_aarch64_trn1v4sf = 2187, + CODE_FOR_aarch64_trn2v4sf = 2188, + CODE_FOR_aarch64_uzp1v4sf = 2189, + CODE_FOR_aarch64_uzp2v4sf = 2190, + CODE_FOR_aarch64_zip1v2df = 2191, + CODE_FOR_aarch64_zip2v2df = 2192, + CODE_FOR_aarch64_trn1v2df = 2193, + CODE_FOR_aarch64_trn2v2df = 2194, + CODE_FOR_aarch64_uzp1v2df = 2195, + CODE_FOR_aarch64_uzp2v2df = 2196, + CODE_FOR_aarch64_extv8qi = 2197, + CODE_FOR_aarch64_extv16qi = 2198, + CODE_FOR_aarch64_extv4hi = 2199, + CODE_FOR_aarch64_extv8hi = 2200, + CODE_FOR_aarch64_extv2si = 2201, + CODE_FOR_aarch64_extv4si = 2202, + CODE_FOR_aarch64_extv2di = 2203, + CODE_FOR_aarch64_extv2sf = 2204, + CODE_FOR_aarch64_extv4sf = 2205, + CODE_FOR_aarch64_extv2df = 2206, + CODE_FOR_aarch64_rev64v8qi = 2207, + CODE_FOR_aarch64_rev32v8qi = 2208, + CODE_FOR_aarch64_rev16v8qi = 2209, + CODE_FOR_aarch64_rev64v16qi = 2210, + CODE_FOR_aarch64_rev32v16qi = 2211, + CODE_FOR_aarch64_rev16v16qi = 2212, + CODE_FOR_aarch64_rev64v4hi = 2213, + CODE_FOR_aarch64_rev32v4hi = 2214, + CODE_FOR_aarch64_rev16v4hi = 2215, + CODE_FOR_aarch64_rev64v8hi = 2216, + CODE_FOR_aarch64_rev32v8hi = 2217, + CODE_FOR_aarch64_rev16v8hi = 2218, + CODE_FOR_aarch64_rev64v2si = 2219, + CODE_FOR_aarch64_rev32v2si = 2220, + CODE_FOR_aarch64_rev16v2si = 2221, + CODE_FOR_aarch64_rev64v4si = 2222, + CODE_FOR_aarch64_rev32v4si = 2223, + CODE_FOR_aarch64_rev16v4si = 2224, + CODE_FOR_aarch64_rev64v2di = 2225, + CODE_FOR_aarch64_rev32v2di = 2226, + CODE_FOR_aarch64_rev16v2di = 2227, + CODE_FOR_aarch64_rev64v2sf = 2228, + CODE_FOR_aarch64_rev32v2sf = 2229, + CODE_FOR_aarch64_rev16v2sf = 2230, + CODE_FOR_aarch64_rev64v4sf = 2231, + CODE_FOR_aarch64_rev32v4sf = 2232, + CODE_FOR_aarch64_rev16v4sf = 2233, + CODE_FOR_aarch64_rev64v2df = 2234, + CODE_FOR_aarch64_rev32v2df = 2235, + CODE_FOR_aarch64_rev16v2df = 2236, + CODE_FOR_aarch64_st2v8qi_dreg = 2237, + CODE_FOR_aarch64_st2v4hi_dreg = 2238, + CODE_FOR_aarch64_st2v2si_dreg = 2239, + CODE_FOR_aarch64_st2v2sf_dreg = 2240, + CODE_FOR_aarch64_st2di_dreg = 2241, + CODE_FOR_aarch64_st2df_dreg = 2242, + CODE_FOR_aarch64_st3v8qi_dreg = 2243, + CODE_FOR_aarch64_st3v4hi_dreg = 2244, + CODE_FOR_aarch64_st3v2si_dreg = 2245, + CODE_FOR_aarch64_st3v2sf_dreg = 2246, + CODE_FOR_aarch64_st3di_dreg = 2247, + CODE_FOR_aarch64_st3df_dreg = 2248, + CODE_FOR_aarch64_st4v8qi_dreg = 2249, + CODE_FOR_aarch64_st4v4hi_dreg = 2250, + CODE_FOR_aarch64_st4v2si_dreg = 2251, + CODE_FOR_aarch64_st4v2sf_dreg = 2252, + CODE_FOR_aarch64_st4di_dreg = 2253, + CODE_FOR_aarch64_st4df_dreg = 2254, + CODE_FOR_aarch64_frecpev2sf = 2266, + CODE_FOR_aarch64_frecpev4sf = 2267, + CODE_FOR_aarch64_frecpev2df = 2268, + CODE_FOR_aarch64_frecpesf = 2269, + CODE_FOR_aarch64_frecpxsf = 2270, + CODE_FOR_aarch64_frecpedf = 2271, + CODE_FOR_aarch64_frecpxdf = 2272, + CODE_FOR_aarch64_frecpsv2sf = 2273, + CODE_FOR_aarch64_frecpsv4sf = 2274, + CODE_FOR_aarch64_frecpsv2df = 2275, + CODE_FOR_aarch64_frecpssf = 2276, + CODE_FOR_aarch64_frecpsdf = 2277, + CODE_FOR_aarch64_crypto_aesev16qi = 2278, + CODE_FOR_aarch64_crypto_aesdv16qi = 2279, + CODE_FOR_aarch64_crypto_aesmcv16qi = 2280, + CODE_FOR_aarch64_crypto_aesimcv16qi = 2281, + CODE_FOR_aarch64_crypto_sha1hsi = 2282, + CODE_FOR_aarch64_crypto_sha1su1v4si = 2283, + CODE_FOR_aarch64_crypto_sha1cv4si = 2284, + CODE_FOR_aarch64_crypto_sha1mv4si = 2285, + CODE_FOR_aarch64_crypto_sha1pv4si = 2286, + CODE_FOR_aarch64_crypto_sha1su0v4si = 2287, + CODE_FOR_aarch64_crypto_sha256hv4si = 2288, + CODE_FOR_aarch64_crypto_sha256h2v4si = 2289, + CODE_FOR_aarch64_crypto_sha256su0v4si = 2290, + CODE_FOR_aarch64_crypto_sha256su1v4si = 2291, + CODE_FOR_aarch64_crypto_pmulldi = 2292, + CODE_FOR_aarch64_crypto_pmullv2di = 2293, + CODE_FOR_atomic_compare_and_swapqi_1 = 2294, + CODE_FOR_atomic_compare_and_swaphi_1 = 2295, + CODE_FOR_atomic_compare_and_swapsi_1 = 2296, + CODE_FOR_atomic_compare_and_swapdi_1 = 2297, + CODE_FOR_atomic_exchangeqi = 2298, + CODE_FOR_atomic_exchangehi = 2299, + CODE_FOR_atomic_exchangesi = 2300, + CODE_FOR_atomic_exchangedi = 2301, + CODE_FOR_atomic_addqi = 2302, + CODE_FOR_atomic_subqi = 2303, + CODE_FOR_atomic_orqi = 2304, + CODE_FOR_atomic_xorqi = 2305, + CODE_FOR_atomic_andqi = 2306, + CODE_FOR_atomic_addhi = 2307, + CODE_FOR_atomic_subhi = 2308, + CODE_FOR_atomic_orhi = 2309, + CODE_FOR_atomic_xorhi = 2310, + CODE_FOR_atomic_andhi = 2311, + CODE_FOR_atomic_addsi = 2312, + CODE_FOR_atomic_subsi = 2313, + CODE_FOR_atomic_orsi = 2314, + CODE_FOR_atomic_xorsi = 2315, + CODE_FOR_atomic_andsi = 2316, + CODE_FOR_atomic_adddi = 2317, + CODE_FOR_atomic_subdi = 2318, + CODE_FOR_atomic_ordi = 2319, + CODE_FOR_atomic_xordi = 2320, + CODE_FOR_atomic_anddi = 2321, + CODE_FOR_atomic_nandqi = 2322, + CODE_FOR_atomic_nandhi = 2323, + CODE_FOR_atomic_nandsi = 2324, + CODE_FOR_atomic_nanddi = 2325, + CODE_FOR_atomic_fetch_addqi = 2326, + CODE_FOR_atomic_fetch_subqi = 2327, + CODE_FOR_atomic_fetch_orqi = 2328, + CODE_FOR_atomic_fetch_xorqi = 2329, + CODE_FOR_atomic_fetch_andqi = 2330, + CODE_FOR_atomic_fetch_addhi = 2331, + CODE_FOR_atomic_fetch_subhi = 2332, + CODE_FOR_atomic_fetch_orhi = 2333, + CODE_FOR_atomic_fetch_xorhi = 2334, + CODE_FOR_atomic_fetch_andhi = 2335, + CODE_FOR_atomic_fetch_addsi = 2336, + CODE_FOR_atomic_fetch_subsi = 2337, + CODE_FOR_atomic_fetch_orsi = 2338, + CODE_FOR_atomic_fetch_xorsi = 2339, + CODE_FOR_atomic_fetch_andsi = 2340, + CODE_FOR_atomic_fetch_adddi = 2341, + CODE_FOR_atomic_fetch_subdi = 2342, + CODE_FOR_atomic_fetch_ordi = 2343, + CODE_FOR_atomic_fetch_xordi = 2344, + CODE_FOR_atomic_fetch_anddi = 2345, + CODE_FOR_atomic_fetch_nandqi = 2346, + CODE_FOR_atomic_fetch_nandhi = 2347, + CODE_FOR_atomic_fetch_nandsi = 2348, + CODE_FOR_atomic_fetch_nanddi = 2349, + CODE_FOR_atomic_add_fetchqi = 2350, + CODE_FOR_atomic_sub_fetchqi = 2351, + CODE_FOR_atomic_or_fetchqi = 2352, + CODE_FOR_atomic_xor_fetchqi = 2353, + CODE_FOR_atomic_and_fetchqi = 2354, + CODE_FOR_atomic_add_fetchhi = 2355, + CODE_FOR_atomic_sub_fetchhi = 2356, + CODE_FOR_atomic_or_fetchhi = 2357, + CODE_FOR_atomic_xor_fetchhi = 2358, + CODE_FOR_atomic_and_fetchhi = 2359, + CODE_FOR_atomic_add_fetchsi = 2360, + CODE_FOR_atomic_sub_fetchsi = 2361, + CODE_FOR_atomic_or_fetchsi = 2362, + CODE_FOR_atomic_xor_fetchsi = 2363, + CODE_FOR_atomic_and_fetchsi = 2364, + CODE_FOR_atomic_add_fetchdi = 2365, + CODE_FOR_atomic_sub_fetchdi = 2366, + CODE_FOR_atomic_or_fetchdi = 2367, + CODE_FOR_atomic_xor_fetchdi = 2368, + CODE_FOR_atomic_and_fetchdi = 2369, + CODE_FOR_atomic_nand_fetchqi = 2370, + CODE_FOR_atomic_nand_fetchhi = 2371, + CODE_FOR_atomic_nand_fetchsi = 2372, + CODE_FOR_atomic_nand_fetchdi = 2373, + CODE_FOR_atomic_loadqi = 2374, + CODE_FOR_atomic_loadhi = 2375, + CODE_FOR_atomic_loadsi = 2376, + CODE_FOR_atomic_loaddi = 2377, + CODE_FOR_atomic_storeqi = 2378, + CODE_FOR_atomic_storehi = 2379, + CODE_FOR_atomic_storesi = 2380, + CODE_FOR_atomic_storedi = 2381, + CODE_FOR_aarch64_load_exclusiveqi = 2382, + CODE_FOR_aarch64_load_exclusivehi = 2383, + CODE_FOR_aarch64_load_exclusivesi = 2384, + CODE_FOR_aarch64_load_exclusivedi = 2385, + CODE_FOR_aarch64_store_exclusiveqi = 2386, + CODE_FOR_aarch64_store_exclusivehi = 2387, + CODE_FOR_aarch64_store_exclusivesi = 2388, + CODE_FOR_aarch64_store_exclusivedi = 2389, + CODE_FOR_cbranchsi4 = 2391, + CODE_FOR_cbranchdi4 = 2392, + CODE_FOR_cbranchsf4 = 2393, + CODE_FOR_cbranchdf4 = 2394, + CODE_FOR_casesi = 2395, + CODE_FOR_prologue = 2396, + CODE_FOR_epilogue = 2397, + CODE_FOR_sibcall_epilogue = 2398, + CODE_FOR_call = 2400, + CODE_FOR_call_value = 2401, + CODE_FOR_sibcall = 2402, + CODE_FOR_sibcall_value = 2403, + CODE_FOR_untyped_call = 2404, + CODE_FOR_movqi = 2405, + CODE_FOR_movhi = 2406, + CODE_FOR_movsi = 2407, + CODE_FOR_movdi = 2408, + CODE_FOR_movti = 2409, + CODE_FOR_movsf = 2411, + CODE_FOR_movdf = 2412, + CODE_FOR_movtf = 2413, + CODE_FOR_movmemdi = 2415, + CODE_FOR_extendsidi2 = 2416, + CODE_FOR_zero_extendsidi2 = 2417, + CODE_FOR_extendqisi2 = 2418, + CODE_FOR_zero_extendqisi2 = 2419, + CODE_FOR_extendqidi2 = 2420, + CODE_FOR_zero_extendqidi2 = 2421, + CODE_FOR_extendhisi2 = 2422, + CODE_FOR_zero_extendhisi2 = 2423, + CODE_FOR_extendhidi2 = 2424, + CODE_FOR_zero_extendhidi2 = 2425, + CODE_FOR_extendqihi2 = 2426, + CODE_FOR_zero_extendqihi2 = 2427, + CODE_FOR_addsi3 = 2428, + CODE_FOR_adddi3 = 2429, + CODE_FOR_addti3 = 2430, + CODE_FOR_subti3 = 2431, + CODE_FOR_mulditi3 = 2433, + CODE_FOR_umulditi3 = 2434, + CODE_FOR_multi3 = 2435, + CODE_FOR_cstoresi4 = 2436, + CODE_FOR_cstoredi4 = 2437, + CODE_FOR_cstoresf4 = 2438, + CODE_FOR_cstoredf4 = 2439, + CODE_FOR_cmovsi6 = 2440, + CODE_FOR_cmovdi6 = 2441, + CODE_FOR_cmovsf6 = 2442, + CODE_FOR_cmovdf6 = 2443, + CODE_FOR_movqicc = 2444, + CODE_FOR_movhicc = 2445, + CODE_FOR_movsicc = 2446, + CODE_FOR_movdicc = 2447, + CODE_FOR_movsfsicc = 2448, + CODE_FOR_movdfsicc = 2449, + CODE_FOR_movsfdicc = 2450, + CODE_FOR_movdfdicc = 2451, + CODE_FOR_movsfcc = 2452, + CODE_FOR_movdfcc = 2453, + CODE_FOR_ffssi2 = 2454, + CODE_FOR_ffsdi2 = 2455, + CODE_FOR_ctzsi2 = 2456, + CODE_FOR_ctzdi2 = 2457, + CODE_FOR_ashlsi3 = 2458, + CODE_FOR_ashrsi3 = 2459, + CODE_FOR_lshrsi3 = 2460, + CODE_FOR_ashldi3 = 2461, + CODE_FOR_ashrdi3 = 2462, + CODE_FOR_lshrdi3 = 2463, + CODE_FOR_ashlqi3 = 2464, + CODE_FOR_ashlhi3 = 2465, + CODE_FOR_rotrsi3 = 2466, + CODE_FOR_rotrdi3 = 2467, + CODE_FOR_rotlsi3 = 2468, + CODE_FOR_rotldi3 = 2469, + CODE_FOR_extv = 2474, + CODE_FOR_extzv = 2475, + CODE_FOR_insvsi = 2476, + CODE_FOR_insvdi = 2477, + CODE_FOR_aarch64_reload_movti = 2478, + CODE_FOR_aarch64_reload_movtf = 2479, + CODE_FOR_add_losym = 2480, + CODE_FOR_tlsgd_small = 2481, + CODE_FOR_tlsle_small = 2482, + CODE_FOR_get_thread_pointerdi = 2483, + CODE_FOR_stack_protect_set = 2484, + CODE_FOR_stack_protect_test = 2485, + CODE_FOR_movv8qi = 2486, + CODE_FOR_movv16qi = 2487, + CODE_FOR_movv4hi = 2488, + CODE_FOR_movv8hi = 2489, + CODE_FOR_movv2si = 2490, + CODE_FOR_movv4si = 2491, + CODE_FOR_movv2di = 2492, + CODE_FOR_movv2sf = 2493, + CODE_FOR_movv4sf = 2494, + CODE_FOR_movv2df = 2495, + CODE_FOR_movmisalignv8qi = 2496, + CODE_FOR_movmisalignv16qi = 2497, + CODE_FOR_movmisalignv4hi = 2498, + CODE_FOR_movmisalignv8hi = 2499, + CODE_FOR_movmisalignv2si = 2500, + CODE_FOR_movmisalignv4si = 2501, + CODE_FOR_movmisalignv2di = 2502, + CODE_FOR_movmisalignv2sf = 2503, + CODE_FOR_movmisalignv4sf = 2504, + CODE_FOR_movmisalignv2df = 2505, + CODE_FOR_aarch64_split_simd_movv16qi = 2518, + CODE_FOR_aarch64_split_simd_movv8hi = 2519, + CODE_FOR_aarch64_split_simd_movv4si = 2520, + CODE_FOR_aarch64_split_simd_movv2di = 2521, + CODE_FOR_aarch64_split_simd_movv4sf = 2522, + CODE_FOR_aarch64_split_simd_movv2df = 2523, + CODE_FOR_ashlv8qi3 = 2524, + CODE_FOR_ashlv16qi3 = 2525, + CODE_FOR_ashlv4hi3 = 2526, + CODE_FOR_ashlv8hi3 = 2527, + CODE_FOR_ashlv2si3 = 2528, + CODE_FOR_ashlv4si3 = 2529, + CODE_FOR_ashlv2di3 = 2530, + CODE_FOR_lshrv8qi3 = 2531, + CODE_FOR_lshrv16qi3 = 2532, + CODE_FOR_lshrv4hi3 = 2533, + CODE_FOR_lshrv8hi3 = 2534, + CODE_FOR_lshrv2si3 = 2535, + CODE_FOR_lshrv4si3 = 2536, + CODE_FOR_lshrv2di3 = 2537, + CODE_FOR_ashrv8qi3 = 2538, + CODE_FOR_ashrv16qi3 = 2539, + CODE_FOR_ashrv4hi3 = 2540, + CODE_FOR_ashrv8hi3 = 2541, + CODE_FOR_ashrv2si3 = 2542, + CODE_FOR_ashrv4si3 = 2543, + CODE_FOR_ashrv2di3 = 2544, + CODE_FOR_vashlv8qi3 = 2545, + CODE_FOR_vashlv16qi3 = 2546, + CODE_FOR_vashlv4hi3 = 2547, + CODE_FOR_vashlv8hi3 = 2548, + CODE_FOR_vashlv2si3 = 2549, + CODE_FOR_vashlv4si3 = 2550, + CODE_FOR_vashlv2di3 = 2551, + CODE_FOR_vashrv8qi3 = 2552, + CODE_FOR_vashrv16qi3 = 2553, + CODE_FOR_vashrv4hi3 = 2554, + CODE_FOR_vashrv8hi3 = 2555, + CODE_FOR_vashrv2si3 = 2556, + CODE_FOR_vashrv4si3 = 2557, + CODE_FOR_aarch64_ashr_simddi = 2558, + CODE_FOR_vlshrv8qi3 = 2559, + CODE_FOR_vlshrv16qi3 = 2560, + CODE_FOR_vlshrv4hi3 = 2561, + CODE_FOR_vlshrv8hi3 = 2562, + CODE_FOR_vlshrv2si3 = 2563, + CODE_FOR_vlshrv4si3 = 2564, + CODE_FOR_aarch64_lshr_simddi = 2565, + CODE_FOR_vec_setv8qi = 2566, + CODE_FOR_vec_setv16qi = 2567, + CODE_FOR_vec_setv4hi = 2568, + CODE_FOR_vec_setv8hi = 2569, + CODE_FOR_vec_setv2si = 2570, + CODE_FOR_vec_setv4si = 2571, + CODE_FOR_vec_setv2di = 2572, + CODE_FOR_vec_setv2sf = 2573, + CODE_FOR_vec_setv4sf = 2574, + CODE_FOR_vec_setv2df = 2575, + CODE_FOR_move_lo_quad_v16qi = 2576, + CODE_FOR_move_lo_quad_v8hi = 2577, + CODE_FOR_move_lo_quad_v4si = 2578, + CODE_FOR_move_lo_quad_v2di = 2579, + CODE_FOR_move_lo_quad_v4sf = 2580, + CODE_FOR_move_lo_quad_v2df = 2581, + CODE_FOR_move_hi_quad_v16qi = 2582, + CODE_FOR_move_hi_quad_v8hi = 2583, + CODE_FOR_move_hi_quad_v4si = 2584, + CODE_FOR_move_hi_quad_v2di = 2585, + CODE_FOR_move_hi_quad_v4sf = 2586, + CODE_FOR_move_hi_quad_v2df = 2587, + CODE_FOR_vec_pack_trunc_v4hi = 2588, + CODE_FOR_vec_pack_trunc_v2si = 2589, + CODE_FOR_vec_pack_trunc_di = 2590, + CODE_FOR_vec_unpacks_hi_v16qi = 2591, + CODE_FOR_vec_unpacku_hi_v16qi = 2592, + CODE_FOR_vec_unpacks_hi_v8hi = 2593, + CODE_FOR_vec_unpacku_hi_v8hi = 2594, + CODE_FOR_vec_unpacks_hi_v4si = 2595, + CODE_FOR_vec_unpacku_hi_v4si = 2596, + CODE_FOR_vec_unpacks_lo_v16qi = 2597, + CODE_FOR_vec_unpacku_lo_v16qi = 2598, + CODE_FOR_vec_unpacks_lo_v8hi = 2599, + CODE_FOR_vec_unpacku_lo_v8hi = 2600, + CODE_FOR_vec_unpacks_lo_v4si = 2601, + CODE_FOR_vec_unpacku_lo_v4si = 2602, + CODE_FOR_vec_widen_smult_lo_v16qi = 2603, + CODE_FOR_vec_widen_umult_lo_v16qi = 2604, + CODE_FOR_vec_widen_smult_lo_v8hi = 2605, + CODE_FOR_vec_widen_umult_lo_v8hi = 2606, + CODE_FOR_vec_widen_smult_lo_v4si = 2607, + CODE_FOR_vec_widen_umult_lo_v4si = 2608, + CODE_FOR_vec_widen_smult_hi_v16qi = 2609, + CODE_FOR_vec_widen_umult_hi_v16qi = 2610, + CODE_FOR_vec_widen_smult_hi_v8hi = 2611, + CODE_FOR_vec_widen_umult_hi_v8hi = 2612, + CODE_FOR_vec_widen_smult_hi_v4si = 2613, + CODE_FOR_vec_widen_umult_hi_v4si = 2614, + CODE_FOR_fixv2sfv2si2 = 2615, + CODE_FOR_fixunsv2sfv2si2 = 2616, + CODE_FOR_fixv4sfv4si2 = 2617, + CODE_FOR_fixunsv4sfv4si2 = 2618, + CODE_FOR_fixv2dfv2di2 = 2619, + CODE_FOR_fixunsv2dfv2di2 = 2620, + CODE_FOR_fix_truncv2sfv2si2 = 2621, + CODE_FOR_fixuns_truncv2sfv2si2 = 2622, + CODE_FOR_fix_truncv4sfv4si2 = 2623, + CODE_FOR_fixuns_truncv4sfv4si2 = 2624, + CODE_FOR_fix_truncv2dfv2di2 = 2625, + CODE_FOR_fixuns_truncv2dfv2di2 = 2626, + CODE_FOR_ftruncv2sf2 = 2627, + CODE_FOR_ftruncv4sf2 = 2628, + CODE_FOR_ftruncv2df2 = 2629, + CODE_FOR_vec_pack_trunc_v2df = 2630, + CODE_FOR_vec_pack_trunc_df = 2631, + CODE_FOR_reduc_splus_v4sf = 2632, + CODE_FOR_aarch64_simd_bslv8qi = 2633, + CODE_FOR_aarch64_simd_bslv16qi = 2634, + CODE_FOR_aarch64_simd_bslv4hi = 2635, + CODE_FOR_aarch64_simd_bslv8hi = 2636, + CODE_FOR_aarch64_simd_bslv2si = 2637, + CODE_FOR_aarch64_simd_bslv4si = 2638, + CODE_FOR_aarch64_simd_bslv2di = 2639, + CODE_FOR_aarch64_simd_bslv2sf = 2640, + CODE_FOR_aarch64_simd_bslv4sf = 2641, + CODE_FOR_aarch64_simd_bslv2df = 2642, + CODE_FOR_aarch64_simd_bsldi = 2643, + CODE_FOR_aarch64_simd_bsldf = 2644, + CODE_FOR_aarch64_vcond_internalv8qiv8qi = 2645, + CODE_FOR_aarch64_vcond_internalv16qiv16qi = 2646, + CODE_FOR_aarch64_vcond_internalv4hiv4hi = 2647, + CODE_FOR_aarch64_vcond_internalv8hiv8hi = 2648, + CODE_FOR_aarch64_vcond_internalv2siv2si = 2649, + CODE_FOR_aarch64_vcond_internalv4siv4si = 2650, + CODE_FOR_aarch64_vcond_internalv2div2di = 2651, + CODE_FOR_aarch64_vcond_internalv2sfv2sf = 2652, + CODE_FOR_aarch64_vcond_internalv2siv2sf = 2653, + CODE_FOR_aarch64_vcond_internalv4sfv2sf = 2654, + CODE_FOR_aarch64_vcond_internalv4siv2sf = 2655, + CODE_FOR_aarch64_vcond_internalv2dfv2sf = 2656, + CODE_FOR_aarch64_vcond_internalv2div2sf = 2657, + CODE_FOR_aarch64_vcond_internalv2sfv4sf = 2658, + CODE_FOR_aarch64_vcond_internalv2siv4sf = 2659, + CODE_FOR_aarch64_vcond_internalv4sfv4sf = 2660, + CODE_FOR_aarch64_vcond_internalv4siv4sf = 2661, + CODE_FOR_aarch64_vcond_internalv2dfv4sf = 2662, + CODE_FOR_aarch64_vcond_internalv2div4sf = 2663, + CODE_FOR_aarch64_vcond_internalv2sfv2df = 2664, + CODE_FOR_aarch64_vcond_internalv2siv2df = 2665, + CODE_FOR_aarch64_vcond_internalv4sfv2df = 2666, + CODE_FOR_aarch64_vcond_internalv4siv2df = 2667, + CODE_FOR_aarch64_vcond_internalv2dfv2df = 2668, + CODE_FOR_aarch64_vcond_internalv2div2df = 2669, + CODE_FOR_vcondv8qiv8qi = 2670, + CODE_FOR_vcondv16qiv16qi = 2671, + CODE_FOR_vcondv4hiv4hi = 2672, + CODE_FOR_vcondv8hiv8hi = 2673, + CODE_FOR_vcondv2siv2si = 2674, + CODE_FOR_vcondv4siv4si = 2675, + CODE_FOR_vcondv2div2di = 2676, + CODE_FOR_vcondv2sfv2sf = 2677, + CODE_FOR_vcondv4sfv4sf = 2678, + CODE_FOR_vcondv2dfv2df = 2679, + CODE_FOR_vcondv2siv2sf = 2680, + CODE_FOR_vcondv4siv4sf = 2681, + CODE_FOR_vcondv2div2df = 2682, + CODE_FOR_vconduv8qiv8qi = 2683, + CODE_FOR_vconduv16qiv16qi = 2684, + CODE_FOR_vconduv4hiv4hi = 2685, + CODE_FOR_vconduv8hiv8hi = 2686, + CODE_FOR_vconduv2siv2si = 2687, + CODE_FOR_vconduv4siv4si = 2688, + CODE_FOR_vconduv2div2di = 2689, + CODE_FOR_aarch64_createv8qi = 2690, + CODE_FOR_aarch64_createv4hi = 2691, + CODE_FOR_aarch64_createv2si = 2692, + CODE_FOR_aarch64_createdi = 2693, + CODE_FOR_aarch64_createdf = 2694, + CODE_FOR_aarch64_createv2sf = 2695, + CODE_FOR_aarch64_be_checked_get_lanev8qi = 2696, + CODE_FOR_aarch64_be_checked_get_lanev16qi = 2697, + CODE_FOR_aarch64_be_checked_get_lanev4hi = 2698, + CODE_FOR_aarch64_be_checked_get_lanev8hi = 2699, + CODE_FOR_aarch64_be_checked_get_lanev2si = 2700, + CODE_FOR_aarch64_be_checked_get_lanev4si = 2701, + CODE_FOR_aarch64_be_checked_get_lanev2di = 2702, + CODE_FOR_aarch64_be_checked_get_lanev2sf = 2703, + CODE_FOR_aarch64_be_checked_get_lanev4sf = 2704, + CODE_FOR_aarch64_be_checked_get_lanev2df = 2705, + CODE_FOR_aarch64_get_lanedi = 2706, + CODE_FOR_aarch64_reinterpretv8qiv8qi = 2707, + CODE_FOR_aarch64_reinterpretv8qiv4hi = 2708, + CODE_FOR_aarch64_reinterpretv8qiv2si = 2709, + CODE_FOR_aarch64_reinterpretv8qiv2sf = 2710, + CODE_FOR_aarch64_reinterpretv8qidi = 2711, + CODE_FOR_aarch64_reinterpretv8qidf = 2712, + CODE_FOR_aarch64_reinterpretv4hiv8qi = 2713, + CODE_FOR_aarch64_reinterpretv4hiv4hi = 2714, + CODE_FOR_aarch64_reinterpretv4hiv2si = 2715, + CODE_FOR_aarch64_reinterpretv4hiv2sf = 2716, + CODE_FOR_aarch64_reinterpretv4hidi = 2717, + CODE_FOR_aarch64_reinterpretv4hidf = 2718, + CODE_FOR_aarch64_reinterpretv2siv8qi = 2719, + CODE_FOR_aarch64_reinterpretv2siv4hi = 2720, + CODE_FOR_aarch64_reinterpretv2siv2si = 2721, + CODE_FOR_aarch64_reinterpretv2siv2sf = 2722, + CODE_FOR_aarch64_reinterpretv2sidi = 2723, + CODE_FOR_aarch64_reinterpretv2sidf = 2724, + CODE_FOR_aarch64_reinterpretv2sfv8qi = 2725, + CODE_FOR_aarch64_reinterpretv2sfv4hi = 2726, + CODE_FOR_aarch64_reinterpretv2sfv2si = 2727, + CODE_FOR_aarch64_reinterpretv2sfv2sf = 2728, + CODE_FOR_aarch64_reinterpretv2sfdi = 2729, + CODE_FOR_aarch64_reinterpretv2sfdf = 2730, + CODE_FOR_aarch64_reinterpretdiv8qi = 2731, + CODE_FOR_aarch64_reinterpretdiv4hi = 2732, + CODE_FOR_aarch64_reinterpretdiv2si = 2733, + CODE_FOR_aarch64_reinterpretdidi = 2734, + CODE_FOR_aarch64_reinterpretdidf = 2735, + CODE_FOR_aarch64_reinterpretdiv2sf = 2736, + CODE_FOR_aarch64_reinterpretdfv8qi = 2737, + CODE_FOR_aarch64_reinterpretdfv4hi = 2738, + CODE_FOR_aarch64_reinterpretdfv2si = 2739, + CODE_FOR_aarch64_reinterpretdfdi = 2740, + CODE_FOR_aarch64_reinterpretdfdf = 2741, + CODE_FOR_aarch64_reinterpretdfv2sf = 2742, + CODE_FOR_aarch64_reinterpretv16qiv16qi = 2743, + CODE_FOR_aarch64_reinterpretv16qiv8hi = 2744, + CODE_FOR_aarch64_reinterpretv16qiv4si = 2745, + CODE_FOR_aarch64_reinterpretv16qiv2di = 2746, + CODE_FOR_aarch64_reinterpretv16qiv4sf = 2747, + CODE_FOR_aarch64_reinterpretv16qiv2df = 2748, + CODE_FOR_aarch64_reinterpretv8hiv16qi = 2749, + CODE_FOR_aarch64_reinterpretv8hiv8hi = 2750, + CODE_FOR_aarch64_reinterpretv8hiv4si = 2751, + CODE_FOR_aarch64_reinterpretv8hiv2di = 2752, + CODE_FOR_aarch64_reinterpretv8hiv4sf = 2753, + CODE_FOR_aarch64_reinterpretv8hiv2df = 2754, + CODE_FOR_aarch64_reinterpretv4siv16qi = 2755, + CODE_FOR_aarch64_reinterpretv4siv8hi = 2756, + CODE_FOR_aarch64_reinterpretv4siv4si = 2757, + CODE_FOR_aarch64_reinterpretv4siv2di = 2758, + CODE_FOR_aarch64_reinterpretv4siv4sf = 2759, + CODE_FOR_aarch64_reinterpretv4siv2df = 2760, + CODE_FOR_aarch64_reinterpretv4sfv16qi = 2761, + CODE_FOR_aarch64_reinterpretv4sfv8hi = 2762, + CODE_FOR_aarch64_reinterpretv4sfv4si = 2763, + CODE_FOR_aarch64_reinterpretv4sfv2di = 2764, + CODE_FOR_aarch64_reinterpretv4sfv4sf = 2765, + CODE_FOR_aarch64_reinterpretv4sfv2df = 2766, + CODE_FOR_aarch64_reinterpretv2div16qi = 2767, + CODE_FOR_aarch64_reinterpretv2div8hi = 2768, + CODE_FOR_aarch64_reinterpretv2div4si = 2769, + CODE_FOR_aarch64_reinterpretv2div2di = 2770, + CODE_FOR_aarch64_reinterpretv2div4sf = 2771, + CODE_FOR_aarch64_reinterpretv2div2df = 2772, + CODE_FOR_aarch64_reinterpretv2dfv16qi = 2773, + CODE_FOR_aarch64_reinterpretv2dfv8hi = 2774, + CODE_FOR_aarch64_reinterpretv2dfv4si = 2775, + CODE_FOR_aarch64_reinterpretv2dfv2di = 2776, + CODE_FOR_aarch64_reinterpretv2dfv4sf = 2777, + CODE_FOR_aarch64_reinterpretv2dfv2df = 2778, + CODE_FOR_aarch64_combinev8qi = 2779, + CODE_FOR_aarch64_combinev4hi = 2780, + CODE_FOR_aarch64_combinev2si = 2781, + CODE_FOR_aarch64_combinev2sf = 2782, + CODE_FOR_aarch64_combinedi = 2783, + CODE_FOR_aarch64_combinedf = 2784, + CODE_FOR_aarch64_simd_combinev8qi = 2791, + CODE_FOR_aarch64_simd_combinev4hi = 2792, + CODE_FOR_aarch64_simd_combinev2si = 2793, + CODE_FOR_aarch64_simd_combinev2sf = 2794, + CODE_FOR_aarch64_simd_combinedi = 2795, + CODE_FOR_aarch64_simd_combinedf = 2796, + CODE_FOR_aarch64_saddl2v16qi = 2797, + CODE_FOR_aarch64_saddl2v8hi = 2798, + CODE_FOR_aarch64_saddl2v4si = 2799, + CODE_FOR_aarch64_uaddl2v16qi = 2800, + CODE_FOR_aarch64_uaddl2v8hi = 2801, + CODE_FOR_aarch64_uaddl2v4si = 2802, + CODE_FOR_aarch64_ssubl2v16qi = 2803, + CODE_FOR_aarch64_ssubl2v8hi = 2804, + CODE_FOR_aarch64_ssubl2v4si = 2805, + CODE_FOR_aarch64_usubl2v16qi = 2806, + CODE_FOR_aarch64_usubl2v8hi = 2807, + CODE_FOR_aarch64_usubl2v4si = 2808, + CODE_FOR_aarch64_saddw2v16qi = 2809, + CODE_FOR_aarch64_saddw2v8hi = 2810, + CODE_FOR_aarch64_saddw2v4si = 2811, + CODE_FOR_aarch64_uaddw2v16qi = 2812, + CODE_FOR_aarch64_uaddw2v8hi = 2813, + CODE_FOR_aarch64_uaddw2v4si = 2814, + CODE_FOR_aarch64_ssubw2v16qi = 2815, + CODE_FOR_aarch64_ssubw2v8hi = 2816, + CODE_FOR_aarch64_ssubw2v4si = 2817, + CODE_FOR_aarch64_usubw2v16qi = 2818, + CODE_FOR_aarch64_usubw2v8hi = 2819, + CODE_FOR_aarch64_usubw2v4si = 2820, + CODE_FOR_aarch64_sqdmulh_lanev4hi = 2821, + CODE_FOR_aarch64_sqdmulh_lanev8hi = 2822, + CODE_FOR_aarch64_sqdmulh_lanev2si = 2823, + CODE_FOR_aarch64_sqdmulh_lanev4si = 2824, + CODE_FOR_aarch64_sqrdmulh_lanev4hi = 2825, + CODE_FOR_aarch64_sqrdmulh_lanev8hi = 2826, + CODE_FOR_aarch64_sqrdmulh_lanev2si = 2827, + CODE_FOR_aarch64_sqrdmulh_lanev4si = 2828, + CODE_FOR_aarch64_sqdmulh_laneqv4hi = 2829, + CODE_FOR_aarch64_sqdmulh_laneqv8hi = 2830, + CODE_FOR_aarch64_sqdmulh_laneqv2si = 2831, + CODE_FOR_aarch64_sqdmulh_laneqv4si = 2832, + CODE_FOR_aarch64_sqrdmulh_laneqv4hi = 2833, + CODE_FOR_aarch64_sqrdmulh_laneqv8hi = 2834, + CODE_FOR_aarch64_sqrdmulh_laneqv2si = 2835, + CODE_FOR_aarch64_sqrdmulh_laneqv4si = 2836, + CODE_FOR_aarch64_sqdmulh_lanehi = 2837, + CODE_FOR_aarch64_sqdmulh_lanesi = 2838, + CODE_FOR_aarch64_sqrdmulh_lanehi = 2839, + CODE_FOR_aarch64_sqrdmulh_lanesi = 2840, + CODE_FOR_aarch64_sqdmlal_lanev4hi = 2841, + CODE_FOR_aarch64_sqdmlal_lanev2si = 2842, + CODE_FOR_aarch64_sqdmlal_lanehi = 2843, + CODE_FOR_aarch64_sqdmlal_lanesi = 2844, + CODE_FOR_aarch64_sqdmlal_laneqv4hi = 2845, + CODE_FOR_aarch64_sqdmlal_laneqv2si = 2846, + CODE_FOR_aarch64_sqdmlal_laneqhi = 2847, + CODE_FOR_aarch64_sqdmlal_laneqsi = 2848, + CODE_FOR_aarch64_sqdmlsl_lanev4hi = 2849, + CODE_FOR_aarch64_sqdmlsl_lanev2si = 2850, + CODE_FOR_aarch64_sqdmlsl_lanehi = 2851, + CODE_FOR_aarch64_sqdmlsl_lanesi = 2852, + CODE_FOR_aarch64_sqdmlsl_laneqv4hi = 2853, + CODE_FOR_aarch64_sqdmlsl_laneqv2si = 2854, + CODE_FOR_aarch64_sqdmlsl_laneqhi = 2855, + CODE_FOR_aarch64_sqdmlsl_laneqsi = 2856, + CODE_FOR_aarch64_sqdmlal2v8hi = 2857, + CODE_FOR_aarch64_sqdmlal2v4si = 2858, + CODE_FOR_aarch64_sqdmlsl2v8hi = 2859, + CODE_FOR_aarch64_sqdmlsl2v4si = 2860, + CODE_FOR_aarch64_sqdmlal2_lanev8hi = 2861, + CODE_FOR_aarch64_sqdmlal2_lanev4si = 2862, + CODE_FOR_aarch64_sqdmlal2_laneqv8hi = 2863, + CODE_FOR_aarch64_sqdmlal2_laneqv4si = 2864, + CODE_FOR_aarch64_sqdmlsl2_lanev8hi = 2865, + CODE_FOR_aarch64_sqdmlsl2_lanev4si = 2866, + CODE_FOR_aarch64_sqdmlsl2_laneqv8hi = 2867, + CODE_FOR_aarch64_sqdmlsl2_laneqv4si = 2868, + CODE_FOR_aarch64_sqdmlal2_nv8hi = 2869, + CODE_FOR_aarch64_sqdmlal2_nv4si = 2870, + CODE_FOR_aarch64_sqdmlsl2_nv8hi = 2871, + CODE_FOR_aarch64_sqdmlsl2_nv4si = 2872, + CODE_FOR_aarch64_sqdmull_lanev4hi = 2873, + CODE_FOR_aarch64_sqdmull_lanev2si = 2874, + CODE_FOR_aarch64_sqdmull_lanehi = 2875, + CODE_FOR_aarch64_sqdmull_lanesi = 2876, + CODE_FOR_aarch64_sqdmull_laneqv4hi = 2877, + CODE_FOR_aarch64_sqdmull_laneqv2si = 2878, + CODE_FOR_aarch64_sqdmull2v8hi = 2879, + CODE_FOR_aarch64_sqdmull2v4si = 2880, + CODE_FOR_aarch64_sqdmull2_lanev8hi = 2881, + CODE_FOR_aarch64_sqdmull2_lanev4si = 2882, + CODE_FOR_aarch64_sqdmull2_laneqv8hi = 2883, + CODE_FOR_aarch64_sqdmull2_laneqv4si = 2884, + CODE_FOR_aarch64_sqdmull2_nv8hi = 2885, + CODE_FOR_aarch64_sqdmull2_nv4si = 2886, + CODE_FOR_movoi = 2897, + CODE_FOR_movci = 2898, + CODE_FOR_movxi = 2899, + CODE_FOR_aarch64_ld2v8qi = 2903, + CODE_FOR_aarch64_ld2v4hi = 2904, + CODE_FOR_aarch64_ld2v2si = 2905, + CODE_FOR_aarch64_ld2v2sf = 2906, + CODE_FOR_aarch64_ld2di = 2907, + CODE_FOR_aarch64_ld2df = 2908, + CODE_FOR_aarch64_ld3v8qi = 2909, + CODE_FOR_aarch64_ld3v4hi = 2910, + CODE_FOR_aarch64_ld3v2si = 2911, + CODE_FOR_aarch64_ld3v2sf = 2912, + CODE_FOR_aarch64_ld3di = 2913, + CODE_FOR_aarch64_ld3df = 2914, + CODE_FOR_aarch64_ld4v8qi = 2915, + CODE_FOR_aarch64_ld4v4hi = 2916, + CODE_FOR_aarch64_ld4v2si = 2917, + CODE_FOR_aarch64_ld4v2sf = 2918, + CODE_FOR_aarch64_ld4di = 2919, + CODE_FOR_aarch64_ld4df = 2920, + CODE_FOR_aarch64_ld1v8qi = 2921, + CODE_FOR_aarch64_ld1v16qi = 2922, + CODE_FOR_aarch64_ld1v4hi = 2923, + CODE_FOR_aarch64_ld1v8hi = 2924, + CODE_FOR_aarch64_ld1v2si = 2925, + CODE_FOR_aarch64_ld1v4si = 2926, + CODE_FOR_aarch64_ld1v2di = 2927, + CODE_FOR_aarch64_ld1v2sf = 2928, + CODE_FOR_aarch64_ld1v4sf = 2929, + CODE_FOR_aarch64_ld1v2df = 2930, + CODE_FOR_aarch64_ld2v16qi = 2931, + CODE_FOR_aarch64_ld3v16qi = 2932, + CODE_FOR_aarch64_ld4v16qi = 2933, + CODE_FOR_aarch64_ld2v8hi = 2934, + CODE_FOR_aarch64_ld3v8hi = 2935, + CODE_FOR_aarch64_ld4v8hi = 2936, + CODE_FOR_aarch64_ld2v4si = 2937, + CODE_FOR_aarch64_ld3v4si = 2938, + CODE_FOR_aarch64_ld4v4si = 2939, + CODE_FOR_aarch64_ld2v2di = 2940, + CODE_FOR_aarch64_ld3v2di = 2941, + CODE_FOR_aarch64_ld4v2di = 2942, + CODE_FOR_aarch64_ld2v4sf = 2943, + CODE_FOR_aarch64_ld3v4sf = 2944, + CODE_FOR_aarch64_ld4v4sf = 2945, + CODE_FOR_aarch64_ld2v2df = 2946, + CODE_FOR_aarch64_ld3v2df = 2947, + CODE_FOR_aarch64_ld4v2df = 2948, + CODE_FOR_aarch64_get_dregoiv8qi = 2949, + CODE_FOR_aarch64_get_dregoiv4hi = 2950, + CODE_FOR_aarch64_get_dregoiv2si = 2951, + CODE_FOR_aarch64_get_dregoiv2sf = 2952, + CODE_FOR_aarch64_get_dregoidi = 2953, + CODE_FOR_aarch64_get_dregoidf = 2954, + CODE_FOR_aarch64_get_dregciv8qi = 2955, + CODE_FOR_aarch64_get_dregciv4hi = 2956, + CODE_FOR_aarch64_get_dregciv2si = 2957, + CODE_FOR_aarch64_get_dregciv2sf = 2958, + CODE_FOR_aarch64_get_dregcidi = 2959, + CODE_FOR_aarch64_get_dregcidf = 2960, + CODE_FOR_aarch64_get_dregxiv8qi = 2961, + CODE_FOR_aarch64_get_dregxiv4hi = 2962, + CODE_FOR_aarch64_get_dregxiv2si = 2963, + CODE_FOR_aarch64_get_dregxiv2sf = 2964, + CODE_FOR_aarch64_get_dregxidi = 2965, + CODE_FOR_aarch64_get_dregxidf = 2966, + CODE_FOR_aarch64_get_qregoiv16qi = 2967, + CODE_FOR_aarch64_get_qregciv16qi = 2968, + CODE_FOR_aarch64_get_qregxiv16qi = 2969, + CODE_FOR_aarch64_get_qregoiv8hi = 2970, + CODE_FOR_aarch64_get_qregciv8hi = 2971, + CODE_FOR_aarch64_get_qregxiv8hi = 2972, + CODE_FOR_aarch64_get_qregoiv4si = 2973, + CODE_FOR_aarch64_get_qregciv4si = 2974, + CODE_FOR_aarch64_get_qregxiv4si = 2975, + CODE_FOR_aarch64_get_qregoiv2di = 2976, + CODE_FOR_aarch64_get_qregciv2di = 2977, + CODE_FOR_aarch64_get_qregxiv2di = 2978, + CODE_FOR_aarch64_get_qregoiv4sf = 2979, + CODE_FOR_aarch64_get_qregciv4sf = 2980, + CODE_FOR_aarch64_get_qregxiv4sf = 2981, + CODE_FOR_aarch64_get_qregoiv2df = 2982, + CODE_FOR_aarch64_get_qregciv2df = 2983, + CODE_FOR_aarch64_get_qregxiv2df = 2984, + CODE_FOR_vec_perm_constv8qi = 2985, + CODE_FOR_vec_perm_constv16qi = 2986, + CODE_FOR_vec_perm_constv4hi = 2987, + CODE_FOR_vec_perm_constv8hi = 2988, + CODE_FOR_vec_perm_constv2si = 2989, + CODE_FOR_vec_perm_constv4si = 2990, + CODE_FOR_vec_perm_constv2di = 2991, + CODE_FOR_vec_perm_constv2sf = 2992, + CODE_FOR_vec_perm_constv4sf = 2993, + CODE_FOR_vec_perm_constv2df = 2994, + CODE_FOR_vec_permv8qi = 2995, + CODE_FOR_vec_permv16qi = 2996, + CODE_FOR_aarch64_im_lane_boundsi = 2998, + CODE_FOR_aarch64_st2v8qi = 2999, + CODE_FOR_aarch64_st2v4hi = 3000, + CODE_FOR_aarch64_st2v2si = 3001, + CODE_FOR_aarch64_st2v2sf = 3002, + CODE_FOR_aarch64_st2di = 3003, + CODE_FOR_aarch64_st2df = 3004, + CODE_FOR_aarch64_st3v8qi = 3005, + CODE_FOR_aarch64_st3v4hi = 3006, + CODE_FOR_aarch64_st3v2si = 3007, + CODE_FOR_aarch64_st3v2sf = 3008, + CODE_FOR_aarch64_st3di = 3009, + CODE_FOR_aarch64_st3df = 3010, + CODE_FOR_aarch64_st4v8qi = 3011, + CODE_FOR_aarch64_st4v4hi = 3012, + CODE_FOR_aarch64_st4v2si = 3013, + CODE_FOR_aarch64_st4v2sf = 3014, + CODE_FOR_aarch64_st4di = 3015, + CODE_FOR_aarch64_st4df = 3016, + CODE_FOR_aarch64_st2v16qi = 3017, + CODE_FOR_aarch64_st3v16qi = 3018, + CODE_FOR_aarch64_st4v16qi = 3019, + CODE_FOR_aarch64_st2v8hi = 3020, + CODE_FOR_aarch64_st3v8hi = 3021, + CODE_FOR_aarch64_st4v8hi = 3022, + CODE_FOR_aarch64_st2v4si = 3023, + CODE_FOR_aarch64_st3v4si = 3024, + CODE_FOR_aarch64_st4v4si = 3025, + CODE_FOR_aarch64_st2v2di = 3026, + CODE_FOR_aarch64_st3v2di = 3027, + CODE_FOR_aarch64_st4v2di = 3028, + CODE_FOR_aarch64_st2v4sf = 3029, + CODE_FOR_aarch64_st3v4sf = 3030, + CODE_FOR_aarch64_st4v4sf = 3031, + CODE_FOR_aarch64_st2v2df = 3032, + CODE_FOR_aarch64_st3v2df = 3033, + CODE_FOR_aarch64_st4v2df = 3034, + CODE_FOR_aarch64_st2_lanev16qi = 3035, + CODE_FOR_aarch64_st2_lanev8hi = 3036, + CODE_FOR_aarch64_st2_lanev4si = 3037, + CODE_FOR_aarch64_st2_lanev2di = 3038, + CODE_FOR_aarch64_st2_lanev4sf = 3039, + CODE_FOR_aarch64_st2_lanev2df = 3040, + CODE_FOR_aarch64_st3_lanev16qi = 3041, + CODE_FOR_aarch64_st3_lanev8hi = 3042, + CODE_FOR_aarch64_st3_lanev4si = 3043, + CODE_FOR_aarch64_st3_lanev2di = 3044, + CODE_FOR_aarch64_st3_lanev4sf = 3045, + CODE_FOR_aarch64_st3_lanev2df = 3046, + CODE_FOR_aarch64_st4_lanev16qi = 3047, + CODE_FOR_aarch64_st4_lanev8hi = 3048, + CODE_FOR_aarch64_st4_lanev4si = 3049, + CODE_FOR_aarch64_st4_lanev2di = 3050, + CODE_FOR_aarch64_st4_lanev4sf = 3051, + CODE_FOR_aarch64_st4_lanev2df = 3052, + CODE_FOR_aarch64_st1v8qi = 3053, + CODE_FOR_aarch64_st1v16qi = 3054, + CODE_FOR_aarch64_st1v4hi = 3055, + CODE_FOR_aarch64_st1v8hi = 3056, + CODE_FOR_aarch64_st1v2si = 3057, + CODE_FOR_aarch64_st1v4si = 3058, + CODE_FOR_aarch64_st1v2di = 3059, + CODE_FOR_aarch64_st1v2sf = 3060, + CODE_FOR_aarch64_st1v4sf = 3061, + CODE_FOR_aarch64_st1v2df = 3062, + CODE_FOR_aarch64_set_qregoiv16qi = 3063, + CODE_FOR_aarch64_set_qregciv16qi = 3064, + CODE_FOR_aarch64_set_qregxiv16qi = 3065, + CODE_FOR_aarch64_set_qregoiv8hi = 3066, + CODE_FOR_aarch64_set_qregciv8hi = 3067, + CODE_FOR_aarch64_set_qregxiv8hi = 3068, + CODE_FOR_aarch64_set_qregoiv4si = 3069, + CODE_FOR_aarch64_set_qregciv4si = 3070, + CODE_FOR_aarch64_set_qregxiv4si = 3071, + CODE_FOR_aarch64_set_qregoiv2di = 3072, + CODE_FOR_aarch64_set_qregciv2di = 3073, + CODE_FOR_aarch64_set_qregxiv2di = 3074, + CODE_FOR_aarch64_set_qregoiv4sf = 3075, + CODE_FOR_aarch64_set_qregciv4sf = 3076, + CODE_FOR_aarch64_set_qregxiv4sf = 3077, + CODE_FOR_aarch64_set_qregoiv2df = 3078, + CODE_FOR_aarch64_set_qregciv2df = 3079, + CODE_FOR_aarch64_set_qregxiv2df = 3080, + CODE_FOR_vec_initv8qi = 3081, + CODE_FOR_vec_initv16qi = 3082, + CODE_FOR_vec_initv4hi = 3083, + CODE_FOR_vec_initv8hi = 3084, + CODE_FOR_vec_initv2si = 3085, + CODE_FOR_vec_initv4si = 3086, + CODE_FOR_vec_initv2di = 3087, + CODE_FOR_vec_initv2sf = 3088, + CODE_FOR_vec_initv4sf = 3089, + CODE_FOR_vec_initv2df = 3090, + CODE_FOR_vec_extractv8qi = 3091, + CODE_FOR_vec_extractv16qi = 3092, + CODE_FOR_vec_extractv4hi = 3093, + CODE_FOR_vec_extractv8hi = 3094, + CODE_FOR_vec_extractv2si = 3095, + CODE_FOR_vec_extractv4si = 3096, + CODE_FOR_vec_extractv2di = 3097, + CODE_FOR_vec_extractv2sf = 3098, + CODE_FOR_vec_extractv4sf = 3099, + CODE_FOR_vec_extractv2df = 3100, + CODE_FOR_atomic_compare_and_swapqi = 3101, + CODE_FOR_atomic_compare_and_swaphi = 3102, + CODE_FOR_atomic_compare_and_swapsi = 3103, + CODE_FOR_atomic_compare_and_swapdi = 3104, + CODE_FOR_mem_thread_fence = 3185, + CODE_FOR_dmb = 3186, LAST_INSN_CODE }; diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-constants.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-constants.h index 6cd61f0..2cd05cb 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-constants.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-constants.h @@ -4,197 +4,213 @@ #ifndef GCC_INSN_CONSTANTS_H #define GCC_INSN_CONSTANTS_H -#define R13_REGNUM 13 -#define R7_REGNUM 7 -#define R16_REGNUM 16 -#define R23_REGNUM 23 -#define R19_REGNUM 19 -#define R26_REGNUM 26 +#define SFP_REGNUM 64 +#define R12_REGNUM 12 +#define CC_REGNUM 66 #define R29_REGNUM 29 -#define R0_REGNUM 0 -#define R3_REGNUM 3 -#define SP_REGNUM 31 -#define R6_REGNUM 6 -#define R9_REGNUM 9 -#define V0_REGNUM 32 -#define R8_REGNUM 8 +#define R15_REGNUM 15 #define V15_REGNUM 47 -#define R11_REGNUM 11 +#define R2_REGNUM 2 +#define R18_REGNUM 18 +#define IP0_REGNUM 16 +#define R5_REGNUM 5 +#define R25_REGNUM 25 #define V31_REGNUM 63 +#define R9_REGNUM 9 +#define V0_REGNUM 32 +#define R10_REGNUM 10 #define R14_REGNUM 14 +#define SP_REGNUM 31 +#define R1_REGNUM 1 #define R21_REGNUM 21 #define R17_REGNUM 17 -#define R24_REGNUM 24 -#define R1_REGNUM 1 -#define R4_REGNUM 4 -#define AP_REGNUM 65 #define R20_REGNUM 20 +#define R4_REGNUM 4 +#define R24_REGNUM 24 +#define R22_REGNUM 22 #define R30_REGNUM 30 -#define LR_REGNUM 30 -#define SFP_REGNUM 64 -#define R12_REGNUM 12 +#define R8_REGNUM 8 #define R27_REGNUM 27 -#define R15_REGNUM 15 -#define R22_REGNUM 22 -#define CC_REGNUM 66 -#define R18_REGNUM 18 -#define R25_REGNUM 25 +#define R13_REGNUM 13 +#define R11_REGNUM 11 +#define AP_REGNUM 65 +#define R0_REGNUM 0 +#define R6_REGNUM 6 +#define LR_REGNUM 30 #define R28_REGNUM 28 -#define R2_REGNUM 2 +#define R16_REGNUM 16 +#define R3_REGNUM 3 +#define R23_REGNUM 23 +#define R19_REGNUM 19 #define IP1_REGNUM 17 -#define R10_REGNUM 10 -#define R5_REGNUM 5 -#define IP0_REGNUM 16 +#define R7_REGNUM 7 +#define R26_REGNUM 26 enum unspec { UNSPEC_CASESI = 0, UNSPEC_CLS = 1, - UNSPEC_FRECPE = 2, - UNSPEC_FRECPS = 3, - UNSPEC_FRECPX = 4, - UNSPEC_FRINTA = 5, - UNSPEC_FRINTI = 6, - UNSPEC_FRINTM = 7, - UNSPEC_FRINTN = 8, - UNSPEC_FRINTP = 9, - UNSPEC_FRINTX = 10, - UNSPEC_FRINTZ = 11, - UNSPEC_GOTSMALLPIC = 12, - UNSPEC_GOTSMALLTLS = 13, - UNSPEC_GOTTINYPIC = 14, - UNSPEC_LD1 = 15, - UNSPEC_LD2 = 16, - UNSPEC_LD3 = 17, - UNSPEC_LD4 = 18, - UNSPEC_MB = 19, - UNSPEC_NOP = 20, - UNSPEC_PRLG_STK = 21, - UNSPEC_RBIT = 22, - UNSPEC_SISD_NEG = 23, - UNSPEC_SISD_SSHL = 24, - UNSPEC_SISD_USHL = 25, - UNSPEC_SSHL_2S = 26, - UNSPEC_SSHR64 = 27, - UNSPEC_ST1 = 28, - UNSPEC_ST2 = 29, - UNSPEC_ST3 = 30, - UNSPEC_ST4 = 31, - UNSPEC_ST2_LANE = 32, - UNSPEC_ST3_LANE = 33, - UNSPEC_ST4_LANE = 34, - UNSPEC_TLS = 35, - UNSPEC_TLSDESC = 36, - UNSPEC_USHL_2S = 37, - UNSPEC_USHR64 = 38, - UNSPEC_VSTRUCTDUMMY = 39, - UNSPEC_SP_SET = 40, - UNSPEC_SP_TEST = 41, - UNSPEC_ASHIFT_SIGNED = 42, - UNSPEC_ASHIFT_UNSIGNED = 43, - UNSPEC_FMAX = 44, - UNSPEC_FMAXNMV = 45, - UNSPEC_FMAXV = 46, - UNSPEC_FMIN = 47, - UNSPEC_FMINNMV = 48, - UNSPEC_FMINV = 49, - UNSPEC_FADDV = 50, - UNSPEC_SADDV = 51, - UNSPEC_UADDV = 52, - UNSPEC_SMAXV = 53, - UNSPEC_SMINV = 54, - UNSPEC_UMAXV = 55, - UNSPEC_UMINV = 56, - UNSPEC_SHADD = 57, - UNSPEC_UHADD = 58, - UNSPEC_SRHADD = 59, - UNSPEC_URHADD = 60, - UNSPEC_SHSUB = 61, - UNSPEC_UHSUB = 62, - UNSPEC_SRHSUB = 63, - UNSPEC_URHSUB = 64, - UNSPEC_ADDHN = 65, - UNSPEC_RADDHN = 66, - UNSPEC_SUBHN = 67, - UNSPEC_RSUBHN = 68, - UNSPEC_ADDHN2 = 69, - UNSPEC_RADDHN2 = 70, - UNSPEC_SUBHN2 = 71, - UNSPEC_RSUBHN2 = 72, - UNSPEC_SQDMULH = 73, - UNSPEC_SQRDMULH = 74, - UNSPEC_PMUL = 75, - UNSPEC_USQADD = 76, - UNSPEC_SUQADD = 77, - UNSPEC_SQXTUN = 78, - UNSPEC_SQXTN = 79, - UNSPEC_UQXTN = 80, - UNSPEC_SSRA = 81, - UNSPEC_USRA = 82, - UNSPEC_SRSRA = 83, - UNSPEC_URSRA = 84, - UNSPEC_SRSHR = 85, - UNSPEC_URSHR = 86, - UNSPEC_SQSHLU = 87, - UNSPEC_SQSHL = 88, - UNSPEC_UQSHL = 89, - UNSPEC_SQSHRUN = 90, - UNSPEC_SQRSHRUN = 91, - UNSPEC_SQSHRN = 92, - UNSPEC_UQSHRN = 93, - UNSPEC_SQRSHRN = 94, - UNSPEC_UQRSHRN = 95, - UNSPEC_SSHL = 96, - UNSPEC_USHL = 97, - UNSPEC_SRSHL = 98, - UNSPEC_URSHL = 99, - UNSPEC_SQRSHL = 100, - UNSPEC_UQRSHL = 101, - UNSPEC_SSLI = 102, - UNSPEC_USLI = 103, - UNSPEC_SSRI = 104, - UNSPEC_USRI = 105, - UNSPEC_SSHLL = 106, - UNSPEC_USHLL = 107, - UNSPEC_ADDP = 108, - UNSPEC_TBL = 109, - UNSPEC_CONCAT = 110, - UNSPEC_ZIP1 = 111, - UNSPEC_ZIP2 = 112, - UNSPEC_UZP1 = 113, - UNSPEC_UZP2 = 114, - UNSPEC_TRN1 = 115, - UNSPEC_TRN2 = 116, - UNSPEC_AESE = 117, - UNSPEC_AESD = 118, - UNSPEC_AESMC = 119, - UNSPEC_AESIMC = 120, - UNSPEC_SHA1C = 121, - UNSPEC_SHA1M = 122, - UNSPEC_SHA1P = 123, - UNSPEC_SHA1H = 124, - UNSPEC_SHA1SU0 = 125, - UNSPEC_SHA1SU1 = 126, - UNSPEC_SHA256H = 127, - UNSPEC_SHA256H2 = 128, - UNSPEC_SHA256SU0 = 129, - UNSPEC_SHA256SU1 = 130, - UNSPEC_PMULL = 131, - UNSPEC_PMULL2 = 132 + UNSPEC_CRC32B = 2, + UNSPEC_CRC32CB = 3, + UNSPEC_CRC32CH = 4, + UNSPEC_CRC32CW = 5, + UNSPEC_CRC32CX = 6, + UNSPEC_CRC32H = 7, + UNSPEC_CRC32W = 8, + UNSPEC_CRC32X = 9, + UNSPEC_FRECPE = 10, + UNSPEC_FRECPS = 11, + UNSPEC_FRECPX = 12, + UNSPEC_FRINTA = 13, + UNSPEC_FRINTI = 14, + UNSPEC_FRINTM = 15, + UNSPEC_FRINTN = 16, + UNSPEC_FRINTP = 17, + UNSPEC_FRINTX = 18, + UNSPEC_FRINTZ = 19, + UNSPEC_GOTSMALLPIC = 20, + UNSPEC_GOTSMALLTLS = 21, + UNSPEC_GOTTINYPIC = 22, + UNSPEC_LD1 = 23, + UNSPEC_LD2 = 24, + UNSPEC_LD3 = 25, + UNSPEC_LD4 = 26, + UNSPEC_MB = 27, + UNSPEC_NOP = 28, + UNSPEC_PRLG_STK = 29, + UNSPEC_RBIT = 30, + UNSPEC_SISD_NEG = 31, + UNSPEC_SISD_SSHL = 32, + UNSPEC_SISD_USHL = 33, + UNSPEC_SSHL_2S = 34, + UNSPEC_SSHR64 = 35, + UNSPEC_ST1 = 36, + UNSPEC_ST2 = 37, + UNSPEC_ST3 = 38, + UNSPEC_ST4 = 39, + UNSPEC_ST2_LANE = 40, + UNSPEC_ST3_LANE = 41, + UNSPEC_ST4_LANE = 42, + UNSPEC_TLS = 43, + UNSPEC_TLSDESC = 44, + UNSPEC_USHL_2S = 45, + UNSPEC_USHR64 = 46, + UNSPEC_VSTRUCTDUMMY = 47, + UNSPEC_SP_SET = 48, + UNSPEC_SP_TEST = 49, + UNSPEC_ASHIFT_SIGNED = 50, + UNSPEC_ASHIFT_UNSIGNED = 51, + UNSPEC_FMAX = 52, + UNSPEC_FMAXNMV = 53, + UNSPEC_FMAXV = 54, + UNSPEC_FMIN = 55, + UNSPEC_FMINNMV = 56, + UNSPEC_FMINV = 57, + UNSPEC_FADDV = 58, + UNSPEC_SADDV = 59, + UNSPEC_UADDV = 60, + UNSPEC_SMAXV = 61, + UNSPEC_SMINV = 62, + UNSPEC_UMAXV = 63, + UNSPEC_UMINV = 64, + UNSPEC_SHADD = 65, + UNSPEC_UHADD = 66, + UNSPEC_SRHADD = 67, + UNSPEC_URHADD = 68, + UNSPEC_SHSUB = 69, + UNSPEC_UHSUB = 70, + UNSPEC_SRHSUB = 71, + UNSPEC_URHSUB = 72, + UNSPEC_ADDHN = 73, + UNSPEC_RADDHN = 74, + UNSPEC_SUBHN = 75, + UNSPEC_RSUBHN = 76, + UNSPEC_ADDHN2 = 77, + UNSPEC_RADDHN2 = 78, + UNSPEC_SUBHN2 = 79, + UNSPEC_RSUBHN2 = 80, + UNSPEC_SQDMULH = 81, + UNSPEC_SQRDMULH = 82, + UNSPEC_PMUL = 83, + UNSPEC_USQADD = 84, + UNSPEC_SUQADD = 85, + UNSPEC_SQXTUN = 86, + UNSPEC_SQXTN = 87, + UNSPEC_UQXTN = 88, + UNSPEC_SSRA = 89, + UNSPEC_USRA = 90, + UNSPEC_SRSRA = 91, + UNSPEC_URSRA = 92, + UNSPEC_SRSHR = 93, + UNSPEC_URSHR = 94, + UNSPEC_SQSHLU = 95, + UNSPEC_SQSHL = 96, + UNSPEC_UQSHL = 97, + UNSPEC_SQSHRUN = 98, + UNSPEC_SQRSHRUN = 99, + UNSPEC_SQSHRN = 100, + UNSPEC_UQSHRN = 101, + UNSPEC_SQRSHRN = 102, + UNSPEC_UQRSHRN = 103, + UNSPEC_SSHL = 104, + UNSPEC_USHL = 105, + UNSPEC_SRSHL = 106, + UNSPEC_URSHL = 107, + UNSPEC_SQRSHL = 108, + UNSPEC_UQRSHL = 109, + UNSPEC_SSLI = 110, + UNSPEC_USLI = 111, + UNSPEC_SSRI = 112, + UNSPEC_USRI = 113, + UNSPEC_SSHLL = 114, + UNSPEC_USHLL = 115, + UNSPEC_ADDP = 116, + UNSPEC_TBL = 117, + UNSPEC_CONCAT = 118, + UNSPEC_ZIP1 = 119, + UNSPEC_ZIP2 = 120, + UNSPEC_UZP1 = 121, + UNSPEC_UZP2 = 122, + UNSPEC_TRN1 = 123, + UNSPEC_TRN2 = 124, + UNSPEC_EXT = 125, + UNSPEC_REV64 = 126, + UNSPEC_REV32 = 127, + UNSPEC_REV16 = 128, + UNSPEC_AESE = 129, + UNSPEC_AESD = 130, + UNSPEC_AESMC = 131, + UNSPEC_AESIMC = 132, + UNSPEC_SHA1C = 133, + UNSPEC_SHA1M = 134, + UNSPEC_SHA1P = 135, + UNSPEC_SHA1H = 136, + UNSPEC_SHA1SU0 = 137, + UNSPEC_SHA1SU1 = 138, + UNSPEC_SHA256H = 139, + UNSPEC_SHA256H2 = 140, + UNSPEC_SHA256SU0 = 141, + UNSPEC_SHA256SU1 = 142, + UNSPEC_PMULL = 143, + UNSPEC_PMULL2 = 144 }; -#define NUM_UNSPEC_VALUES 133 +#define NUM_UNSPEC_VALUES 145 extern const char *const unspec_strings[]; enum unspecv { UNSPECV_EH_RETURN = 0, - UNSPECV_LX = 1, - UNSPECV_SX = 2, - UNSPECV_LDA = 3, - UNSPECV_STL = 4, - UNSPECV_ATOMIC_CMPSW = 5, - UNSPECV_ATOMIC_EXCHG = 6, - UNSPECV_ATOMIC_OP = 7 + UNSPECV_GET_FPCR = 1, + UNSPECV_SET_FPCR = 2, + UNSPECV_GET_FPSR = 3, + UNSPECV_SET_FPSR = 4, + UNSPECV_LX = 5, + UNSPECV_SX = 6, + UNSPECV_LDA = 7, + UNSPECV_STL = 8, + UNSPECV_ATOMIC_CMPSW = 9, + UNSPECV_ATOMIC_EXCHG = 10, + UNSPECV_ATOMIC_OP = 11 }; -#define NUM_UNSPECV_VALUES 8 +#define NUM_UNSPECV_VALUES 12 extern const char *const unspecv_strings[]; #endif /* GCC_INSN_CONSTANTS_H */ diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-flags.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-flags.h index 469874b..28e9777 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-flags.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-flags.h @@ -87,6 +87,14 @@ #define HAVE_cstorehi_neg 1 #define HAVE_cstoresi_neg 1 #define HAVE_cstoredi_neg 1 +#define HAVE_aarch64_crc32b (TARGET_CRC32) +#define HAVE_aarch64_crc32h (TARGET_CRC32) +#define HAVE_aarch64_crc32w (TARGET_CRC32) +#define HAVE_aarch64_crc32x (TARGET_CRC32) +#define HAVE_aarch64_crc32cb (TARGET_CRC32) +#define HAVE_aarch64_crc32ch (TARGET_CRC32) +#define HAVE_aarch64_crc32cw (TARGET_CRC32) +#define HAVE_aarch64_crc32cx (TARGET_CRC32) #define HAVE_csinc3si_insn 1 #define HAVE_csinc3di_insn 1 #define HAVE_andsi3 1 @@ -240,6 +248,10 @@ #define HAVE_stack_protect_set_di (ptr_mode == DImode) #define HAVE_stack_protect_test_si (ptr_mode == SImode) #define HAVE_stack_protect_test_di (ptr_mode == DImode) +#define HAVE_set_fpcr 1 +#define HAVE_get_fpcr 1 +#define HAVE_set_fpsr 1 +#define HAVE_get_fpsr 1 #define HAVE_aarch64_simd_dupv8qi (TARGET_SIMD) #define HAVE_aarch64_simd_dupv16qi (TARGET_SIMD) #define HAVE_aarch64_simd_dupv4hi (TARGET_SIMD) @@ -468,18 +480,30 @@ #define HAVE_sminv4si3 (TARGET_SIMD) #define HAVE_umaxv4si3 (TARGET_SIMD) #define HAVE_uminv4si3 (TARGET_SIMD) -#define HAVE_move_lo_quad_v16qi (TARGET_SIMD) -#define HAVE_move_lo_quad_v8hi (TARGET_SIMD) -#define HAVE_move_lo_quad_v4si (TARGET_SIMD) -#define HAVE_move_lo_quad_v2di (TARGET_SIMD) -#define HAVE_move_lo_quad_v4sf (TARGET_SIMD) -#define HAVE_move_lo_quad_v2df (TARGET_SIMD) -#define HAVE_aarch64_simd_move_hi_quad_v16qi (TARGET_SIMD) -#define HAVE_aarch64_simd_move_hi_quad_v8hi (TARGET_SIMD) -#define HAVE_aarch64_simd_move_hi_quad_v4si (TARGET_SIMD) -#define HAVE_aarch64_simd_move_hi_quad_v2di (TARGET_SIMD) -#define HAVE_aarch64_simd_move_hi_quad_v4sf (TARGET_SIMD) -#define HAVE_aarch64_simd_move_hi_quad_v2df (TARGET_SIMD) +#define HAVE_move_lo_quad_internal_v16qi (TARGET_SIMD && !BYTES_BIG_ENDIAN) +#define HAVE_move_lo_quad_internal_v8hi (TARGET_SIMD && !BYTES_BIG_ENDIAN) +#define HAVE_move_lo_quad_internal_v4si (TARGET_SIMD && !BYTES_BIG_ENDIAN) +#define HAVE_move_lo_quad_internal_v2di (TARGET_SIMD && !BYTES_BIG_ENDIAN) +#define HAVE_move_lo_quad_internal_v4sf (TARGET_SIMD && !BYTES_BIG_ENDIAN) +#define HAVE_move_lo_quad_internal_v2df (TARGET_SIMD && !BYTES_BIG_ENDIAN) +#define HAVE_move_lo_quad_internal_be_v16qi (TARGET_SIMD && BYTES_BIG_ENDIAN) +#define HAVE_move_lo_quad_internal_be_v8hi (TARGET_SIMD && BYTES_BIG_ENDIAN) +#define HAVE_move_lo_quad_internal_be_v4si (TARGET_SIMD && BYTES_BIG_ENDIAN) +#define HAVE_move_lo_quad_internal_be_v2di (TARGET_SIMD && BYTES_BIG_ENDIAN) +#define HAVE_move_lo_quad_internal_be_v4sf (TARGET_SIMD && BYTES_BIG_ENDIAN) +#define HAVE_move_lo_quad_internal_be_v2df (TARGET_SIMD && BYTES_BIG_ENDIAN) +#define HAVE_aarch64_simd_move_hi_quad_v16qi (TARGET_SIMD && !BYTES_BIG_ENDIAN) +#define HAVE_aarch64_simd_move_hi_quad_v8hi (TARGET_SIMD && !BYTES_BIG_ENDIAN) +#define HAVE_aarch64_simd_move_hi_quad_v4si (TARGET_SIMD && !BYTES_BIG_ENDIAN) +#define HAVE_aarch64_simd_move_hi_quad_v2di (TARGET_SIMD && !BYTES_BIG_ENDIAN) +#define HAVE_aarch64_simd_move_hi_quad_v4sf (TARGET_SIMD && !BYTES_BIG_ENDIAN) +#define HAVE_aarch64_simd_move_hi_quad_v2df (TARGET_SIMD && !BYTES_BIG_ENDIAN) +#define HAVE_aarch64_simd_move_hi_quad_be_v16qi (TARGET_SIMD && BYTES_BIG_ENDIAN) +#define HAVE_aarch64_simd_move_hi_quad_be_v8hi (TARGET_SIMD && BYTES_BIG_ENDIAN) +#define HAVE_aarch64_simd_move_hi_quad_be_v4si (TARGET_SIMD && BYTES_BIG_ENDIAN) +#define HAVE_aarch64_simd_move_hi_quad_be_v2di (TARGET_SIMD && BYTES_BIG_ENDIAN) +#define HAVE_aarch64_simd_move_hi_quad_be_v4sf (TARGET_SIMD && BYTES_BIG_ENDIAN) +#define HAVE_aarch64_simd_move_hi_quad_be_v2df (TARGET_SIMD && BYTES_BIG_ENDIAN) #define HAVE_aarch64_simd_vec_pack_trunc_v8hi (TARGET_SIMD) #define HAVE_aarch64_simd_vec_pack_trunc_v4si (TARGET_SIMD) #define HAVE_aarch64_simd_vec_pack_trunc_v2di (TARGET_SIMD) @@ -692,12 +716,12 @@ #define HAVE_aarch64_get_lanev2sf (TARGET_SIMD) #define HAVE_aarch64_get_lanev4sf (TARGET_SIMD) #define HAVE_aarch64_get_lanev2df (TARGET_SIMD) -#define HAVE_aarch64_combinev8qi (TARGET_SIMD) -#define HAVE_aarch64_combinev4hi (TARGET_SIMD) -#define HAVE_aarch64_combinev2si (TARGET_SIMD) -#define HAVE_aarch64_combinev2sf (TARGET_SIMD) -#define HAVE_aarch64_combinedi (TARGET_SIMD) -#define HAVE_aarch64_combinedf (TARGET_SIMD) +#define HAVE_aarch64_combine_internalv8qi (TARGET_SIMD) +#define HAVE_aarch64_combine_internalv4hi (TARGET_SIMD) +#define HAVE_aarch64_combine_internalv2si (TARGET_SIMD) +#define HAVE_aarch64_combine_internalv2sf (TARGET_SIMD) +#define HAVE_aarch64_combine_internaldi (TARGET_SIMD) +#define HAVE_aarch64_combine_internaldf (TARGET_SIMD) #define HAVE_aarch64_saddlv16qi_hi_internal (TARGET_SIMD) #define HAVE_aarch64_ssublv16qi_hi_internal (TARGET_SIMD) #define HAVE_aarch64_uaddlv16qi_hi_internal (TARGET_SIMD) @@ -950,26 +974,26 @@ #define HAVE_aarch64_sqrdmulhhi (TARGET_SIMD) #define HAVE_aarch64_sqdmulhsi (TARGET_SIMD) #define HAVE_aarch64_sqrdmulhsi (TARGET_SIMD) -#define HAVE_aarch64_sqdmulh_lanev4hi (TARGET_SIMD) -#define HAVE_aarch64_sqrdmulh_lanev4hi (TARGET_SIMD) -#define HAVE_aarch64_sqdmulh_lanev8hi (TARGET_SIMD) -#define HAVE_aarch64_sqrdmulh_lanev8hi (TARGET_SIMD) -#define HAVE_aarch64_sqdmulh_lanev2si (TARGET_SIMD) -#define HAVE_aarch64_sqrdmulh_lanev2si (TARGET_SIMD) -#define HAVE_aarch64_sqdmulh_lanev4si (TARGET_SIMD) -#define HAVE_aarch64_sqrdmulh_lanev4si (TARGET_SIMD) -#define HAVE_aarch64_sqdmulh_laneqv4hi (TARGET_SIMD) -#define HAVE_aarch64_sqrdmulh_laneqv4hi (TARGET_SIMD) -#define HAVE_aarch64_sqdmulh_laneqv8hi (TARGET_SIMD) -#define HAVE_aarch64_sqrdmulh_laneqv8hi (TARGET_SIMD) -#define HAVE_aarch64_sqdmulh_laneqv2si (TARGET_SIMD) -#define HAVE_aarch64_sqrdmulh_laneqv2si (TARGET_SIMD) -#define HAVE_aarch64_sqdmulh_laneqv4si (TARGET_SIMD) -#define HAVE_aarch64_sqrdmulh_laneqv4si (TARGET_SIMD) -#define HAVE_aarch64_sqdmulh_lanehi (TARGET_SIMD) -#define HAVE_aarch64_sqrdmulh_lanehi (TARGET_SIMD) -#define HAVE_aarch64_sqdmulh_lanesi (TARGET_SIMD) -#define HAVE_aarch64_sqrdmulh_lanesi (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_lanev4hi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_lanev4hi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_lanev8hi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_lanev8hi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_lanev2si_internal (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_lanev2si_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_lanev4si_internal (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_lanev4si_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_laneqv4hi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_laneqv4hi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_laneqv8hi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_laneqv8hi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_laneqv2si_internal (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_laneqv2si_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_laneqv4si_internal (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_laneqv4si_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_lanehi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_lanehi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_lanesi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_lanesi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmlalv4hi (TARGET_SIMD) #define HAVE_aarch64_sqdmlslv4hi (TARGET_SIMD) #define HAVE_aarch64_sqdmlalv2si (TARGET_SIMD) @@ -982,10 +1006,18 @@ #define HAVE_aarch64_sqdmlsl_lanev4hi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmlal_lanev2si_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmlsl_lanev2si_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmlal_laneqv4hi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmlsl_laneqv4hi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmlal_laneqv2si_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmlsl_laneqv2si_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmlal_lanehi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmlsl_lanehi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmlal_lanesi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmlsl_lanesi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmlal_laneqhi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmlsl_laneqhi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmlal_laneqsi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmlsl_laneqsi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmlal_nv4hi (TARGET_SIMD) #define HAVE_aarch64_sqdmlsl_nv4hi (TARGET_SIMD) #define HAVE_aarch64_sqdmlal_nv2si (TARGET_SIMD) @@ -998,6 +1030,10 @@ #define HAVE_aarch64_sqdmlsl2_lanev8hi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmlal2_lanev4si_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmlsl2_lanev4si_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmlal2_laneqv8hi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmlsl2_laneqv8hi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmlal2_laneqv4si_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmlsl2_laneqv4si_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmlal2_nv8hi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmlsl2_nv8hi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmlal2_nv4si_internal (TARGET_SIMD) @@ -1008,14 +1044,20 @@ #define HAVE_aarch64_sqdmullsi (TARGET_SIMD) #define HAVE_aarch64_sqdmull_lanev4hi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmull_lanev2si_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmull_laneqv4hi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmull_laneqv2si_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmull_lanehi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmull_lanesi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmull_laneqhi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmull_laneqsi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmull_nv4hi (TARGET_SIMD) #define HAVE_aarch64_sqdmull_nv2si (TARGET_SIMD) #define HAVE_aarch64_sqdmull2v8hi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmull2v4si_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmull2_lanev8hi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmull2_lanev4si_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmull2_laneqv8hi_internal (TARGET_SIMD) +#define HAVE_aarch64_sqdmull2_laneqv4si_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmull2_nv8hi_internal (TARGET_SIMD) #define HAVE_aarch64_sqdmull2_nv4si_internal (TARGET_SIMD) #define HAVE_aarch64_sshlv8qi (TARGET_SIMD) @@ -1525,6 +1567,46 @@ #define HAVE_aarch64_trn2v2df (TARGET_SIMD) #define HAVE_aarch64_uzp1v2df (TARGET_SIMD) #define HAVE_aarch64_uzp2v2df (TARGET_SIMD) +#define HAVE_aarch64_extv8qi (TARGET_SIMD) +#define HAVE_aarch64_extv16qi (TARGET_SIMD) +#define HAVE_aarch64_extv4hi (TARGET_SIMD) +#define HAVE_aarch64_extv8hi (TARGET_SIMD) +#define HAVE_aarch64_extv2si (TARGET_SIMD) +#define HAVE_aarch64_extv4si (TARGET_SIMD) +#define HAVE_aarch64_extv2di (TARGET_SIMD) +#define HAVE_aarch64_extv2sf (TARGET_SIMD) +#define HAVE_aarch64_extv4sf (TARGET_SIMD) +#define HAVE_aarch64_extv2df (TARGET_SIMD) +#define HAVE_aarch64_rev64v8qi (TARGET_SIMD) +#define HAVE_aarch64_rev32v8qi (TARGET_SIMD) +#define HAVE_aarch64_rev16v8qi (TARGET_SIMD) +#define HAVE_aarch64_rev64v16qi (TARGET_SIMD) +#define HAVE_aarch64_rev32v16qi (TARGET_SIMD) +#define HAVE_aarch64_rev16v16qi (TARGET_SIMD) +#define HAVE_aarch64_rev64v4hi (TARGET_SIMD) +#define HAVE_aarch64_rev32v4hi (TARGET_SIMD) +#define HAVE_aarch64_rev16v4hi (TARGET_SIMD) +#define HAVE_aarch64_rev64v8hi (TARGET_SIMD) +#define HAVE_aarch64_rev32v8hi (TARGET_SIMD) +#define HAVE_aarch64_rev16v8hi (TARGET_SIMD) +#define HAVE_aarch64_rev64v2si (TARGET_SIMD) +#define HAVE_aarch64_rev32v2si (TARGET_SIMD) +#define HAVE_aarch64_rev16v2si (TARGET_SIMD) +#define HAVE_aarch64_rev64v4si (TARGET_SIMD) +#define HAVE_aarch64_rev32v4si (TARGET_SIMD) +#define HAVE_aarch64_rev16v4si (TARGET_SIMD) +#define HAVE_aarch64_rev64v2di (TARGET_SIMD) +#define HAVE_aarch64_rev32v2di (TARGET_SIMD) +#define HAVE_aarch64_rev16v2di (TARGET_SIMD) +#define HAVE_aarch64_rev64v2sf (TARGET_SIMD) +#define HAVE_aarch64_rev32v2sf (TARGET_SIMD) +#define HAVE_aarch64_rev16v2sf (TARGET_SIMD) +#define HAVE_aarch64_rev64v4sf (TARGET_SIMD) +#define HAVE_aarch64_rev32v4sf (TARGET_SIMD) +#define HAVE_aarch64_rev16v4sf (TARGET_SIMD) +#define HAVE_aarch64_rev64v2df (TARGET_SIMD) +#define HAVE_aarch64_rev32v2df (TARGET_SIMD) +#define HAVE_aarch64_rev16v2df (TARGET_SIMD) #define HAVE_aarch64_st2v8qi_dreg (TARGET_SIMD) #define HAVE_aarch64_st2v4hi_dreg (TARGET_SIMD) #define HAVE_aarch64_st2v2si_dreg (TARGET_SIMD) @@ -1688,6 +1770,7 @@ #define HAVE_movsf 1 #define HAVE_movdf 1 #define HAVE_movtf 1 +#define HAVE_movmemdi (!STRICT_ALIGNMENT) #define HAVE_extendsidi2 1 #define HAVE_zero_extendsidi2 1 #define HAVE_extendqisi2 1 @@ -1831,6 +1914,12 @@ #define HAVE_vec_setv2sf (TARGET_SIMD) #define HAVE_vec_setv4sf (TARGET_SIMD) #define HAVE_vec_setv2df (TARGET_SIMD) +#define HAVE_move_lo_quad_v16qi (TARGET_SIMD) +#define HAVE_move_lo_quad_v8hi (TARGET_SIMD) +#define HAVE_move_lo_quad_v4si (TARGET_SIMD) +#define HAVE_move_lo_quad_v2di (TARGET_SIMD) +#define HAVE_move_lo_quad_v4sf (TARGET_SIMD) +#define HAVE_move_lo_quad_v2df (TARGET_SIMD) #define HAVE_move_hi_quad_v16qi (TARGET_SIMD) #define HAVE_move_hi_quad_v8hi (TARGET_SIMD) #define HAVE_move_hi_quad_v4si (TARGET_SIMD) @@ -2028,6 +2117,12 @@ #define HAVE_aarch64_reinterpretv2dfv2di (TARGET_SIMD) #define HAVE_aarch64_reinterpretv2dfv4sf (TARGET_SIMD) #define HAVE_aarch64_reinterpretv2dfv2df (TARGET_SIMD) +#define HAVE_aarch64_combinev8qi (TARGET_SIMD) +#define HAVE_aarch64_combinev4hi (TARGET_SIMD) +#define HAVE_aarch64_combinev2si (TARGET_SIMD) +#define HAVE_aarch64_combinev2sf (TARGET_SIMD) +#define HAVE_aarch64_combinedi (TARGET_SIMD) +#define HAVE_aarch64_combinedf (TARGET_SIMD) #define HAVE_aarch64_simd_combinev8qi (TARGET_SIMD) #define HAVE_aarch64_simd_combinev4hi (TARGET_SIMD) #define HAVE_aarch64_simd_combinev2si (TARGET_SIMD) @@ -2058,6 +2153,26 @@ #define HAVE_aarch64_usubw2v16qi (TARGET_SIMD) #define HAVE_aarch64_usubw2v8hi (TARGET_SIMD) #define HAVE_aarch64_usubw2v4si (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_lanev4hi (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_lanev8hi (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_lanev2si (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_lanev4si (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_lanev4hi (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_lanev8hi (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_lanev2si (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_lanev4si (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_laneqv4hi (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_laneqv8hi (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_laneqv2si (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_laneqv4si (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_laneqv4hi (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_laneqv8hi (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_laneqv2si (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_laneqv4si (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_lanehi (TARGET_SIMD) +#define HAVE_aarch64_sqdmulh_lanesi (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_lanehi (TARGET_SIMD) +#define HAVE_aarch64_sqrdmulh_lanesi (TARGET_SIMD) #define HAVE_aarch64_sqdmlal_lanev4hi (TARGET_SIMD) #define HAVE_aarch64_sqdmlal_lanev2si (TARGET_SIMD) #define HAVE_aarch64_sqdmlal_lanehi (TARGET_SIMD) @@ -2201,6 +2316,7 @@ #define HAVE_vec_perm_constv2df (TARGET_SIMD) #define HAVE_vec_permv8qi (TARGET_SIMD && !BYTES_BIG_ENDIAN) #define HAVE_vec_permv16qi (TARGET_SIMD && !BYTES_BIG_ENDIAN) +#define HAVE_aarch64_im_lane_boundsi (TARGET_SIMD) #define HAVE_aarch64_st2v8qi (TARGET_SIMD) #define HAVE_aarch64_st2v4hi (TARGET_SIMD) #define HAVE_aarch64_st2v2si (TARGET_SIMD) @@ -2309,2284 +2425,2400 @@ #define HAVE_atomic_compare_and_swapdi 1 #define HAVE_mem_thread_fence 1 #define HAVE_dmb 1 -extern rtx gen_indirect_jump (rtx); -extern rtx gen_jump (rtx); -extern rtx gen_casesi_dispatch (rtx, rtx, rtx); -extern rtx gen_nop (void); -extern rtx gen_trap (void); -extern rtx gen_eh_return (rtx); -extern rtx gen_insv_immsi (rtx, rtx, rtx); -extern rtx gen_insv_immdi (rtx, rtx, rtx); -extern rtx gen_load_pairsi (rtx, rtx, rtx, rtx); -extern rtx gen_load_pairdi (rtx, rtx, rtx, rtx); -extern rtx gen_store_pairsi (rtx, rtx, rtx, rtx); -extern rtx gen_store_pairdi (rtx, rtx, rtx, rtx); -extern rtx gen_load_pairsf (rtx, rtx, rtx, rtx); -extern rtx gen_load_pairdf (rtx, rtx, rtx, rtx); -extern rtx gen_store_pairsf (rtx, rtx, rtx, rtx); -extern rtx gen_store_pairdf (rtx, rtx, rtx, rtx); -extern rtx gen_loadwb_pairsi_si (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_loadwb_pairsi_di (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_loadwb_pairdi_si (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_loadwb_pairdi_di (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_storewb_pairsi_si (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_storewb_pairsi_di (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_storewb_pairdi_si (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_storewb_pairdi_di (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_addsi3_compare0 (rtx, rtx, rtx); -extern rtx gen_adddi3_compare0 (rtx, rtx, rtx); -extern rtx gen_addsi3_carryin (rtx, rtx, rtx); -extern rtx gen_adddi3_carryin (rtx, rtx, rtx); -extern rtx gen_subsi3 (rtx, rtx, rtx); -extern rtx gen_subdi3 (rtx, rtx, rtx); -extern rtx gen_subsi3_compare0 (rtx, rtx, rtx); -extern rtx gen_subdi3_compare0 (rtx, rtx, rtx); -extern rtx gen_subsi3_carryin (rtx, rtx, rtx); -extern rtx gen_subdi3_carryin (rtx, rtx, rtx); -extern rtx gen_absdi2 (rtx, rtx); -extern rtx gen_negsi2 (rtx, rtx); -extern rtx gen_negdi2 (rtx, rtx); -extern rtx gen_mulsi3 (rtx, rtx, rtx); -extern rtx gen_muldi3 (rtx, rtx, rtx); -extern rtx gen_maddsi (rtx, rtx, rtx, rtx); -extern rtx gen_madddi (rtx, rtx, rtx, rtx); -extern rtx gen_mulsidi3 (rtx, rtx, rtx); -extern rtx gen_umulsidi3 (rtx, rtx, rtx); -extern rtx gen_maddsidi4 (rtx, rtx, rtx, rtx); -extern rtx gen_umaddsidi4 (rtx, rtx, rtx, rtx); -extern rtx gen_msubsidi4 (rtx, rtx, rtx, rtx); -extern rtx gen_umsubsidi4 (rtx, rtx, rtx, rtx); -extern rtx gen_smuldi3_highpart (rtx, rtx, rtx); -extern rtx gen_umuldi3_highpart (rtx, rtx, rtx); -extern rtx gen_divsi3 (rtx, rtx, rtx); -extern rtx gen_udivsi3 (rtx, rtx, rtx); -extern rtx gen_divdi3 (rtx, rtx, rtx); -extern rtx gen_udivdi3 (rtx, rtx, rtx); -extern rtx gen_cstoreqi_neg (rtx, rtx, rtx); -extern rtx gen_cstorehi_neg (rtx, rtx, rtx); -extern rtx gen_cstoresi_neg (rtx, rtx, rtx); -extern rtx gen_cstoredi_neg (rtx, rtx, rtx); -extern rtx gen_csinc3si_insn (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_csinc3di_insn (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_andsi3 (rtx, rtx, rtx); -extern rtx gen_iorsi3 (rtx, rtx, rtx); -extern rtx gen_xorsi3 (rtx, rtx, rtx); -extern rtx gen_anddi3 (rtx, rtx, rtx); -extern rtx gen_iordi3 (rtx, rtx, rtx); -extern rtx gen_xordi3 (rtx, rtx, rtx); -extern rtx gen_one_cmplsi2 (rtx, rtx); -extern rtx gen_one_cmpldi2 (rtx, rtx); -extern rtx gen_clzsi2 (rtx, rtx); -extern rtx gen_clzdi2 (rtx, rtx); -extern rtx gen_clrsbsi2 (rtx, rtx); -extern rtx gen_clrsbdi2 (rtx, rtx); -extern rtx gen_rbitsi2 (rtx, rtx); -extern rtx gen_rbitdi2 (rtx, rtx); -extern rtx gen_bswapsi2 (rtx, rtx); -extern rtx gen_bswapdi2 (rtx, rtx); -extern rtx gen_bswaphi2 (rtx, rtx); -extern rtx gen_rev16si2 (rtx, rtx, rtx, rtx); -extern rtx gen_rev16di2 (rtx, rtx, rtx, rtx); -extern rtx gen_rev16si2_alt (rtx, rtx, rtx, rtx); -extern rtx gen_rev16di2_alt (rtx, rtx, rtx, rtx); -extern rtx gen_btruncsf2 (rtx, rtx); -extern rtx gen_ceilsf2 (rtx, rtx); -extern rtx gen_floorsf2 (rtx, rtx); -extern rtx gen_frintnsf2 (rtx, rtx); -extern rtx gen_nearbyintsf2 (rtx, rtx); -extern rtx gen_rintsf2 (rtx, rtx); -extern rtx gen_roundsf2 (rtx, rtx); -extern rtx gen_btruncdf2 (rtx, rtx); -extern rtx gen_ceildf2 (rtx, rtx); -extern rtx gen_floordf2 (rtx, rtx); -extern rtx gen_frintndf2 (rtx, rtx); -extern rtx gen_nearbyintdf2 (rtx, rtx); -extern rtx gen_rintdf2 (rtx, rtx); -extern rtx gen_rounddf2 (rtx, rtx); -extern rtx gen_lbtruncsfsi2 (rtx, rtx); -extern rtx gen_lceilsfsi2 (rtx, rtx); -extern rtx gen_lfloorsfsi2 (rtx, rtx); -extern rtx gen_lroundsfsi2 (rtx, rtx); -extern rtx gen_lfrintnsfsi2 (rtx, rtx); -extern rtx gen_lbtruncusfsi2 (rtx, rtx); -extern rtx gen_lceilusfsi2 (rtx, rtx); -extern rtx gen_lfloorusfsi2 (rtx, rtx); -extern rtx gen_lroundusfsi2 (rtx, rtx); -extern rtx gen_lfrintnusfsi2 (rtx, rtx); -extern rtx gen_lbtruncdfsi2 (rtx, rtx); -extern rtx gen_lceildfsi2 (rtx, rtx); -extern rtx gen_lfloordfsi2 (rtx, rtx); -extern rtx gen_lrounddfsi2 (rtx, rtx); -extern rtx gen_lfrintndfsi2 (rtx, rtx); -extern rtx gen_lbtruncudfsi2 (rtx, rtx); -extern rtx gen_lceiludfsi2 (rtx, rtx); -extern rtx gen_lfloorudfsi2 (rtx, rtx); -extern rtx gen_lroundudfsi2 (rtx, rtx); -extern rtx gen_lfrintnudfsi2 (rtx, rtx); -extern rtx gen_lbtruncsfdi2 (rtx, rtx); -extern rtx gen_lceilsfdi2 (rtx, rtx); -extern rtx gen_lfloorsfdi2 (rtx, rtx); -extern rtx gen_lroundsfdi2 (rtx, rtx); -extern rtx gen_lfrintnsfdi2 (rtx, rtx); -extern rtx gen_lbtruncusfdi2 (rtx, rtx); -extern rtx gen_lceilusfdi2 (rtx, rtx); -extern rtx gen_lfloorusfdi2 (rtx, rtx); -extern rtx gen_lroundusfdi2 (rtx, rtx); -extern rtx gen_lfrintnusfdi2 (rtx, rtx); -extern rtx gen_lbtruncdfdi2 (rtx, rtx); -extern rtx gen_lceildfdi2 (rtx, rtx); -extern rtx gen_lfloordfdi2 (rtx, rtx); -extern rtx gen_lrounddfdi2 (rtx, rtx); -extern rtx gen_lfrintndfdi2 (rtx, rtx); -extern rtx gen_lbtruncudfdi2 (rtx, rtx); -extern rtx gen_lceiludfdi2 (rtx, rtx); -extern rtx gen_lfloorudfdi2 (rtx, rtx); -extern rtx gen_lroundudfdi2 (rtx, rtx); -extern rtx gen_lfrintnudfdi2 (rtx, rtx); -extern rtx gen_fmasf4 (rtx, rtx, rtx, rtx); -extern rtx gen_fmadf4 (rtx, rtx, rtx, rtx); -extern rtx gen_fnmasf4 (rtx, rtx, rtx, rtx); -extern rtx gen_fnmadf4 (rtx, rtx, rtx, rtx); -extern rtx gen_fmssf4 (rtx, rtx, rtx, rtx); -extern rtx gen_fmsdf4 (rtx, rtx, rtx, rtx); -extern rtx gen_fnmssf4 (rtx, rtx, rtx, rtx); -extern rtx gen_fnmsdf4 (rtx, rtx, rtx, rtx); -extern rtx gen_extendsfdf2 (rtx, rtx); -extern rtx gen_truncdfsf2 (rtx, rtx); -extern rtx gen_fix_truncsfsi2 (rtx, rtx); -extern rtx gen_fix_truncdfsi2 (rtx, rtx); -extern rtx gen_fix_truncsfdi2 (rtx, rtx); -extern rtx gen_fix_truncdfdi2 (rtx, rtx); -extern rtx gen_fixuns_truncsfsi2 (rtx, rtx); -extern rtx gen_fixuns_truncdfsi2 (rtx, rtx); -extern rtx gen_fixuns_truncsfdi2 (rtx, rtx); -extern rtx gen_fixuns_truncdfdi2 (rtx, rtx); -extern rtx gen_floatsisf2 (rtx, rtx); -extern rtx gen_floatsidf2 (rtx, rtx); -extern rtx gen_floatdisf2 (rtx, rtx); -extern rtx gen_floatdidf2 (rtx, rtx); -extern rtx gen_floatunssisf2 (rtx, rtx); -extern rtx gen_floatunssidf2 (rtx, rtx); -extern rtx gen_floatunsdisf2 (rtx, rtx); -extern rtx gen_floatunsdidf2 (rtx, rtx); -extern rtx gen_addsf3 (rtx, rtx, rtx); -extern rtx gen_adddf3 (rtx, rtx, rtx); -extern rtx gen_subsf3 (rtx, rtx, rtx); -extern rtx gen_subdf3 (rtx, rtx, rtx); -extern rtx gen_mulsf3 (rtx, rtx, rtx); -extern rtx gen_muldf3 (rtx, rtx, rtx); -extern rtx gen_divsf3 (rtx, rtx, rtx); -extern rtx gen_divdf3 (rtx, rtx, rtx); -extern rtx gen_negsf2 (rtx, rtx); -extern rtx gen_negdf2 (rtx, rtx); -extern rtx gen_sqrtsf2 (rtx, rtx); -extern rtx gen_sqrtdf2 (rtx, rtx); -extern rtx gen_abssf2 (rtx, rtx); -extern rtx gen_absdf2 (rtx, rtx); -extern rtx gen_smaxsf3 (rtx, rtx, rtx); -extern rtx gen_smaxdf3 (rtx, rtx, rtx); -extern rtx gen_sminsf3 (rtx, rtx, rtx); -extern rtx gen_smindf3 (rtx, rtx, rtx); -extern rtx gen_aarch64_movdi_tilow (rtx, rtx); -extern rtx gen_aarch64_movdi_tflow (rtx, rtx); -extern rtx gen_aarch64_movdi_tihigh (rtx, rtx); -extern rtx gen_aarch64_movdi_tfhigh (rtx, rtx); -extern rtx gen_aarch64_movtihigh_di (rtx, rtx); -extern rtx gen_aarch64_movtfhigh_di (rtx, rtx); -extern rtx gen_aarch64_movtilow_di (rtx, rtx); -extern rtx gen_aarch64_movtflow_di (rtx, rtx); -extern rtx gen_aarch64_movtilow_tilow (rtx, rtx); -extern rtx gen_add_losym_si (rtx, rtx, rtx); -extern rtx gen_add_losym_di (rtx, rtx, rtx); -extern rtx gen_ldr_got_small_si (rtx, rtx, rtx); -extern rtx gen_ldr_got_small_di (rtx, rtx, rtx); -extern rtx gen_ldr_got_small_sidi (rtx, rtx, rtx); -extern rtx gen_ldr_got_tiny (rtx, rtx); -extern rtx gen_aarch64_load_tp_hard (rtx); -extern rtx gen_tlsie_small_si (rtx, rtx); -extern rtx gen_tlsie_small_di (rtx, rtx); -extern rtx gen_tlsie_small_sidi (rtx, rtx); -extern rtx gen_tlsle_small_si (rtx, rtx, rtx); -extern rtx gen_tlsle_small_di (rtx, rtx, rtx); -extern rtx gen_tlsdesc_small_si (rtx); -extern rtx gen_tlsdesc_small_di (rtx); -extern rtx gen_stack_tie (rtx, rtx); -extern rtx gen_stack_protect_set_si (rtx, rtx); -extern rtx gen_stack_protect_set_di (rtx, rtx); -extern rtx gen_stack_protect_test_si (rtx, rtx, rtx); -extern rtx gen_stack_protect_test_di (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_dupv8qi (rtx, rtx); -extern rtx gen_aarch64_simd_dupv16qi (rtx, rtx); -extern rtx gen_aarch64_simd_dupv4hi (rtx, rtx); -extern rtx gen_aarch64_simd_dupv8hi (rtx, rtx); -extern rtx gen_aarch64_simd_dupv2si (rtx, rtx); -extern rtx gen_aarch64_simd_dupv4si (rtx, rtx); -extern rtx gen_aarch64_simd_dupv2di (rtx, rtx); -extern rtx gen_aarch64_simd_dupv2sf (rtx, rtx); -extern rtx gen_aarch64_simd_dupv4sf (rtx, rtx); -extern rtx gen_aarch64_simd_dupv2df (rtx, rtx); -extern rtx gen_aarch64_dup_lanev8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lanev16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lanev4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lanev8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lanev2si (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lanev4si (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lanev2di (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lanev2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lanev4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lanev2df (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lane_to_128v8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lane_to_64v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lane_to_128v4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lane_to_64v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lane_to_128v2si (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lane_to_64v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lane_to_64v2di (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lane_to_128v2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lane_to_64v4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_dup_lane_to_64v2df (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_mov_from_v16qilow (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_mov_from_v8hilow (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_mov_from_v4silow (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_mov_from_v2dilow (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_mov_from_v4sflow (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_mov_from_v2dflow (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_mov_from_v16qihigh (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_mov_from_v8hihigh (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_mov_from_v4sihigh (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_mov_from_v2dihigh (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_mov_from_v4sfhigh (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_mov_from_v2dfhigh (rtx, rtx, rtx); -extern rtx gen_ornv8qi3 (rtx, rtx, rtx); -extern rtx gen_ornv16qi3 (rtx, rtx, rtx); -extern rtx gen_ornv4hi3 (rtx, rtx, rtx); -extern rtx gen_ornv8hi3 (rtx, rtx, rtx); -extern rtx gen_ornv2si3 (rtx, rtx, rtx); -extern rtx gen_ornv4si3 (rtx, rtx, rtx); -extern rtx gen_ornv2di3 (rtx, rtx, rtx); -extern rtx gen_bicv8qi3 (rtx, rtx, rtx); -extern rtx gen_bicv16qi3 (rtx, rtx, rtx); -extern rtx gen_bicv4hi3 (rtx, rtx, rtx); -extern rtx gen_bicv8hi3 (rtx, rtx, rtx); -extern rtx gen_bicv2si3 (rtx, rtx, rtx); -extern rtx gen_bicv4si3 (rtx, rtx, rtx); -extern rtx gen_bicv2di3 (rtx, rtx, rtx); -extern rtx gen_addv8qi3 (rtx, rtx, rtx); -extern rtx gen_addv16qi3 (rtx, rtx, rtx); -extern rtx gen_addv4hi3 (rtx, rtx, rtx); -extern rtx gen_addv8hi3 (rtx, rtx, rtx); -extern rtx gen_addv2si3 (rtx, rtx, rtx); -extern rtx gen_addv4si3 (rtx, rtx, rtx); -extern rtx gen_addv2di3 (rtx, rtx, rtx); -extern rtx gen_subv8qi3 (rtx, rtx, rtx); -extern rtx gen_subv16qi3 (rtx, rtx, rtx); -extern rtx gen_subv4hi3 (rtx, rtx, rtx); -extern rtx gen_subv8hi3 (rtx, rtx, rtx); -extern rtx gen_subv2si3 (rtx, rtx, rtx); -extern rtx gen_subv4si3 (rtx, rtx, rtx); -extern rtx gen_subv2di3 (rtx, rtx, rtx); -extern rtx gen_mulv8qi3 (rtx, rtx, rtx); -extern rtx gen_mulv16qi3 (rtx, rtx, rtx); -extern rtx gen_mulv4hi3 (rtx, rtx, rtx); -extern rtx gen_mulv8hi3 (rtx, rtx, rtx); -extern rtx gen_mulv2si3 (rtx, rtx, rtx); -extern rtx gen_mulv4si3 (rtx, rtx, rtx); -extern rtx gen_bswapv4hi (rtx, rtx); -extern rtx gen_bswapv8hi (rtx, rtx); -extern rtx gen_bswapv2si (rtx, rtx); -extern rtx gen_bswapv4si (rtx, rtx); -extern rtx gen_bswapv2di (rtx, rtx); -extern rtx gen_negv8qi2 (rtx, rtx); -extern rtx gen_negv16qi2 (rtx, rtx); -extern rtx gen_negv4hi2 (rtx, rtx); -extern rtx gen_negv8hi2 (rtx, rtx); -extern rtx gen_negv2si2 (rtx, rtx); -extern rtx gen_negv4si2 (rtx, rtx); -extern rtx gen_negv2di2 (rtx, rtx); -extern rtx gen_absv8qi2 (rtx, rtx); -extern rtx gen_absv16qi2 (rtx, rtx); -extern rtx gen_absv4hi2 (rtx, rtx); -extern rtx gen_absv8hi2 (rtx, rtx); -extern rtx gen_absv2si2 (rtx, rtx); -extern rtx gen_absv4si2 (rtx, rtx); -extern rtx gen_absv2di2 (rtx, rtx); -extern rtx gen_abdv8qi_3 (rtx, rtx, rtx); -extern rtx gen_abdv16qi_3 (rtx, rtx, rtx); -extern rtx gen_abdv4hi_3 (rtx, rtx, rtx); -extern rtx gen_abdv8hi_3 (rtx, rtx, rtx); -extern rtx gen_abdv2si_3 (rtx, rtx, rtx); -extern rtx gen_abdv4si_3 (rtx, rtx, rtx); -extern rtx gen_abav8qi_3 (rtx, rtx, rtx, rtx); -extern rtx gen_abav16qi_3 (rtx, rtx, rtx, rtx); -extern rtx gen_abav4hi_3 (rtx, rtx, rtx, rtx); -extern rtx gen_abav8hi_3 (rtx, rtx, rtx, rtx); -extern rtx gen_abav2si_3 (rtx, rtx, rtx, rtx); -extern rtx gen_abav4si_3 (rtx, rtx, rtx, rtx); -extern rtx gen_fabdv2sf_3 (rtx, rtx, rtx); -extern rtx gen_fabdv4sf_3 (rtx, rtx, rtx); -extern rtx gen_fabdv2df_3 (rtx, rtx, rtx); -extern rtx gen_andv8qi3 (rtx, rtx, rtx); -extern rtx gen_andv16qi3 (rtx, rtx, rtx); -extern rtx gen_andv4hi3 (rtx, rtx, rtx); -extern rtx gen_andv8hi3 (rtx, rtx, rtx); -extern rtx gen_andv2si3 (rtx, rtx, rtx); -extern rtx gen_andv4si3 (rtx, rtx, rtx); -extern rtx gen_andv2di3 (rtx, rtx, rtx); -extern rtx gen_iorv8qi3 (rtx, rtx, rtx); -extern rtx gen_iorv16qi3 (rtx, rtx, rtx); -extern rtx gen_iorv4hi3 (rtx, rtx, rtx); -extern rtx gen_iorv8hi3 (rtx, rtx, rtx); -extern rtx gen_iorv2si3 (rtx, rtx, rtx); -extern rtx gen_iorv4si3 (rtx, rtx, rtx); -extern rtx gen_iorv2di3 (rtx, rtx, rtx); -extern rtx gen_xorv8qi3 (rtx, rtx, rtx); -extern rtx gen_xorv16qi3 (rtx, rtx, rtx); -extern rtx gen_xorv4hi3 (rtx, rtx, rtx); -extern rtx gen_xorv8hi3 (rtx, rtx, rtx); -extern rtx gen_xorv2si3 (rtx, rtx, rtx); -extern rtx gen_xorv4si3 (rtx, rtx, rtx); -extern rtx gen_xorv2di3 (rtx, rtx, rtx); -extern rtx gen_one_cmplv8qi2 (rtx, rtx); -extern rtx gen_one_cmplv16qi2 (rtx, rtx); -extern rtx gen_one_cmplv4hi2 (rtx, rtx); -extern rtx gen_one_cmplv8hi2 (rtx, rtx); -extern rtx gen_one_cmplv2si2 (rtx, rtx); -extern rtx gen_one_cmplv4si2 (rtx, rtx); -extern rtx gen_one_cmplv2di2 (rtx, rtx); -extern rtx gen_aarch64_simd_vec_setv8qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_setv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_setv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_setv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_setv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_setv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_lshrv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_lshrv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_lshrv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_lshrv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_lshrv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_lshrv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_lshrv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_ashrv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_ashrv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_ashrv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_ashrv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_ashrv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_ashrv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_ashrv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_imm_shlv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_imm_shlv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_imm_shlv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_imm_shlv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_imm_shlv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_imm_shlv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_imm_shlv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_sshlv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_sshlv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_sshlv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_sshlv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_sshlv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_sshlv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_sshlv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_shlv8qi_unsigned (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_shlv16qi_unsigned (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_shlv4hi_unsigned (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_shlv8hi_unsigned (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_shlv2si_unsigned (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_shlv4si_unsigned (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_shlv2di_unsigned (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_shlv8qi_signed (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_shlv16qi_signed (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_shlv4hi_signed (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_shlv8hi_signed (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_shlv2si_signed (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_shlv4si_signed (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_reg_shlv2di_signed (rtx, rtx, rtx); -extern rtx gen_aarch64_sshr_simddi (rtx, rtx); -extern rtx gen_aarch64_ushr_simddi (rtx, rtx); -extern rtx gen_aarch64_simd_vec_setv2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_setv2sf (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_setv4sf (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_setv2df (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_mlav8qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_mlav16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_mlav4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_mlav8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_mlav2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_mlav4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_mlsv8qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_mlsv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_mlsv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_mlsv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_mlsv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_mlsv4si (rtx, rtx, rtx, rtx); -extern rtx gen_smaxv8qi3 (rtx, rtx, rtx); -extern rtx gen_sminv8qi3 (rtx, rtx, rtx); -extern rtx gen_umaxv8qi3 (rtx, rtx, rtx); -extern rtx gen_uminv8qi3 (rtx, rtx, rtx); -extern rtx gen_smaxv16qi3 (rtx, rtx, rtx); -extern rtx gen_sminv16qi3 (rtx, rtx, rtx); -extern rtx gen_umaxv16qi3 (rtx, rtx, rtx); -extern rtx gen_uminv16qi3 (rtx, rtx, rtx); -extern rtx gen_smaxv4hi3 (rtx, rtx, rtx); -extern rtx gen_sminv4hi3 (rtx, rtx, rtx); -extern rtx gen_umaxv4hi3 (rtx, rtx, rtx); -extern rtx gen_uminv4hi3 (rtx, rtx, rtx); -extern rtx gen_smaxv8hi3 (rtx, rtx, rtx); -extern rtx gen_sminv8hi3 (rtx, rtx, rtx); -extern rtx gen_umaxv8hi3 (rtx, rtx, rtx); -extern rtx gen_uminv8hi3 (rtx, rtx, rtx); -extern rtx gen_smaxv2si3 (rtx, rtx, rtx); -extern rtx gen_sminv2si3 (rtx, rtx, rtx); -extern rtx gen_umaxv2si3 (rtx, rtx, rtx); -extern rtx gen_uminv2si3 (rtx, rtx, rtx); -extern rtx gen_smaxv4si3 (rtx, rtx, rtx); -extern rtx gen_sminv4si3 (rtx, rtx, rtx); -extern rtx gen_umaxv4si3 (rtx, rtx, rtx); -extern rtx gen_uminv4si3 (rtx, rtx, rtx); -extern rtx gen_move_lo_quad_v16qi (rtx, rtx); -extern rtx gen_move_lo_quad_v8hi (rtx, rtx); -extern rtx gen_move_lo_quad_v4si (rtx, rtx); -extern rtx gen_move_lo_quad_v2di (rtx, rtx); -extern rtx gen_move_lo_quad_v4sf (rtx, rtx); -extern rtx gen_move_lo_quad_v2df (rtx, rtx); -extern rtx gen_aarch64_simd_move_hi_quad_v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_move_hi_quad_v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_move_hi_quad_v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_move_hi_quad_v2di (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_move_hi_quad_v4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_move_hi_quad_v2df (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_pack_trunc_v8hi (rtx, rtx); -extern rtx gen_aarch64_simd_vec_pack_trunc_v4si (rtx, rtx); -extern rtx gen_aarch64_simd_vec_pack_trunc_v2di (rtx, rtx); -extern rtx gen_vec_pack_trunc_v8hi (rtx, rtx, rtx); -extern rtx gen_vec_pack_trunc_v4si (rtx, rtx, rtx); -extern rtx gen_vec_pack_trunc_v2di (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_unpacks_lo_v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_unpacku_lo_v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_unpacks_lo_v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_unpacku_lo_v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_unpacks_lo_v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_unpacku_lo_v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_unpacks_hi_v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_unpacku_hi_v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_unpacks_hi_v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_unpacku_hi_v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_unpacks_hi_v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_unpacku_hi_v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_smult_lo_v16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_umult_lo_v16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_smult_lo_v8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_umult_lo_v8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_smult_lo_v4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_umult_lo_v4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_smult_hi_v16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_umult_hi_v16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_smult_hi_v8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_umult_hi_v8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_smult_hi_v4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_vec_umult_hi_v4si (rtx, rtx, rtx, rtx); -extern rtx gen_addv2sf3 (rtx, rtx, rtx); -extern rtx gen_addv4sf3 (rtx, rtx, rtx); -extern rtx gen_addv2df3 (rtx, rtx, rtx); -extern rtx gen_subv2sf3 (rtx, rtx, rtx); -extern rtx gen_subv4sf3 (rtx, rtx, rtx); -extern rtx gen_subv2df3 (rtx, rtx, rtx); -extern rtx gen_mulv2sf3 (rtx, rtx, rtx); -extern rtx gen_mulv4sf3 (rtx, rtx, rtx); -extern rtx gen_mulv2df3 (rtx, rtx, rtx); -extern rtx gen_divv2sf3 (rtx, rtx, rtx); -extern rtx gen_divv4sf3 (rtx, rtx, rtx); -extern rtx gen_divv2df3 (rtx, rtx, rtx); -extern rtx gen_negv2sf2 (rtx, rtx); -extern rtx gen_negv4sf2 (rtx, rtx); -extern rtx gen_negv2df2 (rtx, rtx); -extern rtx gen_absv2sf2 (rtx, rtx); -extern rtx gen_absv4sf2 (rtx, rtx); -extern rtx gen_absv2df2 (rtx, rtx); -extern rtx gen_fmav2sf4 (rtx, rtx, rtx, rtx); -extern rtx gen_fmav4sf4 (rtx, rtx, rtx, rtx); -extern rtx gen_fmav2df4 (rtx, rtx, rtx, rtx); -extern rtx gen_fnmav2sf4 (rtx, rtx, rtx, rtx); -extern rtx gen_fnmav4sf4 (rtx, rtx, rtx, rtx); -extern rtx gen_fnmav2df4 (rtx, rtx, rtx, rtx); -extern rtx gen_btruncv2sf2 (rtx, rtx); -extern rtx gen_ceilv2sf2 (rtx, rtx); -extern rtx gen_floorv2sf2 (rtx, rtx); -extern rtx gen_frintnv2sf2 (rtx, rtx); -extern rtx gen_nearbyintv2sf2 (rtx, rtx); -extern rtx gen_rintv2sf2 (rtx, rtx); -extern rtx gen_roundv2sf2 (rtx, rtx); -extern rtx gen_btruncv4sf2 (rtx, rtx); -extern rtx gen_ceilv4sf2 (rtx, rtx); -extern rtx gen_floorv4sf2 (rtx, rtx); -extern rtx gen_frintnv4sf2 (rtx, rtx); -extern rtx gen_nearbyintv4sf2 (rtx, rtx); -extern rtx gen_rintv4sf2 (rtx, rtx); -extern rtx gen_roundv4sf2 (rtx, rtx); -extern rtx gen_btruncv2df2 (rtx, rtx); -extern rtx gen_ceilv2df2 (rtx, rtx); -extern rtx gen_floorv2df2 (rtx, rtx); -extern rtx gen_frintnv2df2 (rtx, rtx); -extern rtx gen_nearbyintv2df2 (rtx, rtx); -extern rtx gen_rintv2df2 (rtx, rtx); -extern rtx gen_roundv2df2 (rtx, rtx); -extern rtx gen_lbtruncv2sfv2si2 (rtx, rtx); -extern rtx gen_lceilv2sfv2si2 (rtx, rtx); -extern rtx gen_lfloorv2sfv2si2 (rtx, rtx); -extern rtx gen_lroundv2sfv2si2 (rtx, rtx); -extern rtx gen_lfrintnv2sfv2si2 (rtx, rtx); -extern rtx gen_lbtruncuv2sfv2si2 (rtx, rtx); -extern rtx gen_lceiluv2sfv2si2 (rtx, rtx); -extern rtx gen_lflooruv2sfv2si2 (rtx, rtx); -extern rtx gen_lrounduv2sfv2si2 (rtx, rtx); -extern rtx gen_lfrintnuv2sfv2si2 (rtx, rtx); -extern rtx gen_lbtruncv4sfv4si2 (rtx, rtx); -extern rtx gen_lceilv4sfv4si2 (rtx, rtx); -extern rtx gen_lfloorv4sfv4si2 (rtx, rtx); -extern rtx gen_lroundv4sfv4si2 (rtx, rtx); -extern rtx gen_lfrintnv4sfv4si2 (rtx, rtx); -extern rtx gen_lbtruncuv4sfv4si2 (rtx, rtx); -extern rtx gen_lceiluv4sfv4si2 (rtx, rtx); -extern rtx gen_lflooruv4sfv4si2 (rtx, rtx); -extern rtx gen_lrounduv4sfv4si2 (rtx, rtx); -extern rtx gen_lfrintnuv4sfv4si2 (rtx, rtx); -extern rtx gen_lbtruncv2dfv2di2 (rtx, rtx); -extern rtx gen_lceilv2dfv2di2 (rtx, rtx); -extern rtx gen_lfloorv2dfv2di2 (rtx, rtx); -extern rtx gen_lroundv2dfv2di2 (rtx, rtx); -extern rtx gen_lfrintnv2dfv2di2 (rtx, rtx); -extern rtx gen_lbtruncuv2dfv2di2 (rtx, rtx); -extern rtx gen_lceiluv2dfv2di2 (rtx, rtx); -extern rtx gen_lflooruv2dfv2di2 (rtx, rtx); -extern rtx gen_lrounduv2dfv2di2 (rtx, rtx); -extern rtx gen_lfrintnuv2dfv2di2 (rtx, rtx); -extern rtx gen_floatv2siv2sf2 (rtx, rtx); -extern rtx gen_floatunsv2siv2sf2 (rtx, rtx); -extern rtx gen_floatv4siv4sf2 (rtx, rtx); -extern rtx gen_floatunsv4siv4sf2 (rtx, rtx); -extern rtx gen_floatv2div2df2 (rtx, rtx); -extern rtx gen_floatunsv2div2df2 (rtx, rtx); -extern rtx gen_vec_unpacks_lo_v4sf (rtx, rtx); -extern rtx gen_aarch64_float_extend_lo_v2df (rtx, rtx); -extern rtx gen_vec_unpacks_hi_v4sf (rtx, rtx); -extern rtx gen_aarch64_float_truncate_lo_v2sf (rtx, rtx); -extern rtx gen_aarch64_float_truncate_hi_v4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_vmlsv2sf (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vmlsv4sf (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vmlsv2df (rtx, rtx, rtx, rtx); -extern rtx gen_smaxv2sf3 (rtx, rtx, rtx); -extern rtx gen_sminv2sf3 (rtx, rtx, rtx); -extern rtx gen_smaxv4sf3 (rtx, rtx, rtx); -extern rtx gen_sminv4sf3 (rtx, rtx, rtx); -extern rtx gen_smaxv2df3 (rtx, rtx, rtx); -extern rtx gen_sminv2df3 (rtx, rtx, rtx); -extern rtx gen_smax_nanv2sf3 (rtx, rtx, rtx); -extern rtx gen_smin_nanv2sf3 (rtx, rtx, rtx); -extern rtx gen_smax_nanv4sf3 (rtx, rtx, rtx); -extern rtx gen_smin_nanv4sf3 (rtx, rtx, rtx); -extern rtx gen_smax_nanv2df3 (rtx, rtx, rtx); -extern rtx gen_smin_nanv2df3 (rtx, rtx, rtx); -extern rtx gen_reduc_splus_v8qi (rtx, rtx); -extern rtx gen_reduc_uplus_v8qi (rtx, rtx); -extern rtx gen_reduc_splus_v16qi (rtx, rtx); -extern rtx gen_reduc_uplus_v16qi (rtx, rtx); -extern rtx gen_reduc_splus_v4hi (rtx, rtx); -extern rtx gen_reduc_uplus_v4hi (rtx, rtx); -extern rtx gen_reduc_splus_v8hi (rtx, rtx); -extern rtx gen_reduc_uplus_v8hi (rtx, rtx); -extern rtx gen_reduc_splus_v4si (rtx, rtx); -extern rtx gen_reduc_uplus_v4si (rtx, rtx); -extern rtx gen_reduc_splus_v2di (rtx, rtx); -extern rtx gen_reduc_uplus_v2di (rtx, rtx); -extern rtx gen_reduc_splus_v2si (rtx, rtx); -extern rtx gen_reduc_uplus_v2si (rtx, rtx); -extern rtx gen_reduc_splus_v2sf (rtx, rtx); -extern rtx gen_reduc_splus_v2df (rtx, rtx); -extern rtx gen_aarch64_addpv4sf (rtx, rtx); -extern rtx gen_clzv8qi2 (rtx, rtx); -extern rtx gen_clzv16qi2 (rtx, rtx); -extern rtx gen_clzv4hi2 (rtx, rtx); -extern rtx gen_clzv8hi2 (rtx, rtx); -extern rtx gen_clzv2si2 (rtx, rtx); -extern rtx gen_clzv4si2 (rtx, rtx); -extern rtx gen_reduc_umax_v8qi (rtx, rtx); -extern rtx gen_reduc_umin_v8qi (rtx, rtx); -extern rtx gen_reduc_smax_v8qi (rtx, rtx); -extern rtx gen_reduc_smin_v8qi (rtx, rtx); -extern rtx gen_reduc_umax_v16qi (rtx, rtx); -extern rtx gen_reduc_umin_v16qi (rtx, rtx); -extern rtx gen_reduc_smax_v16qi (rtx, rtx); -extern rtx gen_reduc_smin_v16qi (rtx, rtx); -extern rtx gen_reduc_umax_v4hi (rtx, rtx); -extern rtx gen_reduc_umin_v4hi (rtx, rtx); -extern rtx gen_reduc_smax_v4hi (rtx, rtx); -extern rtx gen_reduc_smin_v4hi (rtx, rtx); -extern rtx gen_reduc_umax_v8hi (rtx, rtx); -extern rtx gen_reduc_umin_v8hi (rtx, rtx); -extern rtx gen_reduc_smax_v8hi (rtx, rtx); -extern rtx gen_reduc_smin_v8hi (rtx, rtx); -extern rtx gen_reduc_umax_v4si (rtx, rtx); -extern rtx gen_reduc_umin_v4si (rtx, rtx); -extern rtx gen_reduc_smax_v4si (rtx, rtx); -extern rtx gen_reduc_smin_v4si (rtx, rtx); -extern rtx gen_reduc_umax_v2si (rtx, rtx); -extern rtx gen_reduc_umin_v2si (rtx, rtx); -extern rtx gen_reduc_smax_v2si (rtx, rtx); -extern rtx gen_reduc_smin_v2si (rtx, rtx); -extern rtx gen_reduc_smax_nan_v2sf (rtx, rtx); -extern rtx gen_reduc_smin_nan_v2sf (rtx, rtx); -extern rtx gen_reduc_smax_v2sf (rtx, rtx); -extern rtx gen_reduc_smin_v2sf (rtx, rtx); -extern rtx gen_reduc_smax_nan_v2df (rtx, rtx); -extern rtx gen_reduc_smin_nan_v2df (rtx, rtx); -extern rtx gen_reduc_smax_v2df (rtx, rtx); -extern rtx gen_reduc_smin_v2df (rtx, rtx); -extern rtx gen_reduc_smax_nan_v4sf (rtx, rtx); -extern rtx gen_reduc_smin_nan_v4sf (rtx, rtx); -extern rtx gen_reduc_smax_v4sf (rtx, rtx); -extern rtx gen_reduc_smin_v4sf (rtx, rtx); -extern rtx gen_aarch64_simd_bslv8qi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv16qi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv4hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv8hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv2si_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv4si_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv2di_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv2sf_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv4sf_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv2df_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bsldi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bsldf_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_get_lanev8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_lanev16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_lanev4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_lanev8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_lanev2si (rtx, rtx, rtx); -extern rtx gen_aarch64_get_lanev4si (rtx, rtx, rtx); -extern rtx gen_aarch64_get_lanev2di (rtx, rtx, rtx); -extern rtx gen_aarch64_get_lanev2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_get_lanev4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_get_lanev2df (rtx, rtx, rtx); -extern rtx gen_aarch64_combinev8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_combinev4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_combinev2si (rtx, rtx, rtx); -extern rtx gen_aarch64_combinev2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_combinedi (rtx, rtx, rtx); -extern rtx gen_aarch64_combinedf (rtx, rtx, rtx); -extern rtx gen_aarch64_saddlv16qi_hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssublv16qi_hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_uaddlv16qi_hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usublv16qi_hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_saddlv8hi_hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssublv8hi_hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_uaddlv8hi_hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usublv8hi_hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_saddlv4si_hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssublv4si_hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_uaddlv4si_hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usublv4si_hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_saddlv16qi_lo_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssublv16qi_lo_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_uaddlv16qi_lo_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usublv16qi_lo_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_saddlv8hi_lo_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssublv8hi_lo_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_uaddlv8hi_lo_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usublv8hi_lo_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_saddlv4si_lo_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssublv4si_lo_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_uaddlv4si_lo_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usublv4si_lo_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_saddlv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_ssublv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uaddlv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_usublv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_saddlv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_ssublv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uaddlv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_usublv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_saddlv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_ssublv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_uaddlv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_usublv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_saddwv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_ssubwv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uaddwv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_usubwv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_saddwv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_ssubwv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uaddwv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_usubwv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_saddwv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_ssubwv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_uaddwv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_usubwv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_saddw2v16qi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssubw2v16qi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_uaddw2v16qi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usubw2v16qi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_saddw2v8hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssubw2v8hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_uaddw2v8hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usubw2v8hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_saddw2v4si_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssubw2v4si_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_uaddw2v4si_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usubw2v4si_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_shaddv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uhaddv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_srhaddv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_urhaddv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_shsubv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uhsubv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_srhsubv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_urhsubv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_shaddv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uhaddv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_srhaddv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_urhaddv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_shsubv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uhsubv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_srhsubv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_urhsubv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_shaddv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uhaddv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_srhaddv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_urhaddv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_shsubv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uhsubv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_srhsubv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_urhsubv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_shaddv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uhaddv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_srhaddv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_urhaddv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_shsubv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uhsubv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_srhsubv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_urhsubv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_shaddv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_uhaddv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_srhaddv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_urhaddv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_shsubv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_uhsubv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_srhsubv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_urhsubv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_shaddv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_uhaddv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_srhaddv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_urhaddv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_shsubv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_uhsubv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_srhsubv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_urhsubv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_addhnv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_raddhnv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_subhnv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_rsubhnv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_addhnv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_raddhnv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_subhnv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_rsubhnv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_addhnv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_raddhnv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_subhnv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_rsubhnv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_addhn2v8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_raddhn2v8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_subhn2v8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_rsubhn2v8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_addhn2v4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_raddhn2v4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_subhn2v4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_rsubhn2v4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_addhn2v2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_raddhn2v2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_subhn2v2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_rsubhn2v2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_pmulv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_pmulv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqaddv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqaddv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqsubv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqsubv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqaddv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqaddv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqsubv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqsubv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqaddv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqaddv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqsubv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqsubv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqaddv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqaddv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqsubv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqsubv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqaddv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_uqaddv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqsubv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_uqsubv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqaddv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_uqaddv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqsubv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_uqsubv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqaddv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_uqaddv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_sqsubv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_uqsubv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_sqaddqi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqaddqi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqsubqi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqsubqi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqaddhi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqaddhi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqsubhi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqsubhi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqaddsi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqaddsi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqsubsi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqsubsi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqadddi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqadddi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqsubdi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqsubdi (rtx, rtx, rtx); -extern rtx gen_aarch64_suqaddv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_usqaddv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_suqaddv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_usqaddv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_suqaddv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_usqaddv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_suqaddv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_usqaddv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_suqaddv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_usqaddv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_suqaddv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_usqaddv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_suqaddv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_usqaddv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_suqaddqi (rtx, rtx, rtx); -extern rtx gen_aarch64_usqaddqi (rtx, rtx, rtx); -extern rtx gen_aarch64_suqaddhi (rtx, rtx, rtx); -extern rtx gen_aarch64_usqaddhi (rtx, rtx, rtx); -extern rtx gen_aarch64_suqaddsi (rtx, rtx, rtx); -extern rtx gen_aarch64_usqaddsi (rtx, rtx, rtx); -extern rtx gen_aarch64_suqadddi (rtx, rtx, rtx); -extern rtx gen_aarch64_usqadddi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqmovunv8hi (rtx, rtx); -extern rtx gen_aarch64_sqmovunv4si (rtx, rtx); -extern rtx gen_aarch64_sqmovunv2di (rtx, rtx); -extern rtx gen_aarch64_sqmovunhi (rtx, rtx); -extern rtx gen_aarch64_sqmovunsi (rtx, rtx); -extern rtx gen_aarch64_sqmovundi (rtx, rtx); -extern rtx gen_aarch64_sqmovnv8hi (rtx, rtx); -extern rtx gen_aarch64_uqmovnv8hi (rtx, rtx); -extern rtx gen_aarch64_sqmovnv4si (rtx, rtx); -extern rtx gen_aarch64_uqmovnv4si (rtx, rtx); -extern rtx gen_aarch64_sqmovnv2di (rtx, rtx); -extern rtx gen_aarch64_uqmovnv2di (rtx, rtx); -extern rtx gen_aarch64_sqmovnhi (rtx, rtx); -extern rtx gen_aarch64_uqmovnhi (rtx, rtx); -extern rtx gen_aarch64_sqmovnsi (rtx, rtx); -extern rtx gen_aarch64_uqmovnsi (rtx, rtx); -extern rtx gen_aarch64_sqmovndi (rtx, rtx); -extern rtx gen_aarch64_uqmovndi (rtx, rtx); -extern rtx gen_aarch64_sqnegv8qi (rtx, rtx); -extern rtx gen_aarch64_sqabsv8qi (rtx, rtx); -extern rtx gen_aarch64_sqnegv16qi (rtx, rtx); -extern rtx gen_aarch64_sqabsv16qi (rtx, rtx); -extern rtx gen_aarch64_sqnegv4hi (rtx, rtx); -extern rtx gen_aarch64_sqabsv4hi (rtx, rtx); -extern rtx gen_aarch64_sqnegv8hi (rtx, rtx); -extern rtx gen_aarch64_sqabsv8hi (rtx, rtx); -extern rtx gen_aarch64_sqnegv2si (rtx, rtx); -extern rtx gen_aarch64_sqabsv2si (rtx, rtx); -extern rtx gen_aarch64_sqnegv4si (rtx, rtx); -extern rtx gen_aarch64_sqabsv4si (rtx, rtx); -extern rtx gen_aarch64_sqnegv2di (rtx, rtx); -extern rtx gen_aarch64_sqabsv2di (rtx, rtx); -extern rtx gen_aarch64_sqnegqi (rtx, rtx); -extern rtx gen_aarch64_sqabsqi (rtx, rtx); -extern rtx gen_aarch64_sqneghi (rtx, rtx); -extern rtx gen_aarch64_sqabshi (rtx, rtx); -extern rtx gen_aarch64_sqnegsi (rtx, rtx); -extern rtx gen_aarch64_sqabssi (rtx, rtx); -extern rtx gen_aarch64_sqnegdi (rtx, rtx); -extern rtx gen_aarch64_sqabsdi (rtx, rtx); -extern rtx gen_aarch64_sqdmulhv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulhv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulhv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulhv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulhv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulhv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulhv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulhv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulhhi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulhhi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulhsi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulhsi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulh_lanev4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulh_lanev4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulh_lanev8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulh_lanev8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulh_lanev2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulh_lanev2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulh_lanev4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulh_lanev4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulh_laneqv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulh_laneqv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulh_laneqv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulh_laneqv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulh_laneqv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulh_laneqv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulh_laneqv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulh_laneqv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulh_lanehi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulh_lanehi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmulh_lanesi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqrdmulh_lanesi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlalv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlslv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlalv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlslv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlalhi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlslhi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlalsi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlslsi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal_lanev4hi_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl_lanev4hi_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal_lanev2si_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl_lanev2si_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal_lanehi_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl_lanehi_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal_lanesi_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl_lanesi_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal_nv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl_nv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal_nv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl_nv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal2v8hi_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl2v8hi_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal2v4si_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl2v4si_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal2_lanev8hi_internal (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl2_lanev8hi_internal (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal2_lanev4si_internal (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl2_lanev4si_internal (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal2_nv8hi_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl2_nv8hi_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal2_nv4si_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl2_nv4si_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmullv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmullv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmullhi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmullsi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull_lanev4hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull_lanev2si_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull_lanehi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull_lanesi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull_nv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull_nv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull2v8hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull2v4si_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull2_lanev8hi_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull2_lanev4si_internal (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull2_nv8hi_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull2_nv4si_internal (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sshlv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_ushlv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_srshlv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_urshlv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sshlv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_ushlv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_srshlv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_urshlv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sshlv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_ushlv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_srshlv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_urshlv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sshlv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_ushlv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_srshlv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_urshlv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sshlv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_ushlv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_srshlv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_urshlv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_sshlv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_ushlv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_srshlv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_urshlv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sshlv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_ushlv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_srshlv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_urshlv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_sshldi (rtx, rtx, rtx); -extern rtx gen_aarch64_ushldi (rtx, rtx, rtx); -extern rtx gen_aarch64_srshldi (rtx, rtx, rtx); -extern rtx gen_aarch64_urshldi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshlv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshlv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshlv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshlv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshlv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshlv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshlv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshlv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshlv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshlv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshlv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshlv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshlv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshlv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshlv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshlv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshlv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshlv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshlv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshlv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshlv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlqi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshlqi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshlqi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshlqi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlhi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshlhi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshlhi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshlhi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlsi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshlsi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshlsi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshlsi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshldi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshldi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshldi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshldi (rtx, rtx, rtx); -extern rtx gen_aarch64_sshll_nv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_ushll_nv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sshll_nv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_ushll_nv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sshll_nv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_ushll_nv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_sshll2_nv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_ushll2_nv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sshll2_nv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_ushll2_nv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sshll2_nv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_ushll2_nv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_srshr_nv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_urshr_nv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_srshr_nv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_urshr_nv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_srshr_nv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_urshr_nv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_srshr_nv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_urshr_nv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_srshr_nv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_urshr_nv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_srshr_nv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_urshr_nv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_srshr_nv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_urshr_nv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_srshr_ndi (rtx, rtx, rtx); -extern rtx gen_aarch64_urshr_ndi (rtx, rtx, rtx); -extern rtx gen_aarch64_ssra_nv8qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usra_nv8qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_srsra_nv8qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ursra_nv8qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssra_nv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usra_nv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_srsra_nv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ursra_nv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssra_nv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usra_nv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_srsra_nv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ursra_nv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssra_nv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usra_nv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_srsra_nv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ursra_nv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssra_nv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usra_nv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_srsra_nv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ursra_nv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssra_nv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usra_nv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_srsra_nv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ursra_nv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssra_nv2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usra_nv2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_srsra_nv2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ursra_nv2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssra_ndi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usra_ndi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_srsra_ndi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ursra_ndi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssli_nv8qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usli_nv8qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssri_nv8qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usri_nv8qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssli_nv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usli_nv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssri_nv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usri_nv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssli_nv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usli_nv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssri_nv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usri_nv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssli_nv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usli_nv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssri_nv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usri_nv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssli_nv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usli_nv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssri_nv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usri_nv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssli_nv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usli_nv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssri_nv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usri_nv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssli_nv2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usli_nv2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssri_nv2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usri_nv2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssli_ndi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usli_ndi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_ssri_ndi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_usri_ndi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlu_nv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshl_nv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshl_nv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlu_nv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshl_nv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshl_nv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlu_nv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshl_nv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshl_nv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlu_nv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshl_nv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshl_nv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlu_nv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshl_nv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshl_nv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlu_nv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshl_nv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshl_nv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlu_nv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshl_nv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshl_nv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlu_nqi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshl_nqi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshl_nqi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlu_nhi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshl_nhi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshl_nhi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlu_nsi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshl_nsi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshl_nsi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshlu_ndi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshl_ndi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshl_ndi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshrun_nv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshrun_nv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshrn_nv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshrn_nv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshrn_nv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshrn_nv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshrun_nv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshrun_nv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshrn_nv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshrn_nv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshrn_nv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshrn_nv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshrun_nv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshrun_nv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshrn_nv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshrn_nv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshrn_nv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshrn_nv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshrun_nhi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshrun_nhi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshrn_nhi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshrn_nhi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshrn_nhi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshrn_nhi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshrun_nsi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshrun_nsi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshrn_nsi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshrn_nsi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshrn_nsi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshrn_nsi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshrun_ndi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshrun_ndi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqshrn_ndi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqshrn_ndi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqrshrn_ndi (rtx, rtx, rtx); -extern rtx gen_aarch64_uqrshrn_ndi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmlev8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmeqv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgev8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmlev16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmeqv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgev16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmlev4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmeqv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgev4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmlev8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmeqv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgev8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmlev2si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmeqv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgev2si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmlev4si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmeqv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgev4si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_cmlev2di (rtx, rtx, rtx); -extern rtx gen_aarch64_cmeqv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgev2di (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltdi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmledi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmeqdi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgedi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtdi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltuv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmleuv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgeuv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtuv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltuv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmleuv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgeuv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtuv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltuv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmleuv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgeuv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtuv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltuv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmleuv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgeuv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtuv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltuv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmleuv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgeuv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtuv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltuv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmleuv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgeuv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtuv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltuv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_cmleuv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgeuv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtuv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltudi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmleudi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgeudi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtudi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmtstv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmtstv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmtstv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmtstv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmtstv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmtstv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_cmtstv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_cmtstdi (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltv2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmlev2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmeqv2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgev2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtv2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltv4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmlev4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmeqv4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgev4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtv4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltv2df (rtx, rtx, rtx); -extern rtx gen_aarch64_cmlev2df (rtx, rtx, rtx); -extern rtx gen_aarch64_cmeqv2df (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgev2df (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtv2df (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltsf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmlesf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmeqsf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgesf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtsf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmltdf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmledf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmeqdf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgedf (rtx, rtx, rtx); -extern rtx gen_aarch64_cmgtdf (rtx, rtx, rtx); -extern rtx gen_aarch64_addpv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_addpv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_addpv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_addpdi (rtx, rtx); -extern rtx gen_sqrtv2sf2 (rtx, rtx); -extern rtx gen_sqrtv4sf2 (rtx, rtx); -extern rtx gen_sqrtv2df2 (rtx, rtx); -extern rtx gen_vec_load_lanesoiv16qi (rtx, rtx); -extern rtx gen_vec_load_lanesoiv8hi (rtx, rtx); -extern rtx gen_vec_load_lanesoiv4si (rtx, rtx); -extern rtx gen_vec_load_lanesoiv2di (rtx, rtx); -extern rtx gen_vec_load_lanesoiv4sf (rtx, rtx); -extern rtx gen_vec_load_lanesoiv2df (rtx, rtx); -extern rtx gen_vec_store_lanesoiv16qi (rtx, rtx); -extern rtx gen_vec_store_lanesoiv8hi (rtx, rtx); -extern rtx gen_vec_store_lanesoiv4si (rtx, rtx); -extern rtx gen_vec_store_lanesoiv2di (rtx, rtx); -extern rtx gen_vec_store_lanesoiv4sf (rtx, rtx); -extern rtx gen_vec_store_lanesoiv2df (rtx, rtx); -extern rtx gen_vec_store_lanesoi_lanev16qi (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesoi_lanev8hi (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesoi_lanev4si (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesoi_lanev2di (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesoi_lanev4sf (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesoi_lanev2df (rtx, rtx, rtx); -extern rtx gen_vec_load_lanesciv16qi (rtx, rtx); -extern rtx gen_vec_load_lanesciv8hi (rtx, rtx); -extern rtx gen_vec_load_lanesciv4si (rtx, rtx); -extern rtx gen_vec_load_lanesciv2di (rtx, rtx); -extern rtx gen_vec_load_lanesciv4sf (rtx, rtx); -extern rtx gen_vec_load_lanesciv2df (rtx, rtx); -extern rtx gen_vec_store_lanesciv16qi (rtx, rtx); -extern rtx gen_vec_store_lanesciv8hi (rtx, rtx); -extern rtx gen_vec_store_lanesciv4si (rtx, rtx); -extern rtx gen_vec_store_lanesciv2di (rtx, rtx); -extern rtx gen_vec_store_lanesciv4sf (rtx, rtx); -extern rtx gen_vec_store_lanesciv2df (rtx, rtx); -extern rtx gen_vec_store_lanesci_lanev16qi (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesci_lanev8hi (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesci_lanev4si (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesci_lanev2di (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesci_lanev4sf (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesci_lanev2df (rtx, rtx, rtx); -extern rtx gen_vec_load_lanesxiv16qi (rtx, rtx); -extern rtx gen_vec_load_lanesxiv8hi (rtx, rtx); -extern rtx gen_vec_load_lanesxiv4si (rtx, rtx); -extern rtx gen_vec_load_lanesxiv2di (rtx, rtx); -extern rtx gen_vec_load_lanesxiv4sf (rtx, rtx); -extern rtx gen_vec_load_lanesxiv2df (rtx, rtx); -extern rtx gen_vec_store_lanesxiv16qi (rtx, rtx); -extern rtx gen_vec_store_lanesxiv8hi (rtx, rtx); -extern rtx gen_vec_store_lanesxiv4si (rtx, rtx); -extern rtx gen_vec_store_lanesxiv2di (rtx, rtx); -extern rtx gen_vec_store_lanesxiv4sf (rtx, rtx); -extern rtx gen_vec_store_lanesxiv2df (rtx, rtx); -extern rtx gen_vec_store_lanesxi_lanev16qi (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesxi_lanev8hi (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesxi_lanev4si (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesxi_lanev2di (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesxi_lanev4sf (rtx, rtx, rtx); -extern rtx gen_vec_store_lanesxi_lanev2df (rtx, rtx, rtx); -extern rtx gen_aarch64_be_ld1v8qi (rtx, rtx); -extern rtx gen_aarch64_be_ld1v16qi (rtx, rtx); -extern rtx gen_aarch64_be_ld1v4hi (rtx, rtx); -extern rtx gen_aarch64_be_ld1v8hi (rtx, rtx); -extern rtx gen_aarch64_be_ld1v2si (rtx, rtx); -extern rtx gen_aarch64_be_ld1v4si (rtx, rtx); -extern rtx gen_aarch64_be_ld1v2di (rtx, rtx); -extern rtx gen_aarch64_be_ld1v2sf (rtx, rtx); -extern rtx gen_aarch64_be_ld1v4sf (rtx, rtx); -extern rtx gen_aarch64_be_ld1v2df (rtx, rtx); -extern rtx gen_aarch64_be_ld1di (rtx, rtx); -extern rtx gen_aarch64_be_st1v8qi (rtx, rtx); -extern rtx gen_aarch64_be_st1v16qi (rtx, rtx); -extern rtx gen_aarch64_be_st1v4hi (rtx, rtx); -extern rtx gen_aarch64_be_st1v8hi (rtx, rtx); -extern rtx gen_aarch64_be_st1v2si (rtx, rtx); -extern rtx gen_aarch64_be_st1v4si (rtx, rtx); -extern rtx gen_aarch64_be_st1v2di (rtx, rtx); -extern rtx gen_aarch64_be_st1v2sf (rtx, rtx); -extern rtx gen_aarch64_be_st1v4sf (rtx, rtx); -extern rtx gen_aarch64_be_st1v2df (rtx, rtx); -extern rtx gen_aarch64_be_st1di (rtx, rtx); -extern rtx gen_aarch64_ld2v8qi_dreg (rtx, rtx); -extern rtx gen_aarch64_ld2v4hi_dreg (rtx, rtx); -extern rtx gen_aarch64_ld2v2si_dreg (rtx, rtx); -extern rtx gen_aarch64_ld2v2sf_dreg (rtx, rtx); -extern rtx gen_aarch64_ld2di_dreg (rtx, rtx); -extern rtx gen_aarch64_ld2df_dreg (rtx, rtx); -extern rtx gen_aarch64_ld3v8qi_dreg (rtx, rtx); -extern rtx gen_aarch64_ld3v4hi_dreg (rtx, rtx); -extern rtx gen_aarch64_ld3v2si_dreg (rtx, rtx); -extern rtx gen_aarch64_ld3v2sf_dreg (rtx, rtx); -extern rtx gen_aarch64_ld3di_dreg (rtx, rtx); -extern rtx gen_aarch64_ld3df_dreg (rtx, rtx); -extern rtx gen_aarch64_ld4v8qi_dreg (rtx, rtx); -extern rtx gen_aarch64_ld4v4hi_dreg (rtx, rtx); -extern rtx gen_aarch64_ld4v2si_dreg (rtx, rtx); -extern rtx gen_aarch64_ld4v2sf_dreg (rtx, rtx); -extern rtx gen_aarch64_ld4di_dreg (rtx, rtx); -extern rtx gen_aarch64_ld4df_dreg (rtx, rtx); -extern rtx gen_aarch64_tbl1v8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_tbl1v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_tbl2v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_combinev16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_zip1v8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_zip2v8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_trn1v8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_trn2v8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp1v8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp2v8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_zip1v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_zip2v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_trn1v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_trn2v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp1v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp2v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_zip1v4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_zip2v4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_trn1v4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_trn2v4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp1v4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp2v4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_zip1v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_zip2v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_trn1v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_trn2v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp1v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp2v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_zip1v2si (rtx, rtx, rtx); -extern rtx gen_aarch64_zip2v2si (rtx, rtx, rtx); -extern rtx gen_aarch64_trn1v2si (rtx, rtx, rtx); -extern rtx gen_aarch64_trn2v2si (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp1v2si (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp2v2si (rtx, rtx, rtx); -extern rtx gen_aarch64_zip1v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_zip2v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_trn1v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_trn2v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp1v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp2v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_zip1v2di (rtx, rtx, rtx); -extern rtx gen_aarch64_zip2v2di (rtx, rtx, rtx); -extern rtx gen_aarch64_trn1v2di (rtx, rtx, rtx); -extern rtx gen_aarch64_trn2v2di (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp1v2di (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp2v2di (rtx, rtx, rtx); -extern rtx gen_aarch64_zip1v2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_zip2v2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_trn1v2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_trn2v2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp1v2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp2v2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_zip1v4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_zip2v4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_trn1v4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_trn2v4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp1v4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp2v4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_zip1v2df (rtx, rtx, rtx); -extern rtx gen_aarch64_zip2v2df (rtx, rtx, rtx); -extern rtx gen_aarch64_trn1v2df (rtx, rtx, rtx); -extern rtx gen_aarch64_trn2v2df (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp1v2df (rtx, rtx, rtx); -extern rtx gen_aarch64_uzp2v2df (rtx, rtx, rtx); -extern rtx gen_aarch64_st2v8qi_dreg (rtx, rtx); -extern rtx gen_aarch64_st2v4hi_dreg (rtx, rtx); -extern rtx gen_aarch64_st2v2si_dreg (rtx, rtx); -extern rtx gen_aarch64_st2v2sf_dreg (rtx, rtx); -extern rtx gen_aarch64_st2di_dreg (rtx, rtx); -extern rtx gen_aarch64_st2df_dreg (rtx, rtx); -extern rtx gen_aarch64_st3v8qi_dreg (rtx, rtx); -extern rtx gen_aarch64_st3v4hi_dreg (rtx, rtx); -extern rtx gen_aarch64_st3v2si_dreg (rtx, rtx); -extern rtx gen_aarch64_st3v2sf_dreg (rtx, rtx); -extern rtx gen_aarch64_st3di_dreg (rtx, rtx); -extern rtx gen_aarch64_st3df_dreg (rtx, rtx); -extern rtx gen_aarch64_st4v8qi_dreg (rtx, rtx); -extern rtx gen_aarch64_st4v4hi_dreg (rtx, rtx); -extern rtx gen_aarch64_st4v2si_dreg (rtx, rtx); -extern rtx gen_aarch64_st4v2sf_dreg (rtx, rtx); -extern rtx gen_aarch64_st4di_dreg (rtx, rtx); -extern rtx gen_aarch64_st4df_dreg (rtx, rtx); -extern rtx gen_aarch64_frecpev2sf (rtx, rtx); -extern rtx gen_aarch64_frecpev4sf (rtx, rtx); -extern rtx gen_aarch64_frecpev2df (rtx, rtx); -extern rtx gen_aarch64_frecpesf (rtx, rtx); -extern rtx gen_aarch64_frecpxsf (rtx, rtx); -extern rtx gen_aarch64_frecpedf (rtx, rtx); -extern rtx gen_aarch64_frecpxdf (rtx, rtx); -extern rtx gen_aarch64_frecpsv2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_frecpsv4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_frecpsv2df (rtx, rtx, rtx); -extern rtx gen_aarch64_frecpssf (rtx, rtx, rtx); -extern rtx gen_aarch64_frecpsdf (rtx, rtx, rtx); -extern rtx gen_aarch64_crypto_aesev16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_crypto_aesdv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_crypto_aesmcv16qi (rtx, rtx); -extern rtx gen_aarch64_crypto_aesimcv16qi (rtx, rtx); -extern rtx gen_aarch64_crypto_sha1hsi (rtx, rtx); -extern rtx gen_aarch64_crypto_sha1su1v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_crypto_sha1cv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_crypto_sha1mv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_crypto_sha1pv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_crypto_sha1su0v4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_crypto_sha256hv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_crypto_sha256h2v4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_crypto_sha256su0v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_crypto_sha256su1v4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_crypto_pmulldi (rtx, rtx, rtx); -extern rtx gen_aarch64_crypto_pmullv2di (rtx, rtx, rtx); -extern rtx gen_atomic_compare_and_swapqi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_atomic_compare_and_swaphi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_atomic_compare_and_swapsi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_atomic_compare_and_swapdi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_atomic_exchangeqi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_exchangehi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_exchangesi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_exchangedi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_addqi (rtx, rtx, rtx); -extern rtx gen_atomic_subqi (rtx, rtx, rtx); -extern rtx gen_atomic_orqi (rtx, rtx, rtx); -extern rtx gen_atomic_xorqi (rtx, rtx, rtx); -extern rtx gen_atomic_andqi (rtx, rtx, rtx); -extern rtx gen_atomic_addhi (rtx, rtx, rtx); -extern rtx gen_atomic_subhi (rtx, rtx, rtx); -extern rtx gen_atomic_orhi (rtx, rtx, rtx); -extern rtx gen_atomic_xorhi (rtx, rtx, rtx); -extern rtx gen_atomic_andhi (rtx, rtx, rtx); -extern rtx gen_atomic_addsi (rtx, rtx, rtx); -extern rtx gen_atomic_subsi (rtx, rtx, rtx); -extern rtx gen_atomic_orsi (rtx, rtx, rtx); -extern rtx gen_atomic_xorsi (rtx, rtx, rtx); -extern rtx gen_atomic_andsi (rtx, rtx, rtx); -extern rtx gen_atomic_adddi (rtx, rtx, rtx); -extern rtx gen_atomic_subdi (rtx, rtx, rtx); -extern rtx gen_atomic_ordi (rtx, rtx, rtx); -extern rtx gen_atomic_xordi (rtx, rtx, rtx); -extern rtx gen_atomic_anddi (rtx, rtx, rtx); -extern rtx gen_atomic_nandqi (rtx, rtx, rtx); -extern rtx gen_atomic_nandhi (rtx, rtx, rtx); -extern rtx gen_atomic_nandsi (rtx, rtx, rtx); -extern rtx gen_atomic_nanddi (rtx, rtx, rtx); -extern rtx gen_atomic_fetch_addqi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_subqi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_orqi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_xorqi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_andqi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_addhi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_subhi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_orhi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_xorhi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_andhi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_addsi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_subsi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_orsi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_xorsi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_andsi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_adddi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_subdi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_ordi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_xordi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_anddi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_nandqi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_nandhi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_nandsi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_fetch_nanddi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_add_fetchqi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_sub_fetchqi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_or_fetchqi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_xor_fetchqi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_and_fetchqi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_add_fetchhi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_sub_fetchhi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_or_fetchhi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_xor_fetchhi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_and_fetchhi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_add_fetchsi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_sub_fetchsi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_or_fetchsi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_xor_fetchsi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_and_fetchsi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_add_fetchdi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_sub_fetchdi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_or_fetchdi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_xor_fetchdi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_and_fetchdi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_nand_fetchqi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_nand_fetchhi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_nand_fetchsi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_nand_fetchdi (rtx, rtx, rtx, rtx); -extern rtx gen_atomic_loadqi (rtx, rtx, rtx); -extern rtx gen_atomic_loadhi (rtx, rtx, rtx); -extern rtx gen_atomic_loadsi (rtx, rtx, rtx); -extern rtx gen_atomic_loaddi (rtx, rtx, rtx); -extern rtx gen_atomic_storeqi (rtx, rtx, rtx); -extern rtx gen_atomic_storehi (rtx, rtx, rtx); -extern rtx gen_atomic_storesi (rtx, rtx, rtx); -extern rtx gen_atomic_storedi (rtx, rtx, rtx); -extern rtx gen_aarch64_load_exclusiveqi (rtx, rtx, rtx); -extern rtx gen_aarch64_load_exclusivehi (rtx, rtx, rtx); -extern rtx gen_aarch64_load_exclusivesi (rtx, rtx, rtx); -extern rtx gen_aarch64_load_exclusivedi (rtx, rtx, rtx); -extern rtx gen_aarch64_store_exclusiveqi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_store_exclusivehi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_store_exclusivesi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_store_exclusivedi (rtx, rtx, rtx, rtx); -extern rtx gen_cbranchsi4 (rtx, rtx, rtx, rtx); -extern rtx gen_cbranchdi4 (rtx, rtx, rtx, rtx); -extern rtx gen_cbranchsf4 (rtx, rtx, rtx, rtx); -extern rtx gen_cbranchdf4 (rtx, rtx, rtx, rtx); -extern rtx gen_casesi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_prologue (void); -extern rtx gen_epilogue (void); -extern rtx gen_sibcall_epilogue (void); +extern rtx gen_indirect_jump (rtx); +extern rtx gen_jump (rtx); +extern rtx gen_casesi_dispatch (rtx, rtx, rtx); +extern rtx gen_nop (void); +extern rtx gen_trap (void); +extern rtx gen_eh_return (rtx); +extern rtx gen_insv_immsi (rtx, rtx, rtx); +extern rtx gen_insv_immdi (rtx, rtx, rtx); +extern rtx gen_load_pairsi (rtx, rtx, rtx, rtx); +extern rtx gen_load_pairdi (rtx, rtx, rtx, rtx); +extern rtx gen_store_pairsi (rtx, rtx, rtx, rtx); +extern rtx gen_store_pairdi (rtx, rtx, rtx, rtx); +extern rtx gen_load_pairsf (rtx, rtx, rtx, rtx); +extern rtx gen_load_pairdf (rtx, rtx, rtx, rtx); +extern rtx gen_store_pairsf (rtx, rtx, rtx, rtx); +extern rtx gen_store_pairdf (rtx, rtx, rtx, rtx); +extern rtx gen_loadwb_pairsi_si (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_loadwb_pairsi_di (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_loadwb_pairdi_si (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_loadwb_pairdi_di (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_storewb_pairsi_si (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_storewb_pairsi_di (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_storewb_pairdi_si (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_storewb_pairdi_di (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_addsi3_compare0 (rtx, rtx, rtx); +extern rtx gen_adddi3_compare0 (rtx, rtx, rtx); +extern rtx gen_addsi3_carryin (rtx, rtx, rtx); +extern rtx gen_adddi3_carryin (rtx, rtx, rtx); +extern rtx gen_subsi3 (rtx, rtx, rtx); +extern rtx gen_subdi3 (rtx, rtx, rtx); +extern rtx gen_subsi3_compare0 (rtx, rtx, rtx); +extern rtx gen_subdi3_compare0 (rtx, rtx, rtx); +extern rtx gen_subsi3_carryin (rtx, rtx, rtx); +extern rtx gen_subdi3_carryin (rtx, rtx, rtx); +extern rtx gen_absdi2 (rtx, rtx); +extern rtx gen_negsi2 (rtx, rtx); +extern rtx gen_negdi2 (rtx, rtx); +extern rtx gen_mulsi3 (rtx, rtx, rtx); +extern rtx gen_muldi3 (rtx, rtx, rtx); +extern rtx gen_maddsi (rtx, rtx, rtx, rtx); +extern rtx gen_madddi (rtx, rtx, rtx, rtx); +extern rtx gen_mulsidi3 (rtx, rtx, rtx); +extern rtx gen_umulsidi3 (rtx, rtx, rtx); +extern rtx gen_maddsidi4 (rtx, rtx, rtx, rtx); +extern rtx gen_umaddsidi4 (rtx, rtx, rtx, rtx); +extern rtx gen_msubsidi4 (rtx, rtx, rtx, rtx); +extern rtx gen_umsubsidi4 (rtx, rtx, rtx, rtx); +extern rtx gen_smuldi3_highpart (rtx, rtx, rtx); +extern rtx gen_umuldi3_highpart (rtx, rtx, rtx); +extern rtx gen_divsi3 (rtx, rtx, rtx); +extern rtx gen_udivsi3 (rtx, rtx, rtx); +extern rtx gen_divdi3 (rtx, rtx, rtx); +extern rtx gen_udivdi3 (rtx, rtx, rtx); +extern rtx gen_cstoreqi_neg (rtx, rtx, rtx); +extern rtx gen_cstorehi_neg (rtx, rtx, rtx); +extern rtx gen_cstoresi_neg (rtx, rtx, rtx); +extern rtx gen_cstoredi_neg (rtx, rtx, rtx); +extern rtx gen_aarch64_crc32b (rtx, rtx, rtx); +extern rtx gen_aarch64_crc32h (rtx, rtx, rtx); +extern rtx gen_aarch64_crc32w (rtx, rtx, rtx); +extern rtx gen_aarch64_crc32x (rtx, rtx, rtx); +extern rtx gen_aarch64_crc32cb (rtx, rtx, rtx); +extern rtx gen_aarch64_crc32ch (rtx, rtx, rtx); +extern rtx gen_aarch64_crc32cw (rtx, rtx, rtx); +extern rtx gen_aarch64_crc32cx (rtx, rtx, rtx); +extern rtx gen_csinc3si_insn (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_csinc3di_insn (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_andsi3 (rtx, rtx, rtx); +extern rtx gen_iorsi3 (rtx, rtx, rtx); +extern rtx gen_xorsi3 (rtx, rtx, rtx); +extern rtx gen_anddi3 (rtx, rtx, rtx); +extern rtx gen_iordi3 (rtx, rtx, rtx); +extern rtx gen_xordi3 (rtx, rtx, rtx); +extern rtx gen_one_cmplsi2 (rtx, rtx); +extern rtx gen_one_cmpldi2 (rtx, rtx); +extern rtx gen_clzsi2 (rtx, rtx); +extern rtx gen_clzdi2 (rtx, rtx); +extern rtx gen_clrsbsi2 (rtx, rtx); +extern rtx gen_clrsbdi2 (rtx, rtx); +extern rtx gen_rbitsi2 (rtx, rtx); +extern rtx gen_rbitdi2 (rtx, rtx); +extern rtx gen_bswapsi2 (rtx, rtx); +extern rtx gen_bswapdi2 (rtx, rtx); +extern rtx gen_bswaphi2 (rtx, rtx); +extern rtx gen_rev16si2 (rtx, rtx, rtx, rtx); +extern rtx gen_rev16di2 (rtx, rtx, rtx, rtx); +extern rtx gen_rev16si2_alt (rtx, rtx, rtx, rtx); +extern rtx gen_rev16di2_alt (rtx, rtx, rtx, rtx); +extern rtx gen_btruncsf2 (rtx, rtx); +extern rtx gen_ceilsf2 (rtx, rtx); +extern rtx gen_floorsf2 (rtx, rtx); +extern rtx gen_frintnsf2 (rtx, rtx); +extern rtx gen_nearbyintsf2 (rtx, rtx); +extern rtx gen_rintsf2 (rtx, rtx); +extern rtx gen_roundsf2 (rtx, rtx); +extern rtx gen_btruncdf2 (rtx, rtx); +extern rtx gen_ceildf2 (rtx, rtx); +extern rtx gen_floordf2 (rtx, rtx); +extern rtx gen_frintndf2 (rtx, rtx); +extern rtx gen_nearbyintdf2 (rtx, rtx); +extern rtx gen_rintdf2 (rtx, rtx); +extern rtx gen_rounddf2 (rtx, rtx); +extern rtx gen_lbtruncsfsi2 (rtx, rtx); +extern rtx gen_lceilsfsi2 (rtx, rtx); +extern rtx gen_lfloorsfsi2 (rtx, rtx); +extern rtx gen_lroundsfsi2 (rtx, rtx); +extern rtx gen_lfrintnsfsi2 (rtx, rtx); +extern rtx gen_lbtruncusfsi2 (rtx, rtx); +extern rtx gen_lceilusfsi2 (rtx, rtx); +extern rtx gen_lfloorusfsi2 (rtx, rtx); +extern rtx gen_lroundusfsi2 (rtx, rtx); +extern rtx gen_lfrintnusfsi2 (rtx, rtx); +extern rtx gen_lbtruncdfsi2 (rtx, rtx); +extern rtx gen_lceildfsi2 (rtx, rtx); +extern rtx gen_lfloordfsi2 (rtx, rtx); +extern rtx gen_lrounddfsi2 (rtx, rtx); +extern rtx gen_lfrintndfsi2 (rtx, rtx); +extern rtx gen_lbtruncudfsi2 (rtx, rtx); +extern rtx gen_lceiludfsi2 (rtx, rtx); +extern rtx gen_lfloorudfsi2 (rtx, rtx); +extern rtx gen_lroundudfsi2 (rtx, rtx); +extern rtx gen_lfrintnudfsi2 (rtx, rtx); +extern rtx gen_lbtruncsfdi2 (rtx, rtx); +extern rtx gen_lceilsfdi2 (rtx, rtx); +extern rtx gen_lfloorsfdi2 (rtx, rtx); +extern rtx gen_lroundsfdi2 (rtx, rtx); +extern rtx gen_lfrintnsfdi2 (rtx, rtx); +extern rtx gen_lbtruncusfdi2 (rtx, rtx); +extern rtx gen_lceilusfdi2 (rtx, rtx); +extern rtx gen_lfloorusfdi2 (rtx, rtx); +extern rtx gen_lroundusfdi2 (rtx, rtx); +extern rtx gen_lfrintnusfdi2 (rtx, rtx); +extern rtx gen_lbtruncdfdi2 (rtx, rtx); +extern rtx gen_lceildfdi2 (rtx, rtx); +extern rtx gen_lfloordfdi2 (rtx, rtx); +extern rtx gen_lrounddfdi2 (rtx, rtx); +extern rtx gen_lfrintndfdi2 (rtx, rtx); +extern rtx gen_lbtruncudfdi2 (rtx, rtx); +extern rtx gen_lceiludfdi2 (rtx, rtx); +extern rtx gen_lfloorudfdi2 (rtx, rtx); +extern rtx gen_lroundudfdi2 (rtx, rtx); +extern rtx gen_lfrintnudfdi2 (rtx, rtx); +extern rtx gen_fmasf4 (rtx, rtx, rtx, rtx); +extern rtx gen_fmadf4 (rtx, rtx, rtx, rtx); +extern rtx gen_fnmasf4 (rtx, rtx, rtx, rtx); +extern rtx gen_fnmadf4 (rtx, rtx, rtx, rtx); +extern rtx gen_fmssf4 (rtx, rtx, rtx, rtx); +extern rtx gen_fmsdf4 (rtx, rtx, rtx, rtx); +extern rtx gen_fnmssf4 (rtx, rtx, rtx, rtx); +extern rtx gen_fnmsdf4 (rtx, rtx, rtx, rtx); +extern rtx gen_extendsfdf2 (rtx, rtx); +extern rtx gen_truncdfsf2 (rtx, rtx); +extern rtx gen_fix_truncsfsi2 (rtx, rtx); +extern rtx gen_fix_truncdfsi2 (rtx, rtx); +extern rtx gen_fix_truncsfdi2 (rtx, rtx); +extern rtx gen_fix_truncdfdi2 (rtx, rtx); +extern rtx gen_fixuns_truncsfsi2 (rtx, rtx); +extern rtx gen_fixuns_truncdfsi2 (rtx, rtx); +extern rtx gen_fixuns_truncsfdi2 (rtx, rtx); +extern rtx gen_fixuns_truncdfdi2 (rtx, rtx); +extern rtx gen_floatsisf2 (rtx, rtx); +extern rtx gen_floatsidf2 (rtx, rtx); +extern rtx gen_floatdisf2 (rtx, rtx); +extern rtx gen_floatdidf2 (rtx, rtx); +extern rtx gen_floatunssisf2 (rtx, rtx); +extern rtx gen_floatunssidf2 (rtx, rtx); +extern rtx gen_floatunsdisf2 (rtx, rtx); +extern rtx gen_floatunsdidf2 (rtx, rtx); +extern rtx gen_addsf3 (rtx, rtx, rtx); +extern rtx gen_adddf3 (rtx, rtx, rtx); +extern rtx gen_subsf3 (rtx, rtx, rtx); +extern rtx gen_subdf3 (rtx, rtx, rtx); +extern rtx gen_mulsf3 (rtx, rtx, rtx); +extern rtx gen_muldf3 (rtx, rtx, rtx); +extern rtx gen_divsf3 (rtx, rtx, rtx); +extern rtx gen_divdf3 (rtx, rtx, rtx); +extern rtx gen_negsf2 (rtx, rtx); +extern rtx gen_negdf2 (rtx, rtx); +extern rtx gen_sqrtsf2 (rtx, rtx); +extern rtx gen_sqrtdf2 (rtx, rtx); +extern rtx gen_abssf2 (rtx, rtx); +extern rtx gen_absdf2 (rtx, rtx); +extern rtx gen_smaxsf3 (rtx, rtx, rtx); +extern rtx gen_smaxdf3 (rtx, rtx, rtx); +extern rtx gen_sminsf3 (rtx, rtx, rtx); +extern rtx gen_smindf3 (rtx, rtx, rtx); +extern rtx gen_aarch64_movdi_tilow (rtx, rtx); +extern rtx gen_aarch64_movdi_tflow (rtx, rtx); +extern rtx gen_aarch64_movdi_tihigh (rtx, rtx); +extern rtx gen_aarch64_movdi_tfhigh (rtx, rtx); +extern rtx gen_aarch64_movtihigh_di (rtx, rtx); +extern rtx gen_aarch64_movtfhigh_di (rtx, rtx); +extern rtx gen_aarch64_movtilow_di (rtx, rtx); +extern rtx gen_aarch64_movtflow_di (rtx, rtx); +extern rtx gen_aarch64_movtilow_tilow (rtx, rtx); +extern rtx gen_add_losym_si (rtx, rtx, rtx); +extern rtx gen_add_losym_di (rtx, rtx, rtx); +extern rtx gen_ldr_got_small_si (rtx, rtx, rtx); +extern rtx gen_ldr_got_small_di (rtx, rtx, rtx); +extern rtx gen_ldr_got_small_sidi (rtx, rtx, rtx); +extern rtx gen_ldr_got_tiny (rtx, rtx); +extern rtx gen_aarch64_load_tp_hard (rtx); +extern rtx gen_tlsie_small_si (rtx, rtx); +extern rtx gen_tlsie_small_di (rtx, rtx); +extern rtx gen_tlsie_small_sidi (rtx, rtx); +extern rtx gen_tlsle_small_si (rtx, rtx, rtx); +extern rtx gen_tlsle_small_di (rtx, rtx, rtx); +extern rtx gen_tlsdesc_small_si (rtx); +extern rtx gen_tlsdesc_small_di (rtx); +extern rtx gen_stack_tie (rtx, rtx); +extern rtx gen_stack_protect_set_si (rtx, rtx); +extern rtx gen_stack_protect_set_di (rtx, rtx); +extern rtx gen_stack_protect_test_si (rtx, rtx, rtx); +extern rtx gen_stack_protect_test_di (rtx, rtx, rtx); +extern rtx gen_set_fpcr (rtx); +extern rtx gen_get_fpcr (rtx); +extern rtx gen_set_fpsr (rtx); +extern rtx gen_get_fpsr (rtx); +extern rtx gen_aarch64_simd_dupv8qi (rtx, rtx); +extern rtx gen_aarch64_simd_dupv16qi (rtx, rtx); +extern rtx gen_aarch64_simd_dupv4hi (rtx, rtx); +extern rtx gen_aarch64_simd_dupv8hi (rtx, rtx); +extern rtx gen_aarch64_simd_dupv2si (rtx, rtx); +extern rtx gen_aarch64_simd_dupv4si (rtx, rtx); +extern rtx gen_aarch64_simd_dupv2di (rtx, rtx); +extern rtx gen_aarch64_simd_dupv2sf (rtx, rtx); +extern rtx gen_aarch64_simd_dupv4sf (rtx, rtx); +extern rtx gen_aarch64_simd_dupv2df (rtx, rtx); +extern rtx gen_aarch64_dup_lanev8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lanev16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lanev4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lanev8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lanev2si (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lanev4si (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lanev2di (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lanev2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lanev4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lanev2df (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lane_to_128v8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lane_to_64v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lane_to_128v4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lane_to_64v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lane_to_128v2si (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lane_to_64v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lane_to_64v2di (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lane_to_128v2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lane_to_64v4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_dup_lane_to_64v2df (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_mov_from_v16qilow (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_mov_from_v8hilow (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_mov_from_v4silow (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_mov_from_v2dilow (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_mov_from_v4sflow (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_mov_from_v2dflow (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_mov_from_v16qihigh (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_mov_from_v8hihigh (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_mov_from_v4sihigh (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_mov_from_v2dihigh (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_mov_from_v4sfhigh (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_mov_from_v2dfhigh (rtx, rtx, rtx); +extern rtx gen_ornv8qi3 (rtx, rtx, rtx); +extern rtx gen_ornv16qi3 (rtx, rtx, rtx); +extern rtx gen_ornv4hi3 (rtx, rtx, rtx); +extern rtx gen_ornv8hi3 (rtx, rtx, rtx); +extern rtx gen_ornv2si3 (rtx, rtx, rtx); +extern rtx gen_ornv4si3 (rtx, rtx, rtx); +extern rtx gen_ornv2di3 (rtx, rtx, rtx); +extern rtx gen_bicv8qi3 (rtx, rtx, rtx); +extern rtx gen_bicv16qi3 (rtx, rtx, rtx); +extern rtx gen_bicv4hi3 (rtx, rtx, rtx); +extern rtx gen_bicv8hi3 (rtx, rtx, rtx); +extern rtx gen_bicv2si3 (rtx, rtx, rtx); +extern rtx gen_bicv4si3 (rtx, rtx, rtx); +extern rtx gen_bicv2di3 (rtx, rtx, rtx); +extern rtx gen_addv8qi3 (rtx, rtx, rtx); +extern rtx gen_addv16qi3 (rtx, rtx, rtx); +extern rtx gen_addv4hi3 (rtx, rtx, rtx); +extern rtx gen_addv8hi3 (rtx, rtx, rtx); +extern rtx gen_addv2si3 (rtx, rtx, rtx); +extern rtx gen_addv4si3 (rtx, rtx, rtx); +extern rtx gen_addv2di3 (rtx, rtx, rtx); +extern rtx gen_subv8qi3 (rtx, rtx, rtx); +extern rtx gen_subv16qi3 (rtx, rtx, rtx); +extern rtx gen_subv4hi3 (rtx, rtx, rtx); +extern rtx gen_subv8hi3 (rtx, rtx, rtx); +extern rtx gen_subv2si3 (rtx, rtx, rtx); +extern rtx gen_subv4si3 (rtx, rtx, rtx); +extern rtx gen_subv2di3 (rtx, rtx, rtx); +extern rtx gen_mulv8qi3 (rtx, rtx, rtx); +extern rtx gen_mulv16qi3 (rtx, rtx, rtx); +extern rtx gen_mulv4hi3 (rtx, rtx, rtx); +extern rtx gen_mulv8hi3 (rtx, rtx, rtx); +extern rtx gen_mulv2si3 (rtx, rtx, rtx); +extern rtx gen_mulv4si3 (rtx, rtx, rtx); +extern rtx gen_bswapv4hi (rtx, rtx); +extern rtx gen_bswapv8hi (rtx, rtx); +extern rtx gen_bswapv2si (rtx, rtx); +extern rtx gen_bswapv4si (rtx, rtx); +extern rtx gen_bswapv2di (rtx, rtx); +extern rtx gen_negv8qi2 (rtx, rtx); +extern rtx gen_negv16qi2 (rtx, rtx); +extern rtx gen_negv4hi2 (rtx, rtx); +extern rtx gen_negv8hi2 (rtx, rtx); +extern rtx gen_negv2si2 (rtx, rtx); +extern rtx gen_negv4si2 (rtx, rtx); +extern rtx gen_negv2di2 (rtx, rtx); +extern rtx gen_absv8qi2 (rtx, rtx); +extern rtx gen_absv16qi2 (rtx, rtx); +extern rtx gen_absv4hi2 (rtx, rtx); +extern rtx gen_absv8hi2 (rtx, rtx); +extern rtx gen_absv2si2 (rtx, rtx); +extern rtx gen_absv4si2 (rtx, rtx); +extern rtx gen_absv2di2 (rtx, rtx); +extern rtx gen_abdv8qi_3 (rtx, rtx, rtx); +extern rtx gen_abdv16qi_3 (rtx, rtx, rtx); +extern rtx gen_abdv4hi_3 (rtx, rtx, rtx); +extern rtx gen_abdv8hi_3 (rtx, rtx, rtx); +extern rtx gen_abdv2si_3 (rtx, rtx, rtx); +extern rtx gen_abdv4si_3 (rtx, rtx, rtx); +extern rtx gen_abav8qi_3 (rtx, rtx, rtx, rtx); +extern rtx gen_abav16qi_3 (rtx, rtx, rtx, rtx); +extern rtx gen_abav4hi_3 (rtx, rtx, rtx, rtx); +extern rtx gen_abav8hi_3 (rtx, rtx, rtx, rtx); +extern rtx gen_abav2si_3 (rtx, rtx, rtx, rtx); +extern rtx gen_abav4si_3 (rtx, rtx, rtx, rtx); +extern rtx gen_fabdv2sf_3 (rtx, rtx, rtx); +extern rtx gen_fabdv4sf_3 (rtx, rtx, rtx); +extern rtx gen_fabdv2df_3 (rtx, rtx, rtx); +extern rtx gen_andv8qi3 (rtx, rtx, rtx); +extern rtx gen_andv16qi3 (rtx, rtx, rtx); +extern rtx gen_andv4hi3 (rtx, rtx, rtx); +extern rtx gen_andv8hi3 (rtx, rtx, rtx); +extern rtx gen_andv2si3 (rtx, rtx, rtx); +extern rtx gen_andv4si3 (rtx, rtx, rtx); +extern rtx gen_andv2di3 (rtx, rtx, rtx); +extern rtx gen_iorv8qi3 (rtx, rtx, rtx); +extern rtx gen_iorv16qi3 (rtx, rtx, rtx); +extern rtx gen_iorv4hi3 (rtx, rtx, rtx); +extern rtx gen_iorv8hi3 (rtx, rtx, rtx); +extern rtx gen_iorv2si3 (rtx, rtx, rtx); +extern rtx gen_iorv4si3 (rtx, rtx, rtx); +extern rtx gen_iorv2di3 (rtx, rtx, rtx); +extern rtx gen_xorv8qi3 (rtx, rtx, rtx); +extern rtx gen_xorv16qi3 (rtx, rtx, rtx); +extern rtx gen_xorv4hi3 (rtx, rtx, rtx); +extern rtx gen_xorv8hi3 (rtx, rtx, rtx); +extern rtx gen_xorv2si3 (rtx, rtx, rtx); +extern rtx gen_xorv4si3 (rtx, rtx, rtx); +extern rtx gen_xorv2di3 (rtx, rtx, rtx); +extern rtx gen_one_cmplv8qi2 (rtx, rtx); +extern rtx gen_one_cmplv16qi2 (rtx, rtx); +extern rtx gen_one_cmplv4hi2 (rtx, rtx); +extern rtx gen_one_cmplv8hi2 (rtx, rtx); +extern rtx gen_one_cmplv2si2 (rtx, rtx); +extern rtx gen_one_cmplv4si2 (rtx, rtx); +extern rtx gen_one_cmplv2di2 (rtx, rtx); +extern rtx gen_aarch64_simd_vec_setv8qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_setv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_setv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_setv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_setv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_setv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_lshrv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_lshrv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_lshrv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_lshrv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_lshrv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_lshrv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_lshrv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_ashrv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_ashrv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_ashrv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_ashrv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_ashrv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_ashrv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_ashrv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_imm_shlv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_imm_shlv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_imm_shlv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_imm_shlv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_imm_shlv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_imm_shlv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_imm_shlv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_sshlv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_sshlv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_sshlv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_sshlv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_sshlv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_sshlv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_sshlv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_shlv8qi_unsigned (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_shlv16qi_unsigned (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_shlv4hi_unsigned (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_shlv8hi_unsigned (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_shlv2si_unsigned (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_shlv4si_unsigned (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_shlv2di_unsigned (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_shlv8qi_signed (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_shlv16qi_signed (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_shlv4hi_signed (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_shlv8hi_signed (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_shlv2si_signed (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_shlv4si_signed (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_reg_shlv2di_signed (rtx, rtx, rtx); +extern rtx gen_aarch64_sshr_simddi (rtx, rtx); +extern rtx gen_aarch64_ushr_simddi (rtx, rtx); +extern rtx gen_aarch64_simd_vec_setv2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_setv2sf (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_setv4sf (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_setv2df (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_mlav8qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_mlav16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_mlav4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_mlav8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_mlav2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_mlav4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_mlsv8qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_mlsv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_mlsv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_mlsv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_mlsv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_mlsv4si (rtx, rtx, rtx, rtx); +extern rtx gen_smaxv8qi3 (rtx, rtx, rtx); +extern rtx gen_sminv8qi3 (rtx, rtx, rtx); +extern rtx gen_umaxv8qi3 (rtx, rtx, rtx); +extern rtx gen_uminv8qi3 (rtx, rtx, rtx); +extern rtx gen_smaxv16qi3 (rtx, rtx, rtx); +extern rtx gen_sminv16qi3 (rtx, rtx, rtx); +extern rtx gen_umaxv16qi3 (rtx, rtx, rtx); +extern rtx gen_uminv16qi3 (rtx, rtx, rtx); +extern rtx gen_smaxv4hi3 (rtx, rtx, rtx); +extern rtx gen_sminv4hi3 (rtx, rtx, rtx); +extern rtx gen_umaxv4hi3 (rtx, rtx, rtx); +extern rtx gen_uminv4hi3 (rtx, rtx, rtx); +extern rtx gen_smaxv8hi3 (rtx, rtx, rtx); +extern rtx gen_sminv8hi3 (rtx, rtx, rtx); +extern rtx gen_umaxv8hi3 (rtx, rtx, rtx); +extern rtx gen_uminv8hi3 (rtx, rtx, rtx); +extern rtx gen_smaxv2si3 (rtx, rtx, rtx); +extern rtx gen_sminv2si3 (rtx, rtx, rtx); +extern rtx gen_umaxv2si3 (rtx, rtx, rtx); +extern rtx gen_uminv2si3 (rtx, rtx, rtx); +extern rtx gen_smaxv4si3 (rtx, rtx, rtx); +extern rtx gen_sminv4si3 (rtx, rtx, rtx); +extern rtx gen_umaxv4si3 (rtx, rtx, rtx); +extern rtx gen_uminv4si3 (rtx, rtx, rtx); +extern rtx gen_move_lo_quad_internal_v16qi (rtx, rtx); +extern rtx gen_move_lo_quad_internal_v8hi (rtx, rtx); +extern rtx gen_move_lo_quad_internal_v4si (rtx, rtx); +extern rtx gen_move_lo_quad_internal_v2di (rtx, rtx); +extern rtx gen_move_lo_quad_internal_v4sf (rtx, rtx); +extern rtx gen_move_lo_quad_internal_v2df (rtx, rtx); +extern rtx gen_move_lo_quad_internal_be_v16qi (rtx, rtx); +extern rtx gen_move_lo_quad_internal_be_v8hi (rtx, rtx); +extern rtx gen_move_lo_quad_internal_be_v4si (rtx, rtx); +extern rtx gen_move_lo_quad_internal_be_v2di (rtx, rtx); +extern rtx gen_move_lo_quad_internal_be_v4sf (rtx, rtx); +extern rtx gen_move_lo_quad_internal_be_v2df (rtx, rtx); +extern rtx gen_aarch64_simd_move_hi_quad_v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_move_hi_quad_v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_move_hi_quad_v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_move_hi_quad_v2di (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_move_hi_quad_v4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_move_hi_quad_v2df (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_move_hi_quad_be_v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_move_hi_quad_be_v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_move_hi_quad_be_v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_move_hi_quad_be_v2di (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_move_hi_quad_be_v4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_move_hi_quad_be_v2df (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_pack_trunc_v8hi (rtx, rtx); +extern rtx gen_aarch64_simd_vec_pack_trunc_v4si (rtx, rtx); +extern rtx gen_aarch64_simd_vec_pack_trunc_v2di (rtx, rtx); +extern rtx gen_vec_pack_trunc_v8hi (rtx, rtx, rtx); +extern rtx gen_vec_pack_trunc_v4si (rtx, rtx, rtx); +extern rtx gen_vec_pack_trunc_v2di (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_unpacks_lo_v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_unpacku_lo_v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_unpacks_lo_v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_unpacku_lo_v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_unpacks_lo_v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_unpacku_lo_v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_unpacks_hi_v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_unpacku_hi_v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_unpacks_hi_v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_unpacku_hi_v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_unpacks_hi_v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_unpacku_hi_v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_smult_lo_v16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_umult_lo_v16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_smult_lo_v8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_umult_lo_v8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_smult_lo_v4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_umult_lo_v4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_smult_hi_v16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_umult_hi_v16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_smult_hi_v8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_umult_hi_v8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_smult_hi_v4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_vec_umult_hi_v4si (rtx, rtx, rtx, rtx); +extern rtx gen_addv2sf3 (rtx, rtx, rtx); +extern rtx gen_addv4sf3 (rtx, rtx, rtx); +extern rtx gen_addv2df3 (rtx, rtx, rtx); +extern rtx gen_subv2sf3 (rtx, rtx, rtx); +extern rtx gen_subv4sf3 (rtx, rtx, rtx); +extern rtx gen_subv2df3 (rtx, rtx, rtx); +extern rtx gen_mulv2sf3 (rtx, rtx, rtx); +extern rtx gen_mulv4sf3 (rtx, rtx, rtx); +extern rtx gen_mulv2df3 (rtx, rtx, rtx); +extern rtx gen_divv2sf3 (rtx, rtx, rtx); +extern rtx gen_divv4sf3 (rtx, rtx, rtx); +extern rtx gen_divv2df3 (rtx, rtx, rtx); +extern rtx gen_negv2sf2 (rtx, rtx); +extern rtx gen_negv4sf2 (rtx, rtx); +extern rtx gen_negv2df2 (rtx, rtx); +extern rtx gen_absv2sf2 (rtx, rtx); +extern rtx gen_absv4sf2 (rtx, rtx); +extern rtx gen_absv2df2 (rtx, rtx); +extern rtx gen_fmav2sf4 (rtx, rtx, rtx, rtx); +extern rtx gen_fmav4sf4 (rtx, rtx, rtx, rtx); +extern rtx gen_fmav2df4 (rtx, rtx, rtx, rtx); +extern rtx gen_fnmav2sf4 (rtx, rtx, rtx, rtx); +extern rtx gen_fnmav4sf4 (rtx, rtx, rtx, rtx); +extern rtx gen_fnmav2df4 (rtx, rtx, rtx, rtx); +extern rtx gen_btruncv2sf2 (rtx, rtx); +extern rtx gen_ceilv2sf2 (rtx, rtx); +extern rtx gen_floorv2sf2 (rtx, rtx); +extern rtx gen_frintnv2sf2 (rtx, rtx); +extern rtx gen_nearbyintv2sf2 (rtx, rtx); +extern rtx gen_rintv2sf2 (rtx, rtx); +extern rtx gen_roundv2sf2 (rtx, rtx); +extern rtx gen_btruncv4sf2 (rtx, rtx); +extern rtx gen_ceilv4sf2 (rtx, rtx); +extern rtx gen_floorv4sf2 (rtx, rtx); +extern rtx gen_frintnv4sf2 (rtx, rtx); +extern rtx gen_nearbyintv4sf2 (rtx, rtx); +extern rtx gen_rintv4sf2 (rtx, rtx); +extern rtx gen_roundv4sf2 (rtx, rtx); +extern rtx gen_btruncv2df2 (rtx, rtx); +extern rtx gen_ceilv2df2 (rtx, rtx); +extern rtx gen_floorv2df2 (rtx, rtx); +extern rtx gen_frintnv2df2 (rtx, rtx); +extern rtx gen_nearbyintv2df2 (rtx, rtx); +extern rtx gen_rintv2df2 (rtx, rtx); +extern rtx gen_roundv2df2 (rtx, rtx); +extern rtx gen_lbtruncv2sfv2si2 (rtx, rtx); +extern rtx gen_lceilv2sfv2si2 (rtx, rtx); +extern rtx gen_lfloorv2sfv2si2 (rtx, rtx); +extern rtx gen_lroundv2sfv2si2 (rtx, rtx); +extern rtx gen_lfrintnv2sfv2si2 (rtx, rtx); +extern rtx gen_lbtruncuv2sfv2si2 (rtx, rtx); +extern rtx gen_lceiluv2sfv2si2 (rtx, rtx); +extern rtx gen_lflooruv2sfv2si2 (rtx, rtx); +extern rtx gen_lrounduv2sfv2si2 (rtx, rtx); +extern rtx gen_lfrintnuv2sfv2si2 (rtx, rtx); +extern rtx gen_lbtruncv4sfv4si2 (rtx, rtx); +extern rtx gen_lceilv4sfv4si2 (rtx, rtx); +extern rtx gen_lfloorv4sfv4si2 (rtx, rtx); +extern rtx gen_lroundv4sfv4si2 (rtx, rtx); +extern rtx gen_lfrintnv4sfv4si2 (rtx, rtx); +extern rtx gen_lbtruncuv4sfv4si2 (rtx, rtx); +extern rtx gen_lceiluv4sfv4si2 (rtx, rtx); +extern rtx gen_lflooruv4sfv4si2 (rtx, rtx); +extern rtx gen_lrounduv4sfv4si2 (rtx, rtx); +extern rtx gen_lfrintnuv4sfv4si2 (rtx, rtx); +extern rtx gen_lbtruncv2dfv2di2 (rtx, rtx); +extern rtx gen_lceilv2dfv2di2 (rtx, rtx); +extern rtx gen_lfloorv2dfv2di2 (rtx, rtx); +extern rtx gen_lroundv2dfv2di2 (rtx, rtx); +extern rtx gen_lfrintnv2dfv2di2 (rtx, rtx); +extern rtx gen_lbtruncuv2dfv2di2 (rtx, rtx); +extern rtx gen_lceiluv2dfv2di2 (rtx, rtx); +extern rtx gen_lflooruv2dfv2di2 (rtx, rtx); +extern rtx gen_lrounduv2dfv2di2 (rtx, rtx); +extern rtx gen_lfrintnuv2dfv2di2 (rtx, rtx); +extern rtx gen_floatv2siv2sf2 (rtx, rtx); +extern rtx gen_floatunsv2siv2sf2 (rtx, rtx); +extern rtx gen_floatv4siv4sf2 (rtx, rtx); +extern rtx gen_floatunsv4siv4sf2 (rtx, rtx); +extern rtx gen_floatv2div2df2 (rtx, rtx); +extern rtx gen_floatunsv2div2df2 (rtx, rtx); +extern rtx gen_vec_unpacks_lo_v4sf (rtx, rtx); +extern rtx gen_aarch64_float_extend_lo_v2df (rtx, rtx); +extern rtx gen_vec_unpacks_hi_v4sf (rtx, rtx); +extern rtx gen_aarch64_float_truncate_lo_v2sf (rtx, rtx); +extern rtx gen_aarch64_float_truncate_hi_v4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_vmlsv2sf (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vmlsv4sf (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vmlsv2df (rtx, rtx, rtx, rtx); +extern rtx gen_smaxv2sf3 (rtx, rtx, rtx); +extern rtx gen_sminv2sf3 (rtx, rtx, rtx); +extern rtx gen_smaxv4sf3 (rtx, rtx, rtx); +extern rtx gen_sminv4sf3 (rtx, rtx, rtx); +extern rtx gen_smaxv2df3 (rtx, rtx, rtx); +extern rtx gen_sminv2df3 (rtx, rtx, rtx); +extern rtx gen_smax_nanv2sf3 (rtx, rtx, rtx); +extern rtx gen_smin_nanv2sf3 (rtx, rtx, rtx); +extern rtx gen_smax_nanv4sf3 (rtx, rtx, rtx); +extern rtx gen_smin_nanv4sf3 (rtx, rtx, rtx); +extern rtx gen_smax_nanv2df3 (rtx, rtx, rtx); +extern rtx gen_smin_nanv2df3 (rtx, rtx, rtx); +extern rtx gen_reduc_splus_v8qi (rtx, rtx); +extern rtx gen_reduc_uplus_v8qi (rtx, rtx); +extern rtx gen_reduc_splus_v16qi (rtx, rtx); +extern rtx gen_reduc_uplus_v16qi (rtx, rtx); +extern rtx gen_reduc_splus_v4hi (rtx, rtx); +extern rtx gen_reduc_uplus_v4hi (rtx, rtx); +extern rtx gen_reduc_splus_v8hi (rtx, rtx); +extern rtx gen_reduc_uplus_v8hi (rtx, rtx); +extern rtx gen_reduc_splus_v4si (rtx, rtx); +extern rtx gen_reduc_uplus_v4si (rtx, rtx); +extern rtx gen_reduc_splus_v2di (rtx, rtx); +extern rtx gen_reduc_uplus_v2di (rtx, rtx); +extern rtx gen_reduc_splus_v2si (rtx, rtx); +extern rtx gen_reduc_uplus_v2si (rtx, rtx); +extern rtx gen_reduc_splus_v2sf (rtx, rtx); +extern rtx gen_reduc_splus_v2df (rtx, rtx); +extern rtx gen_aarch64_addpv4sf (rtx, rtx); +extern rtx gen_clzv8qi2 (rtx, rtx); +extern rtx gen_clzv16qi2 (rtx, rtx); +extern rtx gen_clzv4hi2 (rtx, rtx); +extern rtx gen_clzv8hi2 (rtx, rtx); +extern rtx gen_clzv2si2 (rtx, rtx); +extern rtx gen_clzv4si2 (rtx, rtx); +extern rtx gen_reduc_umax_v8qi (rtx, rtx); +extern rtx gen_reduc_umin_v8qi (rtx, rtx); +extern rtx gen_reduc_smax_v8qi (rtx, rtx); +extern rtx gen_reduc_smin_v8qi (rtx, rtx); +extern rtx gen_reduc_umax_v16qi (rtx, rtx); +extern rtx gen_reduc_umin_v16qi (rtx, rtx); +extern rtx gen_reduc_smax_v16qi (rtx, rtx); +extern rtx gen_reduc_smin_v16qi (rtx, rtx); +extern rtx gen_reduc_umax_v4hi (rtx, rtx); +extern rtx gen_reduc_umin_v4hi (rtx, rtx); +extern rtx gen_reduc_smax_v4hi (rtx, rtx); +extern rtx gen_reduc_smin_v4hi (rtx, rtx); +extern rtx gen_reduc_umax_v8hi (rtx, rtx); +extern rtx gen_reduc_umin_v8hi (rtx, rtx); +extern rtx gen_reduc_smax_v8hi (rtx, rtx); +extern rtx gen_reduc_smin_v8hi (rtx, rtx); +extern rtx gen_reduc_umax_v4si (rtx, rtx); +extern rtx gen_reduc_umin_v4si (rtx, rtx); +extern rtx gen_reduc_smax_v4si (rtx, rtx); +extern rtx gen_reduc_smin_v4si (rtx, rtx); +extern rtx gen_reduc_umax_v2si (rtx, rtx); +extern rtx gen_reduc_umin_v2si (rtx, rtx); +extern rtx gen_reduc_smax_v2si (rtx, rtx); +extern rtx gen_reduc_smin_v2si (rtx, rtx); +extern rtx gen_reduc_smax_nan_v2sf (rtx, rtx); +extern rtx gen_reduc_smin_nan_v2sf (rtx, rtx); +extern rtx gen_reduc_smax_v2sf (rtx, rtx); +extern rtx gen_reduc_smin_v2sf (rtx, rtx); +extern rtx gen_reduc_smax_nan_v2df (rtx, rtx); +extern rtx gen_reduc_smin_nan_v2df (rtx, rtx); +extern rtx gen_reduc_smax_v2df (rtx, rtx); +extern rtx gen_reduc_smin_v2df (rtx, rtx); +extern rtx gen_reduc_smax_nan_v4sf (rtx, rtx); +extern rtx gen_reduc_smin_nan_v4sf (rtx, rtx); +extern rtx gen_reduc_smax_v4sf (rtx, rtx); +extern rtx gen_reduc_smin_v4sf (rtx, rtx); +extern rtx gen_aarch64_simd_bslv8qi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv16qi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv4hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv8hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv2si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv4si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv2di_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv2sf_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv4sf_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv2df_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bsldi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bsldf_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_get_lanev8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_lanev16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_lanev4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_lanev8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_lanev2si (rtx, rtx, rtx); +extern rtx gen_aarch64_get_lanev4si (rtx, rtx, rtx); +extern rtx gen_aarch64_get_lanev2di (rtx, rtx, rtx); +extern rtx gen_aarch64_get_lanev2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_get_lanev4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_get_lanev2df (rtx, rtx, rtx); +extern rtx gen_aarch64_combine_internalv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_combine_internalv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_combine_internalv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_combine_internalv2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_combine_internaldi (rtx, rtx, rtx); +extern rtx gen_aarch64_combine_internaldf (rtx, rtx, rtx); +extern rtx gen_aarch64_saddlv16qi_hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssublv16qi_hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_uaddlv16qi_hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usublv16qi_hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_saddlv8hi_hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssublv8hi_hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_uaddlv8hi_hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usublv8hi_hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_saddlv4si_hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssublv4si_hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_uaddlv4si_hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usublv4si_hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_saddlv16qi_lo_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssublv16qi_lo_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_uaddlv16qi_lo_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usublv16qi_lo_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_saddlv8hi_lo_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssublv8hi_lo_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_uaddlv8hi_lo_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usublv8hi_lo_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_saddlv4si_lo_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssublv4si_lo_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_uaddlv4si_lo_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usublv4si_lo_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_saddlv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_ssublv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uaddlv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_usublv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_saddlv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_ssublv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uaddlv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_usublv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_saddlv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_ssublv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_uaddlv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_usublv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_saddwv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_ssubwv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uaddwv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_usubwv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_saddwv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_ssubwv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uaddwv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_usubwv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_saddwv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_ssubwv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_uaddwv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_usubwv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_saddw2v16qi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssubw2v16qi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_uaddw2v16qi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usubw2v16qi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_saddw2v8hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssubw2v8hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_uaddw2v8hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usubw2v8hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_saddw2v4si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssubw2v4si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_uaddw2v4si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usubw2v4si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_shaddv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uhaddv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_srhaddv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_urhaddv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_shsubv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uhsubv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_srhsubv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_urhsubv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_shaddv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uhaddv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_srhaddv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_urhaddv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_shsubv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uhsubv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_srhsubv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_urhsubv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_shaddv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uhaddv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_srhaddv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_urhaddv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_shsubv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uhsubv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_srhsubv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_urhsubv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_shaddv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uhaddv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_srhaddv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_urhaddv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_shsubv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uhsubv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_srhsubv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_urhsubv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_shaddv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_uhaddv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_srhaddv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_urhaddv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_shsubv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_uhsubv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_srhsubv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_urhsubv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_shaddv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_uhaddv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_srhaddv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_urhaddv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_shsubv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_uhsubv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_srhsubv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_urhsubv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_addhnv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_raddhnv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_subhnv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_rsubhnv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_addhnv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_raddhnv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_subhnv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_rsubhnv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_addhnv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_raddhnv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_subhnv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_rsubhnv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_addhn2v8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_raddhn2v8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_subhn2v8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_rsubhn2v8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_addhn2v4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_raddhn2v4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_subhn2v4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_rsubhn2v4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_addhn2v2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_raddhn2v2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_subhn2v2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_rsubhn2v2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_pmulv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_pmulv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqaddv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqaddv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqsubv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqsubv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqaddv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqaddv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqsubv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqsubv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqaddv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqaddv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqsubv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqsubv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqaddv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqaddv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqsubv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqsubv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqaddv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_uqaddv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqsubv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_uqsubv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqaddv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_uqaddv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqsubv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_uqsubv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqaddv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_uqaddv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_sqsubv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_uqsubv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_sqaddqi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqaddqi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqsubqi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqsubqi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqaddhi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqaddhi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqsubhi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqsubhi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqaddsi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqaddsi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqsubsi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqsubsi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqadddi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqadddi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqsubdi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqsubdi (rtx, rtx, rtx); +extern rtx gen_aarch64_suqaddv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_usqaddv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_suqaddv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_usqaddv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_suqaddv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_usqaddv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_suqaddv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_usqaddv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_suqaddv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_usqaddv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_suqaddv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_usqaddv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_suqaddv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_usqaddv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_suqaddqi (rtx, rtx, rtx); +extern rtx gen_aarch64_usqaddqi (rtx, rtx, rtx); +extern rtx gen_aarch64_suqaddhi (rtx, rtx, rtx); +extern rtx gen_aarch64_usqaddhi (rtx, rtx, rtx); +extern rtx gen_aarch64_suqaddsi (rtx, rtx, rtx); +extern rtx gen_aarch64_usqaddsi (rtx, rtx, rtx); +extern rtx gen_aarch64_suqadddi (rtx, rtx, rtx); +extern rtx gen_aarch64_usqadddi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqmovunv8hi (rtx, rtx); +extern rtx gen_aarch64_sqmovunv4si (rtx, rtx); +extern rtx gen_aarch64_sqmovunv2di (rtx, rtx); +extern rtx gen_aarch64_sqmovunhi (rtx, rtx); +extern rtx gen_aarch64_sqmovunsi (rtx, rtx); +extern rtx gen_aarch64_sqmovundi (rtx, rtx); +extern rtx gen_aarch64_sqmovnv8hi (rtx, rtx); +extern rtx gen_aarch64_uqmovnv8hi (rtx, rtx); +extern rtx gen_aarch64_sqmovnv4si (rtx, rtx); +extern rtx gen_aarch64_uqmovnv4si (rtx, rtx); +extern rtx gen_aarch64_sqmovnv2di (rtx, rtx); +extern rtx gen_aarch64_uqmovnv2di (rtx, rtx); +extern rtx gen_aarch64_sqmovnhi (rtx, rtx); +extern rtx gen_aarch64_uqmovnhi (rtx, rtx); +extern rtx gen_aarch64_sqmovnsi (rtx, rtx); +extern rtx gen_aarch64_uqmovnsi (rtx, rtx); +extern rtx gen_aarch64_sqmovndi (rtx, rtx); +extern rtx gen_aarch64_uqmovndi (rtx, rtx); +extern rtx gen_aarch64_sqnegv8qi (rtx, rtx); +extern rtx gen_aarch64_sqabsv8qi (rtx, rtx); +extern rtx gen_aarch64_sqnegv16qi (rtx, rtx); +extern rtx gen_aarch64_sqabsv16qi (rtx, rtx); +extern rtx gen_aarch64_sqnegv4hi (rtx, rtx); +extern rtx gen_aarch64_sqabsv4hi (rtx, rtx); +extern rtx gen_aarch64_sqnegv8hi (rtx, rtx); +extern rtx gen_aarch64_sqabsv8hi (rtx, rtx); +extern rtx gen_aarch64_sqnegv2si (rtx, rtx); +extern rtx gen_aarch64_sqabsv2si (rtx, rtx); +extern rtx gen_aarch64_sqnegv4si (rtx, rtx); +extern rtx gen_aarch64_sqabsv4si (rtx, rtx); +extern rtx gen_aarch64_sqnegv2di (rtx, rtx); +extern rtx gen_aarch64_sqabsv2di (rtx, rtx); +extern rtx gen_aarch64_sqnegqi (rtx, rtx); +extern rtx gen_aarch64_sqabsqi (rtx, rtx); +extern rtx gen_aarch64_sqneghi (rtx, rtx); +extern rtx gen_aarch64_sqabshi (rtx, rtx); +extern rtx gen_aarch64_sqnegsi (rtx, rtx); +extern rtx gen_aarch64_sqabssi (rtx, rtx); +extern rtx gen_aarch64_sqnegdi (rtx, rtx); +extern rtx gen_aarch64_sqabsdi (rtx, rtx); +extern rtx gen_aarch64_sqdmulhv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulhv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulhv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulhv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulhv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulhv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulhv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulhv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulhhi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulhhi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulhsi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulhsi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_lanev4hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_lanev4hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_lanev8hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_lanev8hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_lanev2si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_lanev2si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_lanev4si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_lanev4si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_laneqv4hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_laneqv4hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_laneqv8hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_laneqv8hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_laneqv2si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_laneqv2si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_laneqv4si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_laneqv4si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_lanehi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_lanehi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_lanesi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_lanesi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlalv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlslv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlalv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlslv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlalhi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlslhi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlalsi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlslsi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_lanev4hi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_lanev4hi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_lanev2si_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_lanev2si_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_laneqv4hi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_laneqv4hi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_laneqv2si_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_laneqv2si_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_lanehi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_lanehi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_lanesi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_lanesi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_laneqhi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_laneqhi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_laneqsi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_laneqsi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_nv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_nv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_nv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_nv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2v8hi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2v8hi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2v4si_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2v4si_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2_lanev8hi_internal (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2_lanev8hi_internal (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2_lanev4si_internal (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2_lanev4si_internal (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2_laneqv8hi_internal (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2_laneqv8hi_internal (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2_laneqv4si_internal (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2_laneqv4si_internal (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2_nv8hi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2_nv8hi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2_nv4si_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2_nv4si_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmullv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmullv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmullhi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmullsi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_lanev4hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_lanev2si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_laneqv4hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_laneqv2si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_lanehi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_lanesi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_laneqhi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_laneqsi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_nv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_nv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2v8hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2v4si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2_lanev8hi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2_lanev4si_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2_laneqv8hi_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2_laneqv4si_internal (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2_nv8hi_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2_nv4si_internal (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sshlv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_ushlv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_srshlv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_urshlv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sshlv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_ushlv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_srshlv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_urshlv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sshlv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_ushlv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_srshlv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_urshlv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sshlv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_ushlv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_srshlv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_urshlv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sshlv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_ushlv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_srshlv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_urshlv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_sshlv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_ushlv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_srshlv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_urshlv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sshlv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_ushlv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_srshlv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_urshlv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_sshldi (rtx, rtx, rtx); +extern rtx gen_aarch64_ushldi (rtx, rtx, rtx); +extern rtx gen_aarch64_srshldi (rtx, rtx, rtx); +extern rtx gen_aarch64_urshldi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshlv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshlv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshlv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshlv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshlv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshlv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshlv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshlv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshlv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshlv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshlv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshlv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshlv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshlv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshlv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshlv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshlv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshlv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshlv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshlv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshlv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlqi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshlqi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshlqi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshlqi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlhi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshlhi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshlhi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshlhi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlsi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshlsi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshlsi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshlsi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshldi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshldi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshldi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshldi (rtx, rtx, rtx); +extern rtx gen_aarch64_sshll_nv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_ushll_nv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sshll_nv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_ushll_nv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sshll_nv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_ushll_nv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_sshll2_nv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_ushll2_nv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sshll2_nv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_ushll2_nv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sshll2_nv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_ushll2_nv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_srshr_nv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_urshr_nv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_srshr_nv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_urshr_nv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_srshr_nv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_urshr_nv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_srshr_nv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_urshr_nv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_srshr_nv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_urshr_nv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_srshr_nv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_urshr_nv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_srshr_nv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_urshr_nv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_srshr_ndi (rtx, rtx, rtx); +extern rtx gen_aarch64_urshr_ndi (rtx, rtx, rtx); +extern rtx gen_aarch64_ssra_nv8qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usra_nv8qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_srsra_nv8qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ursra_nv8qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssra_nv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usra_nv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_srsra_nv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ursra_nv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssra_nv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usra_nv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_srsra_nv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ursra_nv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssra_nv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usra_nv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_srsra_nv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ursra_nv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssra_nv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usra_nv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_srsra_nv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ursra_nv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssra_nv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usra_nv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_srsra_nv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ursra_nv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssra_nv2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usra_nv2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_srsra_nv2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ursra_nv2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssra_ndi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usra_ndi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_srsra_ndi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ursra_ndi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssli_nv8qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usli_nv8qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssri_nv8qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usri_nv8qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssli_nv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usli_nv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssri_nv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usri_nv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssli_nv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usli_nv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssri_nv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usri_nv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssli_nv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usli_nv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssri_nv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usri_nv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssli_nv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usli_nv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssri_nv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usri_nv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssli_nv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usli_nv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssri_nv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usri_nv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssli_nv2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usli_nv2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssri_nv2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usri_nv2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssli_ndi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usli_ndi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_ssri_ndi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_usri_ndi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlu_nv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshl_nv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshl_nv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlu_nv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshl_nv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshl_nv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlu_nv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshl_nv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshl_nv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlu_nv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshl_nv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshl_nv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlu_nv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshl_nv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshl_nv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlu_nv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshl_nv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshl_nv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlu_nv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshl_nv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshl_nv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlu_nqi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshl_nqi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshl_nqi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlu_nhi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshl_nhi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshl_nhi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlu_nsi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshl_nsi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshl_nsi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshlu_ndi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshl_ndi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshl_ndi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshrun_nv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshrun_nv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshrn_nv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshrn_nv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshrn_nv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshrn_nv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshrun_nv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshrun_nv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshrn_nv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshrn_nv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshrn_nv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshrn_nv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshrun_nv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshrun_nv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshrn_nv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshrn_nv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshrn_nv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshrn_nv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshrun_nhi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshrun_nhi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshrn_nhi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshrn_nhi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshrn_nhi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshrn_nhi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshrun_nsi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshrun_nsi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshrn_nsi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshrn_nsi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshrn_nsi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshrn_nsi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshrun_ndi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshrun_ndi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqshrn_ndi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqshrn_ndi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqrshrn_ndi (rtx, rtx, rtx); +extern rtx gen_aarch64_uqrshrn_ndi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmlev8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmeqv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgev8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmlev16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmeqv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgev16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmlev4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmeqv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgev4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmlev8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmeqv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgev8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmlev2si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmeqv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgev2si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmlev4si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmeqv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgev4si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_cmlev2di (rtx, rtx, rtx); +extern rtx gen_aarch64_cmeqv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgev2di (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltdi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmledi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmeqdi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgedi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtdi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltuv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmleuv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgeuv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtuv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltuv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmleuv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgeuv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtuv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltuv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmleuv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgeuv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtuv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltuv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmleuv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgeuv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtuv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltuv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmleuv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgeuv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtuv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltuv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmleuv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgeuv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtuv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltuv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_cmleuv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgeuv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtuv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltudi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmleudi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgeudi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtudi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmtstv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmtstv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmtstv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmtstv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmtstv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmtstv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_cmtstv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_cmtstdi (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltv2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmlev2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmeqv2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgev2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtv2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltv4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmlev4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmeqv4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgev4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtv4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltv2df (rtx, rtx, rtx); +extern rtx gen_aarch64_cmlev2df (rtx, rtx, rtx); +extern rtx gen_aarch64_cmeqv2df (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgev2df (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtv2df (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltsf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmlesf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmeqsf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgesf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtsf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmltdf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmledf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmeqdf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgedf (rtx, rtx, rtx); +extern rtx gen_aarch64_cmgtdf (rtx, rtx, rtx); +extern rtx gen_aarch64_addpv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_addpv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_addpv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_addpdi (rtx, rtx); +extern rtx gen_sqrtv2sf2 (rtx, rtx); +extern rtx gen_sqrtv4sf2 (rtx, rtx); +extern rtx gen_sqrtv2df2 (rtx, rtx); +extern rtx gen_vec_load_lanesoiv16qi (rtx, rtx); +extern rtx gen_vec_load_lanesoiv8hi (rtx, rtx); +extern rtx gen_vec_load_lanesoiv4si (rtx, rtx); +extern rtx gen_vec_load_lanesoiv2di (rtx, rtx); +extern rtx gen_vec_load_lanesoiv4sf (rtx, rtx); +extern rtx gen_vec_load_lanesoiv2df (rtx, rtx); +extern rtx gen_vec_store_lanesoiv16qi (rtx, rtx); +extern rtx gen_vec_store_lanesoiv8hi (rtx, rtx); +extern rtx gen_vec_store_lanesoiv4si (rtx, rtx); +extern rtx gen_vec_store_lanesoiv2di (rtx, rtx); +extern rtx gen_vec_store_lanesoiv4sf (rtx, rtx); +extern rtx gen_vec_store_lanesoiv2df (rtx, rtx); +extern rtx gen_vec_store_lanesoi_lanev16qi (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesoi_lanev8hi (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesoi_lanev4si (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesoi_lanev2di (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesoi_lanev4sf (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesoi_lanev2df (rtx, rtx, rtx); +extern rtx gen_vec_load_lanesciv16qi (rtx, rtx); +extern rtx gen_vec_load_lanesciv8hi (rtx, rtx); +extern rtx gen_vec_load_lanesciv4si (rtx, rtx); +extern rtx gen_vec_load_lanesciv2di (rtx, rtx); +extern rtx gen_vec_load_lanesciv4sf (rtx, rtx); +extern rtx gen_vec_load_lanesciv2df (rtx, rtx); +extern rtx gen_vec_store_lanesciv16qi (rtx, rtx); +extern rtx gen_vec_store_lanesciv8hi (rtx, rtx); +extern rtx gen_vec_store_lanesciv4si (rtx, rtx); +extern rtx gen_vec_store_lanesciv2di (rtx, rtx); +extern rtx gen_vec_store_lanesciv4sf (rtx, rtx); +extern rtx gen_vec_store_lanesciv2df (rtx, rtx); +extern rtx gen_vec_store_lanesci_lanev16qi (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesci_lanev8hi (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesci_lanev4si (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesci_lanev2di (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesci_lanev4sf (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesci_lanev2df (rtx, rtx, rtx); +extern rtx gen_vec_load_lanesxiv16qi (rtx, rtx); +extern rtx gen_vec_load_lanesxiv8hi (rtx, rtx); +extern rtx gen_vec_load_lanesxiv4si (rtx, rtx); +extern rtx gen_vec_load_lanesxiv2di (rtx, rtx); +extern rtx gen_vec_load_lanesxiv4sf (rtx, rtx); +extern rtx gen_vec_load_lanesxiv2df (rtx, rtx); +extern rtx gen_vec_store_lanesxiv16qi (rtx, rtx); +extern rtx gen_vec_store_lanesxiv8hi (rtx, rtx); +extern rtx gen_vec_store_lanesxiv4si (rtx, rtx); +extern rtx gen_vec_store_lanesxiv2di (rtx, rtx); +extern rtx gen_vec_store_lanesxiv4sf (rtx, rtx); +extern rtx gen_vec_store_lanesxiv2df (rtx, rtx); +extern rtx gen_vec_store_lanesxi_lanev16qi (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesxi_lanev8hi (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesxi_lanev4si (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesxi_lanev2di (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesxi_lanev4sf (rtx, rtx, rtx); +extern rtx gen_vec_store_lanesxi_lanev2df (rtx, rtx, rtx); +extern rtx gen_aarch64_be_ld1v8qi (rtx, rtx); +extern rtx gen_aarch64_be_ld1v16qi (rtx, rtx); +extern rtx gen_aarch64_be_ld1v4hi (rtx, rtx); +extern rtx gen_aarch64_be_ld1v8hi (rtx, rtx); +extern rtx gen_aarch64_be_ld1v2si (rtx, rtx); +extern rtx gen_aarch64_be_ld1v4si (rtx, rtx); +extern rtx gen_aarch64_be_ld1v2di (rtx, rtx); +extern rtx gen_aarch64_be_ld1v2sf (rtx, rtx); +extern rtx gen_aarch64_be_ld1v4sf (rtx, rtx); +extern rtx gen_aarch64_be_ld1v2df (rtx, rtx); +extern rtx gen_aarch64_be_ld1di (rtx, rtx); +extern rtx gen_aarch64_be_st1v8qi (rtx, rtx); +extern rtx gen_aarch64_be_st1v16qi (rtx, rtx); +extern rtx gen_aarch64_be_st1v4hi (rtx, rtx); +extern rtx gen_aarch64_be_st1v8hi (rtx, rtx); +extern rtx gen_aarch64_be_st1v2si (rtx, rtx); +extern rtx gen_aarch64_be_st1v4si (rtx, rtx); +extern rtx gen_aarch64_be_st1v2di (rtx, rtx); +extern rtx gen_aarch64_be_st1v2sf (rtx, rtx); +extern rtx gen_aarch64_be_st1v4sf (rtx, rtx); +extern rtx gen_aarch64_be_st1v2df (rtx, rtx); +extern rtx gen_aarch64_be_st1di (rtx, rtx); +extern rtx gen_aarch64_ld2v8qi_dreg (rtx, rtx); +extern rtx gen_aarch64_ld2v4hi_dreg (rtx, rtx); +extern rtx gen_aarch64_ld2v2si_dreg (rtx, rtx); +extern rtx gen_aarch64_ld2v2sf_dreg (rtx, rtx); +extern rtx gen_aarch64_ld2di_dreg (rtx, rtx); +extern rtx gen_aarch64_ld2df_dreg (rtx, rtx); +extern rtx gen_aarch64_ld3v8qi_dreg (rtx, rtx); +extern rtx gen_aarch64_ld3v4hi_dreg (rtx, rtx); +extern rtx gen_aarch64_ld3v2si_dreg (rtx, rtx); +extern rtx gen_aarch64_ld3v2sf_dreg (rtx, rtx); +extern rtx gen_aarch64_ld3di_dreg (rtx, rtx); +extern rtx gen_aarch64_ld3df_dreg (rtx, rtx); +extern rtx gen_aarch64_ld4v8qi_dreg (rtx, rtx); +extern rtx gen_aarch64_ld4v4hi_dreg (rtx, rtx); +extern rtx gen_aarch64_ld4v2si_dreg (rtx, rtx); +extern rtx gen_aarch64_ld4v2sf_dreg (rtx, rtx); +extern rtx gen_aarch64_ld4di_dreg (rtx, rtx); +extern rtx gen_aarch64_ld4df_dreg (rtx, rtx); +extern rtx gen_aarch64_tbl1v8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_tbl1v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_tbl2v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_combinev16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_zip1v8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_zip2v8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_trn1v8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_trn2v8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp1v8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp2v8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_zip1v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_zip2v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_trn1v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_trn2v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp1v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp2v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_zip1v4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_zip2v4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_trn1v4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_trn2v4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp1v4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp2v4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_zip1v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_zip2v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_trn1v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_trn2v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp1v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp2v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_zip1v2si (rtx, rtx, rtx); +extern rtx gen_aarch64_zip2v2si (rtx, rtx, rtx); +extern rtx gen_aarch64_trn1v2si (rtx, rtx, rtx); +extern rtx gen_aarch64_trn2v2si (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp1v2si (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp2v2si (rtx, rtx, rtx); +extern rtx gen_aarch64_zip1v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_zip2v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_trn1v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_trn2v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp1v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp2v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_zip1v2di (rtx, rtx, rtx); +extern rtx gen_aarch64_zip2v2di (rtx, rtx, rtx); +extern rtx gen_aarch64_trn1v2di (rtx, rtx, rtx); +extern rtx gen_aarch64_trn2v2di (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp1v2di (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp2v2di (rtx, rtx, rtx); +extern rtx gen_aarch64_zip1v2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_zip2v2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_trn1v2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_trn2v2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp1v2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp2v2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_zip1v4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_zip2v4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_trn1v4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_trn2v4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp1v4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp2v4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_zip1v2df (rtx, rtx, rtx); +extern rtx gen_aarch64_zip2v2df (rtx, rtx, rtx); +extern rtx gen_aarch64_trn1v2df (rtx, rtx, rtx); +extern rtx gen_aarch64_trn2v2df (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp1v2df (rtx, rtx, rtx); +extern rtx gen_aarch64_uzp2v2df (rtx, rtx, rtx); +extern rtx gen_aarch64_extv8qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_extv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_extv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_extv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_extv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_extv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_extv2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_extv2sf (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_extv4sf (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_extv2df (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_rev64v8qi (rtx, rtx); +extern rtx gen_aarch64_rev32v8qi (rtx, rtx); +extern rtx gen_aarch64_rev16v8qi (rtx, rtx); +extern rtx gen_aarch64_rev64v16qi (rtx, rtx); +extern rtx gen_aarch64_rev32v16qi (rtx, rtx); +extern rtx gen_aarch64_rev16v16qi (rtx, rtx); +extern rtx gen_aarch64_rev64v4hi (rtx, rtx); +extern rtx gen_aarch64_rev32v4hi (rtx, rtx); +extern rtx gen_aarch64_rev16v4hi (rtx, rtx); +extern rtx gen_aarch64_rev64v8hi (rtx, rtx); +extern rtx gen_aarch64_rev32v8hi (rtx, rtx); +extern rtx gen_aarch64_rev16v8hi (rtx, rtx); +extern rtx gen_aarch64_rev64v2si (rtx, rtx); +extern rtx gen_aarch64_rev32v2si (rtx, rtx); +extern rtx gen_aarch64_rev16v2si (rtx, rtx); +extern rtx gen_aarch64_rev64v4si (rtx, rtx); +extern rtx gen_aarch64_rev32v4si (rtx, rtx); +extern rtx gen_aarch64_rev16v4si (rtx, rtx); +extern rtx gen_aarch64_rev64v2di (rtx, rtx); +extern rtx gen_aarch64_rev32v2di (rtx, rtx); +extern rtx gen_aarch64_rev16v2di (rtx, rtx); +extern rtx gen_aarch64_rev64v2sf (rtx, rtx); +extern rtx gen_aarch64_rev32v2sf (rtx, rtx); +extern rtx gen_aarch64_rev16v2sf (rtx, rtx); +extern rtx gen_aarch64_rev64v4sf (rtx, rtx); +extern rtx gen_aarch64_rev32v4sf (rtx, rtx); +extern rtx gen_aarch64_rev16v4sf (rtx, rtx); +extern rtx gen_aarch64_rev64v2df (rtx, rtx); +extern rtx gen_aarch64_rev32v2df (rtx, rtx); +extern rtx gen_aarch64_rev16v2df (rtx, rtx); +extern rtx gen_aarch64_st2v8qi_dreg (rtx, rtx); +extern rtx gen_aarch64_st2v4hi_dreg (rtx, rtx); +extern rtx gen_aarch64_st2v2si_dreg (rtx, rtx); +extern rtx gen_aarch64_st2v2sf_dreg (rtx, rtx); +extern rtx gen_aarch64_st2di_dreg (rtx, rtx); +extern rtx gen_aarch64_st2df_dreg (rtx, rtx); +extern rtx gen_aarch64_st3v8qi_dreg (rtx, rtx); +extern rtx gen_aarch64_st3v4hi_dreg (rtx, rtx); +extern rtx gen_aarch64_st3v2si_dreg (rtx, rtx); +extern rtx gen_aarch64_st3v2sf_dreg (rtx, rtx); +extern rtx gen_aarch64_st3di_dreg (rtx, rtx); +extern rtx gen_aarch64_st3df_dreg (rtx, rtx); +extern rtx gen_aarch64_st4v8qi_dreg (rtx, rtx); +extern rtx gen_aarch64_st4v4hi_dreg (rtx, rtx); +extern rtx gen_aarch64_st4v2si_dreg (rtx, rtx); +extern rtx gen_aarch64_st4v2sf_dreg (rtx, rtx); +extern rtx gen_aarch64_st4di_dreg (rtx, rtx); +extern rtx gen_aarch64_st4df_dreg (rtx, rtx); +extern rtx gen_aarch64_frecpev2sf (rtx, rtx); +extern rtx gen_aarch64_frecpev4sf (rtx, rtx); +extern rtx gen_aarch64_frecpev2df (rtx, rtx); +extern rtx gen_aarch64_frecpesf (rtx, rtx); +extern rtx gen_aarch64_frecpxsf (rtx, rtx); +extern rtx gen_aarch64_frecpedf (rtx, rtx); +extern rtx gen_aarch64_frecpxdf (rtx, rtx); +extern rtx gen_aarch64_frecpsv2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_frecpsv4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_frecpsv2df (rtx, rtx, rtx); +extern rtx gen_aarch64_frecpssf (rtx, rtx, rtx); +extern rtx gen_aarch64_frecpsdf (rtx, rtx, rtx); +extern rtx gen_aarch64_crypto_aesev16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_crypto_aesdv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_crypto_aesmcv16qi (rtx, rtx); +extern rtx gen_aarch64_crypto_aesimcv16qi (rtx, rtx); +extern rtx gen_aarch64_crypto_sha1hsi (rtx, rtx); +extern rtx gen_aarch64_crypto_sha1su1v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_crypto_sha1cv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_crypto_sha1mv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_crypto_sha1pv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_crypto_sha1su0v4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_crypto_sha256hv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_crypto_sha256h2v4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_crypto_sha256su0v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_crypto_sha256su1v4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_crypto_pmulldi (rtx, rtx, rtx); +extern rtx gen_aarch64_crypto_pmullv2di (rtx, rtx, rtx); +extern rtx gen_atomic_compare_and_swapqi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_atomic_compare_and_swaphi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_atomic_compare_and_swapsi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_atomic_compare_and_swapdi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_atomic_exchangeqi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_exchangehi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_exchangesi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_exchangedi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_addqi (rtx, rtx, rtx); +extern rtx gen_atomic_subqi (rtx, rtx, rtx); +extern rtx gen_atomic_orqi (rtx, rtx, rtx); +extern rtx gen_atomic_xorqi (rtx, rtx, rtx); +extern rtx gen_atomic_andqi (rtx, rtx, rtx); +extern rtx gen_atomic_addhi (rtx, rtx, rtx); +extern rtx gen_atomic_subhi (rtx, rtx, rtx); +extern rtx gen_atomic_orhi (rtx, rtx, rtx); +extern rtx gen_atomic_xorhi (rtx, rtx, rtx); +extern rtx gen_atomic_andhi (rtx, rtx, rtx); +extern rtx gen_atomic_addsi (rtx, rtx, rtx); +extern rtx gen_atomic_subsi (rtx, rtx, rtx); +extern rtx gen_atomic_orsi (rtx, rtx, rtx); +extern rtx gen_atomic_xorsi (rtx, rtx, rtx); +extern rtx gen_atomic_andsi (rtx, rtx, rtx); +extern rtx gen_atomic_adddi (rtx, rtx, rtx); +extern rtx gen_atomic_subdi (rtx, rtx, rtx); +extern rtx gen_atomic_ordi (rtx, rtx, rtx); +extern rtx gen_atomic_xordi (rtx, rtx, rtx); +extern rtx gen_atomic_anddi (rtx, rtx, rtx); +extern rtx gen_atomic_nandqi (rtx, rtx, rtx); +extern rtx gen_atomic_nandhi (rtx, rtx, rtx); +extern rtx gen_atomic_nandsi (rtx, rtx, rtx); +extern rtx gen_atomic_nanddi (rtx, rtx, rtx); +extern rtx gen_atomic_fetch_addqi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_subqi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_orqi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_xorqi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_andqi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_addhi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_subhi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_orhi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_xorhi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_andhi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_addsi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_subsi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_orsi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_xorsi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_andsi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_adddi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_subdi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_ordi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_xordi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_anddi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_nandqi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_nandhi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_nandsi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_fetch_nanddi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_add_fetchqi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_sub_fetchqi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_or_fetchqi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_xor_fetchqi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_and_fetchqi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_add_fetchhi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_sub_fetchhi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_or_fetchhi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_xor_fetchhi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_and_fetchhi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_add_fetchsi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_sub_fetchsi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_or_fetchsi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_xor_fetchsi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_and_fetchsi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_add_fetchdi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_sub_fetchdi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_or_fetchdi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_xor_fetchdi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_and_fetchdi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_nand_fetchqi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_nand_fetchhi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_nand_fetchsi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_nand_fetchdi (rtx, rtx, rtx, rtx); +extern rtx gen_atomic_loadqi (rtx, rtx, rtx); +extern rtx gen_atomic_loadhi (rtx, rtx, rtx); +extern rtx gen_atomic_loadsi (rtx, rtx, rtx); +extern rtx gen_atomic_loaddi (rtx, rtx, rtx); +extern rtx gen_atomic_storeqi (rtx, rtx, rtx); +extern rtx gen_atomic_storehi (rtx, rtx, rtx); +extern rtx gen_atomic_storesi (rtx, rtx, rtx); +extern rtx gen_atomic_storedi (rtx, rtx, rtx); +extern rtx gen_aarch64_load_exclusiveqi (rtx, rtx, rtx); +extern rtx gen_aarch64_load_exclusivehi (rtx, rtx, rtx); +extern rtx gen_aarch64_load_exclusivesi (rtx, rtx, rtx); +extern rtx gen_aarch64_load_exclusivedi (rtx, rtx, rtx); +extern rtx gen_aarch64_store_exclusiveqi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_store_exclusivehi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_store_exclusivesi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_store_exclusivedi (rtx, rtx, rtx, rtx); +extern rtx gen_cbranchsi4 (rtx, rtx, rtx, rtx); +extern rtx gen_cbranchdi4 (rtx, rtx, rtx, rtx); +extern rtx gen_cbranchsf4 (rtx, rtx, rtx, rtx); +extern rtx gen_cbranchdf4 (rtx, rtx, rtx, rtx); +extern rtx gen_casesi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_prologue (void); +extern rtx gen_epilogue (void); +extern rtx gen_sibcall_epilogue (void); #define GEN_CALL(A, B, C, D) gen_call ((A), (B), (C)) -extern rtx gen_call (rtx, rtx, rtx); +extern rtx gen_call (rtx, rtx, rtx); #define GEN_CALL_VALUE(A, B, C, D, E) gen_call_value ((A), (B), (C), (D)) -extern rtx gen_call_value (rtx, rtx, rtx, rtx); +extern rtx gen_call_value (rtx, rtx, rtx, rtx); #define GEN_SIBCALL(A, B, C, D) gen_sibcall ((A), (B), (C)) -extern rtx gen_sibcall (rtx, rtx, rtx); +extern rtx gen_sibcall (rtx, rtx, rtx); #define GEN_SIBCALL_VALUE(A, B, C, D, E) gen_sibcall_value ((A), (B), (C), (D)) -extern rtx gen_sibcall_value (rtx, rtx, rtx, rtx); -extern rtx gen_untyped_call (rtx, rtx, rtx); -extern rtx gen_movqi (rtx, rtx); -extern rtx gen_movhi (rtx, rtx); -extern rtx gen_movsi (rtx, rtx); -extern rtx gen_movdi (rtx, rtx); -extern rtx gen_movti (rtx, rtx); -extern rtx gen_movsf (rtx, rtx); -extern rtx gen_movdf (rtx, rtx); -extern rtx gen_movtf (rtx, rtx); -extern rtx gen_extendsidi2 (rtx, rtx); -extern rtx gen_zero_extendsidi2 (rtx, rtx); -extern rtx gen_extendqisi2 (rtx, rtx); -extern rtx gen_zero_extendqisi2 (rtx, rtx); -extern rtx gen_extendqidi2 (rtx, rtx); -extern rtx gen_zero_extendqidi2 (rtx, rtx); -extern rtx gen_extendhisi2 (rtx, rtx); -extern rtx gen_zero_extendhisi2 (rtx, rtx); -extern rtx gen_extendhidi2 (rtx, rtx); -extern rtx gen_zero_extendhidi2 (rtx, rtx); -extern rtx gen_extendqihi2 (rtx, rtx); -extern rtx gen_zero_extendqihi2 (rtx, rtx); -extern rtx gen_addsi3 (rtx, rtx, rtx); -extern rtx gen_adddi3 (rtx, rtx, rtx); -extern rtx gen_addti3 (rtx, rtx, rtx); -extern rtx gen_subti3 (rtx, rtx, rtx); -extern rtx gen_mulditi3 (rtx, rtx, rtx); -extern rtx gen_umulditi3 (rtx, rtx, rtx); -extern rtx gen_multi3 (rtx, rtx, rtx); -extern rtx gen_cstoresi4 (rtx, rtx, rtx, rtx); -extern rtx gen_cstoredi4 (rtx, rtx, rtx, rtx); -extern rtx gen_cstoresf4 (rtx, rtx, rtx, rtx); -extern rtx gen_cstoredf4 (rtx, rtx, rtx, rtx); -extern rtx gen_cmovsi6 (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_cmovdi6 (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_cmovsf6 (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_cmovdf6 (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_movqicc (rtx, rtx, rtx, rtx); -extern rtx gen_movhicc (rtx, rtx, rtx, rtx); -extern rtx gen_movsicc (rtx, rtx, rtx, rtx); -extern rtx gen_movdicc (rtx, rtx, rtx, rtx); -extern rtx gen_movsfsicc (rtx, rtx, rtx, rtx); -extern rtx gen_movdfsicc (rtx, rtx, rtx, rtx); -extern rtx gen_movsfdicc (rtx, rtx, rtx, rtx); -extern rtx gen_movdfdicc (rtx, rtx, rtx, rtx); -extern rtx gen_movsfcc (rtx, rtx, rtx, rtx); -extern rtx gen_movdfcc (rtx, rtx, rtx, rtx); -extern rtx gen_ffssi2 (rtx, rtx); -extern rtx gen_ffsdi2 (rtx, rtx); -extern rtx gen_ctzsi2 (rtx, rtx); -extern rtx gen_ctzdi2 (rtx, rtx); -extern rtx gen_ashlsi3 (rtx, rtx, rtx); -extern rtx gen_ashrsi3 (rtx, rtx, rtx); -extern rtx gen_lshrsi3 (rtx, rtx, rtx); -extern rtx gen_ashldi3 (rtx, rtx, rtx); -extern rtx gen_ashrdi3 (rtx, rtx, rtx); -extern rtx gen_lshrdi3 (rtx, rtx, rtx); -extern rtx gen_ashlqi3 (rtx, rtx, rtx); -extern rtx gen_ashlhi3 (rtx, rtx, rtx); -extern rtx gen_rotrsi3 (rtx, rtx, rtx); -extern rtx gen_rotrdi3 (rtx, rtx, rtx); -extern rtx gen_rotlsi3 (rtx, rtx, rtx); -extern rtx gen_rotldi3 (rtx, rtx, rtx); -extern rtx gen_extv (rtx, rtx, rtx, rtx); -extern rtx gen_extzv (rtx, rtx, rtx, rtx); -extern rtx gen_insvsi (rtx, rtx, rtx, rtx); -extern rtx gen_insvdi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_reload_movti (rtx, rtx, rtx); -extern rtx gen_aarch64_reload_movtf (rtx, rtx, rtx); -extern rtx gen_add_losym (rtx, rtx, rtx); -extern rtx gen_tlsgd_small (rtx, rtx); -extern rtx gen_tlsle_small (rtx, rtx, rtx); -extern rtx gen_get_thread_pointerdi (rtx); -extern rtx gen_stack_protect_set (rtx, rtx); -extern rtx gen_stack_protect_test (rtx, rtx, rtx); -extern rtx gen_movv8qi (rtx, rtx); -extern rtx gen_movv16qi (rtx, rtx); -extern rtx gen_movv4hi (rtx, rtx); -extern rtx gen_movv8hi (rtx, rtx); -extern rtx gen_movv2si (rtx, rtx); -extern rtx gen_movv4si (rtx, rtx); -extern rtx gen_movv2di (rtx, rtx); -extern rtx gen_movv2sf (rtx, rtx); -extern rtx gen_movv4sf (rtx, rtx); -extern rtx gen_movv2df (rtx, rtx); -extern rtx gen_movmisalignv8qi (rtx, rtx); -extern rtx gen_movmisalignv16qi (rtx, rtx); -extern rtx gen_movmisalignv4hi (rtx, rtx); -extern rtx gen_movmisalignv8hi (rtx, rtx); -extern rtx gen_movmisalignv2si (rtx, rtx); -extern rtx gen_movmisalignv4si (rtx, rtx); -extern rtx gen_movmisalignv2di (rtx, rtx); -extern rtx gen_movmisalignv2sf (rtx, rtx); -extern rtx gen_movmisalignv4sf (rtx, rtx); -extern rtx gen_movmisalignv2df (rtx, rtx); -extern rtx gen_aarch64_split_simd_movv16qi (rtx, rtx); -extern rtx gen_aarch64_split_simd_movv8hi (rtx, rtx); -extern rtx gen_aarch64_split_simd_movv4si (rtx, rtx); -extern rtx gen_aarch64_split_simd_movv2di (rtx, rtx); -extern rtx gen_aarch64_split_simd_movv4sf (rtx, rtx); -extern rtx gen_aarch64_split_simd_movv2df (rtx, rtx); -extern rtx gen_ashlv8qi3 (rtx, rtx, rtx); -extern rtx gen_ashlv16qi3 (rtx, rtx, rtx); -extern rtx gen_ashlv4hi3 (rtx, rtx, rtx); -extern rtx gen_ashlv8hi3 (rtx, rtx, rtx); -extern rtx gen_ashlv2si3 (rtx, rtx, rtx); -extern rtx gen_ashlv4si3 (rtx, rtx, rtx); -extern rtx gen_ashlv2di3 (rtx, rtx, rtx); -extern rtx gen_lshrv8qi3 (rtx, rtx, rtx); -extern rtx gen_lshrv16qi3 (rtx, rtx, rtx); -extern rtx gen_lshrv4hi3 (rtx, rtx, rtx); -extern rtx gen_lshrv8hi3 (rtx, rtx, rtx); -extern rtx gen_lshrv2si3 (rtx, rtx, rtx); -extern rtx gen_lshrv4si3 (rtx, rtx, rtx); -extern rtx gen_lshrv2di3 (rtx, rtx, rtx); -extern rtx gen_ashrv8qi3 (rtx, rtx, rtx); -extern rtx gen_ashrv16qi3 (rtx, rtx, rtx); -extern rtx gen_ashrv4hi3 (rtx, rtx, rtx); -extern rtx gen_ashrv8hi3 (rtx, rtx, rtx); -extern rtx gen_ashrv2si3 (rtx, rtx, rtx); -extern rtx gen_ashrv4si3 (rtx, rtx, rtx); -extern rtx gen_ashrv2di3 (rtx, rtx, rtx); -extern rtx gen_vashlv8qi3 (rtx, rtx, rtx); -extern rtx gen_vashlv16qi3 (rtx, rtx, rtx); -extern rtx gen_vashlv4hi3 (rtx, rtx, rtx); -extern rtx gen_vashlv8hi3 (rtx, rtx, rtx); -extern rtx gen_vashlv2si3 (rtx, rtx, rtx); -extern rtx gen_vashlv4si3 (rtx, rtx, rtx); -extern rtx gen_vashlv2di3 (rtx, rtx, rtx); -extern rtx gen_vashrv8qi3 (rtx, rtx, rtx); -extern rtx gen_vashrv16qi3 (rtx, rtx, rtx); -extern rtx gen_vashrv4hi3 (rtx, rtx, rtx); -extern rtx gen_vashrv8hi3 (rtx, rtx, rtx); -extern rtx gen_vashrv2si3 (rtx, rtx, rtx); -extern rtx gen_vashrv4si3 (rtx, rtx, rtx); -extern rtx gen_aarch64_ashr_simddi (rtx, rtx, rtx); -extern rtx gen_vlshrv8qi3 (rtx, rtx, rtx); -extern rtx gen_vlshrv16qi3 (rtx, rtx, rtx); -extern rtx gen_vlshrv4hi3 (rtx, rtx, rtx); -extern rtx gen_vlshrv8hi3 (rtx, rtx, rtx); -extern rtx gen_vlshrv2si3 (rtx, rtx, rtx); -extern rtx gen_vlshrv4si3 (rtx, rtx, rtx); -extern rtx gen_aarch64_lshr_simddi (rtx, rtx, rtx); -extern rtx gen_vec_setv8qi (rtx, rtx, rtx); -extern rtx gen_vec_setv16qi (rtx, rtx, rtx); -extern rtx gen_vec_setv4hi (rtx, rtx, rtx); -extern rtx gen_vec_setv8hi (rtx, rtx, rtx); -extern rtx gen_vec_setv2si (rtx, rtx, rtx); -extern rtx gen_vec_setv4si (rtx, rtx, rtx); -extern rtx gen_vec_setv2di (rtx, rtx, rtx); -extern rtx gen_vec_setv2sf (rtx, rtx, rtx); -extern rtx gen_vec_setv4sf (rtx, rtx, rtx); -extern rtx gen_vec_setv2df (rtx, rtx, rtx); -extern rtx gen_move_hi_quad_v16qi (rtx, rtx); -extern rtx gen_move_hi_quad_v8hi (rtx, rtx); -extern rtx gen_move_hi_quad_v4si (rtx, rtx); -extern rtx gen_move_hi_quad_v2di (rtx, rtx); -extern rtx gen_move_hi_quad_v4sf (rtx, rtx); -extern rtx gen_move_hi_quad_v2df (rtx, rtx); -extern rtx gen_vec_pack_trunc_v4hi (rtx, rtx, rtx); -extern rtx gen_vec_pack_trunc_v2si (rtx, rtx, rtx); -extern rtx gen_vec_pack_trunc_di (rtx, rtx, rtx); -extern rtx gen_vec_unpacks_hi_v16qi (rtx, rtx); -extern rtx gen_vec_unpacku_hi_v16qi (rtx, rtx); -extern rtx gen_vec_unpacks_hi_v8hi (rtx, rtx); -extern rtx gen_vec_unpacku_hi_v8hi (rtx, rtx); -extern rtx gen_vec_unpacks_hi_v4si (rtx, rtx); -extern rtx gen_vec_unpacku_hi_v4si (rtx, rtx); -extern rtx gen_vec_unpacks_lo_v16qi (rtx, rtx); -extern rtx gen_vec_unpacku_lo_v16qi (rtx, rtx); -extern rtx gen_vec_unpacks_lo_v8hi (rtx, rtx); -extern rtx gen_vec_unpacku_lo_v8hi (rtx, rtx); -extern rtx gen_vec_unpacks_lo_v4si (rtx, rtx); -extern rtx gen_vec_unpacku_lo_v4si (rtx, rtx); -extern rtx gen_vec_widen_smult_lo_v16qi (rtx, rtx, rtx); -extern rtx gen_vec_widen_umult_lo_v16qi (rtx, rtx, rtx); -extern rtx gen_vec_widen_smult_lo_v8hi (rtx, rtx, rtx); -extern rtx gen_vec_widen_umult_lo_v8hi (rtx, rtx, rtx); -extern rtx gen_vec_widen_smult_lo_v4si (rtx, rtx, rtx); -extern rtx gen_vec_widen_umult_lo_v4si (rtx, rtx, rtx); -extern rtx gen_vec_widen_smult_hi_v16qi (rtx, rtx, rtx); -extern rtx gen_vec_widen_umult_hi_v16qi (rtx, rtx, rtx); -extern rtx gen_vec_widen_smult_hi_v8hi (rtx, rtx, rtx); -extern rtx gen_vec_widen_umult_hi_v8hi (rtx, rtx, rtx); -extern rtx gen_vec_widen_smult_hi_v4si (rtx, rtx, rtx); -extern rtx gen_vec_widen_umult_hi_v4si (rtx, rtx, rtx); -extern rtx gen_fixv2sfv2si2 (rtx, rtx); -extern rtx gen_fixunsv2sfv2si2 (rtx, rtx); -extern rtx gen_fixv4sfv4si2 (rtx, rtx); -extern rtx gen_fixunsv4sfv4si2 (rtx, rtx); -extern rtx gen_fixv2dfv2di2 (rtx, rtx); -extern rtx gen_fixunsv2dfv2di2 (rtx, rtx); -extern rtx gen_fix_truncv2sfv2si2 (rtx, rtx); -extern rtx gen_fixuns_truncv2sfv2si2 (rtx, rtx); -extern rtx gen_fix_truncv4sfv4si2 (rtx, rtx); -extern rtx gen_fixuns_truncv4sfv4si2 (rtx, rtx); -extern rtx gen_fix_truncv2dfv2di2 (rtx, rtx); -extern rtx gen_fixuns_truncv2dfv2di2 (rtx, rtx); -extern rtx gen_ftruncv2sf2 (rtx, rtx); -extern rtx gen_ftruncv4sf2 (rtx, rtx); -extern rtx gen_ftruncv2df2 (rtx, rtx); -extern rtx gen_vec_pack_trunc_v2df (rtx, rtx, rtx); -extern rtx gen_vec_pack_trunc_df (rtx, rtx, rtx); -extern rtx gen_reduc_splus_v4sf (rtx, rtx); -extern rtx gen_aarch64_simd_bslv8qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv2sf (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv4sf (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bslv2df (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bsldi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_simd_bsldf (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv8qiv8qi (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv16qiv16qi (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv4hiv4hi (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv8hiv8hi (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv2siv2si (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv4siv4si (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv2div2di (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv2sfv2sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv2siv2sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv4sfv2sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv4siv2sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv2dfv2sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv2div2sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv2sfv4sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv2siv4sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv4sfv4sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv4siv4sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv2dfv4sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv2div4sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv2sfv2df (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv2siv2df (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv4sfv2df (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv4siv2df (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv2dfv2df (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_vcond_internalv2div2df (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vcondv8qiv8qi (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vcondv16qiv16qi (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vcondv4hiv4hi (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vcondv8hiv8hi (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vcondv2siv2si (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vcondv4siv4si (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vcondv2div2di (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vcondv2sfv2sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vcondv4sfv4sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vcondv2dfv2df (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vcondv2siv2sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vcondv4siv4sf (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vcondv2div2df (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vconduv8qiv8qi (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vconduv16qiv16qi (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vconduv4hiv4hi (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vconduv8hiv8hi (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vconduv2siv2si (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vconduv4siv4si (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_vconduv2div2di (rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_createv8qi (rtx, rtx); -extern rtx gen_aarch64_createv4hi (rtx, rtx); -extern rtx gen_aarch64_createv2si (rtx, rtx); -extern rtx gen_aarch64_createdi (rtx, rtx); -extern rtx gen_aarch64_createdf (rtx, rtx); -extern rtx gen_aarch64_createv2sf (rtx, rtx); -extern rtx gen_aarch64_be_checked_get_lanev8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_be_checked_get_lanev16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_be_checked_get_lanev4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_be_checked_get_lanev8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_be_checked_get_lanev2si (rtx, rtx, rtx); -extern rtx gen_aarch64_be_checked_get_lanev4si (rtx, rtx, rtx); -extern rtx gen_aarch64_be_checked_get_lanev2di (rtx, rtx, rtx); -extern rtx gen_aarch64_be_checked_get_lanev2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_be_checked_get_lanev4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_be_checked_get_lanev2df (rtx, rtx, rtx); -extern rtx gen_aarch64_get_lanedi (rtx, rtx, rtx); -extern rtx gen_aarch64_reinterpretv8qiv8qi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv8qiv4hi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv8qiv2si (rtx, rtx); -extern rtx gen_aarch64_reinterpretv8qiv2sf (rtx, rtx); -extern rtx gen_aarch64_reinterpretv8qidi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv8qidf (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4hiv8qi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4hiv4hi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4hiv2si (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4hiv2sf (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4hidi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4hidf (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2siv8qi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2siv4hi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2siv2si (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2siv2sf (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2sidi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2sidf (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2sfv8qi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2sfv4hi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2sfv2si (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2sfv2sf (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2sfdi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2sfdf (rtx, rtx); -extern rtx gen_aarch64_reinterpretdiv8qi (rtx, rtx); -extern rtx gen_aarch64_reinterpretdiv4hi (rtx, rtx); -extern rtx gen_aarch64_reinterpretdiv2si (rtx, rtx); -extern rtx gen_aarch64_reinterpretdidi (rtx, rtx); -extern rtx gen_aarch64_reinterpretdidf (rtx, rtx); -extern rtx gen_aarch64_reinterpretdiv2sf (rtx, rtx); -extern rtx gen_aarch64_reinterpretdfv8qi (rtx, rtx); -extern rtx gen_aarch64_reinterpretdfv4hi (rtx, rtx); -extern rtx gen_aarch64_reinterpretdfv2si (rtx, rtx); -extern rtx gen_aarch64_reinterpretdfdi (rtx, rtx); -extern rtx gen_aarch64_reinterpretdfdf (rtx, rtx); -extern rtx gen_aarch64_reinterpretdfv2sf (rtx, rtx); -extern rtx gen_aarch64_reinterpretv16qiv16qi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv16qiv8hi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv16qiv4si (rtx, rtx); -extern rtx gen_aarch64_reinterpretv16qiv2di (rtx, rtx); -extern rtx gen_aarch64_reinterpretv16qiv4sf (rtx, rtx); -extern rtx gen_aarch64_reinterpretv16qiv2df (rtx, rtx); -extern rtx gen_aarch64_reinterpretv8hiv16qi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv8hiv8hi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv8hiv4si (rtx, rtx); -extern rtx gen_aarch64_reinterpretv8hiv2di (rtx, rtx); -extern rtx gen_aarch64_reinterpretv8hiv4sf (rtx, rtx); -extern rtx gen_aarch64_reinterpretv8hiv2df (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4siv16qi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4siv8hi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4siv4si (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4siv2di (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4siv4sf (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4siv2df (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4sfv16qi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4sfv8hi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4sfv4si (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4sfv2di (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4sfv4sf (rtx, rtx); -extern rtx gen_aarch64_reinterpretv4sfv2df (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2div16qi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2div8hi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2div4si (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2div2di (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2div4sf (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2div2df (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2dfv16qi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2dfv8hi (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2dfv4si (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2dfv2di (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2dfv4sf (rtx, rtx); -extern rtx gen_aarch64_reinterpretv2dfv2df (rtx, rtx); -extern rtx gen_aarch64_simd_combinev8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_combinev4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_combinev2si (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_combinev2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_combinedi (rtx, rtx, rtx); -extern rtx gen_aarch64_simd_combinedf (rtx, rtx, rtx); -extern rtx gen_aarch64_saddl2v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_saddl2v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_saddl2v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_uaddl2v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uaddl2v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uaddl2v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_ssubl2v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_ssubl2v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_ssubl2v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_usubl2v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_usubl2v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_usubl2v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_saddw2v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_saddw2v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_saddw2v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_uaddw2v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_uaddw2v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_uaddw2v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_ssubw2v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_ssubw2v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_ssubw2v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_usubw2v16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_usubw2v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_usubw2v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal_lanev4hi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal_lanev2si (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal_lanehi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal_lanesi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal_laneqv4hi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal_laneqv2si (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal_laneqhi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal_laneqsi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl_lanev4hi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl_lanev2si (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl_lanehi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl_lanesi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl_laneqv4hi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl_laneqv2si (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl_laneqhi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl_laneqsi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal2v8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal2v4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl2v8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl2v4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal2_lanev8hi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal2_lanev4si (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal2_laneqv8hi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal2_laneqv4si (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl2_lanev8hi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl2_lanev4si (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl2_laneqv8hi (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl2_laneqv4si (rtx, rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal2_nv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlal2_nv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl2_nv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmlsl2_nv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull_lanev4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull_lanev2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull_lanehi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull_lanesi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull_laneqv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull_laneqv2si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull2v8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull2v4si (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull2_lanev8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull2_lanev4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull2_laneqv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull2_laneqv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull2_nv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_sqdmull2_nv4si (rtx, rtx, rtx); -extern rtx gen_movoi (rtx, rtx); -extern rtx gen_movci (rtx, rtx); -extern rtx gen_movxi (rtx, rtx); -extern rtx gen_aarch64_ld2v8qi (rtx, rtx); -extern rtx gen_aarch64_ld2v4hi (rtx, rtx); -extern rtx gen_aarch64_ld2v2si (rtx, rtx); -extern rtx gen_aarch64_ld2v2sf (rtx, rtx); -extern rtx gen_aarch64_ld2di (rtx, rtx); -extern rtx gen_aarch64_ld2df (rtx, rtx); -extern rtx gen_aarch64_ld3v8qi (rtx, rtx); -extern rtx gen_aarch64_ld3v4hi (rtx, rtx); -extern rtx gen_aarch64_ld3v2si (rtx, rtx); -extern rtx gen_aarch64_ld3v2sf (rtx, rtx); -extern rtx gen_aarch64_ld3di (rtx, rtx); -extern rtx gen_aarch64_ld3df (rtx, rtx); -extern rtx gen_aarch64_ld4v8qi (rtx, rtx); -extern rtx gen_aarch64_ld4v4hi (rtx, rtx); -extern rtx gen_aarch64_ld4v2si (rtx, rtx); -extern rtx gen_aarch64_ld4v2sf (rtx, rtx); -extern rtx gen_aarch64_ld4di (rtx, rtx); -extern rtx gen_aarch64_ld4df (rtx, rtx); -extern rtx gen_aarch64_ld1v8qi (rtx, rtx); -extern rtx gen_aarch64_ld1v16qi (rtx, rtx); -extern rtx gen_aarch64_ld1v4hi (rtx, rtx); -extern rtx gen_aarch64_ld1v8hi (rtx, rtx); -extern rtx gen_aarch64_ld1v2si (rtx, rtx); -extern rtx gen_aarch64_ld1v4si (rtx, rtx); -extern rtx gen_aarch64_ld1v2di (rtx, rtx); -extern rtx gen_aarch64_ld1v2sf (rtx, rtx); -extern rtx gen_aarch64_ld1v4sf (rtx, rtx); -extern rtx gen_aarch64_ld1v2df (rtx, rtx); -extern rtx gen_aarch64_ld2v16qi (rtx, rtx); -extern rtx gen_aarch64_ld3v16qi (rtx, rtx); -extern rtx gen_aarch64_ld4v16qi (rtx, rtx); -extern rtx gen_aarch64_ld2v8hi (rtx, rtx); -extern rtx gen_aarch64_ld3v8hi (rtx, rtx); -extern rtx gen_aarch64_ld4v8hi (rtx, rtx); -extern rtx gen_aarch64_ld2v4si (rtx, rtx); -extern rtx gen_aarch64_ld3v4si (rtx, rtx); -extern rtx gen_aarch64_ld4v4si (rtx, rtx); -extern rtx gen_aarch64_ld2v2di (rtx, rtx); -extern rtx gen_aarch64_ld3v2di (rtx, rtx); -extern rtx gen_aarch64_ld4v2di (rtx, rtx); -extern rtx gen_aarch64_ld2v4sf (rtx, rtx); -extern rtx gen_aarch64_ld3v4sf (rtx, rtx); -extern rtx gen_aarch64_ld4v4sf (rtx, rtx); -extern rtx gen_aarch64_ld2v2df (rtx, rtx); -extern rtx gen_aarch64_ld3v2df (rtx, rtx); -extern rtx gen_aarch64_ld4v2df (rtx, rtx); -extern rtx gen_aarch64_get_dregoiv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregoiv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregoiv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregoiv2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregoidi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregoidf (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregciv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregciv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregciv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregciv2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregcidi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregcidf (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregxiv8qi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregxiv4hi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregxiv2si (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregxiv2sf (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregxidi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_dregxidf (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregoiv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregciv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregxiv16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregoiv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregciv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregxiv8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregoiv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregciv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregxiv4si (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregoiv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregciv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregxiv2di (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregoiv4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregciv4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregxiv4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregoiv2df (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregciv2df (rtx, rtx, rtx); -extern rtx gen_aarch64_get_qregxiv2df (rtx, rtx, rtx); -extern rtx gen_vec_perm_constv8qi (rtx, rtx, rtx, rtx); -extern rtx gen_vec_perm_constv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_vec_perm_constv4hi (rtx, rtx, rtx, rtx); -extern rtx gen_vec_perm_constv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_vec_perm_constv2si (rtx, rtx, rtx, rtx); -extern rtx gen_vec_perm_constv4si (rtx, rtx, rtx, rtx); -extern rtx gen_vec_perm_constv2di (rtx, rtx, rtx, rtx); -extern rtx gen_vec_perm_constv2sf (rtx, rtx, rtx, rtx); -extern rtx gen_vec_perm_constv4sf (rtx, rtx, rtx, rtx); -extern rtx gen_vec_perm_constv2df (rtx, rtx, rtx, rtx); -extern rtx gen_vec_permv8qi (rtx, rtx, rtx, rtx); -extern rtx gen_vec_permv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_st2v8qi (rtx, rtx); -extern rtx gen_aarch64_st2v4hi (rtx, rtx); -extern rtx gen_aarch64_st2v2si (rtx, rtx); -extern rtx gen_aarch64_st2v2sf (rtx, rtx); -extern rtx gen_aarch64_st2di (rtx, rtx); -extern rtx gen_aarch64_st2df (rtx, rtx); -extern rtx gen_aarch64_st3v8qi (rtx, rtx); -extern rtx gen_aarch64_st3v4hi (rtx, rtx); -extern rtx gen_aarch64_st3v2si (rtx, rtx); -extern rtx gen_aarch64_st3v2sf (rtx, rtx); -extern rtx gen_aarch64_st3di (rtx, rtx); -extern rtx gen_aarch64_st3df (rtx, rtx); -extern rtx gen_aarch64_st4v8qi (rtx, rtx); -extern rtx gen_aarch64_st4v4hi (rtx, rtx); -extern rtx gen_aarch64_st4v2si (rtx, rtx); -extern rtx gen_aarch64_st4v2sf (rtx, rtx); -extern rtx gen_aarch64_st4di (rtx, rtx); -extern rtx gen_aarch64_st4df (rtx, rtx); -extern rtx gen_aarch64_st2v16qi (rtx, rtx); -extern rtx gen_aarch64_st3v16qi (rtx, rtx); -extern rtx gen_aarch64_st4v16qi (rtx, rtx); -extern rtx gen_aarch64_st2v8hi (rtx, rtx); -extern rtx gen_aarch64_st3v8hi (rtx, rtx); -extern rtx gen_aarch64_st4v8hi (rtx, rtx); -extern rtx gen_aarch64_st2v4si (rtx, rtx); -extern rtx gen_aarch64_st3v4si (rtx, rtx); -extern rtx gen_aarch64_st4v4si (rtx, rtx); -extern rtx gen_aarch64_st2v2di (rtx, rtx); -extern rtx gen_aarch64_st3v2di (rtx, rtx); -extern rtx gen_aarch64_st4v2di (rtx, rtx); -extern rtx gen_aarch64_st2v4sf (rtx, rtx); -extern rtx gen_aarch64_st3v4sf (rtx, rtx); -extern rtx gen_aarch64_st4v4sf (rtx, rtx); -extern rtx gen_aarch64_st2v2df (rtx, rtx); -extern rtx gen_aarch64_st3v2df (rtx, rtx); -extern rtx gen_aarch64_st4v2df (rtx, rtx); -extern rtx gen_aarch64_st2_lanev16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_st2_lanev8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_st2_lanev4si (rtx, rtx, rtx); -extern rtx gen_aarch64_st2_lanev2di (rtx, rtx, rtx); -extern rtx gen_aarch64_st2_lanev4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_st2_lanev2df (rtx, rtx, rtx); -extern rtx gen_aarch64_st3_lanev16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_st3_lanev8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_st3_lanev4si (rtx, rtx, rtx); -extern rtx gen_aarch64_st3_lanev2di (rtx, rtx, rtx); -extern rtx gen_aarch64_st3_lanev4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_st3_lanev2df (rtx, rtx, rtx); -extern rtx gen_aarch64_st4_lanev16qi (rtx, rtx, rtx); -extern rtx gen_aarch64_st4_lanev8hi (rtx, rtx, rtx); -extern rtx gen_aarch64_st4_lanev4si (rtx, rtx, rtx); -extern rtx gen_aarch64_st4_lanev2di (rtx, rtx, rtx); -extern rtx gen_aarch64_st4_lanev4sf (rtx, rtx, rtx); -extern rtx gen_aarch64_st4_lanev2df (rtx, rtx, rtx); -extern rtx gen_aarch64_st1v8qi (rtx, rtx); -extern rtx gen_aarch64_st1v16qi (rtx, rtx); -extern rtx gen_aarch64_st1v4hi (rtx, rtx); -extern rtx gen_aarch64_st1v8hi (rtx, rtx); -extern rtx gen_aarch64_st1v2si (rtx, rtx); -extern rtx gen_aarch64_st1v4si (rtx, rtx); -extern rtx gen_aarch64_st1v2di (rtx, rtx); -extern rtx gen_aarch64_st1v2sf (rtx, rtx); -extern rtx gen_aarch64_st1v4sf (rtx, rtx); -extern rtx gen_aarch64_st1v2df (rtx, rtx); -extern rtx gen_aarch64_set_qregoiv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregciv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregxiv16qi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregoiv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregciv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregxiv8hi (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregoiv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregciv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregxiv4si (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregoiv2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregciv2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregxiv2di (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregoiv4sf (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregciv4sf (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregxiv4sf (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregoiv2df (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregciv2df (rtx, rtx, rtx, rtx); -extern rtx gen_aarch64_set_qregxiv2df (rtx, rtx, rtx, rtx); -extern rtx gen_vec_initv8qi (rtx, rtx); -extern rtx gen_vec_initv16qi (rtx, rtx); -extern rtx gen_vec_initv4hi (rtx, rtx); -extern rtx gen_vec_initv8hi (rtx, rtx); -extern rtx gen_vec_initv2si (rtx, rtx); -extern rtx gen_vec_initv4si (rtx, rtx); -extern rtx gen_vec_initv2di (rtx, rtx); -extern rtx gen_vec_initv2sf (rtx, rtx); -extern rtx gen_vec_initv4sf (rtx, rtx); -extern rtx gen_vec_initv2df (rtx, rtx); -extern rtx gen_vec_extractv8qi (rtx, rtx, rtx); -extern rtx gen_vec_extractv16qi (rtx, rtx, rtx); -extern rtx gen_vec_extractv4hi (rtx, rtx, rtx); -extern rtx gen_vec_extractv8hi (rtx, rtx, rtx); -extern rtx gen_vec_extractv2si (rtx, rtx, rtx); -extern rtx gen_vec_extractv4si (rtx, rtx, rtx); -extern rtx gen_vec_extractv2di (rtx, rtx, rtx); -extern rtx gen_vec_extractv2sf (rtx, rtx, rtx); -extern rtx gen_vec_extractv4sf (rtx, rtx, rtx); -extern rtx gen_vec_extractv2df (rtx, rtx, rtx); -extern rtx gen_atomic_compare_and_swapqi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_atomic_compare_and_swaphi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_atomic_compare_and_swapsi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_atomic_compare_and_swapdi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx); -extern rtx gen_mem_thread_fence (rtx); -extern rtx gen_dmb (rtx); +extern rtx gen_sibcall_value (rtx, rtx, rtx, rtx); +extern rtx gen_untyped_call (rtx, rtx, rtx); +extern rtx gen_movqi (rtx, rtx); +extern rtx gen_movhi (rtx, rtx); +extern rtx gen_movsi (rtx, rtx); +extern rtx gen_movdi (rtx, rtx); +extern rtx gen_movti (rtx, rtx); +extern rtx gen_movsf (rtx, rtx); +extern rtx gen_movdf (rtx, rtx); +extern rtx gen_movtf (rtx, rtx); +extern rtx gen_movmemdi (rtx, rtx, rtx, rtx); +extern rtx gen_extendsidi2 (rtx, rtx); +extern rtx gen_zero_extendsidi2 (rtx, rtx); +extern rtx gen_extendqisi2 (rtx, rtx); +extern rtx gen_zero_extendqisi2 (rtx, rtx); +extern rtx gen_extendqidi2 (rtx, rtx); +extern rtx gen_zero_extendqidi2 (rtx, rtx); +extern rtx gen_extendhisi2 (rtx, rtx); +extern rtx gen_zero_extendhisi2 (rtx, rtx); +extern rtx gen_extendhidi2 (rtx, rtx); +extern rtx gen_zero_extendhidi2 (rtx, rtx); +extern rtx gen_extendqihi2 (rtx, rtx); +extern rtx gen_zero_extendqihi2 (rtx, rtx); +extern rtx gen_addsi3 (rtx, rtx, rtx); +extern rtx gen_adddi3 (rtx, rtx, rtx); +extern rtx gen_addti3 (rtx, rtx, rtx); +extern rtx gen_subti3 (rtx, rtx, rtx); +extern rtx gen_mulditi3 (rtx, rtx, rtx); +extern rtx gen_umulditi3 (rtx, rtx, rtx); +extern rtx gen_multi3 (rtx, rtx, rtx); +extern rtx gen_cstoresi4 (rtx, rtx, rtx, rtx); +extern rtx gen_cstoredi4 (rtx, rtx, rtx, rtx); +extern rtx gen_cstoresf4 (rtx, rtx, rtx, rtx); +extern rtx gen_cstoredf4 (rtx, rtx, rtx, rtx); +extern rtx gen_cmovsi6 (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_cmovdi6 (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_cmovsf6 (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_cmovdf6 (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_movqicc (rtx, rtx, rtx, rtx); +extern rtx gen_movhicc (rtx, rtx, rtx, rtx); +extern rtx gen_movsicc (rtx, rtx, rtx, rtx); +extern rtx gen_movdicc (rtx, rtx, rtx, rtx); +extern rtx gen_movsfsicc (rtx, rtx, rtx, rtx); +extern rtx gen_movdfsicc (rtx, rtx, rtx, rtx); +extern rtx gen_movsfdicc (rtx, rtx, rtx, rtx); +extern rtx gen_movdfdicc (rtx, rtx, rtx, rtx); +extern rtx gen_movsfcc (rtx, rtx, rtx, rtx); +extern rtx gen_movdfcc (rtx, rtx, rtx, rtx); +extern rtx gen_ffssi2 (rtx, rtx); +extern rtx gen_ffsdi2 (rtx, rtx); +extern rtx gen_ctzsi2 (rtx, rtx); +extern rtx gen_ctzdi2 (rtx, rtx); +extern rtx gen_ashlsi3 (rtx, rtx, rtx); +extern rtx gen_ashrsi3 (rtx, rtx, rtx); +extern rtx gen_lshrsi3 (rtx, rtx, rtx); +extern rtx gen_ashldi3 (rtx, rtx, rtx); +extern rtx gen_ashrdi3 (rtx, rtx, rtx); +extern rtx gen_lshrdi3 (rtx, rtx, rtx); +extern rtx gen_ashlqi3 (rtx, rtx, rtx); +extern rtx gen_ashlhi3 (rtx, rtx, rtx); +extern rtx gen_rotrsi3 (rtx, rtx, rtx); +extern rtx gen_rotrdi3 (rtx, rtx, rtx); +extern rtx gen_rotlsi3 (rtx, rtx, rtx); +extern rtx gen_rotldi3 (rtx, rtx, rtx); +extern rtx gen_extv (rtx, rtx, rtx, rtx); +extern rtx gen_extzv (rtx, rtx, rtx, rtx); +extern rtx gen_insvsi (rtx, rtx, rtx, rtx); +extern rtx gen_insvdi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_reload_movti (rtx, rtx, rtx); +extern rtx gen_aarch64_reload_movtf (rtx, rtx, rtx); +extern rtx gen_add_losym (rtx, rtx, rtx); +extern rtx gen_tlsgd_small (rtx, rtx); +extern rtx gen_tlsle_small (rtx, rtx, rtx); +extern rtx gen_get_thread_pointerdi (rtx); +extern rtx gen_stack_protect_set (rtx, rtx); +extern rtx gen_stack_protect_test (rtx, rtx, rtx); +extern rtx gen_movv8qi (rtx, rtx); +extern rtx gen_movv16qi (rtx, rtx); +extern rtx gen_movv4hi (rtx, rtx); +extern rtx gen_movv8hi (rtx, rtx); +extern rtx gen_movv2si (rtx, rtx); +extern rtx gen_movv4si (rtx, rtx); +extern rtx gen_movv2di (rtx, rtx); +extern rtx gen_movv2sf (rtx, rtx); +extern rtx gen_movv4sf (rtx, rtx); +extern rtx gen_movv2df (rtx, rtx); +extern rtx gen_movmisalignv8qi (rtx, rtx); +extern rtx gen_movmisalignv16qi (rtx, rtx); +extern rtx gen_movmisalignv4hi (rtx, rtx); +extern rtx gen_movmisalignv8hi (rtx, rtx); +extern rtx gen_movmisalignv2si (rtx, rtx); +extern rtx gen_movmisalignv4si (rtx, rtx); +extern rtx gen_movmisalignv2di (rtx, rtx); +extern rtx gen_movmisalignv2sf (rtx, rtx); +extern rtx gen_movmisalignv4sf (rtx, rtx); +extern rtx gen_movmisalignv2df (rtx, rtx); +extern rtx gen_aarch64_split_simd_movv16qi (rtx, rtx); +extern rtx gen_aarch64_split_simd_movv8hi (rtx, rtx); +extern rtx gen_aarch64_split_simd_movv4si (rtx, rtx); +extern rtx gen_aarch64_split_simd_movv2di (rtx, rtx); +extern rtx gen_aarch64_split_simd_movv4sf (rtx, rtx); +extern rtx gen_aarch64_split_simd_movv2df (rtx, rtx); +extern rtx gen_ashlv8qi3 (rtx, rtx, rtx); +extern rtx gen_ashlv16qi3 (rtx, rtx, rtx); +extern rtx gen_ashlv4hi3 (rtx, rtx, rtx); +extern rtx gen_ashlv8hi3 (rtx, rtx, rtx); +extern rtx gen_ashlv2si3 (rtx, rtx, rtx); +extern rtx gen_ashlv4si3 (rtx, rtx, rtx); +extern rtx gen_ashlv2di3 (rtx, rtx, rtx); +extern rtx gen_lshrv8qi3 (rtx, rtx, rtx); +extern rtx gen_lshrv16qi3 (rtx, rtx, rtx); +extern rtx gen_lshrv4hi3 (rtx, rtx, rtx); +extern rtx gen_lshrv8hi3 (rtx, rtx, rtx); +extern rtx gen_lshrv2si3 (rtx, rtx, rtx); +extern rtx gen_lshrv4si3 (rtx, rtx, rtx); +extern rtx gen_lshrv2di3 (rtx, rtx, rtx); +extern rtx gen_ashrv8qi3 (rtx, rtx, rtx); +extern rtx gen_ashrv16qi3 (rtx, rtx, rtx); +extern rtx gen_ashrv4hi3 (rtx, rtx, rtx); +extern rtx gen_ashrv8hi3 (rtx, rtx, rtx); +extern rtx gen_ashrv2si3 (rtx, rtx, rtx); +extern rtx gen_ashrv4si3 (rtx, rtx, rtx); +extern rtx gen_ashrv2di3 (rtx, rtx, rtx); +extern rtx gen_vashlv8qi3 (rtx, rtx, rtx); +extern rtx gen_vashlv16qi3 (rtx, rtx, rtx); +extern rtx gen_vashlv4hi3 (rtx, rtx, rtx); +extern rtx gen_vashlv8hi3 (rtx, rtx, rtx); +extern rtx gen_vashlv2si3 (rtx, rtx, rtx); +extern rtx gen_vashlv4si3 (rtx, rtx, rtx); +extern rtx gen_vashlv2di3 (rtx, rtx, rtx); +extern rtx gen_vashrv8qi3 (rtx, rtx, rtx); +extern rtx gen_vashrv16qi3 (rtx, rtx, rtx); +extern rtx gen_vashrv4hi3 (rtx, rtx, rtx); +extern rtx gen_vashrv8hi3 (rtx, rtx, rtx); +extern rtx gen_vashrv2si3 (rtx, rtx, rtx); +extern rtx gen_vashrv4si3 (rtx, rtx, rtx); +extern rtx gen_aarch64_ashr_simddi (rtx, rtx, rtx); +extern rtx gen_vlshrv8qi3 (rtx, rtx, rtx); +extern rtx gen_vlshrv16qi3 (rtx, rtx, rtx); +extern rtx gen_vlshrv4hi3 (rtx, rtx, rtx); +extern rtx gen_vlshrv8hi3 (rtx, rtx, rtx); +extern rtx gen_vlshrv2si3 (rtx, rtx, rtx); +extern rtx gen_vlshrv4si3 (rtx, rtx, rtx); +extern rtx gen_aarch64_lshr_simddi (rtx, rtx, rtx); +extern rtx gen_vec_setv8qi (rtx, rtx, rtx); +extern rtx gen_vec_setv16qi (rtx, rtx, rtx); +extern rtx gen_vec_setv4hi (rtx, rtx, rtx); +extern rtx gen_vec_setv8hi (rtx, rtx, rtx); +extern rtx gen_vec_setv2si (rtx, rtx, rtx); +extern rtx gen_vec_setv4si (rtx, rtx, rtx); +extern rtx gen_vec_setv2di (rtx, rtx, rtx); +extern rtx gen_vec_setv2sf (rtx, rtx, rtx); +extern rtx gen_vec_setv4sf (rtx, rtx, rtx); +extern rtx gen_vec_setv2df (rtx, rtx, rtx); +extern rtx gen_move_lo_quad_v16qi (rtx, rtx); +extern rtx gen_move_lo_quad_v8hi (rtx, rtx); +extern rtx gen_move_lo_quad_v4si (rtx, rtx); +extern rtx gen_move_lo_quad_v2di (rtx, rtx); +extern rtx gen_move_lo_quad_v4sf (rtx, rtx); +extern rtx gen_move_lo_quad_v2df (rtx, rtx); +extern rtx gen_move_hi_quad_v16qi (rtx, rtx); +extern rtx gen_move_hi_quad_v8hi (rtx, rtx); +extern rtx gen_move_hi_quad_v4si (rtx, rtx); +extern rtx gen_move_hi_quad_v2di (rtx, rtx); +extern rtx gen_move_hi_quad_v4sf (rtx, rtx); +extern rtx gen_move_hi_quad_v2df (rtx, rtx); +extern rtx gen_vec_pack_trunc_v4hi (rtx, rtx, rtx); +extern rtx gen_vec_pack_trunc_v2si (rtx, rtx, rtx); +extern rtx gen_vec_pack_trunc_di (rtx, rtx, rtx); +extern rtx gen_vec_unpacks_hi_v16qi (rtx, rtx); +extern rtx gen_vec_unpacku_hi_v16qi (rtx, rtx); +extern rtx gen_vec_unpacks_hi_v8hi (rtx, rtx); +extern rtx gen_vec_unpacku_hi_v8hi (rtx, rtx); +extern rtx gen_vec_unpacks_hi_v4si (rtx, rtx); +extern rtx gen_vec_unpacku_hi_v4si (rtx, rtx); +extern rtx gen_vec_unpacks_lo_v16qi (rtx, rtx); +extern rtx gen_vec_unpacku_lo_v16qi (rtx, rtx); +extern rtx gen_vec_unpacks_lo_v8hi (rtx, rtx); +extern rtx gen_vec_unpacku_lo_v8hi (rtx, rtx); +extern rtx gen_vec_unpacks_lo_v4si (rtx, rtx); +extern rtx gen_vec_unpacku_lo_v4si (rtx, rtx); +extern rtx gen_vec_widen_smult_lo_v16qi (rtx, rtx, rtx); +extern rtx gen_vec_widen_umult_lo_v16qi (rtx, rtx, rtx); +extern rtx gen_vec_widen_smult_lo_v8hi (rtx, rtx, rtx); +extern rtx gen_vec_widen_umult_lo_v8hi (rtx, rtx, rtx); +extern rtx gen_vec_widen_smult_lo_v4si (rtx, rtx, rtx); +extern rtx gen_vec_widen_umult_lo_v4si (rtx, rtx, rtx); +extern rtx gen_vec_widen_smult_hi_v16qi (rtx, rtx, rtx); +extern rtx gen_vec_widen_umult_hi_v16qi (rtx, rtx, rtx); +extern rtx gen_vec_widen_smult_hi_v8hi (rtx, rtx, rtx); +extern rtx gen_vec_widen_umult_hi_v8hi (rtx, rtx, rtx); +extern rtx gen_vec_widen_smult_hi_v4si (rtx, rtx, rtx); +extern rtx gen_vec_widen_umult_hi_v4si (rtx, rtx, rtx); +extern rtx gen_fixv2sfv2si2 (rtx, rtx); +extern rtx gen_fixunsv2sfv2si2 (rtx, rtx); +extern rtx gen_fixv4sfv4si2 (rtx, rtx); +extern rtx gen_fixunsv4sfv4si2 (rtx, rtx); +extern rtx gen_fixv2dfv2di2 (rtx, rtx); +extern rtx gen_fixunsv2dfv2di2 (rtx, rtx); +extern rtx gen_fix_truncv2sfv2si2 (rtx, rtx); +extern rtx gen_fixuns_truncv2sfv2si2 (rtx, rtx); +extern rtx gen_fix_truncv4sfv4si2 (rtx, rtx); +extern rtx gen_fixuns_truncv4sfv4si2 (rtx, rtx); +extern rtx gen_fix_truncv2dfv2di2 (rtx, rtx); +extern rtx gen_fixuns_truncv2dfv2di2 (rtx, rtx); +extern rtx gen_ftruncv2sf2 (rtx, rtx); +extern rtx gen_ftruncv4sf2 (rtx, rtx); +extern rtx gen_ftruncv2df2 (rtx, rtx); +extern rtx gen_vec_pack_trunc_v2df (rtx, rtx, rtx); +extern rtx gen_vec_pack_trunc_df (rtx, rtx, rtx); +extern rtx gen_reduc_splus_v4sf (rtx, rtx); +extern rtx gen_aarch64_simd_bslv8qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv2sf (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv4sf (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bslv2df (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bsldi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_simd_bsldf (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv8qiv8qi (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv16qiv16qi (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv4hiv4hi (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv8hiv8hi (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv2siv2si (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv4siv4si (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv2div2di (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv2sfv2sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv2siv2sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv4sfv2sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv4siv2sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv2dfv2sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv2div2sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv2sfv4sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv2siv4sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv4sfv4sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv4siv4sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv2dfv4sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv2div4sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv2sfv2df (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv2siv2df (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv4sfv2df (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv4siv2df (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv2dfv2df (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_vcond_internalv2div2df (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vcondv8qiv8qi (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vcondv16qiv16qi (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vcondv4hiv4hi (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vcondv8hiv8hi (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vcondv2siv2si (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vcondv4siv4si (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vcondv2div2di (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vcondv2sfv2sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vcondv4sfv4sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vcondv2dfv2df (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vcondv2siv2sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vcondv4siv4sf (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vcondv2div2df (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vconduv8qiv8qi (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vconduv16qiv16qi (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vconduv4hiv4hi (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vconduv8hiv8hi (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vconduv2siv2si (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vconduv4siv4si (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_vconduv2div2di (rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_createv8qi (rtx, rtx); +extern rtx gen_aarch64_createv4hi (rtx, rtx); +extern rtx gen_aarch64_createv2si (rtx, rtx); +extern rtx gen_aarch64_createdi (rtx, rtx); +extern rtx gen_aarch64_createdf (rtx, rtx); +extern rtx gen_aarch64_createv2sf (rtx, rtx); +extern rtx gen_aarch64_be_checked_get_lanev8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_be_checked_get_lanev16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_be_checked_get_lanev4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_be_checked_get_lanev8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_be_checked_get_lanev2si (rtx, rtx, rtx); +extern rtx gen_aarch64_be_checked_get_lanev4si (rtx, rtx, rtx); +extern rtx gen_aarch64_be_checked_get_lanev2di (rtx, rtx, rtx); +extern rtx gen_aarch64_be_checked_get_lanev2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_be_checked_get_lanev4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_be_checked_get_lanev2df (rtx, rtx, rtx); +extern rtx gen_aarch64_get_lanedi (rtx, rtx, rtx); +extern rtx gen_aarch64_reinterpretv8qiv8qi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv8qiv4hi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv8qiv2si (rtx, rtx); +extern rtx gen_aarch64_reinterpretv8qiv2sf (rtx, rtx); +extern rtx gen_aarch64_reinterpretv8qidi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv8qidf (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4hiv8qi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4hiv4hi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4hiv2si (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4hiv2sf (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4hidi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4hidf (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2siv8qi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2siv4hi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2siv2si (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2siv2sf (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2sidi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2sidf (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2sfv8qi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2sfv4hi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2sfv2si (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2sfv2sf (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2sfdi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2sfdf (rtx, rtx); +extern rtx gen_aarch64_reinterpretdiv8qi (rtx, rtx); +extern rtx gen_aarch64_reinterpretdiv4hi (rtx, rtx); +extern rtx gen_aarch64_reinterpretdiv2si (rtx, rtx); +extern rtx gen_aarch64_reinterpretdidi (rtx, rtx); +extern rtx gen_aarch64_reinterpretdidf (rtx, rtx); +extern rtx gen_aarch64_reinterpretdiv2sf (rtx, rtx); +extern rtx gen_aarch64_reinterpretdfv8qi (rtx, rtx); +extern rtx gen_aarch64_reinterpretdfv4hi (rtx, rtx); +extern rtx gen_aarch64_reinterpretdfv2si (rtx, rtx); +extern rtx gen_aarch64_reinterpretdfdi (rtx, rtx); +extern rtx gen_aarch64_reinterpretdfdf (rtx, rtx); +extern rtx gen_aarch64_reinterpretdfv2sf (rtx, rtx); +extern rtx gen_aarch64_reinterpretv16qiv16qi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv16qiv8hi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv16qiv4si (rtx, rtx); +extern rtx gen_aarch64_reinterpretv16qiv2di (rtx, rtx); +extern rtx gen_aarch64_reinterpretv16qiv4sf (rtx, rtx); +extern rtx gen_aarch64_reinterpretv16qiv2df (rtx, rtx); +extern rtx gen_aarch64_reinterpretv8hiv16qi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv8hiv8hi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv8hiv4si (rtx, rtx); +extern rtx gen_aarch64_reinterpretv8hiv2di (rtx, rtx); +extern rtx gen_aarch64_reinterpretv8hiv4sf (rtx, rtx); +extern rtx gen_aarch64_reinterpretv8hiv2df (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4siv16qi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4siv8hi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4siv4si (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4siv2di (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4siv4sf (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4siv2df (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4sfv16qi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4sfv8hi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4sfv4si (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4sfv2di (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4sfv4sf (rtx, rtx); +extern rtx gen_aarch64_reinterpretv4sfv2df (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2div16qi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2div8hi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2div4si (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2div2di (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2div4sf (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2div2df (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2dfv16qi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2dfv8hi (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2dfv4si (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2dfv2di (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2dfv4sf (rtx, rtx); +extern rtx gen_aarch64_reinterpretv2dfv2df (rtx, rtx); +extern rtx gen_aarch64_combinev8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_combinev4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_combinev2si (rtx, rtx, rtx); +extern rtx gen_aarch64_combinev2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_combinedi (rtx, rtx, rtx); +extern rtx gen_aarch64_combinedf (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_combinev8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_combinev4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_combinev2si (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_combinev2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_combinedi (rtx, rtx, rtx); +extern rtx gen_aarch64_simd_combinedf (rtx, rtx, rtx); +extern rtx gen_aarch64_saddl2v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_saddl2v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_saddl2v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_uaddl2v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uaddl2v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uaddl2v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_ssubl2v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_ssubl2v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_ssubl2v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_usubl2v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_usubl2v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_usubl2v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_saddw2v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_saddw2v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_saddw2v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_uaddw2v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_uaddw2v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_uaddw2v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_ssubw2v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_ssubw2v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_ssubw2v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_usubw2v16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_usubw2v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_usubw2v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_lanev4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_lanev8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_lanev2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_lanev4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_lanev4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_lanev8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_lanev2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_lanev4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_laneqv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_laneqv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_laneqv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_laneqv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_laneqv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_laneqv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_laneqv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_laneqv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_lanehi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmulh_lanesi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_lanehi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqrdmulh_lanesi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_lanev4hi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_lanev2si (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_lanehi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_lanesi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_laneqv4hi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_laneqv2si (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_laneqhi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal_laneqsi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_lanev4hi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_lanev2si (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_lanehi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_lanesi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_laneqv4hi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_laneqv2si (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_laneqhi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl_laneqsi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2v8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2v4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2v8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2v4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2_lanev8hi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2_lanev4si (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2_laneqv8hi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2_laneqv4si (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2_lanev8hi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2_lanev4si (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2_laneqv8hi (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2_laneqv4si (rtx, rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2_nv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlal2_nv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2_nv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmlsl2_nv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_lanev4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_lanev2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_lanehi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_lanesi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_laneqv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull_laneqv2si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2v8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2v4si (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2_lanev8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2_lanev4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2_laneqv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2_laneqv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2_nv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_sqdmull2_nv4si (rtx, rtx, rtx); +extern rtx gen_movoi (rtx, rtx); +extern rtx gen_movci (rtx, rtx); +extern rtx gen_movxi (rtx, rtx); +extern rtx gen_aarch64_ld2v8qi (rtx, rtx); +extern rtx gen_aarch64_ld2v4hi (rtx, rtx); +extern rtx gen_aarch64_ld2v2si (rtx, rtx); +extern rtx gen_aarch64_ld2v2sf (rtx, rtx); +extern rtx gen_aarch64_ld2di (rtx, rtx); +extern rtx gen_aarch64_ld2df (rtx, rtx); +extern rtx gen_aarch64_ld3v8qi (rtx, rtx); +extern rtx gen_aarch64_ld3v4hi (rtx, rtx); +extern rtx gen_aarch64_ld3v2si (rtx, rtx); +extern rtx gen_aarch64_ld3v2sf (rtx, rtx); +extern rtx gen_aarch64_ld3di (rtx, rtx); +extern rtx gen_aarch64_ld3df (rtx, rtx); +extern rtx gen_aarch64_ld4v8qi (rtx, rtx); +extern rtx gen_aarch64_ld4v4hi (rtx, rtx); +extern rtx gen_aarch64_ld4v2si (rtx, rtx); +extern rtx gen_aarch64_ld4v2sf (rtx, rtx); +extern rtx gen_aarch64_ld4di (rtx, rtx); +extern rtx gen_aarch64_ld4df (rtx, rtx); +extern rtx gen_aarch64_ld1v8qi (rtx, rtx); +extern rtx gen_aarch64_ld1v16qi (rtx, rtx); +extern rtx gen_aarch64_ld1v4hi (rtx, rtx); +extern rtx gen_aarch64_ld1v8hi (rtx, rtx); +extern rtx gen_aarch64_ld1v2si (rtx, rtx); +extern rtx gen_aarch64_ld1v4si (rtx, rtx); +extern rtx gen_aarch64_ld1v2di (rtx, rtx); +extern rtx gen_aarch64_ld1v2sf (rtx, rtx); +extern rtx gen_aarch64_ld1v4sf (rtx, rtx); +extern rtx gen_aarch64_ld1v2df (rtx, rtx); +extern rtx gen_aarch64_ld2v16qi (rtx, rtx); +extern rtx gen_aarch64_ld3v16qi (rtx, rtx); +extern rtx gen_aarch64_ld4v16qi (rtx, rtx); +extern rtx gen_aarch64_ld2v8hi (rtx, rtx); +extern rtx gen_aarch64_ld3v8hi (rtx, rtx); +extern rtx gen_aarch64_ld4v8hi (rtx, rtx); +extern rtx gen_aarch64_ld2v4si (rtx, rtx); +extern rtx gen_aarch64_ld3v4si (rtx, rtx); +extern rtx gen_aarch64_ld4v4si (rtx, rtx); +extern rtx gen_aarch64_ld2v2di (rtx, rtx); +extern rtx gen_aarch64_ld3v2di (rtx, rtx); +extern rtx gen_aarch64_ld4v2di (rtx, rtx); +extern rtx gen_aarch64_ld2v4sf (rtx, rtx); +extern rtx gen_aarch64_ld3v4sf (rtx, rtx); +extern rtx gen_aarch64_ld4v4sf (rtx, rtx); +extern rtx gen_aarch64_ld2v2df (rtx, rtx); +extern rtx gen_aarch64_ld3v2df (rtx, rtx); +extern rtx gen_aarch64_ld4v2df (rtx, rtx); +extern rtx gen_aarch64_get_dregoiv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregoiv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregoiv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregoiv2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregoidi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregoidf (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregciv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregciv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregciv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregciv2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregcidi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregcidf (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregxiv8qi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregxiv4hi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregxiv2si (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregxiv2sf (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregxidi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_dregxidf (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregoiv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregciv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregxiv16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregoiv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregciv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregxiv8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregoiv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregciv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregxiv4si (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregoiv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregciv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregxiv2di (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregoiv4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregciv4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregxiv4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregoiv2df (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregciv2df (rtx, rtx, rtx); +extern rtx gen_aarch64_get_qregxiv2df (rtx, rtx, rtx); +extern rtx gen_vec_perm_constv8qi (rtx, rtx, rtx, rtx); +extern rtx gen_vec_perm_constv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_vec_perm_constv4hi (rtx, rtx, rtx, rtx); +extern rtx gen_vec_perm_constv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_vec_perm_constv2si (rtx, rtx, rtx, rtx); +extern rtx gen_vec_perm_constv4si (rtx, rtx, rtx, rtx); +extern rtx gen_vec_perm_constv2di (rtx, rtx, rtx, rtx); +extern rtx gen_vec_perm_constv2sf (rtx, rtx, rtx, rtx); +extern rtx gen_vec_perm_constv4sf (rtx, rtx, rtx, rtx); +extern rtx gen_vec_perm_constv2df (rtx, rtx, rtx, rtx); +extern rtx gen_vec_permv8qi (rtx, rtx, rtx, rtx); +extern rtx gen_vec_permv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_im_lane_boundsi (rtx, rtx); +extern rtx gen_aarch64_st2v8qi (rtx, rtx); +extern rtx gen_aarch64_st2v4hi (rtx, rtx); +extern rtx gen_aarch64_st2v2si (rtx, rtx); +extern rtx gen_aarch64_st2v2sf (rtx, rtx); +extern rtx gen_aarch64_st2di (rtx, rtx); +extern rtx gen_aarch64_st2df (rtx, rtx); +extern rtx gen_aarch64_st3v8qi (rtx, rtx); +extern rtx gen_aarch64_st3v4hi (rtx, rtx); +extern rtx gen_aarch64_st3v2si (rtx, rtx); +extern rtx gen_aarch64_st3v2sf (rtx, rtx); +extern rtx gen_aarch64_st3di (rtx, rtx); +extern rtx gen_aarch64_st3df (rtx, rtx); +extern rtx gen_aarch64_st4v8qi (rtx, rtx); +extern rtx gen_aarch64_st4v4hi (rtx, rtx); +extern rtx gen_aarch64_st4v2si (rtx, rtx); +extern rtx gen_aarch64_st4v2sf (rtx, rtx); +extern rtx gen_aarch64_st4di (rtx, rtx); +extern rtx gen_aarch64_st4df (rtx, rtx); +extern rtx gen_aarch64_st2v16qi (rtx, rtx); +extern rtx gen_aarch64_st3v16qi (rtx, rtx); +extern rtx gen_aarch64_st4v16qi (rtx, rtx); +extern rtx gen_aarch64_st2v8hi (rtx, rtx); +extern rtx gen_aarch64_st3v8hi (rtx, rtx); +extern rtx gen_aarch64_st4v8hi (rtx, rtx); +extern rtx gen_aarch64_st2v4si (rtx, rtx); +extern rtx gen_aarch64_st3v4si (rtx, rtx); +extern rtx gen_aarch64_st4v4si (rtx, rtx); +extern rtx gen_aarch64_st2v2di (rtx, rtx); +extern rtx gen_aarch64_st3v2di (rtx, rtx); +extern rtx gen_aarch64_st4v2di (rtx, rtx); +extern rtx gen_aarch64_st2v4sf (rtx, rtx); +extern rtx gen_aarch64_st3v4sf (rtx, rtx); +extern rtx gen_aarch64_st4v4sf (rtx, rtx); +extern rtx gen_aarch64_st2v2df (rtx, rtx); +extern rtx gen_aarch64_st3v2df (rtx, rtx); +extern rtx gen_aarch64_st4v2df (rtx, rtx); +extern rtx gen_aarch64_st2_lanev16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_st2_lanev8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_st2_lanev4si (rtx, rtx, rtx); +extern rtx gen_aarch64_st2_lanev2di (rtx, rtx, rtx); +extern rtx gen_aarch64_st2_lanev4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_st2_lanev2df (rtx, rtx, rtx); +extern rtx gen_aarch64_st3_lanev16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_st3_lanev8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_st3_lanev4si (rtx, rtx, rtx); +extern rtx gen_aarch64_st3_lanev2di (rtx, rtx, rtx); +extern rtx gen_aarch64_st3_lanev4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_st3_lanev2df (rtx, rtx, rtx); +extern rtx gen_aarch64_st4_lanev16qi (rtx, rtx, rtx); +extern rtx gen_aarch64_st4_lanev8hi (rtx, rtx, rtx); +extern rtx gen_aarch64_st4_lanev4si (rtx, rtx, rtx); +extern rtx gen_aarch64_st4_lanev2di (rtx, rtx, rtx); +extern rtx gen_aarch64_st4_lanev4sf (rtx, rtx, rtx); +extern rtx gen_aarch64_st4_lanev2df (rtx, rtx, rtx); +extern rtx gen_aarch64_st1v8qi (rtx, rtx); +extern rtx gen_aarch64_st1v16qi (rtx, rtx); +extern rtx gen_aarch64_st1v4hi (rtx, rtx); +extern rtx gen_aarch64_st1v8hi (rtx, rtx); +extern rtx gen_aarch64_st1v2si (rtx, rtx); +extern rtx gen_aarch64_st1v4si (rtx, rtx); +extern rtx gen_aarch64_st1v2di (rtx, rtx); +extern rtx gen_aarch64_st1v2sf (rtx, rtx); +extern rtx gen_aarch64_st1v4sf (rtx, rtx); +extern rtx gen_aarch64_st1v2df (rtx, rtx); +extern rtx gen_aarch64_set_qregoiv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregciv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregxiv16qi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregoiv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregciv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregxiv8hi (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregoiv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregciv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregxiv4si (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregoiv2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregciv2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregxiv2di (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregoiv4sf (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregciv4sf (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregxiv4sf (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregoiv2df (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregciv2df (rtx, rtx, rtx, rtx); +extern rtx gen_aarch64_set_qregxiv2df (rtx, rtx, rtx, rtx); +extern rtx gen_vec_initv8qi (rtx, rtx); +extern rtx gen_vec_initv16qi (rtx, rtx); +extern rtx gen_vec_initv4hi (rtx, rtx); +extern rtx gen_vec_initv8hi (rtx, rtx); +extern rtx gen_vec_initv2si (rtx, rtx); +extern rtx gen_vec_initv4si (rtx, rtx); +extern rtx gen_vec_initv2di (rtx, rtx); +extern rtx gen_vec_initv2sf (rtx, rtx); +extern rtx gen_vec_initv4sf (rtx, rtx); +extern rtx gen_vec_initv2df (rtx, rtx); +extern rtx gen_vec_extractv8qi (rtx, rtx, rtx); +extern rtx gen_vec_extractv16qi (rtx, rtx, rtx); +extern rtx gen_vec_extractv4hi (rtx, rtx, rtx); +extern rtx gen_vec_extractv8hi (rtx, rtx, rtx); +extern rtx gen_vec_extractv2si (rtx, rtx, rtx); +extern rtx gen_vec_extractv4si (rtx, rtx, rtx); +extern rtx gen_vec_extractv2di (rtx, rtx, rtx); +extern rtx gen_vec_extractv2sf (rtx, rtx, rtx); +extern rtx gen_vec_extractv4sf (rtx, rtx, rtx); +extern rtx gen_vec_extractv2df (rtx, rtx, rtx); +extern rtx gen_atomic_compare_and_swapqi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_atomic_compare_and_swaphi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_atomic_compare_and_swapsi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_atomic_compare_and_swapdi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx); +extern rtx gen_mem_thread_fence (rtx); +extern rtx gen_dmb (rtx); #endif /* GCC_INSN_FLAGS_H */ diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-modes.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-modes.h index e0d2258..e6b5b0a 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-modes.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-modes.h @@ -22,10 +22,10 @@ enum machine_mode SImode, /* machmode.def:189 */ DImode, /* machmode.def:190 */ TImode, /* machmode.def:191 */ - EImode, /* config/aarch64/aarch64-modes.def:40 */ - OImode, /* config/aarch64/aarch64-modes.def:36 */ - CImode, /* config/aarch64/aarch64-modes.def:41 */ - XImode, /* config/aarch64/aarch64-modes.def:42 */ + EImode, /* config/aarch64/aarch64-modes.def:41 */ + OImode, /* config/aarch64/aarch64-modes.def:37 */ + CImode, /* config/aarch64/aarch64-modes.def:42 */ + XImode, /* config/aarch64/aarch64-modes.def:43 */ QQmode, /* machmode.def:210 */ HQmode, /* machmode.def:211 */ SQmode, /* machmode.def:212 */ @@ -46,7 +46,7 @@ enum machine_mode UTAmode, /* machmode.def:230 */ SFmode, /* machmode.def:202 */ DFmode, /* machmode.def:203 */ - TFmode, /* config/aarch64/aarch64-modes.def:55 */ + TFmode, /* config/aarch64/aarch64-modes.def:56 */ SDmode, /* machmode.def:242 */ DDmode, /* machmode.def:243 */ TDmode, /* machmode.def:244 */ @@ -69,32 +69,33 @@ enum machine_mode V8HImode, /* config/aarch64/aarch64-modes.def:31 */ V4SImode, /* config/aarch64/aarch64-modes.def:31 */ V2DImode, /* config/aarch64/aarch64-modes.def:31 */ - V32QImode, /* config/aarch64/aarch64-modes.def:45 */ - V16HImode, /* config/aarch64/aarch64-modes.def:45 */ - V8SImode, /* config/aarch64/aarch64-modes.def:45 */ - V4DImode, /* config/aarch64/aarch64-modes.def:45 */ - V2TImode, /* config/aarch64/aarch64-modes.def:45 */ - V48QImode, /* config/aarch64/aarch64-modes.def:48 */ - V24HImode, /* config/aarch64/aarch64-modes.def:48 */ - V12SImode, /* config/aarch64/aarch64-modes.def:48 */ - V6DImode, /* config/aarch64/aarch64-modes.def:48 */ - V3TImode, /* config/aarch64/aarch64-modes.def:48 */ - V2EImode, /* config/aarch64/aarch64-modes.def:48 */ - V64QImode, /* config/aarch64/aarch64-modes.def:51 */ - V32HImode, /* config/aarch64/aarch64-modes.def:51 */ - V16SImode, /* config/aarch64/aarch64-modes.def:51 */ - V8DImode, /* config/aarch64/aarch64-modes.def:51 */ - V4TImode, /* config/aarch64/aarch64-modes.def:51 */ - V2OImode, /* config/aarch64/aarch64-modes.def:51 */ + V32QImode, /* config/aarch64/aarch64-modes.def:46 */ + V16HImode, /* config/aarch64/aarch64-modes.def:46 */ + V8SImode, /* config/aarch64/aarch64-modes.def:46 */ + V4DImode, /* config/aarch64/aarch64-modes.def:46 */ + V2TImode, /* config/aarch64/aarch64-modes.def:46 */ + V48QImode, /* config/aarch64/aarch64-modes.def:49 */ + V24HImode, /* config/aarch64/aarch64-modes.def:49 */ + V12SImode, /* config/aarch64/aarch64-modes.def:49 */ + V6DImode, /* config/aarch64/aarch64-modes.def:49 */ + V3TImode, /* config/aarch64/aarch64-modes.def:49 */ + V2EImode, /* config/aarch64/aarch64-modes.def:49 */ + V64QImode, /* config/aarch64/aarch64-modes.def:52 */ + V32HImode, /* config/aarch64/aarch64-modes.def:52 */ + V16SImode, /* config/aarch64/aarch64-modes.def:52 */ + V8DImode, /* config/aarch64/aarch64-modes.def:52 */ + V4TImode, /* config/aarch64/aarch64-modes.def:52 */ + V2OImode, /* config/aarch64/aarch64-modes.def:52 */ V2SFmode, /* config/aarch64/aarch64-modes.def:32 */ + V1DFmode, /* config/aarch64/aarch64-modes.def:34 */ V4SFmode, /* config/aarch64/aarch64-modes.def:33 */ V2DFmode, /* config/aarch64/aarch64-modes.def:33 */ - V8SFmode, /* config/aarch64/aarch64-modes.def:46 */ - V4DFmode, /* config/aarch64/aarch64-modes.def:46 */ - V12SFmode, /* config/aarch64/aarch64-modes.def:49 */ - V6DFmode, /* config/aarch64/aarch64-modes.def:49 */ - V16SFmode, /* config/aarch64/aarch64-modes.def:52 */ - V8DFmode, /* config/aarch64/aarch64-modes.def:52 */ + V8SFmode, /* config/aarch64/aarch64-modes.def:47 */ + V4DFmode, /* config/aarch64/aarch64-modes.def:47 */ + V12SFmode, /* config/aarch64/aarch64-modes.def:50 */ + V6DFmode, /* config/aarch64/aarch64-modes.def:50 */ + V16SFmode, /* config/aarch64/aarch64-modes.def:53 */ + V8DFmode, /* config/aarch64/aarch64-modes.def:53 */ MAX_MACHINE_MODE, MIN_MODE_RANDOM = VOIDmode, diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/ipa-prop.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/ipa-prop.h index 8fdd92c..70185b2 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/ipa-prop.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/ipa-prop.h @@ -585,6 +585,7 @@ tree ipa_get_indirect_edge_target (struct cgraph_edge *ie, struct cgraph_edge *ipa_make_edge_direct_to_target (struct cgraph_edge *, tree); tree ipa_binfo_from_known_type_jfunc (struct ipa_jump_func *); tree ipa_intraprocedural_devirtualization (gimple); +tree ipa_impossible_devirt_target (struct cgraph_edge *, tree); /* Functions related to both. */ void ipa_analyze_node (struct cgraph_node *); diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/langhooks.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/langhooks.h index c848b0c..b5997ee 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/langhooks.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/langhooks.h @@ -219,12 +219,16 @@ struct lang_hooks_for_decls /* Similarly, except use an assignment operator instead. */ tree (*omp_clause_assign_op) (tree clause, tree dst, tree src); + /* Build and return code for a constructor of DST that sets it to + SRC + ADD. */ + tree (*omp_clause_linear_ctor) (tree clause, tree dst, tree src, tree add); + /* Build and return code destructing DECL. Return NULL if nothing to be done. */ tree (*omp_clause_dtor) (tree clause, tree decl); /* Do language specific checking on an implicitly determined clause. */ - void (*omp_finish_clause) (tree clause); + void (*omp_finish_clause) (tree clause, gimple_seq *pre_p); }; /* Language hooks related to LTO serialization. */ diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/plugin-version.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/plugin-version.h index 653de0e..591112e 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/plugin-version.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/plugin-version.h @@ -6,7 +6,7 @@ #define GCCPLUGIN_VERSION (GCCPLUGIN_VERSION_MAJOR*1000 + GCCPLUGIN_VERSION_MINOR) static char basever[] = "4.9.1"; -static char datestamp[] = "20140529"; +static char datestamp[] = "20140710"; static char devphase[] = "prerelease"; static char revision[] = ""; diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tm-preds.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tm-preds.h index e31cbef..127ca26 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tm-preds.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tm-preds.h @@ -1,5 +1,5 @@ /* Generated automatically by the program 'build/genpreds' - from the machine description file '/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/config/aarch64/aarch64.md'. */ + from the machine description file '/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/config/aarch64/aarch64.md'. */ #ifndef GCC_TM_PREDS_H #define GCC_TM_PREDS_H @@ -22,6 +22,7 @@ extern int indirect_operand (rtx, enum machine_mode); extern int ordered_comparison_operator (rtx, enum machine_mode); extern int comparison_operator (rtx, enum machine_mode); extern int cc_register (rtx, enum machine_mode); +extern int aarch64_call_insn_operand (rtx, enum machine_mode); extern int aarch64_simd_register (rtx, enum machine_mode); extern int aarch64_reg_or_zero (rtx, enum machine_mode); extern int aarch64_reg_or_fp_zero (rtx, enum machine_mode); @@ -67,6 +68,7 @@ enum constraint_num { CONSTRAINT__UNKNOWN = 0, CONSTRAINT_k, + CONSTRAINT_Ucs, CONSTRAINT_w, CONSTRAINT_x, CONSTRAINT_I, @@ -81,6 +83,7 @@ enum constraint_num CONSTRAINT_Ush, CONSTRAINT_Uss, CONSTRAINT_Usd, + CONSTRAINT_Usf, CONSTRAINT_UsM, CONSTRAINT_Ui1, CONSTRAINT_Ui3, diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree-core.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree-core.h index 1719c7e..40443d4 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree-core.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree-core.h @@ -987,7 +987,7 @@ struct GTY(()) tree_base { SSA_NAME_IN_FREELIST in SSA_NAME - VAR_DECL_NONALIASED in + DECL_NONALIASED in VAR_DECL deprecated_flag: @@ -1127,6 +1127,11 @@ enum omp_clause_map_kind array sections. OMP_CLAUSE_SIZE for these is not the pointer size, which is implicitly POINTER_SIZE / BITS_PER_UNIT, but the bias. */ OMP_CLAUSE_MAP_POINTER, + /* Also internal, behaves like OMP_CLAUS_MAP_TO, but additionally any + OMP_CLAUSE_MAP_POINTER records consecutive after it which have addresses + falling into that range will not be ignored if OMP_CLAUSE_MAP_TO_PSET + wasn't mapped already. */ + OMP_CLAUSE_MAP_TO_PSET, OMP_CLAUSE_MAP_LAST }; diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree.h index 90e5e27..d73bc52 100644 --- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree.h +++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree.h @@ -1327,9 +1327,17 @@ extern void protected_set_expr_location (tree, location_t); #define OMP_CLAUSE_LINEAR_VARIABLE_STRIDE(NODE) \ TREE_PROTECTED (OMP_CLAUSE_SUBCODE_CHECK (NODE, OMP_CLAUSE_LINEAR)) +/* True if a LINEAR clause is for an array or allocatable variable that + needs special handling by the frontend. */ +#define OMP_CLAUSE_LINEAR_ARRAY(NODE) \ + (OMP_CLAUSE_SUBCODE_CHECK (NODE, OMP_CLAUSE_LINEAR)->base.deprecated_flag) + #define OMP_CLAUSE_LINEAR_STEP(NODE) \ OMP_CLAUSE_OPERAND (OMP_CLAUSE_SUBCODE_CHECK (NODE, OMP_CLAUSE_LINEAR), 1) +#define OMP_CLAUSE_LINEAR_STMT(NODE) \ + OMP_CLAUSE_OPERAND (OMP_CLAUSE_SUBCODE_CHECK (NODE, OMP_CLAUSE_LINEAR), 2) + #define OMP_CLAUSE_LINEAR_GIMPLE_SEQ(NODE) \ (OMP_CLAUSE_CHECK (NODE))->omp_clause.gimple_reduction_init |