summaryrefslogtreecommitdiff
path: root/lib/gcc/aarch64-linux-android/4.9.1
diff options
context:
space:
mode:
Diffstat (limited to 'lib/gcc/aarch64-linux-android/4.9.1')
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/crtbegin.obin3576 -> 3552 bytes
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/crtbeginS.obin3960 -> 3936 bytes
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/crtbeginT.obin3576 -> 3552 bytes
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/crtend.obin1317 -> 1317 bytes
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/crtendS.obin1317 -> 1317 bytes
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/include/arm_acle.h90
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/include/arm_neon.h3484
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/libgcc.abin799288 -> 798464 bytes
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/libgcov.abin71624 -> 72008 bytes
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/gtype.state1842
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/auto-host.h2
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/b-header-vars2
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-family/c-common.h1
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-tree.h3
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-linux.h4
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-protos.h17
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64.h26
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/configargs.h2
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/cp/cp-tree.h24
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/gimplify.h1
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-codes.h4552
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-constants.h356
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-flags.h4858
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-modes.h57
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/ipa-prop.h1
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/langhooks.h6
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/plugin-version.h2
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tm-preds.h5
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree-core.h7
-rw-r--r--lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree.h8
30 files changed, 7786 insertions, 7564 deletions
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/crtbegin.o b/lib/gcc/aarch64-linux-android/4.9.1/crtbegin.o
index ac99ac3..cf40b92 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/crtbegin.o
+++ b/lib/gcc/aarch64-linux-android/4.9.1/crtbegin.o
Binary files differ
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/crtbeginS.o b/lib/gcc/aarch64-linux-android/4.9.1/crtbeginS.o
index 09d432c..2fd7372 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/crtbeginS.o
+++ b/lib/gcc/aarch64-linux-android/4.9.1/crtbeginS.o
Binary files differ
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/crtbeginT.o b/lib/gcc/aarch64-linux-android/4.9.1/crtbeginT.o
index ac99ac3..cf40b92 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/crtbeginT.o
+++ b/lib/gcc/aarch64-linux-android/4.9.1/crtbeginT.o
Binary files differ
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/crtend.o b/lib/gcc/aarch64-linux-android/4.9.1/crtend.o
index 602fa94..26d17b0 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/crtend.o
+++ b/lib/gcc/aarch64-linux-android/4.9.1/crtend.o
Binary files differ
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/crtendS.o b/lib/gcc/aarch64-linux-android/4.9.1/crtendS.o
index 602fa94..26d17b0 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/crtendS.o
+++ b/lib/gcc/aarch64-linux-android/4.9.1/crtendS.o
Binary files differ
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/include/arm_acle.h b/lib/gcc/aarch64-linux-android/4.9.1/include/arm_acle.h
new file mode 100644
index 0000000..2e74696
--- /dev/null
+++ b/lib/gcc/aarch64-linux-android/4.9.1/include/arm_acle.h
@@ -0,0 +1,90 @@
+/* AArch64 Non-NEON ACLE intrinsics include file.
+
+ Copyright (C) 2014 Free Software Foundation, Inc.
+ Contributed by ARM Ltd.
+
+ This file is part of GCC.
+
+ GCC is free software; you can redistribute it and/or modify it
+ under the terms of the GNU General Public License as published
+ by the Free Software Foundation; either version 3, or (at your
+ option) any later version.
+
+ GCC is distributed in the hope that it will be useful, but WITHOUT
+ ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
+ or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public
+ License for more details.
+
+ Under Section 7 of GPL version 3, you are granted additional
+ permissions described in the GCC Runtime Library Exception, version
+ 3.1, as published by the Free Software Foundation.
+
+ You should have received a copy of the GNU General Public License and
+ a copy of the GCC Runtime Library Exception along with this program;
+ see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
+ <http://www.gnu.org/licenses/>. */
+
+#ifndef _GCC_ARM_ACLE_H
+#define _GCC_ARM_ACLE_H
+
+#include <stdint.h>
+#ifdef __cplusplus
+extern "C" {
+#endif
+
+#ifdef __ARM_FEATURE_CRC32
+__extension__ static __inline uint32_t __attribute__ ((__always_inline__))
+__crc32b (uint32_t __a, uint8_t __b)
+{
+ return __builtin_aarch64_crc32b (__a, __b);
+}
+
+__extension__ static __inline uint32_t __attribute__ ((__always_inline__))
+__crc32cb (uint32_t __a, uint8_t __b)
+{
+ return __builtin_aarch64_crc32cb (__a, __b);
+}
+
+__extension__ static __inline uint32_t __attribute__ ((__always_inline__))
+__crc32ch (uint32_t __a, uint16_t __b)
+{
+ return __builtin_aarch64_crc32ch (__a, __b);
+}
+
+__extension__ static __inline uint32_t __attribute__ ((__always_inline__))
+__crc32cw (uint32_t __a, uint32_t __b)
+{
+ return __builtin_aarch64_crc32cw (__a, __b);
+}
+
+__extension__ static __inline uint32_t __attribute__ ((__always_inline__))
+__crc32cd (uint32_t __a, uint64_t __b)
+{
+ return __builtin_aarch64_crc32cx (__a, __b);
+}
+
+__extension__ static __inline uint32_t __attribute__ ((__always_inline__))
+__crc32h (uint32_t __a, uint16_t __b)
+{
+ return __builtin_aarch64_crc32h (__a, __b);
+}
+
+__extension__ static __inline uint32_t __attribute__ ((__always_inline__))
+__crc32w (uint32_t __a, uint32_t __b)
+{
+ return __builtin_aarch64_crc32w (__a, __b);
+}
+
+__extension__ static __inline uint32_t __attribute__ ((__always_inline__))
+__crc32d (uint32_t __a, uint64_t __b)
+{
+ return __builtin_aarch64_crc32x (__a, __b);
+}
+
+#endif
+
+#ifdef __cplusplus
+}
+#endif
+
+#endif
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/include/arm_neon.h b/lib/gcc/aarch64-linux-android/4.9.1/include/arm_neon.h
index d4424df..dd6f43c 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/include/arm_neon.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/include/arm_neon.h
@@ -2119,29 +2119,26 @@ vqadd_s64 (int64x1_t __a, int64x1_t __b)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vqadd_u8 (uint8x8_t __a, uint8x8_t __b)
{
- return (uint8x8_t) __builtin_aarch64_uqaddv8qi ((int8x8_t) __a,
- (int8x8_t) __b);
+ return __builtin_aarch64_uqaddv8qi_uuu (__a, __b);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vqadd_u16 (uint16x4_t __a, uint16x4_t __b)
{
- return (uint16x4_t) __builtin_aarch64_uqaddv4hi ((int16x4_t) __a,
- (int16x4_t) __b);
+ return __builtin_aarch64_uqaddv4hi_uuu (__a, __b);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vqadd_u32 (uint32x2_t __a, uint32x2_t __b)
{
- return (uint32x2_t) __builtin_aarch64_uqaddv2si ((int32x2_t) __a,
- (int32x2_t) __b);
+ return __builtin_aarch64_uqaddv2si_uuu (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vqadd_u64 (uint64x1_t __a, uint64x1_t __b)
{
- return (uint64x1_t) __builtin_aarch64_uqadddi ((int64x1_t) __a,
- (int64x1_t) __b);
+ return (uint64x1_t) __builtin_aarch64_uqadddi_uuu ((uint64_t) __a,
+ (uint64_t) __b);
}
__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
@@ -2171,29 +2168,25 @@ vqaddq_s64 (int64x2_t __a, int64x2_t __b)
__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
vqaddq_u8 (uint8x16_t __a, uint8x16_t __b)
{
- return (uint8x16_t) __builtin_aarch64_uqaddv16qi ((int8x16_t) __a,
- (int8x16_t) __b);
+ return __builtin_aarch64_uqaddv16qi_uuu (__a, __b);
}
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vqaddq_u16 (uint16x8_t __a, uint16x8_t __b)
{
- return (uint16x8_t) __builtin_aarch64_uqaddv8hi ((int16x8_t) __a,
- (int16x8_t) __b);
+ return __builtin_aarch64_uqaddv8hi_uuu (__a, __b);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vqaddq_u32 (uint32x4_t __a, uint32x4_t __b)
{
- return (uint32x4_t) __builtin_aarch64_uqaddv4si ((int32x4_t) __a,
- (int32x4_t) __b);
+ return __builtin_aarch64_uqaddv4si_uuu (__a, __b);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vqaddq_u64 (uint64x2_t __a, uint64x2_t __b)
{
- return (uint64x2_t) __builtin_aarch64_uqaddv2di ((int64x2_t) __a,
- (int64x2_t) __b);
+ return __builtin_aarch64_uqaddv2di_uuu (__a, __b);
}
__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
@@ -2223,29 +2216,26 @@ vqsub_s64 (int64x1_t __a, int64x1_t __b)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vqsub_u8 (uint8x8_t __a, uint8x8_t __b)
{
- return (uint8x8_t) __builtin_aarch64_uqsubv8qi ((int8x8_t) __a,
- (int8x8_t) __b);
+ return __builtin_aarch64_uqsubv8qi_uuu (__a, __b);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vqsub_u16 (uint16x4_t __a, uint16x4_t __b)
{
- return (uint16x4_t) __builtin_aarch64_uqsubv4hi ((int16x4_t) __a,
- (int16x4_t) __b);
+ return __builtin_aarch64_uqsubv4hi_uuu (__a, __b);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vqsub_u32 (uint32x2_t __a, uint32x2_t __b)
{
- return (uint32x2_t) __builtin_aarch64_uqsubv2si ((int32x2_t) __a,
- (int32x2_t) __b);
+ return __builtin_aarch64_uqsubv2si_uuu (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vqsub_u64 (uint64x1_t __a, uint64x1_t __b)
{
- return (uint64x1_t) __builtin_aarch64_uqsubdi ((int64x1_t) __a,
- (int64x1_t) __b);
+ return (uint64x1_t) __builtin_aarch64_uqsubdi_uuu ((uint64_t) __a,
+ (uint64_t) __b);
}
__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
@@ -2275,29 +2265,25 @@ vqsubq_s64 (int64x2_t __a, int64x2_t __b)
__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
vqsubq_u8 (uint8x16_t __a, uint8x16_t __b)
{
- return (uint8x16_t) __builtin_aarch64_uqsubv16qi ((int8x16_t) __a,
- (int8x16_t) __b);
+ return __builtin_aarch64_uqsubv16qi_uuu (__a, __b);
}
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vqsubq_u16 (uint16x8_t __a, uint16x8_t __b)
{
- return (uint16x8_t) __builtin_aarch64_uqsubv8hi ((int16x8_t) __a,
- (int16x8_t) __b);
+ return __builtin_aarch64_uqsubv8hi_uuu (__a, __b);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vqsubq_u32 (uint32x4_t __a, uint32x4_t __b)
{
- return (uint32x4_t) __builtin_aarch64_uqsubv4si ((int32x4_t) __a,
- (int32x4_t) __b);
+ return __builtin_aarch64_uqsubv4si_uuu (__a, __b);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vqsubq_u64 (uint64x2_t __a, uint64x2_t __b)
{
- return (uint64x2_t) __builtin_aarch64_uqsubv2di ((int64x2_t) __a,
- (int64x2_t) __b);
+ return __builtin_aarch64_uqsubv2di_uuu (__a, __b);
}
__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
@@ -5661,318 +5647,6 @@ vcvtxd_f32_f64 (float64_t a)
return result;
}
-#define vext_f32(a, b, c) \
- __extension__ \
- ({ \
- float32x2_t b_ = (b); \
- float32x2_t a_ = (a); \
- float32x2_t result; \
- __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*4" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vext_f64(a, b, c) \
- __extension__ \
- ({ \
- float64x1_t b_ = (b); \
- float64x1_t a_ = (a); \
- float64x1_t result; \
- __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*8" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vext_p8(a, b, c) \
- __extension__ \
- ({ \
- poly8x8_t b_ = (b); \
- poly8x8_t a_ = (a); \
- poly8x8_t result; \
- __asm__ ("ext %0.8b,%1.8b,%2.8b,%3" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vext_p16(a, b, c) \
- __extension__ \
- ({ \
- poly16x4_t b_ = (b); \
- poly16x4_t a_ = (a); \
- poly16x4_t result; \
- __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*2" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vext_s8(a, b, c) \
- __extension__ \
- ({ \
- int8x8_t b_ = (b); \
- int8x8_t a_ = (a); \
- int8x8_t result; \
- __asm__ ("ext %0.8b,%1.8b,%2.8b,%3" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vext_s16(a, b, c) \
- __extension__ \
- ({ \
- int16x4_t b_ = (b); \
- int16x4_t a_ = (a); \
- int16x4_t result; \
- __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*2" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vext_s32(a, b, c) \
- __extension__ \
- ({ \
- int32x2_t b_ = (b); \
- int32x2_t a_ = (a); \
- int32x2_t result; \
- __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*4" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vext_s64(a, b, c) \
- __extension__ \
- ({ \
- int64x1_t b_ = (b); \
- int64x1_t a_ = (a); \
- int64x1_t result; \
- __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*8" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vext_u8(a, b, c) \
- __extension__ \
- ({ \
- uint8x8_t b_ = (b); \
- uint8x8_t a_ = (a); \
- uint8x8_t result; \
- __asm__ ("ext %0.8b,%1.8b,%2.8b,%3" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vext_u16(a, b, c) \
- __extension__ \
- ({ \
- uint16x4_t b_ = (b); \
- uint16x4_t a_ = (a); \
- uint16x4_t result; \
- __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*2" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vext_u32(a, b, c) \
- __extension__ \
- ({ \
- uint32x2_t b_ = (b); \
- uint32x2_t a_ = (a); \
- uint32x2_t result; \
- __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*4" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vext_u64(a, b, c) \
- __extension__ \
- ({ \
- uint64x1_t b_ = (b); \
- uint64x1_t a_ = (a); \
- uint64x1_t result; \
- __asm__ ("ext %0.8b, %1.8b, %2.8b, #%3*8" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vextq_f32(a, b, c) \
- __extension__ \
- ({ \
- float32x4_t b_ = (b); \
- float32x4_t a_ = (a); \
- float32x4_t result; \
- __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*4" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vextq_f64(a, b, c) \
- __extension__ \
- ({ \
- float64x2_t b_ = (b); \
- float64x2_t a_ = (a); \
- float64x2_t result; \
- __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*8" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vextq_p8(a, b, c) \
- __extension__ \
- ({ \
- poly8x16_t b_ = (b); \
- poly8x16_t a_ = (a); \
- poly8x16_t result; \
- __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vextq_p16(a, b, c) \
- __extension__ \
- ({ \
- poly16x8_t b_ = (b); \
- poly16x8_t a_ = (a); \
- poly16x8_t result; \
- __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*2" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vextq_s8(a, b, c) \
- __extension__ \
- ({ \
- int8x16_t b_ = (b); \
- int8x16_t a_ = (a); \
- int8x16_t result; \
- __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vextq_s16(a, b, c) \
- __extension__ \
- ({ \
- int16x8_t b_ = (b); \
- int16x8_t a_ = (a); \
- int16x8_t result; \
- __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*2" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vextq_s32(a, b, c) \
- __extension__ \
- ({ \
- int32x4_t b_ = (b); \
- int32x4_t a_ = (a); \
- int32x4_t result; \
- __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*4" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vextq_s64(a, b, c) \
- __extension__ \
- ({ \
- int64x2_t b_ = (b); \
- int64x2_t a_ = (a); \
- int64x2_t result; \
- __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*8" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vextq_u8(a, b, c) \
- __extension__ \
- ({ \
- uint8x16_t b_ = (b); \
- uint8x16_t a_ = (a); \
- uint8x16_t result; \
- __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vextq_u16(a, b, c) \
- __extension__ \
- ({ \
- uint16x8_t b_ = (b); \
- uint16x8_t a_ = (a); \
- uint16x8_t result; \
- __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*2" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vextq_u32(a, b, c) \
- __extension__ \
- ({ \
- uint32x4_t b_ = (b); \
- uint32x4_t a_ = (a); \
- uint32x4_t result; \
- __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*4" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
-#define vextq_u64(a, b, c) \
- __extension__ \
- ({ \
- uint64x2_t b_ = (b); \
- uint64x2_t a_ = (a); \
- uint64x2_t result; \
- __asm__ ("ext %0.16b, %1.16b, %2.16b, #%3*8" \
- : "=w"(result) \
- : "w"(a_), "w"(b_), "i"(c) \
- : /* No clobbers */); \
- result; \
- })
-
__extension__ static __inline float32x2_t __attribute__ ((__always_inline__))
vfma_f32 (float32x2_t a, float32x2_t b, float32x2_t c)
{
@@ -10875,402 +10549,6 @@ vrecpeq_u32 (uint32x4_t a)
return result;
}
-__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
-vrev16_p8 (poly8x8_t a)
-{
- poly8x8_t result;
- __asm__ ("rev16 %0.8b,%1.8b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
-vrev16_s8 (int8x8_t a)
-{
- int8x8_t result;
- __asm__ ("rev16 %0.8b,%1.8b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
-vrev16_u8 (uint8x8_t a)
-{
- uint8x8_t result;
- __asm__ ("rev16 %0.8b,%1.8b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
-vrev16q_p8 (poly8x16_t a)
-{
- poly8x16_t result;
- __asm__ ("rev16 %0.16b,%1.16b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
-vrev16q_s8 (int8x16_t a)
-{
- int8x16_t result;
- __asm__ ("rev16 %0.16b,%1.16b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
-vrev16q_u8 (uint8x16_t a)
-{
- uint8x16_t result;
- __asm__ ("rev16 %0.16b,%1.16b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
-vrev32_p8 (poly8x8_t a)
-{
- poly8x8_t result;
- __asm__ ("rev32 %0.8b,%1.8b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__))
-vrev32_p16 (poly16x4_t a)
-{
- poly16x4_t result;
- __asm__ ("rev32 %0.4h,%1.4h"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
-vrev32_s8 (int8x8_t a)
-{
- int8x8_t result;
- __asm__ ("rev32 %0.8b,%1.8b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
-vrev32_s16 (int16x4_t a)
-{
- int16x4_t result;
- __asm__ ("rev32 %0.4h,%1.4h"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
-vrev32_u8 (uint8x8_t a)
-{
- uint8x8_t result;
- __asm__ ("rev32 %0.8b,%1.8b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
-vrev32_u16 (uint16x4_t a)
-{
- uint16x4_t result;
- __asm__ ("rev32 %0.4h,%1.4h"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
-vrev32q_p8 (poly8x16_t a)
-{
- poly8x16_t result;
- __asm__ ("rev32 %0.16b,%1.16b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__))
-vrev32q_p16 (poly16x8_t a)
-{
- poly16x8_t result;
- __asm__ ("rev32 %0.8h,%1.8h"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
-vrev32q_s8 (int8x16_t a)
-{
- int8x16_t result;
- __asm__ ("rev32 %0.16b,%1.16b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
-vrev32q_s16 (int16x8_t a)
-{
- int16x8_t result;
- __asm__ ("rev32 %0.8h,%1.8h"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
-vrev32q_u8 (uint8x16_t a)
-{
- uint8x16_t result;
- __asm__ ("rev32 %0.16b,%1.16b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
-vrev32q_u16 (uint16x8_t a)
-{
- uint16x8_t result;
- __asm__ ("rev32 %0.8h,%1.8h"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline float32x2_t __attribute__ ((__always_inline__))
-vrev64_f32 (float32x2_t a)
-{
- float32x2_t result;
- __asm__ ("rev64 %0.2s,%1.2s"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
-vrev64_p8 (poly8x8_t a)
-{
- poly8x8_t result;
- __asm__ ("rev64 %0.8b,%1.8b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__))
-vrev64_p16 (poly16x4_t a)
-{
- poly16x4_t result;
- __asm__ ("rev64 %0.4h,%1.4h"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
-vrev64_s8 (int8x8_t a)
-{
- int8x8_t result;
- __asm__ ("rev64 %0.8b,%1.8b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
-vrev64_s16 (int16x4_t a)
-{
- int16x4_t result;
- __asm__ ("rev64 %0.4h,%1.4h"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int32x2_t __attribute__ ((__always_inline__))
-vrev64_s32 (int32x2_t a)
-{
- int32x2_t result;
- __asm__ ("rev64 %0.2s,%1.2s"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
-vrev64_u8 (uint8x8_t a)
-{
- uint8x8_t result;
- __asm__ ("rev64 %0.8b,%1.8b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
-vrev64_u16 (uint16x4_t a)
-{
- uint16x4_t result;
- __asm__ ("rev64 %0.4h,%1.4h"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
-vrev64_u32 (uint32x2_t a)
-{
- uint32x2_t result;
- __asm__ ("rev64 %0.2s,%1.2s"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline float32x4_t __attribute__ ((__always_inline__))
-vrev64q_f32 (float32x4_t a)
-{
- float32x4_t result;
- __asm__ ("rev64 %0.4s,%1.4s"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
-vrev64q_p8 (poly8x16_t a)
-{
- poly8x16_t result;
- __asm__ ("rev64 %0.16b,%1.16b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__))
-vrev64q_p16 (poly16x8_t a)
-{
- poly16x8_t result;
- __asm__ ("rev64 %0.8h,%1.8h"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
-vrev64q_s8 (int8x16_t a)
-{
- int8x16_t result;
- __asm__ ("rev64 %0.16b,%1.16b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
-vrev64q_s16 (int16x8_t a)
-{
- int16x8_t result;
- __asm__ ("rev64 %0.8h,%1.8h"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
-vrev64q_s32 (int32x4_t a)
-{
- int32x4_t result;
- __asm__ ("rev64 %0.4s,%1.4s"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
-vrev64q_u8 (uint8x16_t a)
-{
- uint8x16_t result;
- __asm__ ("rev64 %0.16b,%1.16b"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
-vrev64q_u16 (uint16x8_t a)
-{
- uint16x8_t result;
- __asm__ ("rev64 %0.8h,%1.8h"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
-vrev64q_u32 (uint32x4_t a)
-{
- uint32x4_t result;
- __asm__ ("rev64 %0.4s,%1.4s"
- : "=w"(result)
- : "w"(a)
- : /* No clobbers */);
- return result;
-}
-
#define vrshrn_high_n_s16(a, b, c) \
__extension__ \
({ \
@@ -12694,468 +11972,6 @@ vsubhn_u64 (uint64x2_t a, uint64x2_t b)
return result;
}
-__extension__ static __inline float32x2_t __attribute__ ((__always_inline__))
-vtrn1_f32 (float32x2_t a, float32x2_t b)
-{
- float32x2_t result;
- __asm__ ("trn1 %0.2s,%1.2s,%2.2s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
-vtrn1_p8 (poly8x8_t a, poly8x8_t b)
-{
- poly8x8_t result;
- __asm__ ("trn1 %0.8b,%1.8b,%2.8b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__))
-vtrn1_p16 (poly16x4_t a, poly16x4_t b)
-{
- poly16x4_t result;
- __asm__ ("trn1 %0.4h,%1.4h,%2.4h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
-vtrn1_s8 (int8x8_t a, int8x8_t b)
-{
- int8x8_t result;
- __asm__ ("trn1 %0.8b,%1.8b,%2.8b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
-vtrn1_s16 (int16x4_t a, int16x4_t b)
-{
- int16x4_t result;
- __asm__ ("trn1 %0.4h,%1.4h,%2.4h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int32x2_t __attribute__ ((__always_inline__))
-vtrn1_s32 (int32x2_t a, int32x2_t b)
-{
- int32x2_t result;
- __asm__ ("trn1 %0.2s,%1.2s,%2.2s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
-vtrn1_u8 (uint8x8_t a, uint8x8_t b)
-{
- uint8x8_t result;
- __asm__ ("trn1 %0.8b,%1.8b,%2.8b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
-vtrn1_u16 (uint16x4_t a, uint16x4_t b)
-{
- uint16x4_t result;
- __asm__ ("trn1 %0.4h,%1.4h,%2.4h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
-vtrn1_u32 (uint32x2_t a, uint32x2_t b)
-{
- uint32x2_t result;
- __asm__ ("trn1 %0.2s,%1.2s,%2.2s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline float32x4_t __attribute__ ((__always_inline__))
-vtrn1q_f32 (float32x4_t a, float32x4_t b)
-{
- float32x4_t result;
- __asm__ ("trn1 %0.4s,%1.4s,%2.4s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline float64x2_t __attribute__ ((__always_inline__))
-vtrn1q_f64 (float64x2_t a, float64x2_t b)
-{
- float64x2_t result;
- __asm__ ("trn1 %0.2d,%1.2d,%2.2d"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
-vtrn1q_p8 (poly8x16_t a, poly8x16_t b)
-{
- poly8x16_t result;
- __asm__ ("trn1 %0.16b,%1.16b,%2.16b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__))
-vtrn1q_p16 (poly16x8_t a, poly16x8_t b)
-{
- poly16x8_t result;
- __asm__ ("trn1 %0.8h,%1.8h,%2.8h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
-vtrn1q_s8 (int8x16_t a, int8x16_t b)
-{
- int8x16_t result;
- __asm__ ("trn1 %0.16b,%1.16b,%2.16b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
-vtrn1q_s16 (int16x8_t a, int16x8_t b)
-{
- int16x8_t result;
- __asm__ ("trn1 %0.8h,%1.8h,%2.8h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
-vtrn1q_s32 (int32x4_t a, int32x4_t b)
-{
- int32x4_t result;
- __asm__ ("trn1 %0.4s,%1.4s,%2.4s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
-vtrn1q_s64 (int64x2_t a, int64x2_t b)
-{
- int64x2_t result;
- __asm__ ("trn1 %0.2d,%1.2d,%2.2d"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
-vtrn1q_u8 (uint8x16_t a, uint8x16_t b)
-{
- uint8x16_t result;
- __asm__ ("trn1 %0.16b,%1.16b,%2.16b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
-vtrn1q_u16 (uint16x8_t a, uint16x8_t b)
-{
- uint16x8_t result;
- __asm__ ("trn1 %0.8h,%1.8h,%2.8h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
-vtrn1q_u32 (uint32x4_t a, uint32x4_t b)
-{
- uint32x4_t result;
- __asm__ ("trn1 %0.4s,%1.4s,%2.4s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
-vtrn1q_u64 (uint64x2_t a, uint64x2_t b)
-{
- uint64x2_t result;
- __asm__ ("trn1 %0.2d,%1.2d,%2.2d"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline float32x2_t __attribute__ ((__always_inline__))
-vtrn2_f32 (float32x2_t a, float32x2_t b)
-{
- float32x2_t result;
- __asm__ ("trn2 %0.2s,%1.2s,%2.2s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
-vtrn2_p8 (poly8x8_t a, poly8x8_t b)
-{
- poly8x8_t result;
- __asm__ ("trn2 %0.8b,%1.8b,%2.8b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__))
-vtrn2_p16 (poly16x4_t a, poly16x4_t b)
-{
- poly16x4_t result;
- __asm__ ("trn2 %0.4h,%1.4h,%2.4h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
-vtrn2_s8 (int8x8_t a, int8x8_t b)
-{
- int8x8_t result;
- __asm__ ("trn2 %0.8b,%1.8b,%2.8b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
-vtrn2_s16 (int16x4_t a, int16x4_t b)
-{
- int16x4_t result;
- __asm__ ("trn2 %0.4h,%1.4h,%2.4h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int32x2_t __attribute__ ((__always_inline__))
-vtrn2_s32 (int32x2_t a, int32x2_t b)
-{
- int32x2_t result;
- __asm__ ("trn2 %0.2s,%1.2s,%2.2s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
-vtrn2_u8 (uint8x8_t a, uint8x8_t b)
-{
- uint8x8_t result;
- __asm__ ("trn2 %0.8b,%1.8b,%2.8b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
-vtrn2_u16 (uint16x4_t a, uint16x4_t b)
-{
- uint16x4_t result;
- __asm__ ("trn2 %0.4h,%1.4h,%2.4h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
-vtrn2_u32 (uint32x2_t a, uint32x2_t b)
-{
- uint32x2_t result;
- __asm__ ("trn2 %0.2s,%1.2s,%2.2s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline float32x4_t __attribute__ ((__always_inline__))
-vtrn2q_f32 (float32x4_t a, float32x4_t b)
-{
- float32x4_t result;
- __asm__ ("trn2 %0.4s,%1.4s,%2.4s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline float64x2_t __attribute__ ((__always_inline__))
-vtrn2q_f64 (float64x2_t a, float64x2_t b)
-{
- float64x2_t result;
- __asm__ ("trn2 %0.2d,%1.2d,%2.2d"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
-vtrn2q_p8 (poly8x16_t a, poly8x16_t b)
-{
- poly8x16_t result;
- __asm__ ("trn2 %0.16b,%1.16b,%2.16b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__))
-vtrn2q_p16 (poly16x8_t a, poly16x8_t b)
-{
- poly16x8_t result;
- __asm__ ("trn2 %0.8h,%1.8h,%2.8h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
-vtrn2q_s8 (int8x16_t a, int8x16_t b)
-{
- int8x16_t result;
- __asm__ ("trn2 %0.16b,%1.16b,%2.16b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
-vtrn2q_s16 (int16x8_t a, int16x8_t b)
-{
- int16x8_t result;
- __asm__ ("trn2 %0.8h,%1.8h,%2.8h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
-vtrn2q_s32 (int32x4_t a, int32x4_t b)
-{
- int32x4_t result;
- __asm__ ("trn2 %0.4s,%1.4s,%2.4s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
-vtrn2q_s64 (int64x2_t a, int64x2_t b)
-{
- int64x2_t result;
- __asm__ ("trn2 %0.2d,%1.2d,%2.2d"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
-vtrn2q_u8 (uint8x16_t a, uint8x16_t b)
-{
- uint8x16_t result;
- __asm__ ("trn2 %0.16b,%1.16b,%2.16b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
-vtrn2q_u16 (uint16x8_t a, uint16x8_t b)
-{
- uint16x8_t result;
- __asm__ ("trn2 %0.8h,%1.8h,%2.8h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
-vtrn2q_u32 (uint32x4_t a, uint32x4_t b)
-{
- uint32x4_t result;
- __asm__ ("trn2 %0.4s,%1.4s,%2.4s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
-vtrn2q_u64 (uint64x2_t a, uint64x2_t b)
-{
- uint64x2_t result;
- __asm__ ("trn2 %0.2d,%1.2d,%2.2d"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vtst_p8 (poly8x8_t a, poly8x8_t b)
{
@@ -13199,467 +12015,6 @@ vtstq_p16 (poly16x8_t a, poly16x8_t b)
: /* No clobbers */);
return result;
}
-__extension__ static __inline float32x2_t __attribute__ ((__always_inline__))
-vuzp1_f32 (float32x2_t a, float32x2_t b)
-{
- float32x2_t result;
- __asm__ ("uzp1 %0.2s,%1.2s,%2.2s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
-vuzp1_p8 (poly8x8_t a, poly8x8_t b)
-{
- poly8x8_t result;
- __asm__ ("uzp1 %0.8b,%1.8b,%2.8b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__))
-vuzp1_p16 (poly16x4_t a, poly16x4_t b)
-{
- poly16x4_t result;
- __asm__ ("uzp1 %0.4h,%1.4h,%2.4h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
-vuzp1_s8 (int8x8_t a, int8x8_t b)
-{
- int8x8_t result;
- __asm__ ("uzp1 %0.8b,%1.8b,%2.8b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
-vuzp1_s16 (int16x4_t a, int16x4_t b)
-{
- int16x4_t result;
- __asm__ ("uzp1 %0.4h,%1.4h,%2.4h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int32x2_t __attribute__ ((__always_inline__))
-vuzp1_s32 (int32x2_t a, int32x2_t b)
-{
- int32x2_t result;
- __asm__ ("uzp1 %0.2s,%1.2s,%2.2s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
-vuzp1_u8 (uint8x8_t a, uint8x8_t b)
-{
- uint8x8_t result;
- __asm__ ("uzp1 %0.8b,%1.8b,%2.8b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
-vuzp1_u16 (uint16x4_t a, uint16x4_t b)
-{
- uint16x4_t result;
- __asm__ ("uzp1 %0.4h,%1.4h,%2.4h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
-vuzp1_u32 (uint32x2_t a, uint32x2_t b)
-{
- uint32x2_t result;
- __asm__ ("uzp1 %0.2s,%1.2s,%2.2s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline float32x4_t __attribute__ ((__always_inline__))
-vuzp1q_f32 (float32x4_t a, float32x4_t b)
-{
- float32x4_t result;
- __asm__ ("uzp1 %0.4s,%1.4s,%2.4s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline float64x2_t __attribute__ ((__always_inline__))
-vuzp1q_f64 (float64x2_t a, float64x2_t b)
-{
- float64x2_t result;
- __asm__ ("uzp1 %0.2d,%1.2d,%2.2d"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
-vuzp1q_p8 (poly8x16_t a, poly8x16_t b)
-{
- poly8x16_t result;
- __asm__ ("uzp1 %0.16b,%1.16b,%2.16b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__))
-vuzp1q_p16 (poly16x8_t a, poly16x8_t b)
-{
- poly16x8_t result;
- __asm__ ("uzp1 %0.8h,%1.8h,%2.8h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
-vuzp1q_s8 (int8x16_t a, int8x16_t b)
-{
- int8x16_t result;
- __asm__ ("uzp1 %0.16b,%1.16b,%2.16b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
-vuzp1q_s16 (int16x8_t a, int16x8_t b)
-{
- int16x8_t result;
- __asm__ ("uzp1 %0.8h,%1.8h,%2.8h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
-vuzp1q_s32 (int32x4_t a, int32x4_t b)
-{
- int32x4_t result;
- __asm__ ("uzp1 %0.4s,%1.4s,%2.4s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
-vuzp1q_s64 (int64x2_t a, int64x2_t b)
-{
- int64x2_t result;
- __asm__ ("uzp1 %0.2d,%1.2d,%2.2d"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
-vuzp1q_u8 (uint8x16_t a, uint8x16_t b)
-{
- uint8x16_t result;
- __asm__ ("uzp1 %0.16b,%1.16b,%2.16b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
-vuzp1q_u16 (uint16x8_t a, uint16x8_t b)
-{
- uint16x8_t result;
- __asm__ ("uzp1 %0.8h,%1.8h,%2.8h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
-vuzp1q_u32 (uint32x4_t a, uint32x4_t b)
-{
- uint32x4_t result;
- __asm__ ("uzp1 %0.4s,%1.4s,%2.4s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
-vuzp1q_u64 (uint64x2_t a, uint64x2_t b)
-{
- uint64x2_t result;
- __asm__ ("uzp1 %0.2d,%1.2d,%2.2d"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline float32x2_t __attribute__ ((__always_inline__))
-vuzp2_f32 (float32x2_t a, float32x2_t b)
-{
- float32x2_t result;
- __asm__ ("uzp2 %0.2s,%1.2s,%2.2s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
-vuzp2_p8 (poly8x8_t a, poly8x8_t b)
-{
- poly8x8_t result;
- __asm__ ("uzp2 %0.8b,%1.8b,%2.8b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__))
-vuzp2_p16 (poly16x4_t a, poly16x4_t b)
-{
- poly16x4_t result;
- __asm__ ("uzp2 %0.4h,%1.4h,%2.4h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
-vuzp2_s8 (int8x8_t a, int8x8_t b)
-{
- int8x8_t result;
- __asm__ ("uzp2 %0.8b,%1.8b,%2.8b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
-vuzp2_s16 (int16x4_t a, int16x4_t b)
-{
- int16x4_t result;
- __asm__ ("uzp2 %0.4h,%1.4h,%2.4h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int32x2_t __attribute__ ((__always_inline__))
-vuzp2_s32 (int32x2_t a, int32x2_t b)
-{
- int32x2_t result;
- __asm__ ("uzp2 %0.2s,%1.2s,%2.2s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
-vuzp2_u8 (uint8x8_t a, uint8x8_t b)
-{
- uint8x8_t result;
- __asm__ ("uzp2 %0.8b,%1.8b,%2.8b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
-vuzp2_u16 (uint16x4_t a, uint16x4_t b)
-{
- uint16x4_t result;
- __asm__ ("uzp2 %0.4h,%1.4h,%2.4h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
-vuzp2_u32 (uint32x2_t a, uint32x2_t b)
-{
- uint32x2_t result;
- __asm__ ("uzp2 %0.2s,%1.2s,%2.2s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline float32x4_t __attribute__ ((__always_inline__))
-vuzp2q_f32 (float32x4_t a, float32x4_t b)
-{
- float32x4_t result;
- __asm__ ("uzp2 %0.4s,%1.4s,%2.4s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline float64x2_t __attribute__ ((__always_inline__))
-vuzp2q_f64 (float64x2_t a, float64x2_t b)
-{
- float64x2_t result;
- __asm__ ("uzp2 %0.2d,%1.2d,%2.2d"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
-vuzp2q_p8 (poly8x16_t a, poly8x16_t b)
-{
- poly8x16_t result;
- __asm__ ("uzp2 %0.16b,%1.16b,%2.16b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__))
-vuzp2q_p16 (poly16x8_t a, poly16x8_t b)
-{
- poly16x8_t result;
- __asm__ ("uzp2 %0.8h,%1.8h,%2.8h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
-vuzp2q_s8 (int8x16_t a, int8x16_t b)
-{
- int8x16_t result;
- __asm__ ("uzp2 %0.16b,%1.16b,%2.16b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
-vuzp2q_s16 (int16x8_t a, int16x8_t b)
-{
- int16x8_t result;
- __asm__ ("uzp2 %0.8h,%1.8h,%2.8h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
-vuzp2q_s32 (int32x4_t a, int32x4_t b)
-{
- int32x4_t result;
- __asm__ ("uzp2 %0.4s,%1.4s,%2.4s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
-vuzp2q_s64 (int64x2_t a, int64x2_t b)
-{
- int64x2_t result;
- __asm__ ("uzp2 %0.2d,%1.2d,%2.2d"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
-vuzp2q_u8 (uint8x16_t a, uint8x16_t b)
-{
- uint8x16_t result;
- __asm__ ("uzp2 %0.16b,%1.16b,%2.16b"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
-vuzp2q_u16 (uint16x8_t a, uint16x8_t b)
-{
- uint16x8_t result;
- __asm__ ("uzp2 %0.8h,%1.8h,%2.8h"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
-vuzp2q_u32 (uint32x4_t a, uint32x4_t b)
-{
- uint32x4_t result;
- __asm__ ("uzp2 %0.4s,%1.4s,%2.4s"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
-
-__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
-vuzp2q_u64 (uint64x2_t a, uint64x2_t b)
-{
- uint64x2_t result;
- __asm__ ("uzp2 %0.2d,%1.2d,%2.2d"
- : "=w"(result)
- : "w"(a), "w"(b)
- : /* No clobbers */);
- return result;
-}
/* End of temporary inline asm implementations. */
@@ -18367,6 +16722,292 @@ vdupd_laneq_u64 (uint64x2_t __a, const int __b)
return __aarch64_vgetq_lane_u64 (__a, __b);
}
+/* vext */
+
+__extension__ static __inline float32x2_t __attribute__ ((__always_inline__))
+vext_f32 (float32x2_t __a, float32x2_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 2);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint32x2_t) {2-__c, 3-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {__c, __c+1});
+#endif
+}
+
+__extension__ static __inline float64x1_t __attribute__ ((__always_inline__))
+vext_f64 (float64x1_t __a, float64x1_t __b, __const int __c)
+{
+ /* The only possible index to the assembler instruction returns element 0. */
+ __builtin_aarch64_im_lane_boundsi (__c, 1);
+ return __a;
+}
+__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
+vext_p8 (poly8x8_t __a, poly8x8_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 8);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint8x8_t)
+ {8-__c, 9-__c, 10-__c, 11-__c, 12-__c, 13-__c, 14-__c, 15-__c});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint8x8_t) {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7});
+#endif
+}
+
+__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__))
+vext_p16 (poly16x4_t __a, poly16x4_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 4);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a,
+ (uint16x4_t) {4-__c, 5-__c, 6-__c, 7-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {__c, __c+1, __c+2, __c+3});
+#endif
+}
+
+__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
+vext_s8 (int8x8_t __a, int8x8_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 8);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint8x8_t)
+ {8-__c, 9-__c, 10-__c, 11-__c, 12-__c, 13-__c, 14-__c, 15-__c});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint8x8_t) {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7});
+#endif
+}
+
+__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
+vext_s16 (int16x4_t __a, int16x4_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 4);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a,
+ (uint16x4_t) {4-__c, 5-__c, 6-__c, 7-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {__c, __c+1, __c+2, __c+3});
+#endif
+}
+
+__extension__ static __inline int32x2_t __attribute__ ((__always_inline__))
+vext_s32 (int32x2_t __a, int32x2_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 2);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint32x2_t) {2-__c, 3-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {__c, __c+1});
+#endif
+}
+
+__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
+vext_s64 (int64x1_t __a, int64x1_t __b, __const int __c)
+{
+ /* The only possible index to the assembler instruction returns element 0. */
+ __builtin_aarch64_im_lane_boundsi (__c, 1);
+ return __a;
+}
+
+__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
+vext_u8 (uint8x8_t __a, uint8x8_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 8);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint8x8_t)
+ {8-__c, 9-__c, 10-__c, 11-__c, 12-__c, 13-__c, 14-__c, 15-__c});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint8x8_t) {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7});
+#endif
+}
+
+__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
+vext_u16 (uint16x4_t __a, uint16x4_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 4);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a,
+ (uint16x4_t) {4-__c, 5-__c, 6-__c, 7-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {__c, __c+1, __c+2, __c+3});
+#endif
+}
+
+__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
+vext_u32 (uint32x2_t __a, uint32x2_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 2);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint32x2_t) {2-__c, 3-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {__c, __c+1});
+#endif
+}
+
+__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
+vext_u64 (uint64x1_t __a, uint64x1_t __b, __const int __c)
+{
+ /* The only possible index to the assembler instruction returns element 0. */
+ __builtin_aarch64_im_lane_boundsi (__c, 1);
+ return __a;
+}
+
+__extension__ static __inline float32x4_t __attribute__ ((__always_inline__))
+vextq_f32 (float32x4_t __a, float32x4_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 4);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a,
+ (uint32x4_t) {4-__c, 5-__c, 6-__c, 7-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {__c, __c+1, __c+2, __c+3});
+#endif
+}
+
+__extension__ static __inline float64x2_t __attribute__ ((__always_inline__))
+vextq_f64 (float64x2_t __a, float64x2_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 2);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint64x2_t) {2-__c, 3-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {__c, __c+1});
+#endif
+}
+
+__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
+vextq_p8 (poly8x16_t __a, poly8x16_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 16);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint8x16_t)
+ {16-__c, 17-__c, 18-__c, 19-__c, 20-__c, 21-__c, 22-__c, 23-__c,
+ 24-__c, 25-__c, 26-__c, 27-__c, 28-__c, 29-__c, 30-__c, 31-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x16_t)
+ {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7,
+ __c+8, __c+9, __c+10, __c+11, __c+12, __c+13, __c+14, __c+15});
+#endif
+}
+
+__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__))
+vextq_p16 (poly16x8_t __a, poly16x8_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 8);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint16x8_t)
+ {8-__c, 9-__c, 10-__c, 11-__c, 12-__c, 13-__c, 14-__c, 15-__c});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint16x8_t) {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7});
+#endif
+}
+
+__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
+vextq_s8 (int8x16_t __a, int8x16_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 16);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint8x16_t)
+ {16-__c, 17-__c, 18-__c, 19-__c, 20-__c, 21-__c, 22-__c, 23-__c,
+ 24-__c, 25-__c, 26-__c, 27-__c, 28-__c, 29-__c, 30-__c, 31-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x16_t)
+ {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7,
+ __c+8, __c+9, __c+10, __c+11, __c+12, __c+13, __c+14, __c+15});
+#endif
+}
+
+__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
+vextq_s16 (int16x8_t __a, int16x8_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 8);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint16x8_t)
+ {8-__c, 9-__c, 10-__c, 11-__c, 12-__c, 13-__c, 14-__c, 15-__c});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint16x8_t) {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7});
+#endif
+}
+
+__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
+vextq_s32 (int32x4_t __a, int32x4_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 4);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a,
+ (uint32x4_t) {4-__c, 5-__c, 6-__c, 7-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {__c, __c+1, __c+2, __c+3});
+#endif
+}
+
+__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
+vextq_s64 (int64x2_t __a, int64x2_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 2);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint64x2_t) {2-__c, 3-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {__c, __c+1});
+#endif
+}
+
+__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
+vextq_u8 (uint8x16_t __a, uint8x16_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 16);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint8x16_t)
+ {16-__c, 17-__c, 18-__c, 19-__c, 20-__c, 21-__c, 22-__c, 23-__c,
+ 24-__c, 25-__c, 26-__c, 27-__c, 28-__c, 29-__c, 30-__c, 31-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x16_t)
+ {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7,
+ __c+8, __c+9, __c+10, __c+11, __c+12, __c+13, __c+14, __c+15});
+#endif
+}
+
+__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
+vextq_u16 (uint16x8_t __a, uint16x8_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 8);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint16x8_t)
+ {8-__c, 9-__c, 10-__c, 11-__c, 12-__c, 13-__c, 14-__c, 15-__c});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint16x8_t) {__c, __c+1, __c+2, __c+3, __c+4, __c+5, __c+6, __c+7});
+#endif
+}
+
+__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
+vextq_u32 (uint32x4_t __a, uint32x4_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 4);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a,
+ (uint32x4_t) {4-__c, 5-__c, 6-__c, 7-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {__c, __c+1, __c+2, __c+3});
+#endif
+}
+
+__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
+vextq_u64 (uint64x2_t __a, uint64x2_t __b, __const int __c)
+{
+ __builtin_aarch64_im_lane_boundsi (__c, 2);
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__b, __a, (uint64x2_t) {2-__c, 3-__c});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {__c, __c+1});
+#endif
+}
+
/* vfma_lane */
__extension__ static __inline float32x2_t __attribute__ ((__always_inline__))
@@ -20856,25 +19497,26 @@ vqaddd_s64 (int64x1_t __a, int64x1_t __b)
__extension__ static __inline uint8x1_t __attribute__ ((__always_inline__))
vqaddb_u8 (uint8x1_t __a, uint8x1_t __b)
{
- return (uint8x1_t) __builtin_aarch64_uqaddqi (__a, __b);
+ return (uint8x1_t) __builtin_aarch64_uqaddqi_uuu (__a, __b);
}
__extension__ static __inline uint16x1_t __attribute__ ((__always_inline__))
vqaddh_u16 (uint16x1_t __a, uint16x1_t __b)
{
- return (uint16x1_t) __builtin_aarch64_uqaddhi (__a, __b);
+ return (uint16x1_t) __builtin_aarch64_uqaddhi_uuu (__a, __b);
}
__extension__ static __inline uint32x1_t __attribute__ ((__always_inline__))
vqadds_u32 (uint32x1_t __a, uint32x1_t __b)
{
- return (uint32x1_t) __builtin_aarch64_uqaddsi (__a, __b);
+ return (uint32x1_t) __builtin_aarch64_uqaddsi_uuu (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vqaddd_u64 (uint64x1_t __a, uint64x1_t __b)
{
- return (uint64x1_t) __builtin_aarch64_uqadddi (__a, __b);
+ return (uint64x1_t) __builtin_aarch64_uqadddi_uuu ((uint64_t) __a,
+ (uint64_t) __b);
}
/* vqdmlal */
@@ -20892,7 +19534,7 @@ vqdmlal_high_s16 (int32x4_t __a, int16x8_t __b, int16x8_t __c)
}
__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
-vqdmlal_high_lane_s16 (int32x4_t __a, int16x8_t __b, int16x8_t __c,
+vqdmlal_high_lane_s16 (int32x4_t __a, int16x8_t __b, int16x4_t __c,
int const __d)
{
return __builtin_aarch64_sqdmlal2_lanev8hi (__a, __b, __c, __d);
@@ -20914,8 +19556,7 @@ vqdmlal_high_n_s16 (int32x4_t __a, int16x8_t __b, int16_t __c)
__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
vqdmlal_lane_s16 (int32x4_t __a, int16x4_t __b, int16x4_t __c, int const __d)
{
- int16x8_t __tmp = vcombine_s16 (__c, vcreate_s16 (__AARCH64_INT64_C (0)));
- return __builtin_aarch64_sqdmlal_lanev4hi (__a, __b, __tmp, __d);
+ return __builtin_aarch64_sqdmlal_lanev4hi (__a, __b, __c, __d);
}
__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
@@ -20943,7 +19584,7 @@ vqdmlal_high_s32 (int64x2_t __a, int32x4_t __b, int32x4_t __c)
}
__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
-vqdmlal_high_lane_s32 (int64x2_t __a, int32x4_t __b, int32x4_t __c,
+vqdmlal_high_lane_s32 (int64x2_t __a, int32x4_t __b, int32x2_t __c,
int const __d)
{
return __builtin_aarch64_sqdmlal2_lanev4si (__a, __b, __c, __d);
@@ -20965,8 +19606,7 @@ vqdmlal_high_n_s32 (int64x2_t __a, int32x4_t __b, int32_t __c)
__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
vqdmlal_lane_s32 (int64x2_t __a, int32x2_t __b, int32x2_t __c, int const __d)
{
- int32x4_t __tmp = vcombine_s32 (__c, vcreate_s32 (__AARCH64_INT64_C (0)));
- return __builtin_aarch64_sqdmlal_lanev2si (__a, __b, __tmp, __d);
+ return __builtin_aarch64_sqdmlal_lanev2si (__a, __b, __c, __d);
}
__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
@@ -20988,7 +19628,7 @@ vqdmlalh_s16 (int32x1_t __a, int16x1_t __b, int16x1_t __c)
}
__extension__ static __inline int32x1_t __attribute__ ((__always_inline__))
-vqdmlalh_lane_s16 (int32x1_t __a, int16x1_t __b, int16x8_t __c, const int __d)
+vqdmlalh_lane_s16 (int32x1_t __a, int16x1_t __b, int16x4_t __c, const int __d)
{
return __builtin_aarch64_sqdmlal_lanehi (__a, __b, __c, __d);
}
@@ -21000,7 +19640,7 @@ vqdmlals_s32 (int64x1_t __a, int32x1_t __b, int32x1_t __c)
}
__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
-vqdmlals_lane_s32 (int64x1_t __a, int32x1_t __b, int32x4_t __c, const int __d)
+vqdmlals_lane_s32 (int64x1_t __a, int32x1_t __b, int32x2_t __c, const int __d)
{
return __builtin_aarch64_sqdmlal_lanesi (__a, __b, __c, __d);
}
@@ -21020,7 +19660,7 @@ vqdmlsl_high_s16 (int32x4_t __a, int16x8_t __b, int16x8_t __c)
}
__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
-vqdmlsl_high_lane_s16 (int32x4_t __a, int16x8_t __b, int16x8_t __c,
+vqdmlsl_high_lane_s16 (int32x4_t __a, int16x8_t __b, int16x4_t __c,
int const __d)
{
return __builtin_aarch64_sqdmlsl2_lanev8hi (__a, __b, __c, __d);
@@ -21042,8 +19682,7 @@ vqdmlsl_high_n_s16 (int32x4_t __a, int16x8_t __b, int16_t __c)
__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
vqdmlsl_lane_s16 (int32x4_t __a, int16x4_t __b, int16x4_t __c, int const __d)
{
- int16x8_t __tmp = vcombine_s16 (__c, vcreate_s16 (__AARCH64_INT64_C (0)));
- return __builtin_aarch64_sqdmlsl_lanev4hi (__a, __b, __tmp, __d);
+ return __builtin_aarch64_sqdmlsl_lanev4hi (__a, __b, __c, __d);
}
__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
@@ -21071,7 +19710,7 @@ vqdmlsl_high_s32 (int64x2_t __a, int32x4_t __b, int32x4_t __c)
}
__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
-vqdmlsl_high_lane_s32 (int64x2_t __a, int32x4_t __b, int32x4_t __c,
+vqdmlsl_high_lane_s32 (int64x2_t __a, int32x4_t __b, int32x2_t __c,
int const __d)
{
return __builtin_aarch64_sqdmlsl2_lanev4si (__a, __b, __c, __d);
@@ -21093,8 +19732,7 @@ vqdmlsl_high_n_s32 (int64x2_t __a, int32x4_t __b, int32_t __c)
__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
vqdmlsl_lane_s32 (int64x2_t __a, int32x2_t __b, int32x2_t __c, int const __d)
{
- int32x4_t __tmp = vcombine_s32 (__c, vcreate_s32 (__AARCH64_INT64_C (0)));
- return __builtin_aarch64_sqdmlsl_lanev2si (__a, __b, __tmp, __d);
+ return __builtin_aarch64_sqdmlsl_lanev2si (__a, __b, __c, __d);
}
__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
@@ -21116,7 +19754,7 @@ vqdmlslh_s16 (int32x1_t __a, int16x1_t __b, int16x1_t __c)
}
__extension__ static __inline int32x1_t __attribute__ ((__always_inline__))
-vqdmlslh_lane_s16 (int32x1_t __a, int16x1_t __b, int16x8_t __c, const int __d)
+vqdmlslh_lane_s16 (int32x1_t __a, int16x1_t __b, int16x4_t __c, const int __d)
{
return __builtin_aarch64_sqdmlsl_lanehi (__a, __b, __c, __d);
}
@@ -21128,7 +19766,7 @@ vqdmlsls_s32 (int64x1_t __a, int32x1_t __b, int32x1_t __c)
}
__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
-vqdmlsls_lane_s32 (int64x1_t __a, int32x1_t __b, int32x4_t __c, const int __d)
+vqdmlsls_lane_s32 (int64x1_t __a, int32x1_t __b, int32x2_t __c, const int __d)
{
return __builtin_aarch64_sqdmlsl_lanesi (__a, __b, __c, __d);
}
@@ -21166,7 +19804,7 @@ vqdmulhh_s16 (int16x1_t __a, int16x1_t __b)
}
__extension__ static __inline int16x1_t __attribute__ ((__always_inline__))
-vqdmulhh_lane_s16 (int16x1_t __a, int16x8_t __b, const int __c)
+vqdmulhh_lane_s16 (int16x1_t __a, int16x4_t __b, const int __c)
{
return __builtin_aarch64_sqdmulh_lanehi (__a, __b, __c);
}
@@ -21178,7 +19816,7 @@ vqdmulhs_s32 (int32x1_t __a, int32x1_t __b)
}
__extension__ static __inline int32x1_t __attribute__ ((__always_inline__))
-vqdmulhs_lane_s32 (int32x1_t __a, int32x4_t __b, const int __c)
+vqdmulhs_lane_s32 (int32x1_t __a, int32x2_t __b, const int __c)
{
return __builtin_aarch64_sqdmulh_lanesi (__a, __b, __c);
}
@@ -21198,7 +19836,7 @@ vqdmull_high_s16 (int16x8_t __a, int16x8_t __b)
}
__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
-vqdmull_high_lane_s16 (int16x8_t __a, int16x8_t __b, int const __c)
+vqdmull_high_lane_s16 (int16x8_t __a, int16x4_t __b, int const __c)
{
return __builtin_aarch64_sqdmull2_lanev8hi (__a, __b,__c);
}
@@ -21218,8 +19856,7 @@ vqdmull_high_n_s16 (int16x8_t __a, int16_t __b)
__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
vqdmull_lane_s16 (int16x4_t __a, int16x4_t __b, int const __c)
{
- int16x8_t __tmp = vcombine_s16 (__b, vcreate_s16 (__AARCH64_INT64_C (0)));
- return __builtin_aarch64_sqdmull_lanev4hi (__a, __tmp, __c);
+ return __builtin_aarch64_sqdmull_lanev4hi (__a, __b, __c);
}
__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
@@ -21247,7 +19884,7 @@ vqdmull_high_s32 (int32x4_t __a, int32x4_t __b)
}
__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
-vqdmull_high_lane_s32 (int32x4_t __a, int32x4_t __b, int const __c)
+vqdmull_high_lane_s32 (int32x4_t __a, int32x2_t __b, int const __c)
{
return __builtin_aarch64_sqdmull2_lanev4si (__a, __b, __c);
}
@@ -21267,8 +19904,7 @@ vqdmull_high_n_s32 (int32x4_t __a, int32_t __b)
__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
vqdmull_lane_s32 (int32x2_t __a, int32x2_t __b, int const __c)
{
- int32x4_t __tmp = vcombine_s32 (__b, vcreate_s32 (__AARCH64_INT64_C (0)));
- return __builtin_aarch64_sqdmull_lanev2si (__a, __tmp, __c);
+ return __builtin_aarch64_sqdmull_lanev2si (__a, __b, __c);
}
__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
@@ -21290,7 +19926,7 @@ vqdmullh_s16 (int16x1_t __a, int16x1_t __b)
}
__extension__ static __inline int32x1_t __attribute__ ((__always_inline__))
-vqdmullh_lane_s16 (int16x1_t __a, int16x8_t __b, const int __c)
+vqdmullh_lane_s16 (int16x1_t __a, int16x4_t __b, const int __c)
{
return __builtin_aarch64_sqdmull_lanehi (__a, __b, __c);
}
@@ -21302,7 +19938,7 @@ vqdmulls_s32 (int32x1_t __a, int32x1_t __b)
}
__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
-vqdmulls_lane_s32 (int32x1_t __a, int32x4_t __b, const int __c)
+vqdmulls_lane_s32 (int32x1_t __a, int32x2_t __b, const int __c)
{
return __builtin_aarch64_sqdmull_lanesi (__a, __b, __c);
}
@@ -21484,7 +20120,7 @@ vqrdmulhh_s16 (int16x1_t __a, int16x1_t __b)
}
__extension__ static __inline int16x1_t __attribute__ ((__always_inline__))
-vqrdmulhh_lane_s16 (int16x1_t __a, int16x8_t __b, const int __c)
+vqrdmulhh_lane_s16 (int16x1_t __a, int16x4_t __b, const int __c)
{
return __builtin_aarch64_sqrdmulh_lanehi (__a, __b, __c);
}
@@ -21496,7 +20132,7 @@ vqrdmulhs_s32 (int32x1_t __a, int32x1_t __b)
}
__extension__ static __inline int32x1_t __attribute__ ((__always_inline__))
-vqrdmulhs_lane_s32 (int32x1_t __a, int32x4_t __b, const int __c)
+vqrdmulhs_lane_s32 (int32x1_t __a, int32x2_t __b, const int __c)
{
return __builtin_aarch64_sqrdmulh_lanesi (__a, __b, __c);
}
@@ -21530,25 +20166,25 @@ vqrshl_s64 (int64x1_t __a, int64x1_t __b)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vqrshl_u8 (uint8x8_t __a, int8x8_t __b)
{
- return (uint8x8_t) __builtin_aarch64_uqrshlv8qi ((int8x8_t) __a, __b);
+ return __builtin_aarch64_uqrshlv8qi_uus ( __a, __b);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vqrshl_u16 (uint16x4_t __a, int16x4_t __b)
{
- return (uint16x4_t) __builtin_aarch64_uqrshlv4hi ((int16x4_t) __a, __b);
+ return __builtin_aarch64_uqrshlv4hi_uus ( __a, __b);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vqrshl_u32 (uint32x2_t __a, int32x2_t __b)
{
- return (uint32x2_t) __builtin_aarch64_uqrshlv2si ((int32x2_t) __a, __b);
+ return __builtin_aarch64_uqrshlv2si_uus ( __a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vqrshl_u64 (uint64x1_t __a, int64x1_t __b)
{
- return (uint64x1_t) __builtin_aarch64_uqrshldi ((int64x1_t) __a, __b);
+ return __builtin_aarch64_uqrshldi_uus ( __a, __b);
}
__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
@@ -21578,25 +20214,25 @@ vqrshlq_s64 (int64x2_t __a, int64x2_t __b)
__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
vqrshlq_u8 (uint8x16_t __a, int8x16_t __b)
{
- return (uint8x16_t) __builtin_aarch64_uqrshlv16qi ((int8x16_t) __a, __b);
+ return __builtin_aarch64_uqrshlv16qi_uus ( __a, __b);
}
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vqrshlq_u16 (uint16x8_t __a, int16x8_t __b)
{
- return (uint16x8_t) __builtin_aarch64_uqrshlv8hi ((int16x8_t) __a, __b);
+ return __builtin_aarch64_uqrshlv8hi_uus ( __a, __b);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vqrshlq_u32 (uint32x4_t __a, int32x4_t __b)
{
- return (uint32x4_t) __builtin_aarch64_uqrshlv4si ((int32x4_t) __a, __b);
+ return __builtin_aarch64_uqrshlv4si_uus ( __a, __b);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vqrshlq_u64 (uint64x2_t __a, int64x2_t __b)
{
- return (uint64x2_t) __builtin_aarch64_uqrshlv2di ((int64x2_t) __a, __b);
+ return __builtin_aarch64_uqrshlv2di_uus ( __a, __b);
}
__extension__ static __inline int8x1_t __attribute__ ((__always_inline__))
@@ -21626,25 +20262,25 @@ vqrshld_s64 (int64x1_t __a, int64x1_t __b)
__extension__ static __inline uint8x1_t __attribute__ ((__always_inline__))
vqrshlb_u8 (uint8x1_t __a, uint8x1_t __b)
{
- return (uint8x1_t) __builtin_aarch64_uqrshlqi (__a, __b);
+ return __builtin_aarch64_uqrshlqi_uus (__a, __b);
}
__extension__ static __inline uint16x1_t __attribute__ ((__always_inline__))
vqrshlh_u16 (uint16x1_t __a, uint16x1_t __b)
{
- return (uint16x1_t) __builtin_aarch64_uqrshlhi (__a, __b);
+ return __builtin_aarch64_uqrshlhi_uus (__a, __b);
}
__extension__ static __inline uint32x1_t __attribute__ ((__always_inline__))
vqrshls_u32 (uint32x1_t __a, uint32x1_t __b)
{
- return (uint32x1_t) __builtin_aarch64_uqrshlsi (__a, __b);
+ return __builtin_aarch64_uqrshlsi_uus (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vqrshld_u64 (uint64x1_t __a, uint64x1_t __b)
{
- return (uint64x1_t) __builtin_aarch64_uqrshldi (__a, __b);
+ return __builtin_aarch64_uqrshldi_uus (__a, __b);
}
/* vqrshrn */
@@ -21670,19 +20306,19 @@ vqrshrn_n_s64 (int64x2_t __a, const int __b)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vqrshrn_n_u16 (uint16x8_t __a, const int __b)
{
- return (uint8x8_t) __builtin_aarch64_uqrshrn_nv8hi ((int16x8_t) __a, __b);
+ return __builtin_aarch64_uqrshrn_nv8hi_uus ( __a, __b);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vqrshrn_n_u32 (uint32x4_t __a, const int __b)
{
- return (uint16x4_t) __builtin_aarch64_uqrshrn_nv4si ((int32x4_t) __a, __b);
+ return __builtin_aarch64_uqrshrn_nv4si_uus ( __a, __b);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vqrshrn_n_u64 (uint64x2_t __a, const int __b)
{
- return (uint32x2_t) __builtin_aarch64_uqrshrn_nv2di ((int64x2_t) __a, __b);
+ return __builtin_aarch64_uqrshrn_nv2di_uus ( __a, __b);
}
__extension__ static __inline int8x1_t __attribute__ ((__always_inline__))
@@ -21706,19 +20342,19 @@ vqrshrnd_n_s64 (int64x1_t __a, const int __b)
__extension__ static __inline uint8x1_t __attribute__ ((__always_inline__))
vqrshrnh_n_u16 (uint16x1_t __a, const int __b)
{
- return (uint8x1_t) __builtin_aarch64_uqrshrn_nhi (__a, __b);
+ return __builtin_aarch64_uqrshrn_nhi_uus (__a, __b);
}
__extension__ static __inline uint16x1_t __attribute__ ((__always_inline__))
vqrshrns_n_u32 (uint32x1_t __a, const int __b)
{
- return (uint16x1_t) __builtin_aarch64_uqrshrn_nsi (__a, __b);
+ return __builtin_aarch64_uqrshrn_nsi_uus (__a, __b);
}
__extension__ static __inline uint32x1_t __attribute__ ((__always_inline__))
vqrshrnd_n_u64 (uint64x1_t __a, const int __b)
{
- return (uint32x1_t) __builtin_aarch64_uqrshrn_ndi (__a, __b);
+ return __builtin_aarch64_uqrshrn_ndi_uus (__a, __b);
}
/* vqrshrun */
@@ -21788,25 +20424,25 @@ vqshl_s64 (int64x1_t __a, int64x1_t __b)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vqshl_u8 (uint8x8_t __a, int8x8_t __b)
{
- return (uint8x8_t) __builtin_aarch64_uqshlv8qi ((int8x8_t) __a, __b);
+ return __builtin_aarch64_uqshlv8qi_uus ( __a, __b);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vqshl_u16 (uint16x4_t __a, int16x4_t __b)
{
- return (uint16x4_t) __builtin_aarch64_uqshlv4hi ((int16x4_t) __a, __b);
+ return __builtin_aarch64_uqshlv4hi_uus ( __a, __b);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vqshl_u32 (uint32x2_t __a, int32x2_t __b)
{
- return (uint32x2_t) __builtin_aarch64_uqshlv2si ((int32x2_t) __a, __b);
+ return __builtin_aarch64_uqshlv2si_uus ( __a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vqshl_u64 (uint64x1_t __a, int64x1_t __b)
{
- return (uint64x1_t) __builtin_aarch64_uqshldi ((int64x1_t) __a, __b);
+ return __builtin_aarch64_uqshldi_uus ( __a, __b);
}
__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
@@ -21836,25 +20472,25 @@ vqshlq_s64 (int64x2_t __a, int64x2_t __b)
__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
vqshlq_u8 (uint8x16_t __a, int8x16_t __b)
{
- return (uint8x16_t) __builtin_aarch64_uqshlv16qi ((int8x16_t) __a, __b);
+ return __builtin_aarch64_uqshlv16qi_uus ( __a, __b);
}
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vqshlq_u16 (uint16x8_t __a, int16x8_t __b)
{
- return (uint16x8_t) __builtin_aarch64_uqshlv8hi ((int16x8_t) __a, __b);
+ return __builtin_aarch64_uqshlv8hi_uus ( __a, __b);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vqshlq_u32 (uint32x4_t __a, int32x4_t __b)
{
- return (uint32x4_t) __builtin_aarch64_uqshlv4si ((int32x4_t) __a, __b);
+ return __builtin_aarch64_uqshlv4si_uus ( __a, __b);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vqshlq_u64 (uint64x2_t __a, int64x2_t __b)
{
- return (uint64x2_t) __builtin_aarch64_uqshlv2di ((int64x2_t) __a, __b);
+ return __builtin_aarch64_uqshlv2di_uus ( __a, __b);
}
__extension__ static __inline int8x1_t __attribute__ ((__always_inline__))
@@ -21884,25 +20520,25 @@ vqshld_s64 (int64x1_t __a, int64x1_t __b)
__extension__ static __inline uint8x1_t __attribute__ ((__always_inline__))
vqshlb_u8 (uint8x1_t __a, uint8x1_t __b)
{
- return (uint8x1_t) __builtin_aarch64_uqshlqi (__a, __b);
+ return __builtin_aarch64_uqshlqi_uus (__a, __b);
}
__extension__ static __inline uint16x1_t __attribute__ ((__always_inline__))
vqshlh_u16 (uint16x1_t __a, uint16x1_t __b)
{
- return (uint16x1_t) __builtin_aarch64_uqshlhi (__a, __b);
+ return __builtin_aarch64_uqshlhi_uus (__a, __b);
}
__extension__ static __inline uint32x1_t __attribute__ ((__always_inline__))
vqshls_u32 (uint32x1_t __a, uint32x1_t __b)
{
- return (uint32x1_t) __builtin_aarch64_uqshlsi (__a, __b);
+ return __builtin_aarch64_uqshlsi_uus (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vqshld_u64 (uint64x1_t __a, uint64x1_t __b)
{
- return (uint64x1_t) __builtin_aarch64_uqshldi (__a, __b);
+ return __builtin_aarch64_uqshldi_uus (__a, __b);
}
__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
@@ -21932,25 +20568,25 @@ vqshl_n_s64 (int64x1_t __a, const int __b)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vqshl_n_u8 (uint8x8_t __a, const int __b)
{
- return (uint8x8_t) __builtin_aarch64_uqshl_nv8qi ((int8x8_t) __a, __b);
+ return __builtin_aarch64_uqshl_nv8qi_uus (__a, __b);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vqshl_n_u16 (uint16x4_t __a, const int __b)
{
- return (uint16x4_t) __builtin_aarch64_uqshl_nv4hi ((int16x4_t) __a, __b);
+ return __builtin_aarch64_uqshl_nv4hi_uus (__a, __b);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vqshl_n_u32 (uint32x2_t __a, const int __b)
{
- return (uint32x2_t) __builtin_aarch64_uqshl_nv2si ((int32x2_t) __a, __b);
+ return __builtin_aarch64_uqshl_nv2si_uus (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vqshl_n_u64 (uint64x1_t __a, const int __b)
{
- return (uint64x1_t) __builtin_aarch64_uqshl_ndi ((int64x1_t) __a, __b);
+ return __builtin_aarch64_uqshl_ndi_uus (__a, __b);
}
__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
@@ -21980,25 +20616,25 @@ vqshlq_n_s64 (int64x2_t __a, const int __b)
__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
vqshlq_n_u8 (uint8x16_t __a, const int __b)
{
- return (uint8x16_t) __builtin_aarch64_uqshl_nv16qi ((int8x16_t) __a, __b);
+ return __builtin_aarch64_uqshl_nv16qi_uus (__a, __b);
}
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vqshlq_n_u16 (uint16x8_t __a, const int __b)
{
- return (uint16x8_t) __builtin_aarch64_uqshl_nv8hi ((int16x8_t) __a, __b);
+ return __builtin_aarch64_uqshl_nv8hi_uus (__a, __b);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vqshlq_n_u32 (uint32x4_t __a, const int __b)
{
- return (uint32x4_t) __builtin_aarch64_uqshl_nv4si ((int32x4_t) __a, __b);
+ return __builtin_aarch64_uqshl_nv4si_uus (__a, __b);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vqshlq_n_u64 (uint64x2_t __a, const int __b)
{
- return (uint64x2_t) __builtin_aarch64_uqshl_nv2di ((int64x2_t) __a, __b);
+ return __builtin_aarch64_uqshl_nv2di_uus (__a, __b);
}
__extension__ static __inline int8x1_t __attribute__ ((__always_inline__))
@@ -22028,25 +20664,25 @@ vqshld_n_s64 (int64x1_t __a, const int __b)
__extension__ static __inline uint8x1_t __attribute__ ((__always_inline__))
vqshlb_n_u8 (uint8x1_t __a, const int __b)
{
- return (uint8x1_t) __builtin_aarch64_uqshl_nqi (__a, __b);
+ return __builtin_aarch64_uqshl_nqi_uus (__a, __b);
}
__extension__ static __inline uint16x1_t __attribute__ ((__always_inline__))
vqshlh_n_u16 (uint16x1_t __a, const int __b)
{
- return (uint16x1_t) __builtin_aarch64_uqshl_nhi (__a, __b);
+ return __builtin_aarch64_uqshl_nhi_uus (__a, __b);
}
__extension__ static __inline uint32x1_t __attribute__ ((__always_inline__))
vqshls_n_u32 (uint32x1_t __a, const int __b)
{
- return (uint32x1_t) __builtin_aarch64_uqshl_nsi (__a, __b);
+ return __builtin_aarch64_uqshl_nsi_uus (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vqshld_n_u64 (uint64x1_t __a, const int __b)
{
- return (uint64x1_t) __builtin_aarch64_uqshl_ndi (__a, __b);
+ return __builtin_aarch64_uqshl_ndi_uus (__a, __b);
}
/* vqshlu */
@@ -22054,73 +20690,73 @@ vqshld_n_u64 (uint64x1_t __a, const int __b)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vqshlu_n_s8 (int8x8_t __a, const int __b)
{
- return (uint8x8_t) __builtin_aarch64_sqshlu_nv8qi (__a, __b);
+ return __builtin_aarch64_sqshlu_nv8qi_uss (__a, __b);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vqshlu_n_s16 (int16x4_t __a, const int __b)
{
- return (uint16x4_t) __builtin_aarch64_sqshlu_nv4hi (__a, __b);
+ return __builtin_aarch64_sqshlu_nv4hi_uss (__a, __b);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vqshlu_n_s32 (int32x2_t __a, const int __b)
{
- return (uint32x2_t) __builtin_aarch64_sqshlu_nv2si (__a, __b);
+ return __builtin_aarch64_sqshlu_nv2si_uss (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vqshlu_n_s64 (int64x1_t __a, const int __b)
{
- return (uint64x1_t) __builtin_aarch64_sqshlu_ndi (__a, __b);
+ return __builtin_aarch64_sqshlu_ndi_uss (__a, __b);
}
__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
vqshluq_n_s8 (int8x16_t __a, const int __b)
{
- return (uint8x16_t) __builtin_aarch64_sqshlu_nv16qi (__a, __b);
+ return __builtin_aarch64_sqshlu_nv16qi_uss (__a, __b);
}
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vqshluq_n_s16 (int16x8_t __a, const int __b)
{
- return (uint16x8_t) __builtin_aarch64_sqshlu_nv8hi (__a, __b);
+ return __builtin_aarch64_sqshlu_nv8hi_uss (__a, __b);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vqshluq_n_s32 (int32x4_t __a, const int __b)
{
- return (uint32x4_t) __builtin_aarch64_sqshlu_nv4si (__a, __b);
+ return __builtin_aarch64_sqshlu_nv4si_uss (__a, __b);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vqshluq_n_s64 (int64x2_t __a, const int __b)
{
- return (uint64x2_t) __builtin_aarch64_sqshlu_nv2di (__a, __b);
+ return __builtin_aarch64_sqshlu_nv2di_uss (__a, __b);
}
__extension__ static __inline int8x1_t __attribute__ ((__always_inline__))
vqshlub_n_s8 (int8x1_t __a, const int __b)
{
- return (int8x1_t) __builtin_aarch64_sqshlu_nqi (__a, __b);
+ return (int8x1_t) __builtin_aarch64_sqshlu_nqi_uss (__a, __b);
}
__extension__ static __inline int16x1_t __attribute__ ((__always_inline__))
vqshluh_n_s16 (int16x1_t __a, const int __b)
{
- return (int16x1_t) __builtin_aarch64_sqshlu_nhi (__a, __b);
+ return (int16x1_t) __builtin_aarch64_sqshlu_nhi_uss (__a, __b);
}
__extension__ static __inline int32x1_t __attribute__ ((__always_inline__))
vqshlus_n_s32 (int32x1_t __a, const int __b)
{
- return (int32x1_t) __builtin_aarch64_sqshlu_nsi (__a, __b);
+ return (int32x1_t) __builtin_aarch64_sqshlu_nsi_uss (__a, __b);
}
__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
vqshlud_n_s64 (int64x1_t __a, const int __b)
{
- return (int64x1_t) __builtin_aarch64_sqshlu_ndi (__a, __b);
+ return (int64x1_t) __builtin_aarch64_sqshlu_ndi_uss (__a, __b);
}
/* vqshrn */
@@ -22146,19 +20782,19 @@ vqshrn_n_s64 (int64x2_t __a, const int __b)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vqshrn_n_u16 (uint16x8_t __a, const int __b)
{
- return (uint8x8_t) __builtin_aarch64_uqshrn_nv8hi ((int16x8_t) __a, __b);
+ return __builtin_aarch64_uqshrn_nv8hi_uus ( __a, __b);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vqshrn_n_u32 (uint32x4_t __a, const int __b)
{
- return (uint16x4_t) __builtin_aarch64_uqshrn_nv4si ((int32x4_t) __a, __b);
+ return __builtin_aarch64_uqshrn_nv4si_uus ( __a, __b);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vqshrn_n_u64 (uint64x2_t __a, const int __b)
{
- return (uint32x2_t) __builtin_aarch64_uqshrn_nv2di ((int64x2_t) __a, __b);
+ return __builtin_aarch64_uqshrn_nv2di_uus ( __a, __b);
}
__extension__ static __inline int8x1_t __attribute__ ((__always_inline__))
@@ -22182,19 +20818,19 @@ vqshrnd_n_s64 (int64x1_t __a, const int __b)
__extension__ static __inline uint8x1_t __attribute__ ((__always_inline__))
vqshrnh_n_u16 (uint16x1_t __a, const int __b)
{
- return (uint8x1_t) __builtin_aarch64_uqshrn_nhi (__a, __b);
+ return __builtin_aarch64_uqshrn_nhi_uus (__a, __b);
}
__extension__ static __inline uint16x1_t __attribute__ ((__always_inline__))
vqshrns_n_u32 (uint32x1_t __a, const int __b)
{
- return (uint16x1_t) __builtin_aarch64_uqshrn_nsi (__a, __b);
+ return __builtin_aarch64_uqshrn_nsi_uus (__a, __b);
}
__extension__ static __inline uint32x1_t __attribute__ ((__always_inline__))
vqshrnd_n_u64 (uint64x1_t __a, const int __b)
{
- return (uint32x1_t) __builtin_aarch64_uqshrn_ndi (__a, __b);
+ return __builtin_aarch64_uqshrn_ndi_uus (__a, __b);
}
/* vqshrun */
@@ -22264,25 +20900,26 @@ vqsubd_s64 (int64x1_t __a, int64x1_t __b)
__extension__ static __inline uint8x1_t __attribute__ ((__always_inline__))
vqsubb_u8 (uint8x1_t __a, uint8x1_t __b)
{
- return (uint8x1_t) __builtin_aarch64_uqsubqi (__a, __b);
+ return (uint8x1_t) __builtin_aarch64_uqsubqi_uuu (__a, __b);
}
__extension__ static __inline uint16x1_t __attribute__ ((__always_inline__))
vqsubh_u16 (uint16x1_t __a, uint16x1_t __b)
{
- return (uint16x1_t) __builtin_aarch64_uqsubhi (__a, __b);
+ return (uint16x1_t) __builtin_aarch64_uqsubhi_uuu (__a, __b);
}
__extension__ static __inline uint32x1_t __attribute__ ((__always_inline__))
vqsubs_u32 (uint32x1_t __a, uint32x1_t __b)
{
- return (uint32x1_t) __builtin_aarch64_uqsubsi (__a, __b);
+ return (uint32x1_t) __builtin_aarch64_uqsubsi_uuu (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vqsubd_u64 (uint64x1_t __a, uint64x1_t __b)
{
- return (uint64x1_t) __builtin_aarch64_uqsubdi (__a, __b);
+ return (uint64x1_t) __builtin_aarch64_uqsubdi_uuu ((uint64_t) __a,
+ (uint64_t) __b);
}
/* vrecpe */
@@ -22363,6 +21000,234 @@ vrecpxd_f64 (float64_t __a)
return __builtin_aarch64_frecpxdf (__a);
}
+
+/* vrev */
+
+__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
+vrev16_p8 (poly8x8_t a)
+{
+ return __builtin_shuffle (a, (uint8x8_t) { 1, 0, 3, 2, 5, 4, 7, 6 });
+}
+
+__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
+vrev16_s8 (int8x8_t a)
+{
+ return __builtin_shuffle (a, (uint8x8_t) { 1, 0, 3, 2, 5, 4, 7, 6 });
+}
+
+__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
+vrev16_u8 (uint8x8_t a)
+{
+ return __builtin_shuffle (a, (uint8x8_t) { 1, 0, 3, 2, 5, 4, 7, 6 });
+}
+
+__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
+vrev16q_p8 (poly8x16_t a)
+{
+ return __builtin_shuffle (a,
+ (uint8x16_t) { 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14 });
+}
+
+__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
+vrev16q_s8 (int8x16_t a)
+{
+ return __builtin_shuffle (a,
+ (uint8x16_t) { 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14 });
+}
+
+__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
+vrev16q_u8 (uint8x16_t a)
+{
+ return __builtin_shuffle (a,
+ (uint8x16_t) { 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14 });
+}
+
+__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
+vrev32_p8 (poly8x8_t a)
+{
+ return __builtin_shuffle (a, (uint8x8_t) { 3, 2, 1, 0, 7, 6, 5, 4 });
+}
+
+__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__))
+vrev32_p16 (poly16x4_t a)
+{
+ return __builtin_shuffle (a, (uint16x4_t) { 1, 0, 3, 2 });
+}
+
+__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
+vrev32_s8 (int8x8_t a)
+{
+ return __builtin_shuffle (a, (uint8x8_t) { 3, 2, 1, 0, 7, 6, 5, 4 });
+}
+
+__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
+vrev32_s16 (int16x4_t a)
+{
+ return __builtin_shuffle (a, (uint16x4_t) { 1, 0, 3, 2 });
+}
+
+__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
+vrev32_u8 (uint8x8_t a)
+{
+ return __builtin_shuffle (a, (uint8x8_t) { 3, 2, 1, 0, 7, 6, 5, 4 });
+}
+
+__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
+vrev32_u16 (uint16x4_t a)
+{
+ return __builtin_shuffle (a, (uint16x4_t) { 1, 0, 3, 2 });
+}
+
+__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
+vrev32q_p8 (poly8x16_t a)
+{
+ return __builtin_shuffle (a,
+ (uint8x16_t) { 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12 });
+}
+
+__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__))
+vrev32q_p16 (poly16x8_t a)
+{
+ return __builtin_shuffle (a, (uint16x8_t) { 1, 0, 3, 2, 5, 4, 7, 6 });
+}
+
+__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
+vrev32q_s8 (int8x16_t a)
+{
+ return __builtin_shuffle (a,
+ (uint8x16_t) { 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12 });
+}
+
+__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
+vrev32q_s16 (int16x8_t a)
+{
+ return __builtin_shuffle (a, (uint16x8_t) { 1, 0, 3, 2, 5, 4, 7, 6 });
+}
+
+__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
+vrev32q_u8 (uint8x16_t a)
+{
+ return __builtin_shuffle (a,
+ (uint8x16_t) { 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12 });
+}
+
+__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
+vrev32q_u16 (uint16x8_t a)
+{
+ return __builtin_shuffle (a, (uint16x8_t) { 1, 0, 3, 2, 5, 4, 7, 6 });
+}
+
+__extension__ static __inline float32x2_t __attribute__ ((__always_inline__))
+vrev64_f32 (float32x2_t a)
+{
+ return __builtin_shuffle (a, (uint32x2_t) { 1, 0 });
+}
+
+__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
+vrev64_p8 (poly8x8_t a)
+{
+ return __builtin_shuffle (a, (uint8x8_t) { 7, 6, 5, 4, 3, 2, 1, 0 });
+}
+
+__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__))
+vrev64_p16 (poly16x4_t a)
+{
+ return __builtin_shuffle (a, (uint16x4_t) { 3, 2, 1, 0 });
+}
+
+__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
+vrev64_s8 (int8x8_t a)
+{
+ return __builtin_shuffle (a, (uint8x8_t) { 7, 6, 5, 4, 3, 2, 1, 0 });
+}
+
+__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
+vrev64_s16 (int16x4_t a)
+{
+ return __builtin_shuffle (a, (uint16x4_t) { 3, 2, 1, 0 });
+}
+
+__extension__ static __inline int32x2_t __attribute__ ((__always_inline__))
+vrev64_s32 (int32x2_t a)
+{
+ return __builtin_shuffle (a, (uint32x2_t) { 1, 0 });
+}
+
+__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
+vrev64_u8 (uint8x8_t a)
+{
+ return __builtin_shuffle (a, (uint8x8_t) { 7, 6, 5, 4, 3, 2, 1, 0 });
+}
+
+__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
+vrev64_u16 (uint16x4_t a)
+{
+ return __builtin_shuffle (a, (uint16x4_t) { 3, 2, 1, 0 });
+}
+
+__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
+vrev64_u32 (uint32x2_t a)
+{
+ return __builtin_shuffle (a, (uint32x2_t) { 1, 0 });
+}
+
+__extension__ static __inline float32x4_t __attribute__ ((__always_inline__))
+vrev64q_f32 (float32x4_t a)
+{
+ return __builtin_shuffle (a, (uint32x4_t) { 1, 0, 3, 2 });
+}
+
+__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
+vrev64q_p8 (poly8x16_t a)
+{
+ return __builtin_shuffle (a,
+ (uint8x16_t) { 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8 });
+}
+
+__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__))
+vrev64q_p16 (poly16x8_t a)
+{
+ return __builtin_shuffle (a, (uint16x8_t) { 3, 2, 1, 0, 7, 6, 5, 4 });
+}
+
+__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
+vrev64q_s8 (int8x16_t a)
+{
+ return __builtin_shuffle (a,
+ (uint8x16_t) { 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8 });
+}
+
+__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
+vrev64q_s16 (int16x8_t a)
+{
+ return __builtin_shuffle (a, (uint16x8_t) { 3, 2, 1, 0, 7, 6, 5, 4 });
+}
+
+__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
+vrev64q_s32 (int32x4_t a)
+{
+ return __builtin_shuffle (a, (uint32x4_t) { 1, 0, 3, 2 });
+}
+
+__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
+vrev64q_u8 (uint8x16_t a)
+{
+ return __builtin_shuffle (a,
+ (uint8x16_t) { 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8 });
+}
+
+__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
+vrev64q_u16 (uint16x8_t a)
+{
+ return __builtin_shuffle (a, (uint16x8_t) { 3, 2, 1, 0, 7, 6, 5, 4 });
+}
+
+__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
+vrev64q_u32 (uint32x4_t a)
+{
+ return __builtin_shuffle (a, (uint32x4_t) { 1, 0, 3, 2 });
+}
+
/* vrnd */
__extension__ static __inline float32x2_t __attribute__ ((__always_inline__))
@@ -22574,25 +21439,25 @@ vrshl_s64 (int64x1_t __a, int64x1_t __b)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vrshl_u8 (uint8x8_t __a, int8x8_t __b)
{
- return (uint8x8_t) __builtin_aarch64_urshlv8qi ((int8x8_t) __a, __b);
+ return __builtin_aarch64_urshlv8qi_uus (__a, __b);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vrshl_u16 (uint16x4_t __a, int16x4_t __b)
{
- return (uint16x4_t) __builtin_aarch64_urshlv4hi ((int16x4_t) __a, __b);
+ return __builtin_aarch64_urshlv4hi_uus (__a, __b);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vrshl_u32 (uint32x2_t __a, int32x2_t __b)
{
- return (uint32x2_t) __builtin_aarch64_urshlv2si ((int32x2_t) __a, __b);
+ return __builtin_aarch64_urshlv2si_uus (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vrshl_u64 (uint64x1_t __a, int64x1_t __b)
{
- return (uint64x1_t) __builtin_aarch64_urshldi ((int64x1_t) __a, __b);
+ return __builtin_aarch64_urshldi_uus (__a, __b);
}
__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
@@ -22622,25 +21487,25 @@ vrshlq_s64 (int64x2_t __a, int64x2_t __b)
__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
vrshlq_u8 (uint8x16_t __a, int8x16_t __b)
{
- return (uint8x16_t) __builtin_aarch64_urshlv16qi ((int8x16_t) __a, __b);
+ return __builtin_aarch64_urshlv16qi_uus (__a, __b);
}
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vrshlq_u16 (uint16x8_t __a, int16x8_t __b)
{
- return (uint16x8_t) __builtin_aarch64_urshlv8hi ((int16x8_t) __a, __b);
+ return __builtin_aarch64_urshlv8hi_uus (__a, __b);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vrshlq_u32 (uint32x4_t __a, int32x4_t __b)
{
- return (uint32x4_t) __builtin_aarch64_urshlv4si ((int32x4_t) __a, __b);
+ return __builtin_aarch64_urshlv4si_uus (__a, __b);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vrshlq_u64 (uint64x2_t __a, int64x2_t __b)
{
- return (uint64x2_t) __builtin_aarch64_urshlv2di ((int64x2_t) __a, __b);
+ return __builtin_aarch64_urshlv2di_uus (__a, __b);
}
__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
@@ -22652,7 +21517,7 @@ vrshld_s64 (int64x1_t __a, int64x1_t __b)
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vrshld_u64 (uint64x1_t __a, uint64x1_t __b)
{
- return (uint64x1_t) __builtin_aarch64_urshldi (__a, __b);
+ return __builtin_aarch64_urshldi_uus (__a, __b);
}
/* vrshr */
@@ -22684,25 +21549,25 @@ vrshr_n_s64 (int64x1_t __a, const int __b)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vrshr_n_u8 (uint8x8_t __a, const int __b)
{
- return (uint8x8_t) __builtin_aarch64_urshr_nv8qi ((int8x8_t) __a, __b);
+ return __builtin_aarch64_urshr_nv8qi_uus (__a, __b);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vrshr_n_u16 (uint16x4_t __a, const int __b)
{
- return (uint16x4_t) __builtin_aarch64_urshr_nv4hi ((int16x4_t) __a, __b);
+ return __builtin_aarch64_urshr_nv4hi_uus (__a, __b);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vrshr_n_u32 (uint32x2_t __a, const int __b)
{
- return (uint32x2_t) __builtin_aarch64_urshr_nv2si ((int32x2_t) __a, __b);
+ return __builtin_aarch64_urshr_nv2si_uus (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vrshr_n_u64 (uint64x1_t __a, const int __b)
{
- return (uint64x1_t) __builtin_aarch64_urshr_ndi ((int64x1_t) __a, __b);
+ return __builtin_aarch64_urshr_ndi_uus (__a, __b);
}
__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
@@ -22732,25 +21597,25 @@ vrshrq_n_s64 (int64x2_t __a, const int __b)
__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
vrshrq_n_u8 (uint8x16_t __a, const int __b)
{
- return (uint8x16_t) __builtin_aarch64_urshr_nv16qi ((int8x16_t) __a, __b);
+ return __builtin_aarch64_urshr_nv16qi_uus (__a, __b);
}
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vrshrq_n_u16 (uint16x8_t __a, const int __b)
{
- return (uint16x8_t) __builtin_aarch64_urshr_nv8hi ((int16x8_t) __a, __b);
+ return __builtin_aarch64_urshr_nv8hi_uus (__a, __b);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vrshrq_n_u32 (uint32x4_t __a, const int __b)
{
- return (uint32x4_t) __builtin_aarch64_urshr_nv4si ((int32x4_t) __a, __b);
+ return __builtin_aarch64_urshr_nv4si_uus (__a, __b);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vrshrq_n_u64 (uint64x2_t __a, const int __b)
{
- return (uint64x2_t) __builtin_aarch64_urshr_nv2di ((int64x2_t) __a, __b);
+ return __builtin_aarch64_urshr_nv2di_uus (__a, __b);
}
__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
@@ -22762,7 +21627,7 @@ vrshrd_n_s64 (int64x1_t __a, const int __b)
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vrshrd_n_u64 (uint64x1_t __a, const int __b)
{
- return (uint64x1_t) __builtin_aarch64_urshr_ndi (__a, __b);
+ return __builtin_aarch64_urshr_ndi_uus (__a, __b);
}
/* vrsra */
@@ -22794,29 +21659,25 @@ vrsra_n_s64 (int64x1_t __a, int64x1_t __b, const int __c)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vrsra_n_u8 (uint8x8_t __a, uint8x8_t __b, const int __c)
{
- return (uint8x8_t) __builtin_aarch64_ursra_nv8qi ((int8x8_t) __a,
- (int8x8_t) __b, __c);
+ return __builtin_aarch64_ursra_nv8qi_uuus (__a, __b, __c);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vrsra_n_u16 (uint16x4_t __a, uint16x4_t __b, const int __c)
{
- return (uint16x4_t) __builtin_aarch64_ursra_nv4hi ((int16x4_t) __a,
- (int16x4_t) __b, __c);
+ return __builtin_aarch64_ursra_nv4hi_uuus (__a, __b, __c);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vrsra_n_u32 (uint32x2_t __a, uint32x2_t __b, const int __c)
{
- return (uint32x2_t) __builtin_aarch64_ursra_nv2si ((int32x2_t) __a,
- (int32x2_t) __b, __c);
+ return __builtin_aarch64_ursra_nv2si_uuus (__a, __b, __c);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vrsra_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c)
{
- return (uint64x1_t) __builtin_aarch64_ursra_ndi ((int64x1_t) __a,
- (int64x1_t) __b, __c);
+ return __builtin_aarch64_ursra_ndi_uuus (__a, __b, __c);
}
__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
@@ -22846,29 +21707,25 @@ vrsraq_n_s64 (int64x2_t __a, int64x2_t __b, const int __c)
__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
vrsraq_n_u8 (uint8x16_t __a, uint8x16_t __b, const int __c)
{
- return (uint8x16_t) __builtin_aarch64_ursra_nv16qi ((int8x16_t) __a,
- (int8x16_t) __b, __c);
+ return __builtin_aarch64_ursra_nv16qi_uuus (__a, __b, __c);
}
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vrsraq_n_u16 (uint16x8_t __a, uint16x8_t __b, const int __c)
{
- return (uint16x8_t) __builtin_aarch64_ursra_nv8hi ((int16x8_t) __a,
- (int16x8_t) __b, __c);
+ return __builtin_aarch64_ursra_nv8hi_uuus (__a, __b, __c);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vrsraq_n_u32 (uint32x4_t __a, uint32x4_t __b, const int __c)
{
- return (uint32x4_t) __builtin_aarch64_ursra_nv4si ((int32x4_t) __a,
- (int32x4_t) __b, __c);
+ return __builtin_aarch64_ursra_nv4si_uuus (__a, __b, __c);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vrsraq_n_u64 (uint64x2_t __a, uint64x2_t __b, const int __c)
{
- return (uint64x2_t) __builtin_aarch64_ursra_nv2di ((int64x2_t) __a,
- (int64x2_t) __b, __c);
+ return __builtin_aarch64_ursra_nv2di_uuus (__a, __b, __c);
}
__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
@@ -22880,7 +21737,7 @@ vrsrad_n_s64 (int64x1_t __a, int64x1_t __b, const int __c)
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vrsrad_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c)
{
- return (uint64x1_t) __builtin_aarch64_ursra_ndi (__a, __b, __c);
+ return __builtin_aarch64_ursra_ndi_uuus (__a, __b, __c);
}
#ifdef __ARM_FEATURE_CRYPTO
@@ -23073,109 +21930,109 @@ vshld_n_u64 (uint64x1_t __a, const int __b)
__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
vshl_s8 (int8x8_t __a, int8x8_t __b)
{
- return (int8x8_t) __builtin_aarch64_sshlv8qi (__a, __b);
+ return __builtin_aarch64_sshlv8qi (__a, __b);
}
__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
vshl_s16 (int16x4_t __a, int16x4_t __b)
{
- return (int16x4_t) __builtin_aarch64_sshlv4hi (__a, __b);
+ return __builtin_aarch64_sshlv4hi (__a, __b);
}
__extension__ static __inline int32x2_t __attribute__ ((__always_inline__))
vshl_s32 (int32x2_t __a, int32x2_t __b)
{
- return (int32x2_t) __builtin_aarch64_sshlv2si (__a, __b);
+ return __builtin_aarch64_sshlv2si (__a, __b);
}
__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
vshl_s64 (int64x1_t __a, int64x1_t __b)
{
- return (int64x1_t) __builtin_aarch64_sshldi (__a, __b);
+ return __builtin_aarch64_sshldi (__a, __b);
}
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vshl_u8 (uint8x8_t __a, int8x8_t __b)
{
- return (uint8x8_t) __builtin_aarch64_ushlv8qi ((int8x8_t) __a, __b);
+ return __builtin_aarch64_ushlv8qi_uus (__a, __b);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vshl_u16 (uint16x4_t __a, int16x4_t __b)
{
- return (uint16x4_t) __builtin_aarch64_ushlv4hi ((int16x4_t) __a, __b);
+ return __builtin_aarch64_ushlv4hi_uus (__a, __b);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vshl_u32 (uint32x2_t __a, int32x2_t __b)
{
- return (uint32x2_t) __builtin_aarch64_ushlv2si ((int32x2_t) __a, __b);
+ return __builtin_aarch64_ushlv2si_uus (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vshl_u64 (uint64x1_t __a, int64x1_t __b)
{
- return (uint64x1_t) __builtin_aarch64_ushldi ((int64x1_t) __a, __b);
+ return __builtin_aarch64_ushldi_uus (__a, __b);
}
__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
vshlq_s8 (int8x16_t __a, int8x16_t __b)
{
- return (int8x16_t) __builtin_aarch64_sshlv16qi (__a, __b);
+ return __builtin_aarch64_sshlv16qi (__a, __b);
}
__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
vshlq_s16 (int16x8_t __a, int16x8_t __b)
{
- return (int16x8_t) __builtin_aarch64_sshlv8hi (__a, __b);
+ return __builtin_aarch64_sshlv8hi (__a, __b);
}
__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
vshlq_s32 (int32x4_t __a, int32x4_t __b)
{
- return (int32x4_t) __builtin_aarch64_sshlv4si (__a, __b);
+ return __builtin_aarch64_sshlv4si (__a, __b);
}
__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
vshlq_s64 (int64x2_t __a, int64x2_t __b)
{
- return (int64x2_t) __builtin_aarch64_sshlv2di (__a, __b);
+ return __builtin_aarch64_sshlv2di (__a, __b);
}
__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
vshlq_u8 (uint8x16_t __a, int8x16_t __b)
{
- return (uint8x16_t) __builtin_aarch64_ushlv16qi ((int8x16_t) __a, __b);
+ return __builtin_aarch64_ushlv16qi_uus (__a, __b);
}
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vshlq_u16 (uint16x8_t __a, int16x8_t __b)
{
- return (uint16x8_t) __builtin_aarch64_ushlv8hi ((int16x8_t) __a, __b);
+ return __builtin_aarch64_ushlv8hi_uus (__a, __b);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vshlq_u32 (uint32x4_t __a, int32x4_t __b)
{
- return (uint32x4_t) __builtin_aarch64_ushlv4si ((int32x4_t) __a, __b);
+ return __builtin_aarch64_ushlv4si_uus (__a, __b);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vshlq_u64 (uint64x2_t __a, int64x2_t __b)
{
- return (uint64x2_t) __builtin_aarch64_ushlv2di ((int64x2_t) __a, __b);
+ return __builtin_aarch64_ushlv2di_uus (__a, __b);
}
__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
vshld_s64 (int64x1_t __a, int64x1_t __b)
{
- return (int64x1_t) __builtin_aarch64_sshldi (__a, __b);
+ return __builtin_aarch64_sshldi (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vshld_u64 (uint64x1_t __a, uint64x1_t __b)
{
- return (uint64x1_t) __builtin_aarch64_ushldi (__a, __b);
+ return __builtin_aarch64_ushldi_uus (__a, __b);
}
__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
@@ -23235,19 +22092,19 @@ vshll_n_s32 (int32x2_t __a, const int __b)
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vshll_n_u8 (uint8x8_t __a, const int __b)
{
- return (uint16x8_t) __builtin_aarch64_ushll_nv8qi ((int8x8_t) __a, __b);
+ return __builtin_aarch64_ushll_nv8qi_uus (__a, __b);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vshll_n_u16 (uint16x4_t __a, const int __b)
{
- return (uint32x4_t) __builtin_aarch64_ushll_nv4hi ((int16x4_t) __a, __b);
+ return __builtin_aarch64_ushll_nv4hi_uus (__a, __b);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vshll_n_u32 (uint32x2_t __a, const int __b)
{
- return (uint64x2_t) __builtin_aarch64_ushll_nv2si ((int32x2_t) __a, __b);
+ return __builtin_aarch64_ushll_nv2si_uus (__a, __b);
}
/* vshr */
@@ -23389,29 +22246,25 @@ vsli_n_s64 (int64x1_t __a, int64x1_t __b, const int __c)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vsli_n_u8 (uint8x8_t __a, uint8x8_t __b, const int __c)
{
- return (uint8x8_t) __builtin_aarch64_usli_nv8qi ((int8x8_t) __a,
- (int8x8_t) __b, __c);
+ return __builtin_aarch64_usli_nv8qi_uuus (__a, __b, __c);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vsli_n_u16 (uint16x4_t __a, uint16x4_t __b, const int __c)
{
- return (uint16x4_t) __builtin_aarch64_usli_nv4hi ((int16x4_t) __a,
- (int16x4_t) __b, __c);
+ return __builtin_aarch64_usli_nv4hi_uuus (__a, __b, __c);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vsli_n_u32 (uint32x2_t __a, uint32x2_t __b, const int __c)
{
- return (uint32x2_t) __builtin_aarch64_usli_nv2si ((int32x2_t) __a,
- (int32x2_t) __b, __c);
+ return __builtin_aarch64_usli_nv2si_uuus (__a, __b, __c);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vsli_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c)
{
- return (uint64x1_t) __builtin_aarch64_usli_ndi ((int64x1_t) __a,
- (int64x1_t) __b, __c);
+ return __builtin_aarch64_usli_ndi_uuus (__a, __b, __c);
}
__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
@@ -23441,29 +22294,25 @@ vsliq_n_s64 (int64x2_t __a, int64x2_t __b, const int __c)
__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
vsliq_n_u8 (uint8x16_t __a, uint8x16_t __b, const int __c)
{
- return (uint8x16_t) __builtin_aarch64_usli_nv16qi ((int8x16_t) __a,
- (int8x16_t) __b, __c);
+ return __builtin_aarch64_usli_nv16qi_uuus (__a, __b, __c);
}
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vsliq_n_u16 (uint16x8_t __a, uint16x8_t __b, const int __c)
{
- return (uint16x8_t) __builtin_aarch64_usli_nv8hi ((int16x8_t) __a,
- (int16x8_t) __b, __c);
+ return __builtin_aarch64_usli_nv8hi_uuus (__a, __b, __c);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vsliq_n_u32 (uint32x4_t __a, uint32x4_t __b, const int __c)
{
- return (uint32x4_t) __builtin_aarch64_usli_nv4si ((int32x4_t) __a,
- (int32x4_t) __b, __c);
+ return __builtin_aarch64_usli_nv4si_uuus (__a, __b, __c);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vsliq_n_u64 (uint64x2_t __a, uint64x2_t __b, const int __c)
{
- return (uint64x2_t) __builtin_aarch64_usli_nv2di ((int64x2_t) __a,
- (int64x2_t) __b, __c);
+ return __builtin_aarch64_usli_nv2di_uuus (__a, __b, __c);
}
__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
@@ -23475,7 +22324,7 @@ vslid_n_s64 (int64x1_t __a, int64x1_t __b, const int __c)
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vslid_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c)
{
- return (uint64x1_t) __builtin_aarch64_usli_ndi (__a, __b, __c);
+ return __builtin_aarch64_usli_ndi_uuus (__a, __b, __c);
}
/* vsqadd */
@@ -23483,80 +22332,73 @@ vslid_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vsqadd_u8 (uint8x8_t __a, int8x8_t __b)
{
- return (uint8x8_t) __builtin_aarch64_usqaddv8qi ((int8x8_t) __a,
- (int8x8_t) __b);
+ return __builtin_aarch64_usqaddv8qi_uus (__a, __b);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vsqadd_u16 (uint16x4_t __a, int16x4_t __b)
{
- return (uint16x4_t) __builtin_aarch64_usqaddv4hi ((int16x4_t) __a,
- (int16x4_t) __b);
+ return __builtin_aarch64_usqaddv4hi_uus (__a, __b);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vsqadd_u32 (uint32x2_t __a, int32x2_t __b)
{
- return (uint32x2_t) __builtin_aarch64_usqaddv2si ((int32x2_t) __a,
- (int32x2_t) __b);
+ return __builtin_aarch64_usqaddv2si_uus (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vsqadd_u64 (uint64x1_t __a, int64x1_t __b)
{
- return (uint64x1_t) __builtin_aarch64_usqadddi ((int64x1_t) __a, __b);
+ return __builtin_aarch64_usqadddi_uus (__a, __b);
}
__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
vsqaddq_u8 (uint8x16_t __a, int8x16_t __b)
{
- return (uint8x16_t) __builtin_aarch64_usqaddv16qi ((int8x16_t) __a,
- (int8x16_t) __b);
+ return __builtin_aarch64_usqaddv16qi_uus (__a, __b);
}
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vsqaddq_u16 (uint16x8_t __a, int16x8_t __b)
{
- return (uint16x8_t) __builtin_aarch64_usqaddv8hi ((int16x8_t) __a,
- (int16x8_t) __b);
+ return __builtin_aarch64_usqaddv8hi_uus (__a, __b);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vsqaddq_u32 (uint32x4_t __a, int32x4_t __b)
{
- return (uint32x4_t) __builtin_aarch64_usqaddv4si ((int32x4_t) __a,
- (int32x4_t) __b);
+ return __builtin_aarch64_usqaddv4si_uus (__a, __b);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vsqaddq_u64 (uint64x2_t __a, int64x2_t __b)
{
- return (uint64x2_t) __builtin_aarch64_usqaddv2di ((int64x2_t) __a,
- (int64x2_t) __b);
+ return __builtin_aarch64_usqaddv2di_uus (__a, __b);
}
__extension__ static __inline uint8x1_t __attribute__ ((__always_inline__))
vsqaddb_u8 (uint8x1_t __a, int8x1_t __b)
{
- return (uint8x1_t) __builtin_aarch64_usqaddqi ((int8x1_t) __a, __b);
+ return __builtin_aarch64_usqaddqi_uus (__a, __b);
}
__extension__ static __inline uint16x1_t __attribute__ ((__always_inline__))
vsqaddh_u16 (uint16x1_t __a, int16x1_t __b)
{
- return (uint16x1_t) __builtin_aarch64_usqaddhi ((int16x1_t) __a, __b);
+ return __builtin_aarch64_usqaddhi_uus (__a, __b);
}
__extension__ static __inline uint32x1_t __attribute__ ((__always_inline__))
vsqadds_u32 (uint32x1_t __a, int32x1_t __b)
{
- return (uint32x1_t) __builtin_aarch64_usqaddsi ((int32x1_t) __a, __b);
+ return __builtin_aarch64_usqaddsi_uus (__a, __b);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vsqaddd_u64 (uint64x1_t __a, int64x1_t __b)
{
- return (uint64x1_t) __builtin_aarch64_usqadddi ((int64x1_t) __a, __b);
+ return __builtin_aarch64_usqadddi_uus (__a, __b);
}
/* vsqrt */
@@ -23607,29 +22449,25 @@ vsra_n_s64 (int64x1_t __a, int64x1_t __b, const int __c)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vsra_n_u8 (uint8x8_t __a, uint8x8_t __b, const int __c)
{
- return (uint8x8_t) __builtin_aarch64_usra_nv8qi ((int8x8_t) __a,
- (int8x8_t) __b, __c);
+ return __builtin_aarch64_usra_nv8qi_uuus (__a, __b, __c);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vsra_n_u16 (uint16x4_t __a, uint16x4_t __b, const int __c)
{
- return (uint16x4_t) __builtin_aarch64_usra_nv4hi ((int16x4_t) __a,
- (int16x4_t) __b, __c);
+ return __builtin_aarch64_usra_nv4hi_uuus (__a, __b, __c);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vsra_n_u32 (uint32x2_t __a, uint32x2_t __b, const int __c)
{
- return (uint32x2_t) __builtin_aarch64_usra_nv2si ((int32x2_t) __a,
- (int32x2_t) __b, __c);
+ return __builtin_aarch64_usra_nv2si_uuus (__a, __b, __c);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vsra_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c)
{
- return (uint64x1_t) __builtin_aarch64_usra_ndi ((int64x1_t) __a,
- (int64x1_t) __b, __c);
+ return __builtin_aarch64_usra_ndi_uuus (__a, __b, __c);
}
__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
@@ -23659,29 +22497,25 @@ vsraq_n_s64 (int64x2_t __a, int64x2_t __b, const int __c)
__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
vsraq_n_u8 (uint8x16_t __a, uint8x16_t __b, const int __c)
{
- return (uint8x16_t) __builtin_aarch64_usra_nv16qi ((int8x16_t) __a,
- (int8x16_t) __b, __c);
+ return __builtin_aarch64_usra_nv16qi_uuus (__a, __b, __c);
}
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vsraq_n_u16 (uint16x8_t __a, uint16x8_t __b, const int __c)
{
- return (uint16x8_t) __builtin_aarch64_usra_nv8hi ((int16x8_t) __a,
- (int16x8_t) __b, __c);
+ return __builtin_aarch64_usra_nv8hi_uuus (__a, __b, __c);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vsraq_n_u32 (uint32x4_t __a, uint32x4_t __b, const int __c)
{
- return (uint32x4_t) __builtin_aarch64_usra_nv4si ((int32x4_t) __a,
- (int32x4_t) __b, __c);
+ return __builtin_aarch64_usra_nv4si_uuus (__a, __b, __c);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vsraq_n_u64 (uint64x2_t __a, uint64x2_t __b, const int __c)
{
- return (uint64x2_t) __builtin_aarch64_usra_nv2di ((int64x2_t) __a,
- (int64x2_t) __b, __c);
+ return __builtin_aarch64_usra_nv2di_uuus (__a, __b, __c);
}
__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
@@ -23693,7 +22527,7 @@ vsrad_n_s64 (int64x1_t __a, int64x1_t __b, const int __c)
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vsrad_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c)
{
- return (uint64x1_t) __builtin_aarch64_usra_ndi (__a, __b, __c);
+ return __builtin_aarch64_usra_ndi_uuus (__a, __b, __c);
}
/* vsri */
@@ -23725,29 +22559,25 @@ vsri_n_s64 (int64x1_t __a, int64x1_t __b, const int __c)
__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
vsri_n_u8 (uint8x8_t __a, uint8x8_t __b, const int __c)
{
- return (uint8x8_t) __builtin_aarch64_usri_nv8qi ((int8x8_t) __a,
- (int8x8_t) __b, __c);
+ return __builtin_aarch64_usri_nv8qi_uuus (__a, __b, __c);
}
__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
vsri_n_u16 (uint16x4_t __a, uint16x4_t __b, const int __c)
{
- return (uint16x4_t) __builtin_aarch64_usri_nv4hi ((int16x4_t) __a,
- (int16x4_t) __b, __c);
+ return __builtin_aarch64_usri_nv4hi_uuus (__a, __b, __c);
}
__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
vsri_n_u32 (uint32x2_t __a, uint32x2_t __b, const int __c)
{
- return (uint32x2_t) __builtin_aarch64_usri_nv2si ((int32x2_t) __a,
- (int32x2_t) __b, __c);
+ return __builtin_aarch64_usri_nv2si_uuus (__a, __b, __c);
}
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vsri_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c)
{
- return (uint64x1_t) __builtin_aarch64_usri_ndi ((int64x1_t) __a,
- (int64x1_t) __b, __c);
+ return __builtin_aarch64_usri_ndi_uuus (__a, __b, __c);
}
__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
@@ -23777,29 +22607,25 @@ vsriq_n_s64 (int64x2_t __a, int64x2_t __b, const int __c)
__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
vsriq_n_u8 (uint8x16_t __a, uint8x16_t __b, const int __c)
{
- return (uint8x16_t) __builtin_aarch64_usri_nv16qi ((int8x16_t) __a,
- (int8x16_t) __b, __c);
+ return __builtin_aarch64_usri_nv16qi_uuus (__a, __b, __c);
}
__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
vsriq_n_u16 (uint16x8_t __a, uint16x8_t __b, const int __c)
{
- return (uint16x8_t) __builtin_aarch64_usri_nv8hi ((int16x8_t) __a,
- (int16x8_t) __b, __c);
+ return __builtin_aarch64_usri_nv8hi_uuus (__a, __b, __c);
}
__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
vsriq_n_u32 (uint32x4_t __a, uint32x4_t __b, const int __c)
{
- return (uint32x4_t) __builtin_aarch64_usri_nv4si ((int32x4_t) __a,
- (int32x4_t) __b, __c);
+ return __builtin_aarch64_usri_nv4si_uuus (__a, __b, __c);
}
__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
vsriq_n_u64 (uint64x2_t __a, uint64x2_t __b, const int __c)
{
- return (uint64x2_t) __builtin_aarch64_usri_nv2di ((int64x2_t) __a,
- (int64x2_t) __b, __c);
+ return __builtin_aarch64_usri_nv2di_uuus (__a, __b, __c);
}
__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
@@ -23811,7 +22637,7 @@ vsrid_n_s64 (int64x1_t __a, int64x1_t __b, const int __c)
__extension__ static __inline uint64x1_t __attribute__ ((__always_inline__))
vsrid_n_u64 (uint64x1_t __a, uint64x1_t __b, const int __c)
{
- return (uint64x1_t) __builtin_aarch64_usri_ndi (__a, __b, __c);
+ return __builtin_aarch64_usri_ndi_uuus (__a, __b, __c);
}
/* vst1 */
@@ -24915,6 +23741,438 @@ vtbx3_p8 (poly8x8_t __r, poly8x8x3_t __tab, uint8x8_t __idx)
/* vtrn */
+__extension__ static __inline float32x2_t __attribute__ ((__always_inline__))
+vtrn1_f32 (float32x2_t __a, float32x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {3, 1});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {0, 2});
+#endif
+}
+
+__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
+vtrn1_p8 (poly8x8_t __a, poly8x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {9, 1, 11, 3, 13, 5, 15, 7});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {0, 8, 2, 10, 4, 12, 6, 14});
+#endif
+}
+
+__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__))
+vtrn1_p16 (poly16x4_t __a, poly16x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {5, 1, 7, 3});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {0, 4, 2, 6});
+#endif
+}
+
+__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
+vtrn1_s8 (int8x8_t __a, int8x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {9, 1, 11, 3, 13, 5, 15, 7});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {0, 8, 2, 10, 4, 12, 6, 14});
+#endif
+}
+
+__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
+vtrn1_s16 (int16x4_t __a, int16x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {5, 1, 7, 3});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {0, 4, 2, 6});
+#endif
+}
+
+__extension__ static __inline int32x2_t __attribute__ ((__always_inline__))
+vtrn1_s32 (int32x2_t __a, int32x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {3, 1});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {0, 2});
+#endif
+}
+
+__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
+vtrn1_u8 (uint8x8_t __a, uint8x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {9, 1, 11, 3, 13, 5, 15, 7});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {0, 8, 2, 10, 4, 12, 6, 14});
+#endif
+}
+
+__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
+vtrn1_u16 (uint16x4_t __a, uint16x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {5, 1, 7, 3});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {0, 4, 2, 6});
+#endif
+}
+
+__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
+vtrn1_u32 (uint32x2_t __a, uint32x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {3, 1});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {0, 2});
+#endif
+}
+
+__extension__ static __inline float32x4_t __attribute__ ((__always_inline__))
+vtrn1q_f32 (float32x4_t __a, float32x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {5, 1, 7, 3});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {0, 4, 2, 6});
+#endif
+}
+
+__extension__ static __inline float64x2_t __attribute__ ((__always_inline__))
+vtrn1q_f64 (float64x2_t __a, float64x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {3, 1});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {0, 2});
+#endif
+}
+
+__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
+vtrn1q_p8 (poly8x16_t __a, poly8x16_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {17, 1, 19, 3, 21, 5, 23, 7, 25, 9, 27, 11, 29, 13, 31, 15});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30});
+#endif
+}
+
+__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__))
+vtrn1q_p16 (poly16x8_t __a, poly16x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {9, 1, 11, 3, 13, 5, 15, 7});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {0, 8, 2, 10, 4, 12, 6, 14});
+#endif
+}
+
+__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
+vtrn1q_s8 (int8x16_t __a, int8x16_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {17, 1, 19, 3, 21, 5, 23, 7, 25, 9, 27, 11, 29, 13, 31, 15});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30});
+#endif
+}
+
+__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
+vtrn1q_s16 (int16x8_t __a, int16x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {9, 1, 11, 3, 13, 5, 15, 7});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {0, 8, 2, 10, 4, 12, 6, 14});
+#endif
+}
+
+__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
+vtrn1q_s32 (int32x4_t __a, int32x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {5, 1, 7, 3});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {0, 4, 2, 6});
+#endif
+}
+
+__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
+vtrn1q_s64 (int64x2_t __a, int64x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {3, 1});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {0, 2});
+#endif
+}
+
+__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
+vtrn1q_u8 (uint8x16_t __a, uint8x16_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {17, 1, 19, 3, 21, 5, 23, 7, 25, 9, 27, 11, 29, 13, 31, 15});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30});
+#endif
+}
+
+__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
+vtrn1q_u16 (uint16x8_t __a, uint16x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {9, 1, 11, 3, 13, 5, 15, 7});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {0, 8, 2, 10, 4, 12, 6, 14});
+#endif
+}
+
+__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
+vtrn1q_u32 (uint32x4_t __a, uint32x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {5, 1, 7, 3});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {0, 4, 2, 6});
+#endif
+}
+
+__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
+vtrn1q_u64 (uint64x2_t __a, uint64x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {3, 1});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {0, 2});
+#endif
+}
+
+__extension__ static __inline float32x2_t __attribute__ ((__always_inline__))
+vtrn2_f32 (float32x2_t __a, float32x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {2, 0});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {1, 3});
+#endif
+}
+
+__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
+vtrn2_p8 (poly8x8_t __a, poly8x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {8, 0, 10, 2, 12, 4, 14, 6});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {1, 9, 3, 11, 5, 13, 7, 15});
+#endif
+}
+
+__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__))
+vtrn2_p16 (poly16x4_t __a, poly16x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {4, 0, 6, 2});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {1, 5, 3, 7});
+#endif
+}
+
+__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
+vtrn2_s8 (int8x8_t __a, int8x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {8, 0, 10, 2, 12, 4, 14, 6});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {1, 9, 3, 11, 5, 13, 7, 15});
+#endif
+}
+
+__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
+vtrn2_s16 (int16x4_t __a, int16x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {4, 0, 6, 2});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {1, 5, 3, 7});
+#endif
+}
+
+__extension__ static __inline int32x2_t __attribute__ ((__always_inline__))
+vtrn2_s32 (int32x2_t __a, int32x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {2, 0});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {1, 3});
+#endif
+}
+
+__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
+vtrn2_u8 (uint8x8_t __a, uint8x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {8, 0, 10, 2, 12, 4, 14, 6});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {1, 9, 3, 11, 5, 13, 7, 15});
+#endif
+}
+
+__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
+vtrn2_u16 (uint16x4_t __a, uint16x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {4, 0, 6, 2});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {1, 5, 3, 7});
+#endif
+}
+
+__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
+vtrn2_u32 (uint32x2_t __a, uint32x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {2, 0});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {1, 3});
+#endif
+}
+
+__extension__ static __inline float32x4_t __attribute__ ((__always_inline__))
+vtrn2q_f32 (float32x4_t __a, float32x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {4, 0, 6, 2});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {1, 5, 3, 7});
+#endif
+}
+
+__extension__ static __inline float64x2_t __attribute__ ((__always_inline__))
+vtrn2q_f64 (float64x2_t __a, float64x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {2, 0});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {1, 3});
+#endif
+}
+
+__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
+vtrn2q_p8 (poly8x16_t __a, poly8x16_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {16, 0, 18, 2, 20, 4, 22, 6, 24, 8, 26, 10, 28, 12, 30, 14});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31});
+#endif
+}
+
+__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__))
+vtrn2q_p16 (poly16x8_t __a, poly16x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {8, 0, 10, 2, 12, 4, 14, 6});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {1, 9, 3, 11, 5, 13, 7, 15});
+#endif
+}
+
+__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
+vtrn2q_s8 (int8x16_t __a, int8x16_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {16, 0, 18, 2, 20, 4, 22, 6, 24, 8, 26, 10, 28, 12, 30, 14});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31});
+#endif
+}
+
+__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
+vtrn2q_s16 (int16x8_t __a, int16x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {8, 0, 10, 2, 12, 4, 14, 6});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {1, 9, 3, 11, 5, 13, 7, 15});
+#endif
+}
+
+__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
+vtrn2q_s32 (int32x4_t __a, int32x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {4, 0, 6, 2});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {1, 5, 3, 7});
+#endif
+}
+
+__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
+vtrn2q_s64 (int64x2_t __a, int64x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {2, 0});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {1, 3});
+#endif
+}
+
+__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
+vtrn2q_u8 (uint8x16_t __a, uint8x16_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {16, 0, 18, 2, 20, 4, 22, 6, 24, 8, 26, 10, 28, 12, 30, 14});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31});
+#endif
+}
+
+__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
+vtrn2q_u16 (uint16x8_t __a, uint16x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {8, 0, 10, 2, 12, 4, 14, 6});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {1, 9, 3, 11, 5, 13, 7, 15});
+#endif
+}
+
+__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
+vtrn2q_u32 (uint32x4_t __a, uint32x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {4, 0, 6, 2});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {1, 5, 3, 7});
+#endif
+}
+
+__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
+vtrn2q_u64 (uint64x2_t __a, uint64x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {2, 0});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {1, 3});
+#endif
+}
+
__extension__ static __inline float32x2x2_t __attribute__ ((__always_inline__))
vtrn_f32 (float32x2_t a, float32x2_t b)
{
@@ -25145,73 +24403,73 @@ vtstd_u64 (uint64x1_t __a, uint64x1_t __b)
__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
vuqadd_s8 (int8x8_t __a, uint8x8_t __b)
{
- return (int8x8_t) __builtin_aarch64_suqaddv8qi (__a, (int8x8_t) __b);
+ return __builtin_aarch64_suqaddv8qi_ssu (__a, __b);
}
__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
vuqadd_s16 (int16x4_t __a, uint16x4_t __b)
{
- return (int16x4_t) __builtin_aarch64_suqaddv4hi (__a, (int16x4_t) __b);
+ return __builtin_aarch64_suqaddv4hi_ssu (__a, __b);
}
__extension__ static __inline int32x2_t __attribute__ ((__always_inline__))
vuqadd_s32 (int32x2_t __a, uint32x2_t __b)
{
- return (int32x2_t) __builtin_aarch64_suqaddv2si (__a, (int32x2_t) __b);
+ return __builtin_aarch64_suqaddv2si_ssu (__a, __b);
}
__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
vuqadd_s64 (int64x1_t __a, uint64x1_t __b)
{
- return (int64x1_t) __builtin_aarch64_suqadddi (__a, (int64x1_t) __b);
+ return __builtin_aarch64_suqadddi_ssu (__a, __b);
}
__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
vuqaddq_s8 (int8x16_t __a, uint8x16_t __b)
{
- return (int8x16_t) __builtin_aarch64_suqaddv16qi (__a, (int8x16_t) __b);
+ return __builtin_aarch64_suqaddv16qi_ssu (__a, __b);
}
__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
vuqaddq_s16 (int16x8_t __a, uint16x8_t __b)
{
- return (int16x8_t) __builtin_aarch64_suqaddv8hi (__a, (int16x8_t) __b);
+ return __builtin_aarch64_suqaddv8hi_ssu (__a, __b);
}
__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
vuqaddq_s32 (int32x4_t __a, uint32x4_t __b)
{
- return (int32x4_t) __builtin_aarch64_suqaddv4si (__a, (int32x4_t) __b);
+ return __builtin_aarch64_suqaddv4si_ssu (__a, __b);
}
__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
vuqaddq_s64 (int64x2_t __a, uint64x2_t __b)
{
- return (int64x2_t) __builtin_aarch64_suqaddv2di (__a, (int64x2_t) __b);
+ return __builtin_aarch64_suqaddv2di_ssu (__a, __b);
}
__extension__ static __inline int8x1_t __attribute__ ((__always_inline__))
vuqaddb_s8 (int8x1_t __a, uint8x1_t __b)
{
- return (int8x1_t) __builtin_aarch64_suqaddqi (__a, (int8x1_t) __b);
+ return __builtin_aarch64_suqaddqi_ssu (__a, __b);
}
__extension__ static __inline int16x1_t __attribute__ ((__always_inline__))
vuqaddh_s16 (int16x1_t __a, uint16x1_t __b)
{
- return (int16x1_t) __builtin_aarch64_suqaddhi (__a, (int16x1_t) __b);
+ return __builtin_aarch64_suqaddhi_ssu (__a, __b);
}
__extension__ static __inline int32x1_t __attribute__ ((__always_inline__))
vuqadds_s32 (int32x1_t __a, uint32x1_t __b)
{
- return (int32x1_t) __builtin_aarch64_suqaddsi (__a, (int32x1_t) __b);
+ return __builtin_aarch64_suqaddsi_ssu (__a, __b);
}
__extension__ static __inline int64x1_t __attribute__ ((__always_inline__))
vuqaddd_s64 (int64x1_t __a, uint64x1_t __b)
{
- return (int64x1_t) __builtin_aarch64_suqadddi (__a, (int64x1_t) __b);
+ return __builtin_aarch64_suqadddi_ssu (__a, __b);
}
#define __DEFINTERLEAVE(op, rettype, intype, funcsuffix, Q) \
@@ -25245,6 +24503,438 @@ vuqaddd_s64 (int64x1_t __a, uint64x1_t __b)
/* vuzp */
+__extension__ static __inline float32x2_t __attribute__ ((__always_inline__))
+vuzp1_f32 (float32x2_t __a, float32x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {3, 1});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {0, 2});
+#endif
+}
+
+__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
+vuzp1_p8 (poly8x8_t __a, poly8x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {9, 11, 13, 15, 1, 3, 5, 7});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {0, 2, 4, 6, 8, 10, 12, 14});
+#endif
+}
+
+__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__))
+vuzp1_p16 (poly16x4_t __a, poly16x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {5, 7, 1, 3});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {0, 2, 4, 6});
+#endif
+}
+
+__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
+vuzp1_s8 (int8x8_t __a, int8x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {9, 11, 13, 15, 1, 3, 5, 7});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {0, 2, 4, 6, 8, 10, 12, 14});
+#endif
+}
+
+__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
+vuzp1_s16 (int16x4_t __a, int16x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {5, 7, 1, 3});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {0, 2, 4, 6});
+#endif
+}
+
+__extension__ static __inline int32x2_t __attribute__ ((__always_inline__))
+vuzp1_s32 (int32x2_t __a, int32x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {3, 1});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {0, 2});
+#endif
+}
+
+__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
+vuzp1_u8 (uint8x8_t __a, uint8x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {9, 11, 13, 15, 1, 3, 5, 7});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {0, 2, 4, 6, 8, 10, 12, 14});
+#endif
+}
+
+__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
+vuzp1_u16 (uint16x4_t __a, uint16x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {5, 7, 1, 3});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {0, 2, 4, 6});
+#endif
+}
+
+__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
+vuzp1_u32 (uint32x2_t __a, uint32x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {3, 1});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {0, 2});
+#endif
+}
+
+__extension__ static __inline float32x4_t __attribute__ ((__always_inline__))
+vuzp1q_f32 (float32x4_t __a, float32x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {5, 7, 1, 3});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {0, 2, 4, 6});
+#endif
+}
+
+__extension__ static __inline float64x2_t __attribute__ ((__always_inline__))
+vuzp1q_f64 (float64x2_t __a, float64x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {3, 1});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {0, 2});
+#endif
+}
+
+__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
+vuzp1q_p8 (poly8x16_t __a, poly8x16_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint8x16_t)
+ {17, 19, 21, 23, 25, 27, 29, 31, 1, 3, 5, 7, 9, 11, 13, 15});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x16_t)
+ {0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30});
+#endif
+}
+
+__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__))
+vuzp1q_p16 (poly16x8_t __a, poly16x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {9, 11, 13, 15, 1, 3, 5, 7});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {0, 2, 4, 6, 8, 10, 12, 14});
+#endif
+}
+
+__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
+vuzp1q_s8 (int8x16_t __a, int8x16_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {17, 19, 21, 23, 25, 27, 29, 31, 1, 3, 5, 7, 9, 11, 13, 15});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30});
+#endif
+}
+
+__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
+vuzp1q_s16 (int16x8_t __a, int16x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {9, 11, 13, 15, 1, 3, 5, 7});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {0, 2, 4, 6, 8, 10, 12, 14});
+#endif
+}
+
+__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
+vuzp1q_s32 (int32x4_t __a, int32x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {5, 7, 1, 3});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {0, 2, 4, 6});
+#endif
+}
+
+__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
+vuzp1q_s64 (int64x2_t __a, int64x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {3, 1});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {0, 2});
+#endif
+}
+
+__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
+vuzp1q_u8 (uint8x16_t __a, uint8x16_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {17, 19, 21, 23, 25, 27, 29, 31, 1, 3, 5, 7, 9, 11, 13, 15});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30});
+#endif
+}
+
+__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
+vuzp1q_u16 (uint16x8_t __a, uint16x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {9, 11, 13, 15, 1, 3, 5, 7});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {0, 2, 4, 6, 8, 10, 12, 14});
+#endif
+}
+
+__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
+vuzp1q_u32 (uint32x4_t __a, uint32x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {5, 7, 1, 3});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {0, 2, 4, 6});
+#endif
+}
+
+__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
+vuzp1q_u64 (uint64x2_t __a, uint64x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {3, 1});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {0, 2});
+#endif
+}
+
+__extension__ static __inline float32x2_t __attribute__ ((__always_inline__))
+vuzp2_f32 (float32x2_t __a, float32x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {2, 0});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {1, 3});
+#endif
+}
+
+__extension__ static __inline poly8x8_t __attribute__ ((__always_inline__))
+vuzp2_p8 (poly8x8_t __a, poly8x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {8, 10, 12, 14, 0, 2, 4, 6});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {1, 3, 5, 7, 9, 11, 13, 15});
+#endif
+}
+
+__extension__ static __inline poly16x4_t __attribute__ ((__always_inline__))
+vuzp2_p16 (poly16x4_t __a, poly16x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {4, 6, 0, 2});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {1, 3, 5, 7});
+#endif
+}
+
+__extension__ static __inline int8x8_t __attribute__ ((__always_inline__))
+vuzp2_s8 (int8x8_t __a, int8x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {8, 10, 12, 14, 0, 2, 4, 6});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {1, 3, 5, 7, 9, 11, 13, 15});
+#endif
+}
+
+__extension__ static __inline int16x4_t __attribute__ ((__always_inline__))
+vuzp2_s16 (int16x4_t __a, int16x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {4, 6, 0, 2});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {1, 3, 5, 7});
+#endif
+}
+
+__extension__ static __inline int32x2_t __attribute__ ((__always_inline__))
+vuzp2_s32 (int32x2_t __a, int32x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {2, 0});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {1, 3});
+#endif
+}
+
+__extension__ static __inline uint8x8_t __attribute__ ((__always_inline__))
+vuzp2_u8 (uint8x8_t __a, uint8x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {8, 10, 12, 14, 0, 2, 4, 6});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x8_t) {1, 3, 5, 7, 9, 11, 13, 15});
+#endif
+}
+
+__extension__ static __inline uint16x4_t __attribute__ ((__always_inline__))
+vuzp2_u16 (uint16x4_t __a, uint16x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {4, 6, 0, 2});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x4_t) {1, 3, 5, 7});
+#endif
+}
+
+__extension__ static __inline uint32x2_t __attribute__ ((__always_inline__))
+vuzp2_u32 (uint32x2_t __a, uint32x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {2, 0});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x2_t) {1, 3});
+#endif
+}
+
+__extension__ static __inline float32x4_t __attribute__ ((__always_inline__))
+vuzp2q_f32 (float32x4_t __a, float32x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {4, 6, 0, 2});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {1, 3, 5, 7});
+#endif
+}
+
+__extension__ static __inline float64x2_t __attribute__ ((__always_inline__))
+vuzp2q_f64 (float64x2_t __a, float64x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {2, 0});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {1, 3});
+#endif
+}
+
+__extension__ static __inline poly8x16_t __attribute__ ((__always_inline__))
+vuzp2q_p8 (poly8x16_t __a, poly8x16_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {16, 18, 20, 22, 24, 26, 28, 30, 0, 2, 4, 6, 8, 10, 12, 14});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31});
+#endif
+}
+
+__extension__ static __inline poly16x8_t __attribute__ ((__always_inline__))
+vuzp2q_p16 (poly16x8_t __a, poly16x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {8, 10, 12, 14, 0, 2, 4, 6});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {1, 3, 5, 7, 9, 11, 13, 15});
+#endif
+}
+
+__extension__ static __inline int8x16_t __attribute__ ((__always_inline__))
+vuzp2q_s8 (int8x16_t __a, int8x16_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {16, 18, 20, 22, 24, 26, 28, 30, 0, 2, 4, 6, 8, 10, 12, 14});
+#else
+ return __builtin_shuffle (__a, __b,
+ (uint8x16_t) {1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31});
+#endif
+}
+
+__extension__ static __inline int16x8_t __attribute__ ((__always_inline__))
+vuzp2q_s16 (int16x8_t __a, int16x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {8, 10, 12, 14, 0, 2, 4, 6});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {1, 3, 5, 7, 9, 11, 13, 15});
+#endif
+}
+
+__extension__ static __inline int32x4_t __attribute__ ((__always_inline__))
+vuzp2q_s32 (int32x4_t __a, int32x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {4, 6, 0, 2});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {1, 3, 5, 7});
+#endif
+}
+
+__extension__ static __inline int64x2_t __attribute__ ((__always_inline__))
+vuzp2q_s64 (int64x2_t __a, int64x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {2, 0});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {1, 3});
+#endif
+}
+
+__extension__ static __inline uint8x16_t __attribute__ ((__always_inline__))
+vuzp2q_u8 (uint8x16_t __a, uint8x16_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint8x16_t)
+ {16, 18, 20, 22, 24, 26, 28, 30, 0, 2, 4, 6, 8, 10, 12, 14});
+#else
+ return __builtin_shuffle (__a, __b, (uint8x16_t)
+ {1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31});
+#endif
+}
+
+__extension__ static __inline uint16x8_t __attribute__ ((__always_inline__))
+vuzp2q_u16 (uint16x8_t __a, uint16x8_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {8, 10, 12, 14, 0, 2, 4, 6});
+#else
+ return __builtin_shuffle (__a, __b, (uint16x8_t) {1, 3, 5, 7, 9, 11, 13, 15});
+#endif
+}
+
+__extension__ static __inline uint32x4_t __attribute__ ((__always_inline__))
+vuzp2q_u32 (uint32x4_t __a, uint32x4_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {4, 6, 0, 2});
+#else
+ return __builtin_shuffle (__a, __b, (uint32x4_t) {1, 3, 5, 7});
+#endif
+}
+
+__extension__ static __inline uint64x2_t __attribute__ ((__always_inline__))
+vuzp2q_u64 (uint64x2_t __a, uint64x2_t __b)
+{
+#ifdef __AARCH64EB__
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {2, 0});
+#else
+ return __builtin_shuffle (__a, __b, (uint64x2_t) {1, 3});
+#endif
+}
+
__INTERLEAVE_LIST (uzp)
/* vzip */
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/libgcc.a b/lib/gcc/aarch64-linux-android/4.9.1/libgcc.a
index c56bcd0..a02e0f2 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/libgcc.a
+++ b/lib/gcc/aarch64-linux-android/4.9.1/libgcc.a
Binary files differ
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/libgcov.a b/lib/gcc/aarch64-linux-android/4.9.1/libgcov.a
index 1cde3fe..40e11aa 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/libgcov.a
+++ b/lib/gcc/aarch64-linux-android/4.9.1/libgcov.a
Binary files differ
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/gtype.state b/lib/gcc/aarch64-linux-android/4.9.1/plugin/gtype.state
index 3923d7d..2806ccf 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/gtype.state
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/gtype.state
@@ -4,8 +4,8 @@
;;; Don't parse this file wihout knowing GCC gengtype internals.
;;; This file should be parsed by the same gengtype which wrote it.
-(!version "4.9.1 20140529 (prerelease)")
-(!srcdir "/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc")
+(!version "4.9.1 20140710 (prerelease)")
+(!srcdir "/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc")
(!languages 9 ada c cp fortran go java lto objc objcp)
(!fileslist 190
(!srcfile 511 "../libcpp/include/line-map.h")
@@ -227,7 +227,7 @@
(!type union 7
(!type already_seen 5)
gc_pointed_to "tree_node"
- (!srcfileloc "tree-core.h" 1626)
+ (!srcfileloc "tree-core.h" 1631)
(!fields 38
(!pair "base"
(!type struct 8 nil gc_used "tree_base"
@@ -310,11 +310,11 @@
(!srcfileloc "tree-core.h" 729)
nil )
(!pair "u"
- (!type union 12 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/tree-core.h:731"
+ (!type union 12 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/tree-core.h:731"
(!srcfileloc "tree-core.h" 765)
(!fields 3
(!pair "bits"
- (!type struct 13 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/tree-core.h:734"
+ (!type struct 13 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/tree-core.h:734"
(!srcfileloc "tree-core.h" 757)
(!fields 16
(!pair "lang_flag_0"
@@ -402,7 +402,7 @@
)
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1584)
+ (!srcfileloc "tree-core.h" 1589)
(!options
(!option tag string "TS_BASE")
)
@@ -421,7 +421,7 @@
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1585)
+ (!srcfileloc "tree-core.h" 1590)
(!options
(!option tag string "TS_TYPED")
)
@@ -440,7 +440,7 @@
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1586)
+ (!srcfileloc "tree-core.h" 1591)
(!options
(!option tag string "TS_COMMON")
)
@@ -459,7 +459,7 @@
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1587)
+ (!srcfileloc "tree-core.h" 1592)
(!options
(!option tag string "TS_INT_CST")
)
@@ -516,7 +516,7 @@
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1588)
+ (!srcfileloc "tree-core.h" 1593)
(!options
(!option tag string "TS_REAL_CST")
)
@@ -551,7 +551,7 @@
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1589)
+ (!srcfileloc "tree-core.h" 1594)
(!options
(!option tag string "TS_FIXED_CST")
)
@@ -575,7 +575,7 @@
)
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1590)
+ (!srcfileloc "tree-core.h" 1595)
(!options
(!option tag string "TS_VECTOR")
)
@@ -608,7 +608,7 @@
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1591)
+ (!srcfileloc "tree-core.h" 1596)
(!options
(!option tag string "TS_STRING")
)
@@ -631,7 +631,7 @@
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1592)
+ (!srcfileloc "tree-core.h" 1597)
(!options
(!option tag string "TS_COMPLEX")
)
@@ -674,181 +674,181 @@
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1593)
+ (!srcfileloc "tree-core.h" 1598)
(!options
(!option tag string "TS_IDENTIFIER")
)
)
(!pair "decl_minimal"
(!type struct 37 nil gc_used "tree_decl_minimal"
- (!srcfileloc "tree-core.h" 1314)
+ (!srcfileloc "tree-core.h" 1319)
(!fields 5
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "tree-core.h" 1309)
+ (!srcfileloc "tree-core.h" 1314)
nil )
(!pair "locus"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1310)
+ (!srcfileloc "tree-core.h" 1315)
nil )
(!pair "uid"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1311)
+ (!srcfileloc "tree-core.h" 1316)
nil )
(!pair "name"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1312)
+ (!srcfileloc "tree-core.h" 1317)
nil )
(!pair "context"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1313)
+ (!srcfileloc "tree-core.h" 1318)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1594)
+ (!srcfileloc "tree-core.h" 1599)
(!options
(!option tag string "TS_DECL_MINIMAL")
)
)
(!pair "decl_common"
(!type struct 38 nil gc_used "tree_decl_common"
- (!srcfileloc "tree-core.h" 1380)
+ (!srcfileloc "tree-core.h" 1385)
(!fields 35
(!pair "common"
(!type already_seen 37)
- (!srcfileloc "tree-core.h" 1317)
+ (!srcfileloc "tree-core.h" 1322)
nil )
(!pair "size"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1318)
+ (!srcfileloc "tree-core.h" 1323)
nil )
(!pair "mode"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1320)
+ (!srcfileloc "tree-core.h" 1325)
nil )
(!pair "nonlocal_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1322)
+ (!srcfileloc "tree-core.h" 1327)
nil )
(!pair "virtual_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1323)
+ (!srcfileloc "tree-core.h" 1328)
nil )
(!pair "ignored_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1324)
+ (!srcfileloc "tree-core.h" 1329)
nil )
(!pair "abstract_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1325)
+ (!srcfileloc "tree-core.h" 1330)
nil )
(!pair "artificial_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1326)
+ (!srcfileloc "tree-core.h" 1331)
nil )
(!pair "preserve_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1327)
+ (!srcfileloc "tree-core.h" 1332)
nil )
(!pair "debug_expr_is_from"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1328)
+ (!srcfileloc "tree-core.h" 1333)
nil )
(!pair "lang_flag_0"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1330)
+ (!srcfileloc "tree-core.h" 1335)
nil )
(!pair "lang_flag_1"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1331)
+ (!srcfileloc "tree-core.h" 1336)
nil )
(!pair "lang_flag_2"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1332)
+ (!srcfileloc "tree-core.h" 1337)
nil )
(!pair "lang_flag_3"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1333)
+ (!srcfileloc "tree-core.h" 1338)
nil )
(!pair "lang_flag_4"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1334)
+ (!srcfileloc "tree-core.h" 1339)
nil )
(!pair "lang_flag_5"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1335)
+ (!srcfileloc "tree-core.h" 1340)
nil )
(!pair "lang_flag_6"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1336)
+ (!srcfileloc "tree-core.h" 1341)
nil )
(!pair "lang_flag_7"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1337)
+ (!srcfileloc "tree-core.h" 1342)
nil )
(!pair "lang_flag_8"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1338)
+ (!srcfileloc "tree-core.h" 1343)
nil )
(!pair "decl_flag_0"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1341)
+ (!srcfileloc "tree-core.h" 1346)
nil )
(!pair "decl_flag_1"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1345)
+ (!srcfileloc "tree-core.h" 1350)
nil )
(!pair "decl_flag_2"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1349)
+ (!srcfileloc "tree-core.h" 1354)
nil )
(!pair "decl_flag_3"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1351)
+ (!srcfileloc "tree-core.h" 1356)
nil )
(!pair "gimple_reg_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1354)
+ (!srcfileloc "tree-core.h" 1359)
nil )
(!pair "decl_by_reference_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1356)
+ (!srcfileloc "tree-core.h" 1361)
nil )
(!pair "decl_read_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1358)
+ (!srcfileloc "tree-core.h" 1363)
nil )
(!pair "decl_nonshareable_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1360)
+ (!srcfileloc "tree-core.h" 1365)
nil )
(!pair "off_align"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1363)
+ (!srcfileloc "tree-core.h" 1368)
nil )
(!pair "align"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1368)
+ (!srcfileloc "tree-core.h" 1373)
nil )
(!pair "pt_uid"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1371)
+ (!srcfileloc "tree-core.h" 1376)
nil )
(!pair "size_unit"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1373)
+ (!srcfileloc "tree-core.h" 1378)
nil )
(!pair "initial"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1374)
+ (!srcfileloc "tree-core.h" 1379)
nil )
(!pair "attributes"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1375)
+ (!srcfileloc "tree-core.h" 1380)
nil )
(!pair "abstract_origin"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1376)
+ (!srcfileloc "tree-core.h" 1381)
nil )
(!pair "lang_specific"
(!type pointer 39 nil gc_used
@@ -933,15 +933,15 @@
(!type user_struct 46
(!type already_seen 45)
gc_pointed_to "vec<tree,va_gc>"
- (!srcfileloc "tree-core.h" 1228)
+ (!srcfileloc "tree-core.h" 1233)
(!fields 2
(!pair "va_gc"
(!type already_seen 3)
- (!srcfileloc "tree-core.h" 1228)
+ (!srcfileloc "tree-core.h" 1233)
nil )
(!pair "tree"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1228)
+ (!srcfileloc "tree-core.h" 1233)
nil )
)
)
@@ -1166,7 +1166,7 @@
)
nil 32
(!type lang_struct 56 nil gc_used "lang_decl_u"
- (!srcfileloc "cp/cp-tree.h" 2069)
+ (!srcfileloc "cp/cp-tree.h" 2071)
(!fields 0 )
(!options
(!option desc string "%h.base.selector")
@@ -1176,223 +1176,223 @@
(!type already_seen 43)
(!type union 57 nil gc_used "lang_decl_u"
- (!srcfileloc "cp/cp-tree.h" 2069)
+ (!srcfileloc "cp/cp-tree.h" 2071)
(!fields 5
(!pair "base"
(!type struct 58 nil gc_used "lang_decl_base"
- (!srcfileloc "cp/cp-tree.h" 1948)
+ (!srcfileloc "cp/cp-tree.h" 1950)
(!fields 12
(!pair "selector"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1935)
+ (!srcfileloc "cp/cp-tree.h" 1937)
nil )
(!pair "language"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1936)
+ (!srcfileloc "cp/cp-tree.h" 1938)
nil )
(!pair "use_template"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1937)
+ (!srcfileloc "cp/cp-tree.h" 1939)
nil )
(!pair "not_really_extern"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1938)
+ (!srcfileloc "cp/cp-tree.h" 1940)
nil )
(!pair "initialized_in_class"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1939)
+ (!srcfileloc "cp/cp-tree.h" 1941)
nil )
(!pair "repo_available_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1940)
+ (!srcfileloc "cp/cp-tree.h" 1942)
nil )
(!pair "threadprivate_or_deleted_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1941)
+ (!srcfileloc "cp/cp-tree.h" 1943)
nil )
(!pair "anticipated_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1942)
+ (!srcfileloc "cp/cp-tree.h" 1944)
nil )
(!pair "friend_attr"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1943)
+ (!srcfileloc "cp/cp-tree.h" 1945)
nil )
(!pair "template_conv_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1944)
+ (!srcfileloc "cp/cp-tree.h" 1946)
nil )
(!pair "odr_used"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1945)
+ (!srcfileloc "cp/cp-tree.h" 1947)
nil )
(!pair "u2sel"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1946)
+ (!srcfileloc "cp/cp-tree.h" 1948)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 2064)
+ (!srcfileloc "cp/cp-tree.h" 2066)
(!options
(!option default string "")
)
)
(!pair "min"
(!type struct 59 nil gc_used "lang_decl_min"
- (!srcfileloc "cp/cp-tree.h" 1980)
+ (!srcfileloc "cp/cp-tree.h" 1982)
(!fields 3
(!pair "base"
(!type already_seen 58)
- (!srcfileloc "cp/cp-tree.h" 1962)
+ (!srcfileloc "cp/cp-tree.h" 1964)
nil )
(!pair "template_info"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1969)
+ (!srcfileloc "cp/cp-tree.h" 1971)
nil )
(!pair "u2"
(!type union 60 nil gc_used "lang_decl_u2"
- (!srcfileloc "cp/cp-tree.h" 1979)
+ (!srcfileloc "cp/cp-tree.h" 1981)
(!fields 2
(!pair "access"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1975)
+ (!srcfileloc "cp/cp-tree.h" 1977)
(!options
(!option tag string "0")
)
)
(!pair "discriminator"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1978)
+ (!srcfileloc "cp/cp-tree.h" 1980)
(!options
(!option tag string "1")
)
)
)
nil 260 nil )
- (!srcfileloc "cp/cp-tree.h" 1979)
+ (!srcfileloc "cp/cp-tree.h" 1981)
(!options
(!option desc string "%0.u.base.u2sel")
)
)
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 2065)
+ (!srcfileloc "cp/cp-tree.h" 2067)
(!options
(!option tag string "0")
)
)
(!pair "fn"
(!type struct 61 nil gc_used "lang_decl_fn"
- (!srcfileloc "cp/cp-tree.h" 2041)
+ (!srcfileloc "cp/cp-tree.h" 2043)
(!fields 22
(!pair "min"
(!type already_seen 59)
- (!srcfileloc "cp/cp-tree.h" 1985)
+ (!srcfileloc "cp/cp-tree.h" 1987)
nil )
(!pair "operator_code"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1989)
+ (!srcfileloc "cp/cp-tree.h" 1991)
nil )
(!pair "global_ctor_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1991)
+ (!srcfileloc "cp/cp-tree.h" 1993)
nil )
(!pair "global_dtor_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1992)
+ (!srcfileloc "cp/cp-tree.h" 1994)
nil )
(!pair "constructor_attr"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1993)
+ (!srcfileloc "cp/cp-tree.h" 1995)
nil )
(!pair "destructor_attr"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1994)
+ (!srcfileloc "cp/cp-tree.h" 1996)
nil )
(!pair "assignment_operator_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1995)
+ (!srcfileloc "cp/cp-tree.h" 1997)
nil )
(!pair "static_function"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1996)
+ (!srcfileloc "cp/cp-tree.h" 1998)
nil )
(!pair "pure_virtual"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1997)
+ (!srcfileloc "cp/cp-tree.h" 1999)
nil )
(!pair "defaulted_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1998)
+ (!srcfileloc "cp/cp-tree.h" 2000)
nil )
(!pair "has_in_charge_parm_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 2000)
+ (!srcfileloc "cp/cp-tree.h" 2002)
nil )
(!pair "has_vtt_parm_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 2001)
+ (!srcfileloc "cp/cp-tree.h" 2003)
nil )
(!pair "pending_inline_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 2002)
+ (!srcfileloc "cp/cp-tree.h" 2004)
nil )
(!pair "nonconverting"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 2003)
+ (!srcfileloc "cp/cp-tree.h" 2005)
nil )
(!pair "thunk_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 2004)
+ (!srcfileloc "cp/cp-tree.h" 2006)
nil )
(!pair "this_thunk_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 2005)
+ (!srcfileloc "cp/cp-tree.h" 2007)
nil )
(!pair "hidden_friend_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 2006)
+ (!srcfileloc "cp/cp-tree.h" 2008)
nil )
(!pair "omp_declare_reduction_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 2007)
+ (!srcfileloc "cp/cp-tree.h" 2009)
nil )
(!pair "befriending_classes"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 2013)
+ (!srcfileloc "cp/cp-tree.h" 2015)
nil )
(!pair "context"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 2021)
+ (!srcfileloc "cp/cp-tree.h" 2023)
nil )
(!pair "u5"
(!type union 62 nil gc_used "lang_decl_u5"
- (!srcfileloc "cp/cp-tree.h" 2032)
+ (!srcfileloc "cp/cp-tree.h" 2034)
(!fields 2
(!pair "cloned_function"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 2027)
+ (!srcfileloc "cp/cp-tree.h" 2029)
(!options
(!option tag string "0")
)
)
(!pair "fixed_offset"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 2031)
+ (!srcfileloc "cp/cp-tree.h" 2033)
(!options
(!option tag string "1")
)
)
)
nil 260 nil )
- (!srcfileloc "cp/cp-tree.h" 2032)
+ (!srcfileloc "cp/cp-tree.h" 2034)
(!options
(!option desc string "%1.thunk_p")
)
)
(!pair "u"
(!type union 63 nil gc_used "lang_decl_u3"
- (!srcfileloc "cp/cp-tree.h" 2039)
+ (!srcfileloc "cp/cp-tree.h" 2041)
(!fields 2
(!pair "pending_inline_info"
(!type pointer 64 nil gc_used
@@ -1468,23 +1468,23 @@
nil )
(!pair "deferred_access_check"
(!type struct 73 nil gc_used "deferred_access_check"
- (!srcfileloc "cp/cp-tree.h" 5686)
+ (!srcfileloc "cp/cp-tree.h" 5697)
(!fields 4
(!pair "binfo"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 5679)
+ (!srcfileloc "cp/cp-tree.h" 5690)
nil )
(!pair "decl"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 5681)
+ (!srcfileloc "cp/cp-tree.h" 5692)
nil )
(!pair "diag_decl"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 5683)
+ (!srcfileloc "cp/cp-tree.h" 5694)
nil )
(!pair "loc"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 5685)
+ (!srcfileloc "cp/cp-tree.h" 5696)
nil )
)
nil 260 nil nil )
@@ -1539,7 +1539,7 @@
)
nil 260 nil nil )
)
- (!srcfileloc "cp/cp-tree.h" 2036)
+ (!srcfileloc "cp/cp-tree.h" 2038)
(!options
(!option tag string "1")
)
@@ -1597,7 +1597,7 @@
(!type lang_struct 81
(!type already_seen 80)
gc_pointed_to "binding_level"
- (!srcfileloc "fortran/f95-lang.c" 294)
+ (!srcfileloc "fortran/f95-lang.c" 318)
(!fields 0 )
nil 40
(!homotypes 2
@@ -1659,19 +1659,19 @@
nil )
(!type struct 85 nil gc_pointed_to "binding_level"
- (!srcfileloc "fortran/f95-lang.c" 294)
+ (!srcfileloc "fortran/f95-lang.c" 318)
(!fields 3
(!pair "names"
(!type already_seen 5)
- (!srcfileloc "fortran/f95-lang.c" 288)
+ (!srcfileloc "fortran/f95-lang.c" 312)
nil )
(!pair "blocks"
(!type already_seen 5)
- (!srcfileloc "fortran/f95-lang.c" 291)
+ (!srcfileloc "fortran/f95-lang.c" 315)
nil )
(!pair "level_chain"
(!type already_seen 80)
- (!srcfileloc "fortran/f95-lang.c" 293)
+ (!srcfileloc "fortran/f95-lang.c" 317)
nil )
)
nil 8
@@ -1688,7 +1688,7 @@
nil )
(!type struct 86 nil gc_pointed_to "language_function"
- (!srcfileloc "cp/cp-tree.h" 1153)
+ (!srcfileloc "cp/cp-tree.h" 1155)
(!fields 21
(!pair "base"
(!type struct 87 nil gc_used "c_language_function"
@@ -1720,67 +1720,67 @@
nil )
)
nil 390 nil nil )
- (!srcfileloc "cp/cp-tree.h" 1125)
+ (!srcfileloc "cp/cp-tree.h" 1127)
nil )
(!pair "x_cdtor_label"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1127)
+ (!srcfileloc "cp/cp-tree.h" 1129)
nil )
(!pair "x_current_class_ptr"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1128)
+ (!srcfileloc "cp/cp-tree.h" 1130)
nil )
(!pair "x_current_class_ref"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1129)
+ (!srcfileloc "cp/cp-tree.h" 1131)
nil )
(!pair "x_eh_spec_block"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1130)
+ (!srcfileloc "cp/cp-tree.h" 1132)
nil )
(!pair "x_in_charge_parm"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1131)
+ (!srcfileloc "cp/cp-tree.h" 1133)
nil )
(!pair "x_vtt_parm"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1132)
+ (!srcfileloc "cp/cp-tree.h" 1134)
nil )
(!pair "x_return_value"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1133)
+ (!srcfileloc "cp/cp-tree.h" 1135)
nil )
(!pair "x_auto_return_pattern"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1134)
+ (!srcfileloc "cp/cp-tree.h" 1136)
nil )
(!pair "returns_value"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1136)
+ (!srcfileloc "cp/cp-tree.h" 1138)
nil )
(!pair "returns_null"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1137)
+ (!srcfileloc "cp/cp-tree.h" 1139)
nil )
(!pair "returns_abnormally"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1138)
+ (!srcfileloc "cp/cp-tree.h" 1140)
nil )
(!pair "infinite_loop"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1139)
+ (!srcfileloc "cp/cp-tree.h" 1141)
nil )
(!pair "x_in_function_try_handler"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1140)
+ (!srcfileloc "cp/cp-tree.h" 1142)
nil )
(!pair "x_in_base_initializer"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1141)
+ (!srcfileloc "cp/cp-tree.h" 1143)
nil )
(!pair "can_throw"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1144)
+ (!srcfileloc "cp/cp-tree.h" 1146)
nil )
(!pair "x_named_labels"
(!type pointer 90 nil gc_used
@@ -2039,7 +2039,7 @@
)
nil nil nil nil nil nil nil nil nil nil )
)
- (!srcfileloc "cp/cp-tree.h" 1146)
+ (!srcfileloc "cp/cp-tree.h" 1148)
(!options
(!option param_is type
(!type already_seen 93)
@@ -2048,15 +2048,15 @@
)
(!pair "bindings"
(!type already_seen 94)
- (!srcfileloc "cp/cp-tree.h" 1147)
+ (!srcfileloc "cp/cp-tree.h" 1149)
nil )
(!pair "x_local_names"
(!type already_seen 45)
- (!srcfileloc "cp/cp-tree.h" 1148)
+ (!srcfileloc "cp/cp-tree.h" 1150)
nil )
(!pair "infinite_loops"
(!type already_seen 45)
- (!srcfileloc "cp/cp-tree.h" 1151)
+ (!srcfileloc "cp/cp-tree.h" 1153)
nil )
(!pair "extern_decl_map"
(!type pointer 106 nil gc_used
@@ -2069,22 +2069,22 @@
(!type struct 109
(!type already_seen 108)
gc_pointed_to "cxx_int_tree_map"
- (!srcfileloc "cp/cp-tree.h" 1117)
+ (!srcfileloc "cp/cp-tree.h" 1119)
(!fields 2
(!pair "uid"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1115)
+ (!srcfileloc "cp/cp-tree.h" 1117)
nil )
(!pair "to"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1116)
+ (!srcfileloc "cp/cp-tree.h" 1118)
nil )
)
nil 260 nil nil )
)
nil nil nil nil nil nil nil nil nil nil )
)
- (!srcfileloc "cp/cp-tree.h" 1152)
+ (!srcfileloc "cp/cp-tree.h" 1154)
(!options
(!option param_is type
(!type already_seen 109)
@@ -2314,63 +2314,63 @@
)
)
)
- (!srcfileloc "cp/cp-tree.h" 2038)
+ (!srcfileloc "cp/cp-tree.h" 2040)
(!options
(!option tag string "0")
)
)
)
nil 260 nil )
- (!srcfileloc "cp/cp-tree.h" 2039)
+ (!srcfileloc "cp/cp-tree.h" 2041)
(!options
(!option desc string "%1.pending_inline_p")
)
)
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 2066)
+ (!srcfileloc "cp/cp-tree.h" 2068)
(!options
(!option tag string "1")
)
)
(!pair "ns"
(!type struct 128 nil gc_used "lang_decl_ns"
- (!srcfileloc "cp/cp-tree.h" 2048)
+ (!srcfileloc "cp/cp-tree.h" 2050)
(!fields 2
(!pair "base"
(!type already_seen 58)
- (!srcfileloc "cp/cp-tree.h" 2046)
+ (!srcfileloc "cp/cp-tree.h" 2048)
nil )
(!pair "level"
(!type already_seen 94)
- (!srcfileloc "cp/cp-tree.h" 2047)
+ (!srcfileloc "cp/cp-tree.h" 2049)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 2067)
+ (!srcfileloc "cp/cp-tree.h" 2069)
(!options
(!option tag string "2")
)
)
(!pair "parm"
(!type struct 129 nil gc_used "lang_decl_parm"
- (!srcfileloc "cp/cp-tree.h" 2056)
+ (!srcfileloc "cp/cp-tree.h" 2058)
(!fields 3
(!pair "base"
(!type already_seen 58)
- (!srcfileloc "cp/cp-tree.h" 2053)
+ (!srcfileloc "cp/cp-tree.h" 2055)
nil )
(!pair "level"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 2054)
+ (!srcfileloc "cp/cp-tree.h" 2056)
nil )
(!pair "index"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 2055)
+ (!srcfileloc "cp/cp-tree.h" 2057)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 2068)
+ (!srcfileloc "cp/cp-tree.h" 2070)
(!options
(!option tag string "3")
)
@@ -2414,31 +2414,39 @@
nil )
(!type struct 131 nil gc_pointed_to "lang_decl"
- (!srcfileloc "fortran/trans.h" 827)
- (!fields 6
+ (!srcfileloc "fortran/trans.h" 834)
+ (!fields 8
(!pair "saved_descriptor"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 817)
+ (!srcfileloc "fortran/trans.h" 822)
nil )
(!pair "stringlen"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 822)
+ (!srcfileloc "fortran/trans.h" 827)
nil )
(!pair "addr"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 823)
+ (!srcfileloc "fortran/trans.h" 828)
nil )
(!pair "span"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 824)
+ (!srcfileloc "fortran/trans.h" 829)
nil )
(!pair "token"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 826)
+ (!srcfileloc "fortran/trans.h" 831)
nil )
(!pair "caf_offset"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 826)
+ (!srcfileloc "fortran/trans.h" 831)
+ nil )
+ (!pair "scalar_allocatable"
+ (!type already_seen 9)
+ (!srcfileloc "fortran/trans.h" 832)
+ nil )
+ (!pair "scalar_pointer"
+ (!type already_seen 9)
+ (!srcfileloc "fortran/trans.h" 833)
nil )
)
(!options
@@ -2449,11 +2457,11 @@
nil )
(!type struct 132 nil gc_pointed_to "lang_decl"
- (!srcfileloc "cp/cp-tree.h" 2070)
+ (!srcfileloc "cp/cp-tree.h" 2072)
(!fields 1
(!pair "u"
(!type already_seen 56)
- (!srcfileloc "cp/cp-tree.h" 2069)
+ (!srcfileloc "cp/cp-tree.h" 2071)
nil )
)
(!options
@@ -2495,22 +2503,22 @@
)
)
)
- (!srcfileloc "tree-core.h" 1379)
+ (!srcfileloc "tree-core.h" 1384)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1595)
+ (!srcfileloc "tree-core.h" 1600)
(!options
(!option tag string "TS_DECL_COMMON")
)
)
(!pair "decl_with_rtl"
(!type struct 135 nil gc_used "tree_decl_with_rtl"
- (!srcfileloc "tree-core.h" 1385)
+ (!srcfileloc "tree-core.h" 1390)
(!fields 2
(!pair "common"
(!type already_seen 38)
- (!srcfileloc "tree-core.h" 1383)
+ (!srcfileloc "tree-core.h" 1388)
nil )
(!pair "rtl"
(!type pointer 136
@@ -6604,7 +6612,7 @@
)
)
(!pair "x"
- (!type struct 321 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/basic-block.h:187"
+ (!type struct 321 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/basic-block.h:187"
(!srcfileloc "basic-block.h" 190)
(!fields 2
(!pair "head_"
@@ -7399,289 +7407,289 @@
)
511 nil nil )
)
- (!srcfileloc "tree-core.h" 1384)
+ (!srcfileloc "tree-core.h" 1389)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1596)
+ (!srcfileloc "tree-core.h" 1601)
(!options
(!option tag string "TS_DECL_WRTL")
)
)
(!pair "decl_non_common"
(!type struct 339 nil gc_used "tree_decl_non_common"
- (!srcfileloc "tree-core.h" 1468)
+ (!srcfileloc "tree-core.h" 1473)
(!fields 5
(!pair "common"
(!type struct 340 nil gc_used "tree_decl_with_vis"
- (!srcfileloc "tree-core.h" 1452)
+ (!srcfileloc "tree-core.h" 1457)
(!fields 22
(!pair "common"
(!type already_seen 135)
- (!srcfileloc "tree-core.h" 1417)
+ (!srcfileloc "tree-core.h" 1422)
nil )
(!pair "assembler_name"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1418)
+ (!srcfileloc "tree-core.h" 1423)
nil )
(!pair "section_name"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1419)
+ (!srcfileloc "tree-core.h" 1424)
nil )
(!pair "comdat_group"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1420)
+ (!srcfileloc "tree-core.h" 1425)
nil )
(!pair "defer_output"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1423)
+ (!srcfileloc "tree-core.h" 1428)
nil )
(!pair "hard_register"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1424)
+ (!srcfileloc "tree-core.h" 1429)
nil )
(!pair "common_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1425)
+ (!srcfileloc "tree-core.h" 1430)
nil )
(!pair "in_text_section"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1426)
+ (!srcfileloc "tree-core.h" 1431)
nil )
(!pair "in_constant_pool"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1427)
+ (!srcfileloc "tree-core.h" 1432)
nil )
(!pair "dllimport_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1428)
+ (!srcfileloc "tree-core.h" 1433)
nil )
(!pair "weak_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1430)
+ (!srcfileloc "tree-core.h" 1435)
nil )
(!pair "implicit_section_name_p"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1432)
+ (!srcfileloc "tree-core.h" 1437)
nil )
(!pair "seen_in_bind_expr"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1434)
+ (!srcfileloc "tree-core.h" 1439)
nil )
(!pair "comdat_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1435)
+ (!srcfileloc "tree-core.h" 1440)
nil )
(!pair "visibility"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1436)
+ (!srcfileloc "tree-core.h" 1441)
nil )
(!pair "visibility_specified"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1437)
+ (!srcfileloc "tree-core.h" 1442)
nil )
(!pair "tls_model"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1439)
+ (!srcfileloc "tree-core.h" 1444)
nil )
(!pair "init_priority_p"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1442)
+ (!srcfileloc "tree-core.h" 1447)
nil )
(!pair "shadowed_for_var_p"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1444)
+ (!srcfileloc "tree-core.h" 1449)
nil )
(!pair "cxx_constructor"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1446)
+ (!srcfileloc "tree-core.h" 1451)
nil )
(!pair "cxx_destructor"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1448)
+ (!srcfileloc "tree-core.h" 1453)
nil )
(!pair "final"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1450)
+ (!srcfileloc "tree-core.h" 1455)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1459)
+ (!srcfileloc "tree-core.h" 1464)
nil )
(!pair "saved_tree"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1461)
+ (!srcfileloc "tree-core.h" 1466)
nil )
(!pair "arguments"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1463)
+ (!srcfileloc "tree-core.h" 1468)
nil )
(!pair "result"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1465)
+ (!srcfileloc "tree-core.h" 1470)
nil )
(!pair "vindex"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1467)
+ (!srcfileloc "tree-core.h" 1472)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1598)
+ (!srcfileloc "tree-core.h" 1603)
(!options
(!option tag string "TS_DECL_NON_COMMON")
)
)
(!pair "parm_decl"
(!type struct 341 nil gc_used "tree_parm_decl"
- (!srcfileloc "tree-core.h" 1414)
+ (!srcfileloc "tree-core.h" 1419)
(!fields 2
(!pair "common"
(!type already_seen 135)
- (!srcfileloc "tree-core.h" 1412)
+ (!srcfileloc "tree-core.h" 1417)
nil )
(!pair "incoming_rtl"
(!type already_seen 136)
- (!srcfileloc "tree-core.h" 1413)
+ (!srcfileloc "tree-core.h" 1418)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1599)
+ (!srcfileloc "tree-core.h" 1604)
(!options
(!option tag string "TS_PARM_DECL")
)
)
(!pair "decl_with_vis"
(!type already_seen 340)
- (!srcfileloc "tree-core.h" 1600)
+ (!srcfileloc "tree-core.h" 1605)
(!options
(!option tag string "TS_DECL_WITH_VIS")
)
)
(!pair "var_decl"
(!type struct 342 nil gc_used "tree_var_decl"
- (!srcfileloc "tree-core.h" 1456)
+ (!srcfileloc "tree-core.h" 1461)
(!fields 1
(!pair "common"
(!type already_seen 340)
- (!srcfileloc "tree-core.h" 1455)
+ (!srcfileloc "tree-core.h" 1460)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1601)
+ (!srcfileloc "tree-core.h" 1606)
(!options
(!option tag string "TS_VAR_DECL")
)
)
(!pair "field_decl"
(!type struct 343 nil gc_used "tree_field_decl"
- (!srcfileloc "tree-core.h" 1395)
+ (!srcfileloc "tree-core.h" 1400)
(!fields 6
(!pair "common"
(!type already_seen 38)
- (!srcfileloc "tree-core.h" 1388)
+ (!srcfileloc "tree-core.h" 1393)
nil )
(!pair "offset"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1390)
+ (!srcfileloc "tree-core.h" 1395)
nil )
(!pair "bit_field_type"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1391)
+ (!srcfileloc "tree-core.h" 1396)
nil )
(!pair "qualifier"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1392)
+ (!srcfileloc "tree-core.h" 1397)
nil )
(!pair "bit_offset"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1393)
+ (!srcfileloc "tree-core.h" 1398)
nil )
(!pair "fcontext"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1394)
+ (!srcfileloc "tree-core.h" 1399)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1602)
+ (!srcfileloc "tree-core.h" 1607)
(!options
(!option tag string "TS_FIELD_DECL")
)
)
(!pair "label_decl"
(!type struct 344 nil gc_used "tree_label_decl"
- (!srcfileloc "tree-core.h" 1401)
+ (!srcfileloc "tree-core.h" 1406)
(!fields 3
(!pair "common"
(!type already_seen 135)
- (!srcfileloc "tree-core.h" 1398)
+ (!srcfileloc "tree-core.h" 1403)
nil )
(!pair "label_decl_uid"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1399)
+ (!srcfileloc "tree-core.h" 1404)
nil )
(!pair "eh_landing_pad_nr"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1400)
+ (!srcfileloc "tree-core.h" 1405)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1603)
+ (!srcfileloc "tree-core.h" 1608)
(!options
(!option tag string "TS_LABEL_DECL")
)
)
(!pair "result_decl"
(!type struct 345 nil gc_used "tree_result_decl"
- (!srcfileloc "tree-core.h" 1405)
+ (!srcfileloc "tree-core.h" 1410)
(!fields 1
(!pair "common"
(!type already_seen 135)
- (!srcfileloc "tree-core.h" 1404)
+ (!srcfileloc "tree-core.h" 1409)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1604)
+ (!srcfileloc "tree-core.h" 1609)
(!options
(!option tag string "TS_RESULT_DECL")
)
)
(!pair "const_decl"
(!type struct 346 nil gc_used "tree_const_decl"
- (!srcfileloc "tree-core.h" 1409)
+ (!srcfileloc "tree-core.h" 1414)
(!fields 1
(!pair "common"
(!type already_seen 38)
- (!srcfileloc "tree-core.h" 1408)
+ (!srcfileloc "tree-core.h" 1413)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1605)
+ (!srcfileloc "tree-core.h" 1610)
(!options
(!option tag string "TS_CONST_DECL")
)
)
(!pair "type_decl"
(!type struct 347 nil gc_used "tree_type_decl"
- (!srcfileloc "tree-core.h" 1529)
+ (!srcfileloc "tree-core.h" 1534)
(!fields 1
(!pair "common"
(!type already_seen 339)
- (!srcfileloc "tree-core.h" 1527)
+ (!srcfileloc "tree-core.h" 1532)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1606)
+ (!srcfileloc "tree-core.h" 1611)
(!options
(!option tag string "TS_TYPE_DECL")
)
)
(!pair "function_decl"
(!type struct 348 nil gc_used "tree_function_decl"
- (!srcfileloc "tree-core.h" 1516)
+ (!srcfileloc "tree-core.h" 1521)
(!fields 26
(!pair "common"
(!type already_seen 339)
- (!srcfileloc "tree-core.h" 1476)
+ (!srcfileloc "tree-core.h" 1481)
nil )
(!pair "f"
(!type pointer 349 nil gc_used
@@ -8400,41 +8408,41 @@
(!type struct 404
(!type already_seen 403)
gc_maybe_pointed_to "machine_function"
- (!srcfileloc "config/aarch64/aarch64.h" 534)
+ (!srcfileloc "config/aarch64/aarch64.h" 543)
(!fields 2
(!pair "frame"
(!type struct 405 nil gc_used "aarch64_frame"
- (!srcfileloc "config/aarch64/aarch64.h" 524)
+ (!srcfileloc "config/aarch64/aarch64.h" 533)
(!fields 5
(!pair "reg_offset"
(!type array 406 nil gc_used "FIRST_PSEUDO_REGISTER"
(!type already_seen 9)
)
- (!srcfileloc "config/aarch64/aarch64.h" 516)
+ (!srcfileloc "config/aarch64/aarch64.h" 525)
nil )
(!pair "saved_regs_size"
(!type already_seen 9)
- (!srcfileloc "config/aarch64/aarch64.h" 517)
+ (!srcfileloc "config/aarch64/aarch64.h" 526)
nil )
(!pair "padding0"
(!type already_seen 9)
- (!srcfileloc "config/aarch64/aarch64.h" 520)
+ (!srcfileloc "config/aarch64/aarch64.h" 529)
nil )
(!pair "hardfp_offset"
(!type already_seen 9)
- (!srcfileloc "config/aarch64/aarch64.h" 521)
+ (!srcfileloc "config/aarch64/aarch64.h" 530)
nil )
(!pair "laid_out"
(!type already_seen 9)
- (!srcfileloc "config/aarch64/aarch64.h" 523)
+ (!srcfileloc "config/aarch64/aarch64.h" 532)
nil )
)
nil 511 nil nil )
- (!srcfileloc "config/aarch64/aarch64.h" 528)
+ (!srcfileloc "config/aarch64/aarch64.h" 537)
nil )
(!pair "saved_varargs_size"
(!type already_seen 9)
- (!srcfileloc "config/aarch64/aarch64.h" 532)
+ (!srcfileloc "config/aarch64/aarch64.h" 541)
nil )
)
nil 511 nil nil )
@@ -8519,7 +8527,7 @@
(!srcfileloc "dwarf2out.h" 55)
nil )
(!pair "dw_cfi_oprnd1"
- (!type union 413 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/dwarf2out.h:47"
+ (!type union 413 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/dwarf2out.h:47"
(!srcfileloc "dwarf2out.h" 52)
(!fields 4
(!pair "dw_cfi_reg_num"
@@ -9263,299 +9271,299 @@
)
nil 511 nil nil )
)
- (!srcfileloc "tree-core.h" 1478)
+ (!srcfileloc "tree-core.h" 1483)
nil )
(!pair "personality"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1481)
+ (!srcfileloc "tree-core.h" 1486)
nil )
(!pair "function_specific_target"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1484)
+ (!srcfileloc "tree-core.h" 1489)
nil )
(!pair "function_specific_optimization"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1485)
+ (!srcfileloc "tree-core.h" 1490)
nil )
(!pair "function_code"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1491)
+ (!srcfileloc "tree-core.h" 1496)
nil )
(!pair "built_in_class"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1492)
+ (!srcfileloc "tree-core.h" 1497)
nil )
(!pair "static_ctor_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1494)
+ (!srcfileloc "tree-core.h" 1499)
nil )
(!pair "static_dtor_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1495)
+ (!srcfileloc "tree-core.h" 1500)
nil )
(!pair "uninlinable"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1496)
+ (!srcfileloc "tree-core.h" 1501)
nil )
(!pair "possibly_inlined"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1498)
+ (!srcfileloc "tree-core.h" 1503)
nil )
(!pair "novops_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1499)
+ (!srcfileloc "tree-core.h" 1504)
nil )
(!pair "returns_twice_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1500)
+ (!srcfileloc "tree-core.h" 1505)
nil )
(!pair "malloc_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1501)
+ (!srcfileloc "tree-core.h" 1506)
nil )
(!pair "operator_new_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1502)
+ (!srcfileloc "tree-core.h" 1507)
nil )
(!pair "declared_inline_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1503)
+ (!srcfileloc "tree-core.h" 1508)
nil )
(!pair "regdecl_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1504)
+ (!srcfileloc "tree-core.h" 1509)
nil )
(!pair "no_inline_warning_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1505)
+ (!srcfileloc "tree-core.h" 1510)
nil )
(!pair "no_instrument_function_entry_exit"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1507)
+ (!srcfileloc "tree-core.h" 1512)
nil )
(!pair "no_limit_stack"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1508)
+ (!srcfileloc "tree-core.h" 1513)
nil )
(!pair "disregard_inline_limits"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1509)
+ (!srcfileloc "tree-core.h" 1514)
nil )
(!pair "pure_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1510)
+ (!srcfileloc "tree-core.h" 1515)
nil )
(!pair "looping_const_or_pure_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1511)
+ (!srcfileloc "tree-core.h" 1516)
nil )
(!pair "has_debug_args_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1512)
+ (!srcfileloc "tree-core.h" 1517)
nil )
(!pair "tm_clone_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1513)
+ (!srcfileloc "tree-core.h" 1518)
nil )
(!pair "versioned_function"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1514)
+ (!srcfileloc "tree-core.h" 1519)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1607)
+ (!srcfileloc "tree-core.h" 1612)
(!options
(!option tag string "TS_FUNCTION_DECL")
)
)
(!pair "translation_unit_decl"
(!type struct 443 nil gc_used "tree_translation_unit_decl"
- (!srcfileloc "tree-core.h" 1524)
+ (!srcfileloc "tree-core.h" 1529)
(!fields 2
(!pair "common"
(!type already_seen 38)
- (!srcfileloc "tree-core.h" 1519)
+ (!srcfileloc "tree-core.h" 1524)
nil )
(!pair "language"
(!type already_seen 36)
- (!srcfileloc "tree-core.h" 1521)
+ (!srcfileloc "tree-core.h" 1526)
(!options
(!option skip string "")
)
)
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1609)
+ (!srcfileloc "tree-core.h" 1614)
(!options
(!option tag string "TS_TRANSLATION_UNIT_DECL")
)
)
(!pair "type_common"
(!type struct 444 nil gc_used "tree_type_common"
- (!srcfileloc "tree-core.h" 1276)
+ (!srcfileloc "tree-core.h" 1281)
(!fields 30
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "tree-core.h" 1238)
+ (!srcfileloc "tree-core.h" 1243)
nil )
(!pair "size"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1239)
+ (!srcfileloc "tree-core.h" 1244)
nil )
(!pair "size_unit"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1240)
+ (!srcfileloc "tree-core.h" 1245)
nil )
(!pair "attributes"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1241)
+ (!srcfileloc "tree-core.h" 1246)
nil )
(!pair "uid"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1242)
+ (!srcfileloc "tree-core.h" 1247)
nil )
(!pair "precision"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1244)
+ (!srcfileloc "tree-core.h" 1249)
nil )
(!pair "no_force_blk_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1245)
+ (!srcfileloc "tree-core.h" 1250)
nil )
(!pair "needs_constructing_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1246)
+ (!srcfileloc "tree-core.h" 1251)
nil )
(!pair "transparent_aggr_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1247)
+ (!srcfileloc "tree-core.h" 1252)
nil )
(!pair "restrict_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1248)
+ (!srcfileloc "tree-core.h" 1253)
nil )
(!pair "contains_placeholder_bits"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1249)
+ (!srcfileloc "tree-core.h" 1254)
nil )
(!pair "mode"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1251)
+ (!srcfileloc "tree-core.h" 1256)
nil )
(!pair "string_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1253)
+ (!srcfileloc "tree-core.h" 1258)
nil )
(!pair "lang_flag_0"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1254)
+ (!srcfileloc "tree-core.h" 1259)
nil )
(!pair "lang_flag_1"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1255)
+ (!srcfileloc "tree-core.h" 1260)
nil )
(!pair "lang_flag_2"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1256)
+ (!srcfileloc "tree-core.h" 1261)
nil )
(!pair "lang_flag_3"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1257)
+ (!srcfileloc "tree-core.h" 1262)
nil )
(!pair "lang_flag_4"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1258)
+ (!srcfileloc "tree-core.h" 1263)
nil )
(!pair "lang_flag_5"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1259)
+ (!srcfileloc "tree-core.h" 1264)
nil )
(!pair "lang_flag_6"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1260)
+ (!srcfileloc "tree-core.h" 1265)
nil )
(!pair "align"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1262)
+ (!srcfileloc "tree-core.h" 1267)
nil )
(!pair "alias_set"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1263)
+ (!srcfileloc "tree-core.h" 1268)
nil )
(!pair "pointer_to"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1264)
+ (!srcfileloc "tree-core.h" 1269)
nil )
(!pair "reference_to"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1265)
+ (!srcfileloc "tree-core.h" 1270)
nil )
(!pair "symtab"
(!type union 445 nil gc_used "tree_type_symtab"
- (!srcfileloc "tree-core.h" 1270)
+ (!srcfileloc "tree-core.h" 1275)
(!fields 3
(!pair "address"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1267)
+ (!srcfileloc "tree-core.h" 1272)
(!options
(!option tag string "TYPE_SYMTAB_IS_ADDRESS")
)
)
(!pair "pointer"
(!type already_seen 36)
- (!srcfileloc "tree-core.h" 1268)
+ (!srcfileloc "tree-core.h" 1273)
(!options
(!option tag string "TYPE_SYMTAB_IS_POINTER")
)
)
(!pair "die"
(!type already_seen 426)
- (!srcfileloc "tree-core.h" 1269)
+ (!srcfileloc "tree-core.h" 1274)
(!options
(!option tag string "TYPE_SYMTAB_IS_DIE")
)
)
)
nil 511 nil )
- (!srcfileloc "tree-core.h" 1270)
+ (!srcfileloc "tree-core.h" 1275)
(!options
(!option desc string "debug_hooks->tree_type_symtab_field")
)
)
(!pair "canonical"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1271)
+ (!srcfileloc "tree-core.h" 1276)
nil )
(!pair "next_variant"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1272)
+ (!srcfileloc "tree-core.h" 1277)
nil )
(!pair "main_variant"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1273)
+ (!srcfileloc "tree-core.h" 1278)
nil )
(!pair "context"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1274)
+ (!srcfileloc "tree-core.h" 1279)
nil )
(!pair "name"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1275)
+ (!srcfileloc "tree-core.h" 1280)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1610)
+ (!srcfileloc "tree-core.h" 1615)
(!options
(!option tag string "TS_TYPE_COMMON")
)
)
(!pair "type_with_lang_specific"
(!type struct 446 nil gc_used "tree_type_with_lang_specific"
- (!srcfileloc "tree-core.h" 1282)
+ (!srcfileloc "tree-core.h" 1287)
(!fields 2
(!pair "common"
(!type already_seen 444)
- (!srcfileloc "tree-core.h" 1279)
+ (!srcfileloc "tree-core.h" 1284)
nil )
(!pair "lang_specific"
(!type pointer 447 nil gc_used
@@ -10011,75 +10019,75 @@
nil )
(!type struct 473 nil gc_pointed_to "lang_type"
- (!srcfileloc "fortran/trans.h" 813)
+ (!srcfileloc "fortran/trans.h" 818)
(!fields 15
(!pair "rank"
(!type already_seen 9)
- (!srcfileloc "fortran/trans.h" 799)
+ (!srcfileloc "fortran/trans.h" 804)
nil )
(!pair "corank"
(!type already_seen 9)
- (!srcfileloc "fortran/trans.h" 799)
+ (!srcfileloc "fortran/trans.h" 804)
nil )
(!pair "akind"
(!type already_seen 9)
- (!srcfileloc "fortran/trans.h" 800)
+ (!srcfileloc "fortran/trans.h" 805)
nil )
(!pair "lbound"
(!type array 474 nil gc_used "GFC_MAX_DIMENSIONS"
(!type already_seen 5)
)
- (!srcfileloc "fortran/trans.h" 801)
+ (!srcfileloc "fortran/trans.h" 806)
nil )
(!pair "ubound"
(!type array 475 nil gc_used "GFC_MAX_DIMENSIONS"
(!type already_seen 5)
)
- (!srcfileloc "fortran/trans.h" 802)
+ (!srcfileloc "fortran/trans.h" 807)
nil )
(!pair "stride"
(!type array 476 nil gc_used "GFC_MAX_DIMENSIONS"
(!type already_seen 5)
)
- (!srcfileloc "fortran/trans.h" 803)
+ (!srcfileloc "fortran/trans.h" 808)
nil )
(!pair "size"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 804)
+ (!srcfileloc "fortran/trans.h" 809)
nil )
(!pair "offset"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 805)
+ (!srcfileloc "fortran/trans.h" 810)
nil )
(!pair "dtype"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 806)
+ (!srcfileloc "fortran/trans.h" 811)
nil )
(!pair "dataptr_type"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 807)
+ (!srcfileloc "fortran/trans.h" 812)
nil )
(!pair "span"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 808)
+ (!srcfileloc "fortran/trans.h" 813)
nil )
(!pair "base_decl"
(!type array 477 nil gc_used "2"
(!type already_seen 5)
)
- (!srcfileloc "fortran/trans.h" 809)
+ (!srcfileloc "fortran/trans.h" 814)
nil )
(!pair "nonrestricted_type"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 810)
+ (!srcfileloc "fortran/trans.h" 815)
nil )
(!pair "caf_token"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 811)
+ (!srcfileloc "fortran/trans.h" 816)
nil )
(!pair "caf_offset"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 812)
+ (!srcfileloc "fortran/trans.h" 817)
nil )
)
(!options
@@ -10090,257 +10098,257 @@
nil )
(!type struct 478 nil gc_pointed_to "lang_type"
- (!srcfileloc "cp/cp-tree.h" 1491)
+ (!srcfileloc "cp/cp-tree.h" 1493)
(!fields 1
(!pair "u"
(!type union 479 nil gc_used "lang_type_u"
- (!srcfileloc "cp/cp-tree.h" 1490)
+ (!srcfileloc "cp/cp-tree.h" 1492)
(!fields 3
(!pair "h"
(!type struct 480 nil gc_used "lang_type_header"
- (!srcfileloc "cp/cp-tree.h" 1375)
+ (!srcfileloc "cp/cp-tree.h" 1377)
(!fields 8
(!pair "is_lang_type_class"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1365)
+ (!srcfileloc "cp/cp-tree.h" 1367)
nil )
(!pair "has_type_conversion"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1367)
+ (!srcfileloc "cp/cp-tree.h" 1369)
nil )
(!pair "has_copy_ctor"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1368)
+ (!srcfileloc "cp/cp-tree.h" 1370)
nil )
(!pair "has_default_ctor"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1369)
+ (!srcfileloc "cp/cp-tree.h" 1371)
nil )
(!pair "const_needs_init"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1370)
+ (!srcfileloc "cp/cp-tree.h" 1372)
nil )
(!pair "ref_needs_init"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1371)
+ (!srcfileloc "cp/cp-tree.h" 1373)
nil )
(!pair "has_const_copy_assign"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1372)
+ (!srcfileloc "cp/cp-tree.h" 1374)
nil )
(!pair "spare"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1374)
+ (!srcfileloc "cp/cp-tree.h" 1376)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 1487)
+ (!srcfileloc "cp/cp-tree.h" 1489)
(!options
(!option skip string "")
)
)
(!pair "c"
(!type struct 481 nil gc_used "lang_type_class"
- (!srcfileloc "cp/cp-tree.h" 1477)
+ (!srcfileloc "cp/cp-tree.h" 1479)
(!fields 63
(!pair "h"
(!type already_seen 480)
- (!srcfileloc "cp/cp-tree.h" 1391)
+ (!srcfileloc "cp/cp-tree.h" 1393)
nil )
(!pair "align"
(!type already_seen 28)
- (!srcfileloc "cp/cp-tree.h" 1393)
+ (!srcfileloc "cp/cp-tree.h" 1395)
nil )
(!pair "has_mutable"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1395)
+ (!srcfileloc "cp/cp-tree.h" 1397)
nil )
(!pair "com_interface"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1396)
+ (!srcfileloc "cp/cp-tree.h" 1398)
nil )
(!pair "non_pod_class"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1397)
+ (!srcfileloc "cp/cp-tree.h" 1399)
nil )
(!pair "nearly_empty_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1398)
+ (!srcfileloc "cp/cp-tree.h" 1400)
nil )
(!pair "user_align"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1399)
+ (!srcfileloc "cp/cp-tree.h" 1401)
nil )
(!pair "has_copy_assign"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1400)
+ (!srcfileloc "cp/cp-tree.h" 1402)
nil )
(!pair "has_new"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1401)
+ (!srcfileloc "cp/cp-tree.h" 1403)
nil )
(!pair "has_array_new"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1402)
+ (!srcfileloc "cp/cp-tree.h" 1404)
nil )
(!pair "gets_delete"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1404)
+ (!srcfileloc "cp/cp-tree.h" 1406)
nil )
(!pair "interface_only"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1405)
+ (!srcfileloc "cp/cp-tree.h" 1407)
nil )
(!pair "interface_unknown"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1406)
+ (!srcfileloc "cp/cp-tree.h" 1408)
nil )
(!pair "contains_empty_class_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1407)
+ (!srcfileloc "cp/cp-tree.h" 1409)
nil )
(!pair "anon_aggr"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1408)
+ (!srcfileloc "cp/cp-tree.h" 1410)
nil )
(!pair "non_zero_init"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1409)
+ (!srcfileloc "cp/cp-tree.h" 1411)
nil )
(!pair "empty_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1410)
+ (!srcfileloc "cp/cp-tree.h" 1412)
nil )
(!pair "vec_new_uses_cookie"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1412)
+ (!srcfileloc "cp/cp-tree.h" 1414)
nil )
(!pair "declared_class"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1413)
+ (!srcfileloc "cp/cp-tree.h" 1415)
nil )
(!pair "diamond_shaped"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1414)
+ (!srcfileloc "cp/cp-tree.h" 1416)
nil )
(!pair "repeated_base"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1415)
+ (!srcfileloc "cp/cp-tree.h" 1417)
nil )
(!pair "being_defined"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1416)
+ (!srcfileloc "cp/cp-tree.h" 1418)
nil )
(!pair "java_interface"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1417)
+ (!srcfileloc "cp/cp-tree.h" 1419)
nil )
(!pair "debug_requested"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1418)
+ (!srcfileloc "cp/cp-tree.h" 1420)
nil )
(!pair "fields_readonly"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1419)
+ (!srcfileloc "cp/cp-tree.h" 1421)
nil )
(!pair "use_template"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1421)
+ (!srcfileloc "cp/cp-tree.h" 1423)
nil )
(!pair "ptrmemfunc_flag"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1422)
+ (!srcfileloc "cp/cp-tree.h" 1424)
nil )
(!pair "was_anonymous"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1423)
+ (!srcfileloc "cp/cp-tree.h" 1425)
nil )
(!pair "lazy_default_ctor"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1424)
+ (!srcfileloc "cp/cp-tree.h" 1426)
nil )
(!pair "lazy_copy_ctor"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1425)
+ (!srcfileloc "cp/cp-tree.h" 1427)
nil )
(!pair "lazy_copy_assign"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1426)
+ (!srcfileloc "cp/cp-tree.h" 1428)
nil )
(!pair "lazy_destructor"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1427)
+ (!srcfileloc "cp/cp-tree.h" 1429)
nil )
(!pair "has_const_copy_ctor"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1429)
+ (!srcfileloc "cp/cp-tree.h" 1431)
nil )
(!pair "has_complex_copy_ctor"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1430)
+ (!srcfileloc "cp/cp-tree.h" 1432)
nil )
(!pair "has_complex_copy_assign"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1431)
+ (!srcfileloc "cp/cp-tree.h" 1433)
nil )
(!pair "non_aggregate"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1432)
+ (!srcfileloc "cp/cp-tree.h" 1434)
nil )
(!pair "has_complex_dflt"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1433)
+ (!srcfileloc "cp/cp-tree.h" 1435)
nil )
(!pair "has_list_ctor"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1434)
+ (!srcfileloc "cp/cp-tree.h" 1436)
nil )
(!pair "non_std_layout"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1435)
+ (!srcfileloc "cp/cp-tree.h" 1437)
nil )
(!pair "is_literal"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1436)
+ (!srcfileloc "cp/cp-tree.h" 1438)
nil )
(!pair "lazy_move_ctor"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1438)
+ (!srcfileloc "cp/cp-tree.h" 1440)
nil )
(!pair "lazy_move_assign"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1439)
+ (!srcfileloc "cp/cp-tree.h" 1441)
nil )
(!pair "has_complex_move_ctor"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1440)
+ (!srcfileloc "cp/cp-tree.h" 1442)
nil )
(!pair "has_complex_move_assign"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1441)
+ (!srcfileloc "cp/cp-tree.h" 1443)
nil )
(!pair "has_constexpr_ctor"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1442)
+ (!srcfileloc "cp/cp-tree.h" 1444)
nil )
(!pair "dummy"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1451)
+ (!srcfileloc "cp/cp-tree.h" 1453)
nil )
(!pair "primary_base"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1453)
+ (!srcfileloc "cp/cp-tree.h" 1455)
nil )
(!pair "vcall_indices"
(!type pointer 482 nil gc_used
(!type user_struct 483
(!type already_seen 482)
gc_pointed_to "vec<tree_pair_s,va_gc>"
- (!srcfileloc "cp/cp-tree.h" 1454)
+ (!srcfileloc "cp/cp-tree.h" 1456)
(!fields 2
(!pair "va_gc"
(!type already_seen 3)
- (!srcfileloc "cp/cp-tree.h" 1454)
+ (!srcfileloc "cp/cp-tree.h" 1456)
nil )
(!pair "tree_pair_s"
(!type struct 484
@@ -10348,36 +10356,36 @@
(!type already_seen 484)
)
gc_used "tree_pair_s"
- (!srcfileloc "cp/cp-tree.h" 1358)
+ (!srcfileloc "cp/cp-tree.h" 1360)
(!fields 2
(!pair "purpose"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1356)
+ (!srcfileloc "cp/cp-tree.h" 1358)
nil )
(!pair "value"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1357)
+ (!srcfileloc "cp/cp-tree.h" 1359)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 1454)
+ (!srcfileloc "cp/cp-tree.h" 1456)
nil )
)
)
)
- (!srcfileloc "cp/cp-tree.h" 1454)
+ (!srcfileloc "cp/cp-tree.h" 1456)
nil )
(!pair "vtables"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1455)
+ (!srcfileloc "cp/cp-tree.h" 1457)
nil )
(!pair "typeinfo_var"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1456)
+ (!srcfileloc "cp/cp-tree.h" 1458)
nil )
(!pair "vbases"
(!type already_seen 45)
- (!srcfileloc "cp/cp-tree.h" 1457)
+ (!srcfileloc "cp/cp-tree.h" 1459)
nil )
(!pair "nested_udts"
(!type pointer 486 nil gc_used
@@ -10428,46 +10436,46 @@
)
nil 260 nil nil )
)
- (!srcfileloc "cp/cp-tree.h" 1458)
+ (!srcfileloc "cp/cp-tree.h" 1460)
nil )
(!pair "as_base"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1459)
+ (!srcfileloc "cp/cp-tree.h" 1461)
nil )
(!pair "pure_virtuals"
(!type already_seen 45)
- (!srcfileloc "cp/cp-tree.h" 1460)
+ (!srcfileloc "cp/cp-tree.h" 1462)
nil )
(!pair "friend_classes"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1461)
+ (!srcfileloc "cp/cp-tree.h" 1463)
nil )
(!pair "methods"
(!type already_seen 45)
- (!srcfileloc "cp/cp-tree.h" 1462)
+ (!srcfileloc "cp/cp-tree.h" 1464)
(!options
(!option reorder string "resort_type_method_vec")
)
)
(!pair "key_method"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1463)
+ (!srcfileloc "cp/cp-tree.h" 1465)
nil )
(!pair "decl_list"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1464)
+ (!srcfileloc "cp/cp-tree.h" 1466)
nil )
(!pair "template_info"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1465)
+ (!srcfileloc "cp/cp-tree.h" 1467)
nil )
(!pair "befriending_classes"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1466)
+ (!srcfileloc "cp/cp-tree.h" 1468)
nil )
(!pair "objc_info"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1470)
+ (!srcfileloc "cp/cp-tree.h" 1472)
nil )
(!pair "sorted_fields"
(!type pointer 491 nil gc_used
@@ -10495,44 +10503,44 @@
)
390 nil nil )
)
- (!srcfileloc "cp/cp-tree.h" 1474)
+ (!srcfileloc "cp/cp-tree.h" 1476)
(!options
(!option reorder string "resort_sorted_fields")
)
)
(!pair "lambda_expr"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1476)
+ (!srcfileloc "cp/cp-tree.h" 1478)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 1488)
+ (!srcfileloc "cp/cp-tree.h" 1490)
(!options
(!option tag string "1")
)
)
(!pair "ptrmem"
(!type struct 494 nil gc_used "lang_type_ptrmem"
- (!srcfileloc "cp/cp-tree.h" 1482)
+ (!srcfileloc "cp/cp-tree.h" 1484)
(!fields 2
(!pair "h"
(!type already_seen 480)
- (!srcfileloc "cp/cp-tree.h" 1480)
+ (!srcfileloc "cp/cp-tree.h" 1482)
nil )
(!pair "record"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1481)
+ (!srcfileloc "cp/cp-tree.h" 1483)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 1489)
+ (!srcfileloc "cp/cp-tree.h" 1491)
(!options
(!option tag string "0")
)
)
)
nil 260 nil )
- (!srcfileloc "cp/cp-tree.h" 1490)
+ (!srcfileloc "cp/cp-tree.h" 1492)
(!options
(!option desc string "%h.h.is_lang_type_class")
)
@@ -10592,42 +10600,42 @@
)
)
)
- (!srcfileloc "tree-core.h" 1281)
+ (!srcfileloc "tree-core.h" 1286)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1612)
+ (!srcfileloc "tree-core.h" 1617)
(!options
(!option tag string "TS_TYPE_WITH_LANG_SPECIFIC")
)
)
(!pair "type_non_common"
(!type struct 497 nil gc_used "tree_type_non_common"
- (!srcfileloc "tree-core.h" 1290)
+ (!srcfileloc "tree-core.h" 1295)
(!fields 5
(!pair "with_lang_specific"
(!type already_seen 446)
- (!srcfileloc "tree-core.h" 1285)
+ (!srcfileloc "tree-core.h" 1290)
nil )
(!pair "values"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1286)
+ (!srcfileloc "tree-core.h" 1291)
nil )
(!pair "minval"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1287)
+ (!srcfileloc "tree-core.h" 1292)
nil )
(!pair "maxval"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1288)
+ (!srcfileloc "tree-core.h" 1293)
nil )
(!pair "binfo"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1289)
+ (!srcfileloc "tree-core.h" 1294)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1614)
+ (!srcfileloc "tree-core.h" 1619)
(!options
(!option tag string "TS_TYPE_NON_COMMON")
)
@@ -10650,7 +10658,7 @@
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1615)
+ (!srcfileloc "tree-core.h" 1620)
(!options
(!option tag string "TS_LIST")
)
@@ -10674,26 +10682,26 @@
)
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1616)
+ (!srcfileloc "tree-core.h" 1621)
(!options
(!option tag string "TS_VEC")
)
)
(!pair "exp"
(!type struct 501 nil gc_used "tree_exp"
- (!srcfileloc "tree-core.h" 1150)
+ (!srcfileloc "tree-core.h" 1155)
(!fields 3
(!pair "typed"
(!type already_seen 14)
- (!srcfileloc "tree-core.h" 1145)
+ (!srcfileloc "tree-core.h" 1150)
nil )
(!pair "locus"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1146)
+ (!srcfileloc "tree-core.h" 1151)
nil )
(!pair "operands"
(!type union 502 nil gc_used "tree_exp_subunion"
- (!srcfileloc "tree-core.h" 1149)
+ (!srcfileloc "tree-core.h" 1154)
(!fields 1
(!pair ""
(!type array 503 nil gc_used "1"
@@ -10711,7 +10719,7 @@
(!option dot string "")
)
511 nil )
- (!srcfileloc "tree-core.h" 1149)
+ (!srcfileloc "tree-core.h" 1154)
(!options
(!option desc string "TREE_CODE ((tree) &%0)")
(!option special string "tree_exp")
@@ -10719,30 +10727,30 @@
)
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1617)
+ (!srcfileloc "tree-core.h" 1622)
(!options
(!option tag string "TS_EXP")
)
)
(!pair "ssa_name"
(!type struct 504 nil gc_used "tree_ssa_name"
- (!srcfileloc "tree-core.h" 1186)
+ (!srcfileloc "tree-core.h" 1191)
(!fields 5
(!pair "typed"
(!type already_seen 14)
- (!srcfileloc "tree-core.h" 1167)
+ (!srcfileloc "tree-core.h" 1172)
nil )
(!pair "var"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1170)
+ (!srcfileloc "tree-core.h" 1175)
nil )
(!pair "def_stmt"
(!type already_seen 310)
- (!srcfileloc "tree-core.h" 1173)
+ (!srcfileloc "tree-core.h" 1178)
nil )
(!pair "info"
(!type union 505 nil gc_used "ssa_name_info_type"
- (!srcfileloc "tree-core.h" 1181)
+ (!srcfileloc "tree-core.h" 1186)
(!fields 2
(!pair "ptr_info"
(!type pointer 506 nil gc_used
@@ -10766,7 +10774,7 @@
)
nil 511 nil nil )
)
- (!srcfileloc "tree-core.h" 1178)
+ (!srcfileloc "tree-core.h" 1183)
(!options
(!option tag string "0")
)
@@ -10793,14 +10801,14 @@
)
nil 511 nil nil )
)
- (!srcfileloc "tree-core.h" 1180)
+ (!srcfileloc "tree-core.h" 1185)
(!options
(!option tag string "1")
)
)
)
nil 511 nil )
- (!srcfileloc "tree-core.h" 1182)
+ (!srcfileloc "tree-core.h" 1187)
(!options
(!option desc string "%1.typed.type ?!POINTER_TYPE_P (TREE_TYPE ((tree)&%1)) : 2")
)
@@ -10815,191 +10823,191 @@
(!fields 4
(!pair "prev"
(!type already_seen 511)
- (!srcfileloc "tree-core.h" 1155)
+ (!srcfileloc "tree-core.h" 1160)
(!options
(!option skip string "")
)
)
(!pair "next"
(!type already_seen 511)
- (!srcfileloc "tree-core.h" 1156)
+ (!srcfileloc "tree-core.h" 1161)
(!options
(!option skip string "")
)
)
(!pair "loc"
- (!type union 512 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/tree-core.h:1162"
- (!srcfileloc "tree-core.h" 1162)
+ (!type union 512 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/tree-core.h:1167"
+ (!srcfileloc "tree-core.h" 1167)
(!fields 2
(!pair "stmt"
(!type already_seen 310)
- (!srcfileloc "tree-core.h" 1162)
+ (!srcfileloc "tree-core.h" 1167)
nil )
(!pair "ssa_name"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1162)
+ (!srcfileloc "tree-core.h" 1167)
nil )
)
nil 511 nil )
- (!srcfileloc "tree-core.h" 1162)
+ (!srcfileloc "tree-core.h" 1167)
(!options
(!option skip string "")
)
)
(!pair "use"
(!type already_seen 6)
- (!srcfileloc "tree-core.h" 1163)
+ (!srcfileloc "tree-core.h" 1168)
(!options
(!option skip string "")
)
)
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1185)
+ (!srcfileloc "tree-core.h" 1190)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1618)
+ (!srcfileloc "tree-core.h" 1623)
(!options
(!option tag string "TS_SSA_NAME")
)
)
(!pair "block"
(!type struct 513 nil gc_used "tree_block"
- (!srcfileloc "tree-core.h" 1235)
+ (!srcfileloc "tree-core.h" 1240)
(!fields 12
(!pair "base"
(!type already_seen 8)
- (!srcfileloc "tree-core.h" 1219)
+ (!srcfileloc "tree-core.h" 1224)
nil )
(!pair "chain"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1220)
+ (!srcfileloc "tree-core.h" 1225)
nil )
(!pair "abstract_flag"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1222)
+ (!srcfileloc "tree-core.h" 1227)
nil )
(!pair "block_num"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1223)
+ (!srcfileloc "tree-core.h" 1228)
nil )
(!pair "locus"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1225)
+ (!srcfileloc "tree-core.h" 1230)
nil )
(!pair "vars"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1227)
+ (!srcfileloc "tree-core.h" 1232)
nil )
(!pair "nonlocalized_vars"
(!type already_seen 45)
- (!srcfileloc "tree-core.h" 1228)
+ (!srcfileloc "tree-core.h" 1233)
nil )
(!pair "subblocks"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1230)
+ (!srcfileloc "tree-core.h" 1235)
nil )
(!pair "supercontext"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1231)
+ (!srcfileloc "tree-core.h" 1236)
nil )
(!pair "abstract_origin"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1232)
+ (!srcfileloc "tree-core.h" 1237)
nil )
(!pair "fragment_origin"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1233)
+ (!srcfileloc "tree-core.h" 1238)
nil )
(!pair "fragment_chain"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1234)
+ (!srcfileloc "tree-core.h" 1239)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1619)
+ (!srcfileloc "tree-core.h" 1624)
(!options
(!option tag string "TS_BLOCK")
)
)
(!pair "binfo"
(!type struct 514 nil gc_used "tree_binfo"
- (!srcfileloc "tree-core.h" 1306)
+ (!srcfileloc "tree-core.h" 1311)
(!fields 10
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "tree-core.h" 1293)
+ (!srcfileloc "tree-core.h" 1298)
nil )
(!pair "offset"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1295)
+ (!srcfileloc "tree-core.h" 1300)
nil )
(!pair "vtable"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1296)
+ (!srcfileloc "tree-core.h" 1301)
nil )
(!pair "virtuals"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1297)
+ (!srcfileloc "tree-core.h" 1302)
nil )
(!pair "vptr_field"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1298)
+ (!srcfileloc "tree-core.h" 1303)
nil )
(!pair "base_accesses"
(!type already_seen 45)
- (!srcfileloc "tree-core.h" 1299)
+ (!srcfileloc "tree-core.h" 1304)
nil )
(!pair "inheritance"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1300)
+ (!srcfileloc "tree-core.h" 1305)
nil )
(!pair "vtt_subvtt"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1302)
+ (!srcfileloc "tree-core.h" 1307)
nil )
(!pair "vtt_vptr"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1303)
+ (!srcfileloc "tree-core.h" 1308)
nil )
(!pair "base_binfos"
(!type already_seen 46)
- (!srcfileloc "tree-core.h" 1305)
+ (!srcfileloc "tree-core.h" 1310)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1620)
+ (!srcfileloc "tree-core.h" 1625)
(!options
(!option tag string "TS_BINFO")
)
)
(!pair "stmt_list"
(!type struct 515 nil gc_used "tree_statement_list"
- (!srcfileloc "tree-core.h" 1543)
+ (!srcfileloc "tree-core.h" 1548)
(!fields 3
(!pair "typed"
(!type already_seen 14)
- (!srcfileloc "tree-core.h" 1540)
+ (!srcfileloc "tree-core.h" 1545)
nil )
(!pair "head"
(!type pointer 516 nil gc_used
(!type struct 517
(!type already_seen 516)
gc_pointed_to "tree_statement_list_node"
- (!srcfileloc "tree-core.h" 1536)
+ (!srcfileloc "tree-core.h" 1541)
(!fields 3
(!pair "prev"
(!type already_seen 516)
- (!srcfileloc "tree-core.h" 1533)
+ (!srcfileloc "tree-core.h" 1538)
nil )
(!pair "next"
(!type already_seen 516)
- (!srcfileloc "tree-core.h" 1534)
+ (!srcfileloc "tree-core.h" 1539)
nil )
(!pair "stmt"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1535)
+ (!srcfileloc "tree-core.h" 1540)
nil )
)
(!options
@@ -11008,15 +11016,15 @@
)
511 nil nil )
)
- (!srcfileloc "tree-core.h" 1541)
+ (!srcfileloc "tree-core.h" 1546)
nil )
(!pair "tail"
(!type already_seen 516)
- (!srcfileloc "tree-core.h" 1542)
+ (!srcfileloc "tree-core.h" 1547)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1621)
+ (!srcfileloc "tree-core.h" 1626)
(!options
(!option tag string "TS_STATEMENT_LIST")
)
@@ -11035,93 +11043,93 @@
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1622)
+ (!srcfileloc "tree-core.h" 1627)
(!options
(!option tag string "TS_CONSTRUCTOR")
)
)
(!pair "omp_clause"
(!type struct 519 nil gc_used "tree_omp_clause"
- (!srcfileloc "tree-core.h" 1216)
+ (!srcfileloc "tree-core.h" 1221)
(!fields 7
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "tree-core.h" 1197)
+ (!srcfileloc "tree-core.h" 1202)
nil )
(!pair "locus"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1198)
+ (!srcfileloc "tree-core.h" 1203)
nil )
(!pair "code"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1199)
+ (!srcfileloc "tree-core.h" 1204)
nil )
(!pair "subcode"
(!type union 520 nil gc_unused "omp_clause_subcode"
- (!srcfileloc "tree-core.h" 1207)
+ (!srcfileloc "tree-core.h" 1212)
(!fields 6
(!pair "default_kind"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1201)
+ (!srcfileloc "tree-core.h" 1206)
nil )
(!pair "schedule_kind"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1202)
+ (!srcfileloc "tree-core.h" 1207)
nil )
(!pair "depend_kind"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1203)
+ (!srcfileloc "tree-core.h" 1208)
nil )
(!pair "map_kind"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1204)
+ (!srcfileloc "tree-core.h" 1209)
nil )
(!pair "proc_bind_kind"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1205)
+ (!srcfileloc "tree-core.h" 1210)
nil )
(!pair "reduction_code"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1206)
+ (!srcfileloc "tree-core.h" 1211)
nil )
)
nil 511 nil )
- (!srcfileloc "tree-core.h" 1207)
+ (!srcfileloc "tree-core.h" 1212)
(!options
(!option skip string "")
)
)
(!pair "gimple_reduction_init"
(!type already_seen 310)
- (!srcfileloc "tree-core.h" 1211)
+ (!srcfileloc "tree-core.h" 1216)
nil )
(!pair "gimple_reduction_merge"
(!type already_seen 310)
- (!srcfileloc "tree-core.h" 1212)
+ (!srcfileloc "tree-core.h" 1217)
nil )
(!pair "ops"
(!type array 521 nil gc_used "1"
(!type already_seen 5)
)
- (!srcfileloc "tree-core.h" 1215)
+ (!srcfileloc "tree-core.h" 1220)
(!options
(!option length string "omp_clause_num_ops[OMP_CLAUSE_CODE ((tree)&%h)]")
)
)
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1623)
+ (!srcfileloc "tree-core.h" 1628)
(!options
(!option tag string "TS_OMP_CLAUSE")
)
)
(!pair "optimization"
(!type struct 522 nil gc_used "tree_optimization_option"
- (!srcfileloc "tree-core.h" 1561)
+ (!srcfileloc "tree-core.h" 1566)
(!fields 4
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "tree-core.h" 1549)
+ (!srcfileloc "tree-core.h" 1554)
nil )
(!pair "opts"
(!type struct 523 nil gc_used "cl_optimization"
@@ -11881,11 +11889,11 @@
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1552)
+ (!srcfileloc "tree-core.h" 1557)
nil )
(!pair "optabs"
(!type already_seen 10)
- (!srcfileloc "tree-core.h" 1556)
+ (!srcfileloc "tree-core.h" 1561)
(!options
(!option atomic string "")
)
@@ -11898,25 +11906,25 @@
(!fields 0 )
nil 0 nil nil )
)
- (!srcfileloc "tree-core.h" 1560)
+ (!srcfileloc "tree-core.h" 1565)
(!options
(!option skip string "")
)
)
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1624)
+ (!srcfileloc "tree-core.h" 1629)
(!options
(!option tag string "TS_OPTIMIZATION")
)
)
(!pair "target_option"
(!type struct 526 nil gc_used "tree_target_option"
- (!srcfileloc "tree-core.h" 1577)
+ (!srcfileloc "tree-core.h" 1582)
(!fields 3
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "tree-core.h" 1570)
+ (!srcfileloc "tree-core.h" 1575)
nil )
(!pair "globals"
(!type pointer 527 nil gc_used
@@ -12171,7 +12179,7 @@
)
nil 511 nil nil )
)
- (!srcfileloc "tree-core.h" 1573)
+ (!srcfileloc "tree-core.h" 1578)
nil )
(!pair "opts"
(!type struct 555 nil gc_used "cl_target_option"
@@ -12187,11 +12195,11 @@
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1576)
+ (!srcfileloc "tree-core.h" 1581)
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1625)
+ (!srcfileloc "tree-core.h" 1630)
(!options
(!option tag string "TS_TARGET_OPTION")
)
@@ -12308,7 +12316,7 @@
nil )
(!type struct 564 nil gc_used "lang_identifier"
- (!srcfileloc "cp/cp-tree.h" 251)
+ (!srcfileloc "cp/cp-tree.h" 253)
(!fields 5
(!pair "c_common"
(!type struct 565 nil gc_used "c_common_identifier"
@@ -12610,23 +12618,23 @@
nil )
)
nil 390 nil nil )
- (!srcfileloc "cp/cp-tree.h" 246)
+ (!srcfileloc "cp/cp-tree.h" 248)
nil )
(!pair "namespace_bindings"
(!type already_seen 99)
- (!srcfileloc "cp/cp-tree.h" 247)
+ (!srcfileloc "cp/cp-tree.h" 249)
nil )
(!pair "bindings"
(!type already_seen 99)
- (!srcfileloc "cp/cp-tree.h" 248)
+ (!srcfileloc "cp/cp-tree.h" 250)
nil )
(!pair "class_template_info"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 249)
+ (!srcfileloc "cp/cp-tree.h" 251)
nil )
(!pair "label_value"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 250)
+ (!srcfileloc "cp/cp-tree.h" 252)
nil )
)
nil 260
@@ -12648,7 +12656,7 @@
(!srcfileloc "c/c-decl.c" 212)
(!fields 11
(!pair "u"
- (!type union 585 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/c/c-decl.c:198"
+ (!type union 585 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/c/c-decl.c:198"
(!srcfileloc "c/c-decl.c" 201)
(!fields 2
(!pair "type"
@@ -12946,11 +12954,11 @@
)
(!type union 597 nil gc_pointed_to "lang_tree_node"
- (!srcfileloc "cp/cp-tree.h" 829)
+ (!srcfileloc "cp/cp-tree.h" 831)
(!fields 14
(!pair "generic"
(!type already_seen 7)
- (!srcfileloc "cp/cp-tree.h" 809)
+ (!srcfileloc "cp/cp-tree.h" 811)
(!options
(!option desc string "tree_node_structure (&%h)")
(!option tag string "TS_CP_GENERIC")
@@ -12958,31 +12966,31 @@
)
(!pair "tpi"
(!type struct 598 nil gc_used "template_parm_index_s"
- (!srcfileloc "cp/cp-tree.h" 278)
+ (!srcfileloc "cp/cp-tree.h" 280)
(!fields 5
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "cp/cp-tree.h" 273)
+ (!srcfileloc "cp/cp-tree.h" 275)
nil )
(!pair "index"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 274)
+ (!srcfileloc "cp/cp-tree.h" 276)
nil )
(!pair "level"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 275)
+ (!srcfileloc "cp/cp-tree.h" 277)
nil )
(!pair "orig_level"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 276)
+ (!srcfileloc "cp/cp-tree.h" 278)
nil )
(!pair "decl"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 277)
+ (!srcfileloc "cp/cp-tree.h" 279)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 810)
+ (!srcfileloc "cp/cp-tree.h" 812)
(!options
(!option tag string "TS_CP_TPI")
)
@@ -12993,328 +13001,328 @@
(!type already_seen 599)
)
gc_used "ptrmem_cst"
- (!srcfileloc "cp/cp-tree.h" 284)
+ (!srcfileloc "cp/cp-tree.h" 286)
(!fields 2
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "cp/cp-tree.h" 282)
+ (!srcfileloc "cp/cp-tree.h" 284)
nil )
(!pair "member"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 283)
+ (!srcfileloc "cp/cp-tree.h" 285)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 811)
+ (!srcfileloc "cp/cp-tree.h" 813)
(!options
(!option tag string "TS_CP_PTRMEM")
)
)
(!pair "overload"
(!type struct 601 nil gc_used "tree_overload"
- (!srcfileloc "cp/cp-tree.h" 359)
+ (!srcfileloc "cp/cp-tree.h" 361)
(!fields 2
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "cp/cp-tree.h" 357)
+ (!srcfileloc "cp/cp-tree.h" 359)
nil )
(!pair "function"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 358)
+ (!srcfileloc "cp/cp-tree.h" 360)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 812)
+ (!srcfileloc "cp/cp-tree.h" 814)
(!options
(!option tag string "TS_CP_OVERLOAD")
)
)
(!pair "baselink"
(!type struct 602 nil gc_used "tree_baselink"
- (!srcfileloc "cp/cp-tree.h" 393)
+ (!srcfileloc "cp/cp-tree.h" 395)
(!fields 4
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "cp/cp-tree.h" 389)
+ (!srcfileloc "cp/cp-tree.h" 391)
nil )
(!pair "binfo"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 390)
+ (!srcfileloc "cp/cp-tree.h" 392)
nil )
(!pair "functions"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 391)
+ (!srcfileloc "cp/cp-tree.h" 393)
nil )
(!pair "access_binfo"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 392)
+ (!srcfileloc "cp/cp-tree.h" 394)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 813)
+ (!srcfileloc "cp/cp-tree.h" 815)
(!options
(!option tag string "TS_CP_BASELINK")
)
)
(!pair "default_arg"
(!type struct 603 nil gc_used "tree_default_arg"
- (!srcfileloc "cp/cp-tree.h" 571)
+ (!srcfileloc "cp/cp-tree.h" 573)
(!fields 3
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "cp/cp-tree.h" 568)
+ (!srcfileloc "cp/cp-tree.h" 570)
nil )
(!pair "tokens"
(!type already_seen 64)
- (!srcfileloc "cp/cp-tree.h" 569)
+ (!srcfileloc "cp/cp-tree.h" 571)
nil )
(!pair "instantiations"
(!type already_seen 45)
- (!srcfileloc "cp/cp-tree.h" 570)
+ (!srcfileloc "cp/cp-tree.h" 572)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 814)
+ (!srcfileloc "cp/cp-tree.h" 816)
(!options
(!option tag string "TS_CP_DEFAULT_ARG")
)
)
(!pair "deferred_noexcept"
(!type struct 604 nil gc_used "tree_deferred_noexcept"
- (!srcfileloc "cp/cp-tree.h" 587)
+ (!srcfileloc "cp/cp-tree.h" 589)
(!fields 3
(!pair "base"
(!type already_seen 8)
- (!srcfileloc "cp/cp-tree.h" 584)
+ (!srcfileloc "cp/cp-tree.h" 586)
nil )
(!pair "pattern"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 585)
+ (!srcfileloc "cp/cp-tree.h" 587)
nil )
(!pair "args"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 586)
+ (!srcfileloc "cp/cp-tree.h" 588)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 815)
+ (!srcfileloc "cp/cp-tree.h" 817)
(!options
(!option tag string "TS_CP_DEFERRED_NOEXCEPT")
)
)
(!pair "identifier"
(!type already_seen 559)
- (!srcfileloc "cp/cp-tree.h" 816)
+ (!srcfileloc "cp/cp-tree.h" 818)
(!options
(!option tag string "TS_CP_IDENTIFIER")
)
)
(!pair "static_assertion"
(!type struct 605 nil gc_used "tree_static_assert"
- (!srcfileloc "cp/cp-tree.h" 610)
+ (!srcfileloc "cp/cp-tree.h" 612)
(!fields 4
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "cp/cp-tree.h" 606)
+ (!srcfileloc "cp/cp-tree.h" 608)
nil )
(!pair "condition"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 607)
+ (!srcfileloc "cp/cp-tree.h" 609)
nil )
(!pair "message"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 608)
+ (!srcfileloc "cp/cp-tree.h" 610)
nil )
(!pair "location"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 609)
+ (!srcfileloc "cp/cp-tree.h" 611)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 818)
+ (!srcfileloc "cp/cp-tree.h" 820)
(!options
(!option tag string "TS_CP_STATIC_ASSERT")
)
)
(!pair "argument_pack_select"
(!type struct 606 nil gc_used "tree_argument_pack_select"
- (!srcfileloc "cp/cp-tree.h" 616)
+ (!srcfileloc "cp/cp-tree.h" 618)
(!fields 3
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "cp/cp-tree.h" 613)
+ (!srcfileloc "cp/cp-tree.h" 615)
nil )
(!pair "argument_pack"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 614)
+ (!srcfileloc "cp/cp-tree.h" 616)
nil )
(!pair "index"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 615)
+ (!srcfileloc "cp/cp-tree.h" 617)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 820)
+ (!srcfileloc "cp/cp-tree.h" 822)
(!options
(!option tag string "TS_CP_ARGUMENT_PACK_SELECT")
)
)
(!pair "trait_expression"
(!type struct 607 nil gc_used "tree_trait_expr"
- (!srcfileloc "cp/cp-tree.h" 664)
+ (!srcfileloc "cp/cp-tree.h" 666)
(!fields 4
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "cp/cp-tree.h" 660)
+ (!srcfileloc "cp/cp-tree.h" 662)
nil )
(!pair "type1"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 661)
+ (!srcfileloc "cp/cp-tree.h" 663)
nil )
(!pair "type2"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 662)
+ (!srcfileloc "cp/cp-tree.h" 664)
nil )
(!pair "kind"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 663)
+ (!srcfileloc "cp/cp-tree.h" 665)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 822)
+ (!srcfileloc "cp/cp-tree.h" 824)
(!options
(!option tag string "TS_CP_TRAIT_EXPR")
)
)
(!pair "lambda_expression"
(!type struct 608 nil gc_used "tree_lambda_expr"
- (!srcfileloc "cp/cp-tree.h" 749)
+ (!srcfileloc "cp/cp-tree.h" 751)
(!fields 10
(!pair "typed"
(!type already_seen 14)
- (!srcfileloc "cp/cp-tree.h" 739)
+ (!srcfileloc "cp/cp-tree.h" 741)
nil )
(!pair "capture_list"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 740)
+ (!srcfileloc "cp/cp-tree.h" 742)
nil )
(!pair "this_capture"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 741)
+ (!srcfileloc "cp/cp-tree.h" 743)
nil )
(!pair "return_type"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 742)
+ (!srcfileloc "cp/cp-tree.h" 744)
nil )
(!pair "extra_scope"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 743)
+ (!srcfileloc "cp/cp-tree.h" 745)
nil )
(!pair "closure"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 744)
+ (!srcfileloc "cp/cp-tree.h" 746)
nil )
(!pair "pending_proxies"
(!type already_seen 45)
- (!srcfileloc "cp/cp-tree.h" 745)
+ (!srcfileloc "cp/cp-tree.h" 747)
nil )
(!pair "locus"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 746)
+ (!srcfileloc "cp/cp-tree.h" 748)
nil )
(!pair "default_capture_mode"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 747)
+ (!srcfileloc "cp/cp-tree.h" 749)
nil )
(!pair "discriminator"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 748)
+ (!srcfileloc "cp/cp-tree.h" 750)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 824)
+ (!srcfileloc "cp/cp-tree.h" 826)
(!options
(!option tag string "TS_CP_LAMBDA_EXPR")
)
)
(!pair "template_info"
(!type struct 609 nil gc_used "tree_template_info"
- (!srcfileloc "cp/cp-tree.h" 783)
+ (!srcfileloc "cp/cp-tree.h" 785)
(!fields 2
(!pair "common"
(!type already_seen 15)
- (!srcfileloc "cp/cp-tree.h" 781)
+ (!srcfileloc "cp/cp-tree.h" 783)
nil )
(!pair "typedefs_needing_access_checking"
(!type pointer 610 nil gc_used
(!type user_struct 611
(!type already_seen 610)
gc_pointed_to "vec<qualified_typedef_usage_t,va_gc>"
- (!srcfileloc "cp/cp-tree.h" 782)
+ (!srcfileloc "cp/cp-tree.h" 784)
(!fields 2
(!pair "va_gc"
(!type already_seen 3)
- (!srcfileloc "cp/cp-tree.h" 782)
+ (!srcfileloc "cp/cp-tree.h" 784)
nil )
(!pair "qualified_typedef_usage_t"
(!type struct 612 nil gc_used "qualified_typedef_usage_s"
- (!srcfileloc "cp/cp-tree.h" 769)
+ (!srcfileloc "cp/cp-tree.h" 771)
(!fields 3
(!pair "typedef_decl"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 766)
+ (!srcfileloc "cp/cp-tree.h" 768)
nil )
(!pair "context"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 767)
+ (!srcfileloc "cp/cp-tree.h" 769)
nil )
(!pair "locus"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 768)
+ (!srcfileloc "cp/cp-tree.h" 770)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 782)
+ (!srcfileloc "cp/cp-tree.h" 784)
nil )
)
)
)
- (!srcfileloc "cp/cp-tree.h" 782)
+ (!srcfileloc "cp/cp-tree.h" 784)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 826)
+ (!srcfileloc "cp/cp-tree.h" 828)
(!options
(!option tag string "TS_CP_TEMPLATE_INFO")
)
)
(!pair "userdef_literal"
(!type struct 613 nil gc_used "tree_userdef_literal"
- (!srcfileloc "c-family/c-common.h" 1267)
+ (!srcfileloc "c-family/c-common.h" 1268)
(!fields 5
(!pair "base"
(!type already_seen 8)
- (!srcfileloc "c-family/c-common.h" 1262)
+ (!srcfileloc "c-family/c-common.h" 1263)
nil )
(!pair "suffix_id"
(!type already_seen 5)
- (!srcfileloc "c-family/c-common.h" 1263)
+ (!srcfileloc "c-family/c-common.h" 1264)
nil )
(!pair "value"
(!type already_seen 5)
- (!srcfileloc "c-family/c-common.h" 1264)
+ (!srcfileloc "c-family/c-common.h" 1265)
nil )
(!pair "num_string"
(!type already_seen 5)
- (!srcfileloc "c-family/c-common.h" 1265)
+ (!srcfileloc "c-family/c-common.h" 1266)
nil )
(!pair "overflow"
(!type already_seen 9)
- (!srcfileloc "c-family/c-common.h" 1266)
+ (!srcfileloc "c-family/c-common.h" 1267)
nil )
)
nil 390 nil nil )
- (!srcfileloc "cp/cp-tree.h" 828)
+ (!srcfileloc "cp/cp-tree.h" 830)
(!options
(!option tag string "TS_CP_USERDEF_LITERAL")
)
@@ -13667,7 +13675,7 @@
(!fields 0 )
nil 0 nil nil )
- (!type struct 644 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/java/expr.c:2060"nil
+ (!type struct 644 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/java/expr.c:2060"nil
(!fields 0 )
nil 0 nil nil )
@@ -13773,19 +13781,19 @@
nil 0 nil nil )
(!type struct 653 nil gc_used "gfc_powdecl_list"
- (!srcfileloc "fortran/trans.h" 723)
+ (!srcfileloc "fortran/trans.h" 728)
(!fields 3
(!pair "integer"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 720)
+ (!srcfileloc "fortran/trans.h" 725)
nil )
(!pair "real"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 721)
+ (!srcfileloc "fortran/trans.h" 726)
nil )
(!pair "cmplx"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 722)
+ (!srcfileloc "fortran/trans.h" 727)
nil )
)
nil 8 nil nil )
@@ -13794,11 +13802,11 @@
(!fields 0 )
nil 0 nil nil )
- (!type struct 655 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/fortran/trans.h:340"nil
+ (!type struct 655 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/fortran/trans.h:340"nil
(!fields 0 )
nil 0 nil nil )
- (!type struct 656 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/fortran/trans.h:328"nil
+ (!type struct 656 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/fortran/trans.h:328"nil
(!fields 0 )
nil 0 nil nil )
@@ -13822,7 +13830,7 @@
(!fields 0 )
nil 0 nil nil )
- (!type struct 662 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/fortran/trans.h:33"nil
+ (!type struct 662 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/fortran/trans.h:33"nil
(!fields 0 )
nil 0 nil nil )
@@ -13965,19 +13973,19 @@
(!type already_seen 668)
)
gc_pointed_to "module_htab_entry"
- (!srcfileloc "fortran/trans.h" 571)
+ (!srcfileloc "fortran/trans.h" 574)
(!fields 3
(!pair "name"
(!type already_seen 36)
- (!srcfileloc "fortran/trans.h" 568)
+ (!srcfileloc "fortran/trans.h" 571)
nil )
(!pair "namespace_decl"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 569)
+ (!srcfileloc "fortran/trans.h" 572)
nil )
(!pair "decls"
(!type already_seen 53)
- (!srcfileloc "fortran/trans.h" 570)
+ (!srcfileloc "fortran/trans.h" 573)
(!options
(!option param_is type
(!type already_seen 7)
@@ -14095,27 +14103,27 @@
(!type already_seen 684)
)
gc_pointed_to "vec<tree_int,va_gc>"
- (!srcfileloc "cp/parser.c" 8654)
+ (!srcfileloc "cp/parser.c" 8652)
(!fields 2
(!pair "va_gc"
(!type already_seen 3)
- (!srcfileloc "cp/parser.c" 8654)
+ (!srcfileloc "cp/parser.c" 8652)
nil )
(!pair "tree_int"
(!type struct 686 nil gc_used "tree_int"
- (!srcfileloc "cp/parser.c" 8653)
+ (!srcfileloc "cp/parser.c" 8651)
(!fields 2
(!pair "t"
(!type already_seen 5)
- (!srcfileloc "cp/parser.c" 8651)
+ (!srcfileloc "cp/parser.c" 8649)
nil )
(!pair "i"
(!type already_seen 9)
- (!srcfileloc "cp/parser.c" 8652)
+ (!srcfileloc "cp/parser.c" 8650)
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/parser.c" 8654)
+ (!srcfileloc "cp/parser.c" 8652)
nil )
)
)
@@ -14479,39 +14487,39 @@
(!type already_seen 709)
)
gc_pointed_to "constexpr_call"
- (!srcfileloc "cp/semantics.c" 8081)
+ (!srcfileloc "cp/semantics.c" 8083)
(!fields 4
(!pair "fundef"
(!type pointer 711 nil gc_used
(!type struct 712
(!type already_seen 711)
gc_pointed_to "constexpr_fundef"
- (!srcfileloc "cp/semantics.c" 7398)
+ (!srcfileloc "cp/semantics.c" 7400)
(!fields 2
(!pair "decl"
(!type already_seen 5)
- (!srcfileloc "cp/semantics.c" 7396)
+ (!srcfileloc "cp/semantics.c" 7398)
nil )
(!pair "body"
(!type already_seen 5)
- (!srcfileloc "cp/semantics.c" 7397)
+ (!srcfileloc "cp/semantics.c" 7399)
nil )
)
nil 260 nil nil )
)
- (!srcfileloc "cp/semantics.c" 8065)
+ (!srcfileloc "cp/semantics.c" 8067)
nil )
(!pair "bindings"
(!type already_seen 5)
- (!srcfileloc "cp/semantics.c" 8072)
+ (!srcfileloc "cp/semantics.c" 8074)
nil )
(!pair "result"
(!type already_seen 5)
- (!srcfileloc "cp/semantics.c" 8077)
+ (!srcfileloc "cp/semantics.c" 8079)
nil )
(!pair "hash"
(!type already_seen 9)
- (!srcfileloc "cp/semantics.c" 8080)
+ (!srcfileloc "cp/semantics.c" 8082)
nil )
)
nil 260 nil nil )
@@ -14610,27 +14618,27 @@
(!type struct 726
(!type already_seen 725)
gc_pointed_to "tinst_level"
- (!srcfileloc "cp/cp-tree.h" 4985)
+ (!srcfileloc "cp/cp-tree.h" 4995)
(!fields 5
(!pair "next"
(!type already_seen 725)
- (!srcfileloc "cp/cp-tree.h" 4970)
+ (!srcfileloc "cp/cp-tree.h" 4980)
nil )
(!pair "decl"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 4975)
+ (!srcfileloc "cp/cp-tree.h" 4985)
nil )
(!pair "locus"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4978)
+ (!srcfileloc "cp/cp-tree.h" 4988)
nil )
(!pair "errors"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4981)
+ (!srcfileloc "cp/cp-tree.h" 4991)
nil )
(!pair "in_system_header_p"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4984)
+ (!srcfileloc "cp/cp-tree.h" 4994)
nil )
)
(!options
@@ -14743,23 +14751,23 @@
nil 0 nil nil )
(!type struct 745 nil gc_used "operator_name_info_t"
- (!srcfileloc "cp/cp-tree.h" 4729)
+ (!srcfileloc "cp/cp-tree.h" 4744)
(!fields 4
(!pair "identifier"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 4722)
+ (!srcfileloc "cp/cp-tree.h" 4737)
nil )
(!pair "name"
(!type already_seen 36)
- (!srcfileloc "cp/cp-tree.h" 4724)
+ (!srcfileloc "cp/cp-tree.h" 4739)
nil )
(!pair "mangled_name"
(!type already_seen 36)
- (!srcfileloc "cp/cp-tree.h" 4726)
+ (!srcfileloc "cp/cp-tree.h" 4741)
nil )
(!pair "arity"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4728)
+ (!srcfileloc "cp/cp-tree.h" 4743)
nil )
)
nil 260 nil nil )
@@ -14805,11 +14813,11 @@
(!type already_seen 747)
)
gc_pointed_to "vec<cxx_saved_binding,va_gc>"
- (!srcfileloc "cp/cp-tree.h" 1020)
+ (!srcfileloc "cp/cp-tree.h" 1022)
(!fields 2
(!pair "va_gc"
(!type already_seen 3)
- (!srcfileloc "cp/cp-tree.h" 1020)
+ (!srcfileloc "cp/cp-tree.h" 1022)
nil )
(!pair "cxx_saved_binding"
(!type struct 749 nil gc_used "cxx_saved_binding"
@@ -14829,7 +14837,7 @@
nil )
)
nil 260 nil nil )
- (!srcfileloc "cp/cp-tree.h" 1020)
+ (!srcfileloc "cp/cp-tree.h" 1022)
nil )
)
)
@@ -14867,111 +14875,111 @@
(!type already_seen 750)
)
gc_pointed_to "saved_scope"
- (!srcfileloc "cp/cp-tree.h" 1056)
+ (!srcfileloc "cp/cp-tree.h" 1058)
(!fields 26
(!pair "old_bindings"
(!type already_seen 748)
- (!srcfileloc "cp/cp-tree.h" 1020)
+ (!srcfileloc "cp/cp-tree.h" 1022)
nil )
(!pair "old_namespace"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1021)
+ (!srcfileloc "cp/cp-tree.h" 1023)
nil )
(!pair "decl_ns_list"
(!type already_seen 45)
- (!srcfileloc "cp/cp-tree.h" 1022)
+ (!srcfileloc "cp/cp-tree.h" 1024)
nil )
(!pair "class_name"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1023)
+ (!srcfileloc "cp/cp-tree.h" 1025)
nil )
(!pair "class_type"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1024)
+ (!srcfileloc "cp/cp-tree.h" 1026)
nil )
(!pair "access_specifier"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1025)
+ (!srcfileloc "cp/cp-tree.h" 1027)
nil )
(!pair "function_decl"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1026)
+ (!srcfileloc "cp/cp-tree.h" 1028)
nil )
(!pair "lang_base"
(!type already_seen 45)
- (!srcfileloc "cp/cp-tree.h" 1027)
+ (!srcfileloc "cp/cp-tree.h" 1029)
nil )
(!pair "lang_name"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1028)
+ (!srcfileloc "cp/cp-tree.h" 1030)
nil )
(!pair "template_parms"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1029)
+ (!srcfileloc "cp/cp-tree.h" 1031)
nil )
(!pair "x_previous_class_level"
(!type already_seen 94)
- (!srcfileloc "cp/cp-tree.h" 1030)
+ (!srcfileloc "cp/cp-tree.h" 1032)
nil )
(!pair "x_saved_tree"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1031)
+ (!srcfileloc "cp/cp-tree.h" 1033)
nil )
(!pair "x_current_class_ptr"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1034)
+ (!srcfileloc "cp/cp-tree.h" 1036)
nil )
(!pair "x_current_class_ref"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 1035)
+ (!srcfileloc "cp/cp-tree.h" 1037)
nil )
(!pair "x_processing_template_decl"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1037)
+ (!srcfileloc "cp/cp-tree.h" 1039)
nil )
(!pair "x_processing_specialization"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1038)
+ (!srcfileloc "cp/cp-tree.h" 1040)
nil )
(!pair "x_processing_explicit_instantiation"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1039)
+ (!srcfileloc "cp/cp-tree.h" 1041)
nil )
(!pair "need_pop_function_context"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1040)
+ (!srcfileloc "cp/cp-tree.h" 1042)
nil )
(!pair "unevaluated_operand"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1042)
+ (!srcfileloc "cp/cp-tree.h" 1044)
nil )
(!pair "inhibit_evaluation_warnings"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1043)
+ (!srcfileloc "cp/cp-tree.h" 1045)
nil )
(!pair "omp_declare_target_attribute"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 1046)
+ (!srcfileloc "cp/cp-tree.h" 1048)
nil )
(!pair "x_stmt_tree"
(!type already_seen 88)
- (!srcfileloc "cp/cp-tree.h" 1048)
+ (!srcfileloc "cp/cp-tree.h" 1050)
nil )
(!pair "class_bindings"
(!type already_seen 94)
- (!srcfileloc "cp/cp-tree.h" 1050)
+ (!srcfileloc "cp/cp-tree.h" 1052)
nil )
(!pair "bindings"
(!type already_seen 94)
- (!srcfileloc "cp/cp-tree.h" 1051)
+ (!srcfileloc "cp/cp-tree.h" 1053)
nil )
(!pair "x_local_specializations"
(!type already_seen 385)
- (!srcfileloc "cp/cp-tree.h" 1053)
+ (!srcfileloc "cp/cp-tree.h" 1055)
nil )
(!pair "prev"
(!type already_seen 751)
- (!srcfileloc "cp/cp-tree.h" 1055)
+ (!srcfileloc "cp/cp-tree.h" 1057)
nil )
)
nil 260 nil nil )
@@ -15073,7 +15081,7 @@
(!fields 0 )
nil 0 nil nil )
- (!type struct 759 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/c/c-parser.c:6077"nil
+ (!type struct 759 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/c/c-parser.c:6071"nil
(!fields 0 )
nil 0 nil nil )
@@ -15200,7 +15208,7 @@
(!fields 0 )
nil 0 nil nil )
- (!type struct 768 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/c-family/c-pragma.c:1169"nil
+ (!type struct 768 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/c-family/c-pragma.c:1169"nil
(!fields 0 )
nil 0 nil nil )
@@ -15385,15 +15393,15 @@
(!type already_seen 788)
)
gc_pointed_to "vec<tree_gc_vec,va_gc>"
- (!srcfileloc "c-family/c-common.c" 11540)
+ (!srcfileloc "c-family/c-common.c" 11546)
(!fields 2
(!pair "va_gc"
(!type already_seen 3)
- (!srcfileloc "c-family/c-common.c" 11540)
+ (!srcfileloc "c-family/c-common.c" 11546)
nil )
(!pair "tree_gc_vec"
(!type already_seen 45)
- (!srcfileloc "c-family/c-common.c" 11540)
+ (!srcfileloc "c-family/c-common.c" 11546)
nil )
)
)
@@ -15409,15 +15417,15 @@
(!type already_seen 791)
)
gc_pointed_to "vec<const_char_p,va_gc>"
- (!srcfileloc "c-family/c-common.c" 9005)
+ (!srcfileloc "c-family/c-common.c" 9011)
(!fields 2
(!pair "va_gc"
(!type already_seen 3)
- (!srcfileloc "c-family/c-common.c" 9005)
+ (!srcfileloc "c-family/c-common.c" 9011)
nil )
(!pair "const_char_p"
(!type already_seen 36)
- (!srcfileloc "c-family/c-common.c" 9005)
+ (!srcfileloc "c-family/c-common.c" 9011)
nil )
)
)
@@ -15732,7 +15740,7 @@
(!fields 0 )
nil 0 nil nil )
- (!type struct 828 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/ada/gcc-interface/decl.c:178"nil
+ (!type struct 828 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/ada/gcc-interface/decl.c:178"nil
(!fields 0 )
nil 0 nil nil )
@@ -15919,11 +15927,11 @@
(!fields 2
(!pair "type"
(!type struct 858 nil gc_used "tree_map_base"
- (!srcfileloc "tree-core.h" 1715)
+ (!srcfileloc "tree-core.h" 1720)
(!fields 1
(!pair "from"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1714)
+ (!srcfileloc "tree-core.h" 1719)
nil )
)
nil 511 nil nil )
@@ -17279,14 +17287,14 @@
nil 0 nil nil )
(!type user_struct 970 nil gc_unused "vec<ipa_parm_adjustment>"
- (!srcfileloc "ipa-prop.h" 685)
+ (!srcfileloc "ipa-prop.h" 686)
(!fields 1
(!pair "ipa_parm_adjustment"
(!type struct 971 nil gc_unused "ipa_parm_adjustment"
- (!srcfileloc "ipa-prop.h" 685)
+ (!srcfileloc "ipa-prop.h" 686)
(!fields 0 )
nil 511 nil nil )
- (!srcfileloc "ipa-prop.h" 685)
+ (!srcfileloc "ipa-prop.h" 686)
nil )
)
)
@@ -17653,7 +17661,7 @@
(!fields 0 )
nil 0 nil nil )
- (!type struct 1014 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/config/aarch64/aarch64.c:6436"nil
+ (!type struct 1014 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/config/aarch64/aarch64.c:7357"nil
(!fields 0 )
nil 0 nil nil )
@@ -17983,7 +17991,7 @@
(!fields 0 )
nil 0 nil nil )
- (!type union 1079 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/tree-eh.c:59"nil
+ (!type union 1079 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/tree-eh.c:59"nil
(!fields 0 )
nil 0 nil )
@@ -18474,19 +18482,19 @@
(!pair "args"
(!type array 1112 nil gc_used "1"
(!type struct 1113 nil gc_used "phi_arg_d"
- (!srcfileloc "tree-core.h" 1194)
+ (!srcfileloc "tree-core.h" 1199)
(!fields 3
(!pair "imm_use"
(!type already_seen 510)
- (!srcfileloc "tree-core.h" 1191)
+ (!srcfileloc "tree-core.h" 1196)
nil )
(!pair "def"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1192)
+ (!srcfileloc "tree-core.h" 1197)
nil )
(!pair "locus"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1193)
+ (!srcfileloc "tree-core.h" 1198)
nil )
)
nil 511 nil nil )
@@ -18613,7 +18621,7 @@
(!srcfileloc "gimple.h" 277)
nil )
(!pair "u"
- (!type union 1120 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/gimple.h:280"
+ (!type union 1120 nil gc_used "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/gimple.h:280"
(!srcfileloc "gimple.h" 283)
(!fields 2
(!pair "fntype"
@@ -18718,11 +18726,11 @@
(!fields 0 )
nil 0 nil nil )
- (!type struct 1128 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/varasm.c:4818"nil
+ (!type struct 1128 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/varasm.c:4818"nil
(!fields 0 )
nil 0 nil nil )
- (!type struct 1129 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/varasm.c:4574"nil
+ (!type struct 1129 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/varasm.c:4574"nil
(!fields 0 )
nil 0 nil nil )
@@ -18828,7 +18836,7 @@
(!fields 0 )
nil 0 nil nil )
- (!type union 1148 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/mcf.c:344"nil
+ (!type union 1148 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/mcf.c:344"nil
(!fields 0 )
nil 0 nil )
@@ -19744,7 +19752,7 @@
(!fields 0 )
nil 0 nil nil )
- (!type struct 1268 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/dwarf2cfi.c:218"nil
+ (!type struct 1268 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/dwarf2cfi.c:218"nil
(!fields 0 )
nil 0 nil nil )
@@ -19756,7 +19764,7 @@
(!type pointer 1271 nil gc_unused
(!type already_seen 1270)
)
- gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/dwarf2cfi.c:101"nil
+ gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/dwarf2cfi.c:101"nil
(!fields 0 )
nil 0 nil nil )
@@ -21059,15 +21067,15 @@
(!type already_seen 1384)
)
gc_pointed_to "vec<alias_pair,va_gc>"
- (!srcfileloc "tree-core.h" 1793)
+ (!srcfileloc "tree-core.h" 1798)
(!fields 2
(!pair "va_gc"
(!type already_seen 3)
- (!srcfileloc "tree-core.h" 1793)
+ (!srcfileloc "tree-core.h" 1798)
nil )
(!pair "alias_pair"
(!type struct 1386 nil gc_used "alias_pair"
- (!srcfileloc "tree-core.h" 1793)
+ (!srcfileloc "tree-core.h" 1798)
(!fields 2
(!pair "decl"
(!type already_seen 5)
@@ -21079,25 +21087,25 @@
nil )
)
nil 511 nil nil )
- (!srcfileloc "tree-core.h" 1793)
+ (!srcfileloc "tree-core.h" 1798)
nil )
)
)
(!type struct 1387 nil gc_used "builtin_info_type"
- (!srcfileloc "tree-core.h" 1828)
+ (!srcfileloc "tree-core.h" 1833)
(!fields 2
(!pair "decl"
(!type array 1388 nil gc_used "(int)END_BUILTINS"
(!type already_seen 5)
)
- (!srcfileloc "tree-core.h" 1771)
+ (!srcfileloc "tree-core.h" 1776)
nil )
(!pair "implicit_p"
(!type array 1389 nil gc_used "(int)END_BUILTINS"
(!type already_seen 9)
)
- (!srcfileloc "tree-core.h" 1772)
+ (!srcfileloc "tree-core.h" 1777)
nil )
)
nil 511 nil nil )
@@ -21115,15 +21123,15 @@
(!type already_seen 1392)
)
gc_pointed_to "tree_vec_map"
- (!srcfileloc "tree-core.h" 1748)
+ (!srcfileloc "tree-core.h" 1753)
(!fields 2
(!pair "base"
(!type already_seen 858)
- (!srcfileloc "tree-core.h" 1746)
+ (!srcfileloc "tree-core.h" 1751)
nil )
(!pair "to"
(!type already_seen 45)
- (!srcfileloc "tree-core.h" 1747)
+ (!srcfileloc "tree-core.h" 1752)
nil )
)
nil 511 nil nil )
@@ -21133,19 +21141,19 @@
(!type already_seen 1394)
)
gc_pointed_to "tree_priority_map"
- (!srcfileloc "tree-core.h" 1742)
+ (!srcfileloc "tree-core.h" 1747)
(!fields 3
(!pair "base"
(!type already_seen 858)
- (!srcfileloc "tree-core.h" 1739)
+ (!srcfileloc "tree-core.h" 1744)
nil )
(!pair "init"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1740)
+ (!srcfileloc "tree-core.h" 1745)
nil )
(!pair "fini"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1741)
+ (!srcfileloc "tree-core.h" 1746)
nil )
)
nil 511 nil nil )
@@ -21155,15 +21163,15 @@
(!type already_seen 1396)
)
gc_pointed_to "tree_int_map"
- (!srcfileloc "tree-core.h" 1735)
+ (!srcfileloc "tree-core.h" 1740)
(!fields 2
(!pair "base"
(!type already_seen 858)
- (!srcfileloc "tree-core.h" 1733)
+ (!srcfileloc "tree-core.h" 1738)
nil )
(!pair "to"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1734)
+ (!srcfileloc "tree-core.h" 1739)
nil )
)
nil 511 nil nil )
@@ -21173,15 +21181,15 @@
(!type already_seen 1398)
)
gc_pointed_to "tree_decl_map"
- (!srcfileloc "tree-core.h" 1729)
+ (!srcfileloc "tree-core.h" 1734)
(!fields 2
(!pair "base"
(!type already_seen 858)
- (!srcfileloc "tree-core.h" 1727)
+ (!srcfileloc "tree-core.h" 1732)
nil )
(!pair "to"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1728)
+ (!srcfileloc "tree-core.h" 1733)
nil )
)
nil 511 nil nil )
@@ -21191,19 +21199,19 @@
(!type already_seen 1400)
)
gc_pointed_to "tree_map"
- (!srcfileloc "tree-core.h" 1723)
+ (!srcfileloc "tree-core.h" 1728)
(!fields 3
(!pair "base"
(!type already_seen 858)
- (!srcfileloc "tree-core.h" 1720)
+ (!srcfileloc "tree-core.h" 1725)
nil )
(!pair "hash"
(!type already_seen 9)
- (!srcfileloc "tree-core.h" 1721)
+ (!srcfileloc "tree-core.h" 1726)
nil )
(!pair "to"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1722)
+ (!srcfileloc "tree-core.h" 1727)
nil )
)
nil 511 nil nil )
@@ -21809,7 +21817,7 @@
(!type already_seen 1318)
- (!type struct 1426 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/config/aarch64/aarch64.h:576"nil
+ (!type struct 1426 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/config/aarch64/aarch64.h:585"nil
(!fields 0 )
nil 0 nil nil )
@@ -21897,11 +21905,11 @@
(!fields 0 )
nil 0 nil nil )
- (!type struct 1442 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/../libcpp/include/cpplib.h:999"nil
+ (!type struct 1442 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/../libcpp/include/cpplib.h:999"nil
(!fields 0 )
nil 0 nil nil )
- (!type struct 1443 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/../libcpp/include/cpplib.h:988"nil
+ (!type struct 1443 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/../libcpp/include/cpplib.h:988"nil
(!fields 0 )
nil 0 nil nil )
@@ -21955,7 +21963,7 @@
(!fields 0 )
nil 0 nil nil )
- (!type struct 1452 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/../libcpp/include/line-map.h:627"nil
+ (!type struct 1452 nil gc_unused "anonymous:/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/../libcpp/include/line-map.h:627"nil
(!fields 0 )
nil 0 nil nil )
@@ -22317,15 +22325,15 @@
nil )
(!pair "gfc_interface_mapping"
(!type already_seen 651)
- (!srcfileloc "fortran/trans.h" 949)
+ (!srcfileloc "fortran/trans.h" 963)
nil )
(!pair "gfc_interface_sym_mapping"
(!type already_seen 652)
- (!srcfileloc "fortran/trans.h" 934)
+ (!srcfileloc "fortran/trans.h" 948)
nil )
(!pair "gfc_powdecl_list"
(!type already_seen 653)
- (!srcfileloc "fortran/trans.h" 724)
+ (!srcfileloc "fortran/trans.h" 729)
nil )
(!pair "gfc_wrapped_block"
(!type already_seen 655)
@@ -22425,11 +22433,11 @@
nil )
(!pair "vec<tree_int,va_gc>"
(!type already_seen 684)
- (!srcfileloc "cp/parser.c" 8654)
+ (!srcfileloc "cp/parser.c" 8652)
nil )
(!pair "tree_int"
(!type already_seen 686)
- (!srcfileloc "cp/parser.c" 8653)
+ (!srcfileloc "cp/parser.c" 8651)
nil )
(!pair "cp_parser_expression_stack"
(!type array 1464 nil gc_unused "NUM_PREC_VALUES"
@@ -22531,15 +22539,15 @@
nil )
(!pair "cplus_array_info"
(!type already_seen 708)
- (!srcfileloc "cp/tree.c" 730)
+ (!srcfileloc "cp/tree.c" 741)
nil )
(!pair "constexpr_call"
(!type already_seen 709)
- (!srcfileloc "cp/semantics.c" 8081)
+ (!srcfileloc "cp/semantics.c" 8083)
nil )
(!pair "constexpr_fundef"
(!type already_seen 712)
- (!srcfileloc "cp/semantics.c" 7398)
+ (!srcfileloc "cp/semantics.c" 7400)
nil )
(!pair "vec<deferred_access,va_gc>"
(!type already_seen 714)
@@ -22607,159 +22615,159 @@
nil )
(!pair "deferred_access_check"
(!type already_seen 73)
- (!srcfileloc "cp/cp-tree.h" 5686)
+ (!srcfileloc "cp/cp-tree.h" 5697)
nil )
(!pair "walk_namespaces_fn"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 5267)
+ (!srcfileloc "cp/cp-tree.h" 5277)
nil )
(!pair "cp_parameter_declarator"
(!type already_seen 742)
- (!srcfileloc "cp/cp-tree.h" 4880)
+ (!srcfileloc "cp/cp-tree.h" 4890)
nil )
(!pair "cp_declarator"
(!type already_seen 743)
- (!srcfileloc "cp/cp-tree.h" 4878)
+ (!srcfileloc "cp/cp-tree.h" 4888)
nil )
(!pair "cp_declarator_kind"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4874)
+ (!srcfileloc "cp/cp-tree.h" 4884)
nil )
(!pair "cp_decl_specifier_seq"
(!type already_seen 744)
- (!srcfileloc "cp/cp-tree.h" 4862)
+ (!srcfileloc "cp/cp-tree.h" 4872)
nil )
(!pair "cp_decl_spec"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4818)
+ (!srcfileloc "cp/cp-tree.h" 4828)
nil )
(!pair "cp_storage_class"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4787)
+ (!srcfileloc "cp/cp-tree.h" 4797)
nil )
(!pair "cp_virt_specifiers"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4757)
- nil )
- (!pair "cp_cv_quals"
- (!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4741)
+ (!srcfileloc "cp/cp-tree.h" 4767)
nil )
(!pair "operator_name_info_t"
(!type already_seen 745)
- (!srcfileloc "cp/cp-tree.h" 4729)
+ (!srcfileloc "cp/cp-tree.h" 4744)
nil )
(!pair "unification_kind_t"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4370)
+ (!srcfileloc "cp/cp-tree.h" 4385)
+ nil )
+ (!pair "cp_cv_quals"
+ (!type already_seen 9)
+ (!srcfileloc "cp/cp-tree.h" 4362)
nil )
(!pair "base_kind"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4331)
+ (!srcfileloc "cp/cp-tree.h" 4340)
nil )
(!pair "deferring_kind"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4318)
+ (!srcfileloc "cp/cp-tree.h" 4327)
nil )
(!pair "base_access"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4311)
+ (!srcfileloc "cp/cp-tree.h" 4320)
nil )
(!pair "tsubst_flags_t"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4297)
+ (!srcfileloc "cp/cp-tree.h" 4306)
nil )
(!pair "duration_kind"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4268)
+ (!srcfileloc "cp/cp-tree.h" 4277)
nil )
(!pair "linkage_kind"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4261)
+ (!srcfileloc "cp/cp-tree.h" 4270)
nil )
(!pair "special_function_kind"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4238)
+ (!srcfileloc "cp/cp-tree.h" 4247)
nil )
(!pair "access_kind"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4217)
+ (!srcfileloc "cp/cp-tree.h" 4226)
nil )
(!pair "tmpl_spec_kind"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4206)
+ (!srcfileloc "cp/cp-tree.h" 4215)
nil )
(!pair "cp_lvalue_kind"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 4188)
+ (!srcfileloc "cp/cp-tree.h" 4197)
nil )
(!pair "aggr_init_expr_arg_iterator"
(!type already_seen 746)
- (!srcfileloc "cp/cp-tree.h" 3078)
+ (!srcfileloc "cp/cp-tree.h" 3084)
nil )
(!pair "vec<tree_pair_s,va_gc>"
(!type already_seen 483)
- (!srcfileloc "cp/cp-tree.h" 1454)
+ (!srcfileloc "cp/cp-tree.h" 1456)
nil )
(!pair "tree_pair_p"
(!type already_seen 485)
- (!srcfileloc "cp/cp-tree.h" 1359)
+ (!srcfileloc "cp/cp-tree.h" 1361)
nil )
(!pair "tree_pair_s"
(!type already_seen 484)
- (!srcfileloc "cp/cp-tree.h" 1358)
+ (!srcfileloc "cp/cp-tree.h" 1360)
nil )
(!pair "vec<cxx_saved_binding,va_gc>"
(!type already_seen 747)
- (!srcfileloc "cp/cp-tree.h" 1020)
+ (!srcfileloc "cp/cp-tree.h" 1022)
nil )
(!pair "vec<qualified_typedef_usage_t,va_gc>"
(!type already_seen 611)
- (!srcfileloc "cp/cp-tree.h" 782)
+ (!srcfileloc "cp/cp-tree.h" 784)
nil )
(!pair "qualified_typedef_usage_t"
(!type already_seen 612)
- (!srcfileloc "cp/cp-tree.h" 770)
+ (!srcfileloc "cp/cp-tree.h" 772)
nil )
(!pair "cp_trait_kind"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 646)
+ (!srcfileloc "cp/cp-tree.h" 648)
nil )
(!pair "abstract_class_use"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 499)
+ (!srcfileloc "cp/cp-tree.h" 501)
nil )
(!pair "impl_conv_void"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 486)
+ (!srcfileloc "cp/cp-tree.h" 488)
nil )
(!pair "impl_conv_rhs"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 475)
+ (!srcfileloc "cp/cp-tree.h" 477)
nil )
(!pair "expr_list_kind"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 465)
+ (!srcfileloc "cp/cp-tree.h" 467)
nil )
(!pair "composite_pointer_operation"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 458)
+ (!srcfileloc "cp/cp-tree.h" 460)
nil )
(!pair "cpp0x_warn_str"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 446)
+ (!srcfileloc "cp/cp-tree.h" 448)
nil )
(!pair "cp_id_kind"
(!type already_seen 9)
- (!srcfileloc "cp/cp-tree.h" 409)
+ (!srcfileloc "cp/cp-tree.h" 411)
nil )
(!pair "ptrmem_cst_t"
(!type already_seen 600)
- (!srcfileloc "cp/cp-tree.h" 285)
+ (!srcfileloc "cp/cp-tree.h" 287)
nil )
(!pair "template_parm_index"
(!type already_seen 598)
- (!srcfileloc "cp/cp-tree.h" 279)
+ (!srcfileloc "cp/cp-tree.h" 281)
nil )
(!pair "vec<cp_label_binding,va_gc>"
(!type already_seen 102)
@@ -22903,15 +22911,15 @@
nil )
(!pair "an_loop_parts"
(!type already_seen 782)
- (!srcfileloc "c-family/c-common.h" 1351)
+ (!srcfileloc "c-family/c-common.h" 1352)
nil )
(!pair "an_parts"
(!type already_seen 783)
- (!srcfileloc "c-family/c-common.h" 1340)
+ (!srcfileloc "c-family/c-common.h" 1341)
nil )
(!pair "omp_clause_mask"
(!type already_seen 9)
- (!srcfileloc "c-family/c-common.h" 1050)
+ (!srcfileloc "c-family/c-common.h" 1051)
nil )
(!pair "stmt_tree"
(!type already_seen 89)
@@ -22927,23 +22935,23 @@
nil )
(!pair "vec<tree_gc_vec,va_gc>"
(!type already_seen 788)
- (!srcfileloc "c-family/c-common.c" 11540)
+ (!srcfileloc "c-family/c-common.c" 11546)
nil )
(!pair "tree_gc_vec"
(!type already_seen 45)
- (!srcfileloc "c-family/c-common.c" 11539)
+ (!srcfileloc "c-family/c-common.c" 11545)
nil )
(!pair "vec<const_char_p,va_gc>"
(!type already_seen 791)
- (!srcfileloc "c-family/c-common.c" 9005)
+ (!srcfileloc "c-family/c-common.c" 9011)
nil )
(!pair "const_char_p"
(!type already_seen 29)
- (!srcfileloc "c-family/c-common.c" 9004)
+ (!srcfileloc "c-family/c-common.c" 9010)
nil )
(!pair "disabled_builtin"
(!type already_seen 793)
- (!srcfileloc "c-family/c-common.c" 5755)
+ (!srcfileloc "c-family/c-common.c" 5761)
nil )
(!pair "c_binding_ptr"
(!type already_seen 583)
@@ -23175,15 +23183,15 @@
nil )
(!pair "ipa_parm_adjustment_vec"
(!type already_seen 970)
- (!srcfileloc "ipa-prop.h" 685)
+ (!srcfileloc "ipa-prop.h" 686)
nil )
(!pair "ipa_parm_adjustment"
(!type already_seen 971)
- (!srcfileloc "ipa-prop.h" 685)
+ (!srcfileloc "ipa-prop.h" 686)
nil )
(!pair "vec<ipa_parm_adjustment>"
(!type already_seen 970)
- (!srcfileloc "ipa-prop.h" 685)
+ (!srcfileloc "ipa-prop.h" 686)
nil )
(!pair "ipa_edge_args"
(!type already_seen 974)
@@ -23231,11 +23239,11 @@
nil )
(!pair "aarch64_simd_mangle_map_entry"
(!type already_seen 1014)
- (!srcfileloc "config/aarch64/aarch64.c" 6440)
+ (!srcfileloc "config/aarch64/aarch64.c" 7361)
nil )
(!pair "aarch64_cc"
(!type already_seen 9)
- (!srcfileloc "config/aarch64/aarch64.c" 324)
+ (!srcfileloc "config/aarch64/aarch64.c" 376)
nil )
(!pair "omp_context"
(!type already_seen 1029)
@@ -23937,23 +23945,23 @@
nil )
(!pair "builtin_info_type"
(!type already_seen 1387)
- (!srcfileloc "tree-core.h" 1828)
+ (!srcfileloc "tree-core.h" 1833)
nil )
(!pair "alias_pair"
(!type already_seen 1386)
- (!srcfileloc "tree-core.h" 1793)
+ (!srcfileloc "tree-core.h" 1798)
nil )
(!pair "vec<alias_pair,va_gc>"
(!type already_seen 1384)
- (!srcfileloc "tree-core.h" 1793)
+ (!srcfileloc "tree-core.h" 1798)
nil )
(!pair "record_layout_info"
(!type already_seen 1404)
- (!srcfileloc "tree-core.h" 1705)
+ (!srcfileloc "tree-core.h" 1710)
nil )
(!pair "vec<tree,va_gc>"
(!type already_seen 46)
- (!srcfileloc "tree-core.h" 1228)
+ (!srcfileloc "tree-core.h" 1233)
nil )
(!pair "constructor_elt"
(!type already_seen 467)
@@ -24117,7 +24125,7 @@
nil )
(!pair "machine_function"
(!type already_seen 404)
- (!srcfileloc "config/aarch64/aarch64.h" 534)
+ (!srcfileloc "config/aarch64/aarch64.h" 543)
nil )
(!pair "gt_pointer_operator"
(!type already_seen 9)
@@ -25188,151 +25196,151 @@
nil )
(!pair "gfor_fndecl_sr_kind"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 773)
+ (!srcfileloc "fortran/trans.h" 778)
nil )
(!pair "gfor_fndecl_si_kind"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 772)
+ (!srcfileloc "fortran/trans.h" 777)
nil )
(!pair "gfor_fndecl_sc_kind"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 771)
+ (!srcfileloc "fortran/trans.h" 776)
nil )
(!pair "gfor_fndecl_iargc"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 768)
+ (!srcfileloc "fortran/trans.h" 773)
nil )
(!pair "gfor_fndecl_size1"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 767)
+ (!srcfileloc "fortran/trans.h" 772)
nil )
(!pair "gfor_fndecl_size0"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 766)
+ (!srcfileloc "fortran/trans.h" 771)
nil )
(!pair "gfor_fndecl_convert_char4_to_char1"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 763)
+ (!srcfileloc "fortran/trans.h" 768)
nil )
(!pair "gfor_fndecl_convert_char1_to_char4"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 762)
+ (!srcfileloc "fortran/trans.h" 767)
nil )
(!pair "gfor_fndecl_select_string_char4"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 759)
+ (!srcfileloc "fortran/trans.h" 764)
nil )
(!pair "gfor_fndecl_adjustr_char4"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 758)
+ (!srcfileloc "fortran/trans.h" 763)
nil )
(!pair "gfor_fndecl_adjustl_char4"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 757)
+ (!srcfileloc "fortran/trans.h" 762)
nil )
(!pair "gfor_fndecl_string_minmax_char4"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 756)
+ (!srcfileloc "fortran/trans.h" 761)
nil )
(!pair "gfor_fndecl_string_trim_char4"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 755)
+ (!srcfileloc "fortran/trans.h" 760)
nil )
(!pair "gfor_fndecl_string_verify_char4"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 754)
+ (!srcfileloc "fortran/trans.h" 759)
nil )
(!pair "gfor_fndecl_string_scan_char4"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 753)
+ (!srcfileloc "fortran/trans.h" 758)
nil )
(!pair "gfor_fndecl_string_index_char4"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 752)
+ (!srcfileloc "fortran/trans.h" 757)
nil )
(!pair "gfor_fndecl_string_len_trim_char4"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 751)
+ (!srcfileloc "fortran/trans.h" 756)
nil )
(!pair "gfor_fndecl_concat_string_char4"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 750)
+ (!srcfileloc "fortran/trans.h" 755)
nil )
(!pair "gfor_fndecl_compare_string_char4"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 749)
+ (!srcfileloc "fortran/trans.h" 754)
nil )
(!pair "gfor_fndecl_select_string"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 748)
+ (!srcfileloc "fortran/trans.h" 753)
nil )
(!pair "gfor_fndecl_adjustr"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 747)
+ (!srcfileloc "fortran/trans.h" 752)
nil )
(!pair "gfor_fndecl_adjustl"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 746)
+ (!srcfileloc "fortran/trans.h" 751)
nil )
(!pair "gfor_fndecl_string_minmax"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 745)
+ (!srcfileloc "fortran/trans.h" 750)
nil )
(!pair "gfor_fndecl_string_trim"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 744)
+ (!srcfileloc "fortran/trans.h" 749)
nil )
(!pair "gfor_fndecl_string_verify"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 743)
+ (!srcfileloc "fortran/trans.h" 748)
nil )
(!pair "gfor_fndecl_string_scan"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 742)
+ (!srcfileloc "fortran/trans.h" 747)
nil )
(!pair "gfor_fndecl_string_index"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 741)
+ (!srcfileloc "fortran/trans.h" 746)
nil )
(!pair "gfor_fndecl_string_len_trim"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 740)
+ (!srcfileloc "fortran/trans.h" 745)
nil )
(!pair "gfor_fndecl_concat_string"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 739)
+ (!srcfileloc "fortran/trans.h" 744)
nil )
(!pair "gfor_fndecl_compare_string"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 738)
+ (!srcfileloc "fortran/trans.h" 743)
nil )
(!pair "gfor_fndecl_zgemm"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 735)
+ (!srcfileloc "fortran/trans.h" 740)
nil )
(!pair "gfor_fndecl_cgemm"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 734)
+ (!srcfileloc "fortran/trans.h" 739)
nil )
(!pair "gfor_fndecl_dgemm"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 733)
+ (!srcfileloc "fortran/trans.h" 738)
nil )
(!pair "gfor_fndecl_sgemm"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 732)
+ (!srcfileloc "fortran/trans.h" 737)
nil )
(!pair "gfor_fndecl_math_ishftc16"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 729)
+ (!srcfileloc "fortran/trans.h" 734)
nil )
(!pair "gfor_fndecl_math_ishftc8"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 728)
+ (!srcfileloc "fortran/trans.h" 733)
nil )
(!pair "gfor_fndecl_math_ishftc4"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 727)
+ (!srcfileloc "fortran/trans.h" 732)
nil )
(!pair "gfor_fndecl_math_powi"
(!type array 1547 nil gc_used "4"
@@ -25340,139 +25348,139 @@
(!type already_seen 653)
)
)
- (!srcfileloc "fortran/trans.h" 726)
+ (!srcfileloc "fortran/trans.h" 731)
nil )
(!pair "gfort_gvar_caf_this_image"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 713)
+ (!srcfileloc "fortran/trans.h" 718)
nil )
(!pair "gfort_gvar_caf_num_images"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 712)
+ (!srcfileloc "fortran/trans.h" 717)
nil )
(!pair "gfor_fndecl_caf_error_stop_str"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 709)
+ (!srcfileloc "fortran/trans.h" 714)
nil )
(!pair "gfor_fndecl_caf_error_stop"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 708)
+ (!srcfileloc "fortran/trans.h" 713)
nil )
(!pair "gfor_fndecl_caf_sync_images"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 707)
+ (!srcfileloc "fortran/trans.h" 712)
nil )
(!pair "gfor_fndecl_caf_sync_all"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 706)
+ (!srcfileloc "fortran/trans.h" 711)
nil )
(!pair "gfor_fndecl_caf_end_critical"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 705)
+ (!srcfileloc "fortran/trans.h" 710)
nil )
(!pair "gfor_fndecl_caf_critical"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 704)
+ (!srcfileloc "fortran/trans.h" 709)
nil )
(!pair "gfor_fndecl_caf_deregister"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 703)
+ (!srcfileloc "fortran/trans.h" 708)
nil )
(!pair "gfor_fndecl_caf_register"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 702)
+ (!srcfileloc "fortran/trans.h" 707)
nil )
(!pair "gfor_fndecl_caf_finalize"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 701)
+ (!srcfileloc "fortran/trans.h" 706)
nil )
(!pair "gfor_fndecl_caf_init"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 700)
+ (!srcfileloc "fortran/trans.h" 705)
nil )
(!pair "gfor_fndecl_associated"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 696)
+ (!srcfileloc "fortran/trans.h" 701)
nil )
(!pair "gfor_fndecl_in_unpack"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 695)
+ (!srcfileloc "fortran/trans.h" 700)
nil )
(!pair "gfor_fndecl_in_pack"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 694)
+ (!srcfileloc "fortran/trans.h" 699)
nil )
(!pair "gfor_fndecl_fdate"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 693)
+ (!srcfileloc "fortran/trans.h" 698)
nil )
(!pair "gfor_fndecl_ctime"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 692)
+ (!srcfileloc "fortran/trans.h" 697)
nil )
(!pair "gfor_fndecl_ttynam"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 691)
+ (!srcfileloc "fortran/trans.h" 696)
nil )
(!pair "gfor_fndecl_set_options"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 690)
+ (!srcfileloc "fortran/trans.h" 695)
nil )
(!pair "gfor_fndecl_set_fpe"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 689)
+ (!srcfileloc "fortran/trans.h" 694)
nil )
(!pair "gfor_fndecl_generate_error"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 688)
+ (!srcfileloc "fortran/trans.h" 693)
nil )
(!pair "gfor_fndecl_os_error"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 687)
+ (!srcfileloc "fortran/trans.h" 692)
nil )
(!pair "gfor_fndecl_runtime_warning_at"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 686)
+ (!srcfileloc "fortran/trans.h" 691)
nil )
(!pair "gfor_fndecl_runtime_error_at"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 685)
+ (!srcfileloc "fortran/trans.h" 690)
nil )
(!pair "gfor_fndecl_runtime_error"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 684)
+ (!srcfileloc "fortran/trans.h" 689)
nil )
(!pair "gfor_fndecl_error_stop_string"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 683)
+ (!srcfileloc "fortran/trans.h" 688)
nil )
(!pair "gfor_fndecl_error_stop_numeric"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 682)
+ (!srcfileloc "fortran/trans.h" 687)
nil )
(!pair "gfor_fndecl_stop_string"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 681)
+ (!srcfileloc "fortran/trans.h" 686)
nil )
(!pair "gfor_fndecl_stop_numeric_f08"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 680)
+ (!srcfileloc "fortran/trans.h" 685)
nil )
(!pair "gfor_fndecl_stop_numeric"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 679)
+ (!srcfileloc "fortran/trans.h" 684)
nil )
(!pair "gfor_fndecl_pause_string"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 678)
+ (!srcfileloc "fortran/trans.h" 683)
nil )
(!pair "gfor_fndecl_pause_numeric"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 677)
+ (!srcfileloc "fortran/trans.h" 682)
nil )
(!pair "gfc_static_ctors"
(!type already_seen 5)
- (!srcfileloc "fortran/trans.h" 582)
+ (!srcfileloc "fortran/trans.h" 585)
nil )
(!pair "gfc_charlen_type_node"
(!type already_seen 5)
@@ -25574,7 +25582,7 @@
(!type array 1557 nil gc_used "2"
(!type already_seen 5)
)
- (!srcfileloc "fortran/trans-stmt.c" 2127)
+ (!srcfileloc "fortran/trans-stmt.c" 2130)
nil )
(!pair "dt_parm"
(!type already_seen 5)
@@ -25606,7 +25614,7 @@
nil )
(!pair "module_htab"
(!type already_seen 1472)
- (!srcfileloc "fortran/trans-decl.c" 4069)
+ (!srcfileloc "fortran/trans-decl.c" 4112)
(!options
(!option param_is type
(!type already_seen 668)
@@ -25639,19 +25647,19 @@
nil )
(!pair "global_binding_level"
(!type already_seen 80)
- (!srcfileloc "fortran/f95-lang.c" 301)
+ (!srcfileloc "fortran/f95-lang.c" 325)
nil )
(!pair "current_binding_level"
(!type already_seen 80)
- (!srcfileloc "fortran/f95-lang.c" 297)
+ (!srcfileloc "fortran/f95-lang.c" 321)
nil )
(!pair "current_translation_unit"
(!type already_seen 5)
- (!srcfileloc "fortran/f95-lang.c" 162)
+ (!srcfileloc "fortran/f95-lang.c" 186)
nil )
(!pair "free_binding_level"
(!type already_seen 80)
- (!srcfileloc "fortran/f95-lang.c" 156)
+ (!srcfileloc "fortran/f95-lang.c" 180)
nil )
(!pair "vlt_register_set_fndecl"
(!type already_seen 5)
@@ -25706,23 +25714,23 @@
nil )
(!pair "generic_parm_count"
(!type already_seen 9)
- (!srcfileloc "cp/parser.c" 31972)
+ (!srcfileloc "cp/parser.c" 31989)
nil )
(!pair "the_parser"
(!type already_seen 690)
- (!srcfileloc "cp/parser.c" 31395)
+ (!srcfileloc "cp/parser.c" 31412)
nil )
(!pair "lambda_scope_stack"
(!type already_seen 685)
- (!srcfileloc "cp/parser.c" 8654)
+ (!srcfileloc "cp/parser.c" 8652)
nil )
(!pair "lambda_count"
(!type already_seen 9)
- (!srcfileloc "cp/parser.c" 8648)
+ (!srcfileloc "cp/parser.c" 8646)
nil )
(!pair "lambda_scope"
(!type already_seen 5)
- (!srcfileloc "cp/parser.c" 8647)
+ (!srcfileloc "cp/parser.c" 8645)
nil )
(!pair "cp_parser_context_free_list"
(!type already_seen 696)
@@ -25733,7 +25741,7 @@
)
(!pair "list_hash_table"
(!type already_seen 53)
- (!srcfileloc "cp/tree.c" 1598)
+ (!srcfileloc "cp/tree.c" 1609)
(!options
(!option param_is type
(!type already_seen 7)
@@ -25742,7 +25750,7 @@
)
(!pair "cplus_array_htab"
(!type already_seen 53)
- (!srcfileloc "cp/tree.c" 746)
+ (!srcfileloc "cp/tree.c" 757)
(!options
(!option param_is type
(!type already_seen 7)
@@ -25751,7 +25759,7 @@
)
(!pair "constexpr_call_table"
(!type already_seen 1476)
- (!srcfileloc "cp/semantics.c" 8086)
+ (!srcfileloc "cp/semantics.c" 8088)
(!options
(!option param_is type
(!type already_seen 709)
@@ -25760,7 +25768,7 @@
)
(!pair "constexpr_fundef_table"
(!type already_seen 1478)
- (!srcfileloc "cp/semantics.c" 7403)
+ (!srcfileloc "cp/semantics.c" 7405)
(!options
(!option param_is type
(!type already_seen 712)
@@ -25781,7 +25789,7 @@
nil )
(!pair "last_error_tinst_level"
(!type already_seen 725)
- (!srcfileloc "cp/pt.c" 8230)
+ (!srcfileloc "cp/pt.c" 8235)
nil )
(!pair "canonical_template_parms"
(!type already_seen 45)
@@ -25855,7 +25863,7 @@
nil )
(!pair "start_cleanup_cnt"
(!type already_seen 9)
- (!srcfileloc "cp/decl.c" 6739)
+ (!srcfileloc "cp/decl.c" 6746)
nil )
(!pair "typename_htab"
(!type already_seen 53)
@@ -25872,57 +25880,57 @@
nil )
(!pair "java_iface_lookup_fn"
(!type already_seen 5)
- (!srcfileloc "cp/call.c" 7318)
+ (!srcfileloc "cp/call.c" 7341)
nil )
(!pair "default_arg_context"
(!type already_seen 45)
- (!srcfileloc "cp/call.c" 6491)
+ (!srcfileloc "cp/call.c" 6505)
nil )
(!pair "unemitted_tinfo_decls"
(!type already_seen 45)
- (!srcfileloc "cp/cp-tree.h" 5620)
+ (!srcfileloc "cp/cp-tree.h" 5631)
nil )
(!pair "deferred_mark_used_calls"
(!type already_seen 45)
- (!srcfileloc "cp/cp-tree.h" 5291)
+ (!srcfileloc "cp/cp-tree.h" 5301)
nil )
(!pair "assignment_operator_name_info"
(!type array 1562 nil gc_used "(int) MAX_TREE_CODES"
(!type already_seen 745)
)
- (!srcfileloc "cp/cp-tree.h" 4736)
+ (!srcfileloc "cp/cp-tree.h" 4751)
nil )
(!pair "operator_name_info"
(!type array 1563 nil gc_used "(int) MAX_TREE_CODES"
(!type already_seen 745)
)
- (!srcfileloc "cp/cp-tree.h" 4733)
+ (!srcfileloc "cp/cp-tree.h" 4748)
nil )
(!pair "tls_aggregates"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 4478)
+ (!srcfileloc "cp/cp-tree.h" 4493)
nil )
(!pair "static_aggregates"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 4476)
+ (!srcfileloc "cp/cp-tree.h" 4491)
nil )
(!pair "local_classes"
(!type already_seen 45)
- (!srcfileloc "cp/cp-tree.h" 4378)
+ (!srcfileloc "cp/cp-tree.h" 4393)
nil )
(!pair "integer_two_node"
(!type already_seen 5)
- (!srcfileloc "cp/cp-tree.h" 4339)
+ (!srcfileloc "cp/cp-tree.h" 4348)
nil )
(!pair "scope_chain"
(!type already_seen 751)
- (!srcfileloc "cp/cp-tree.h" 1112)
+ (!srcfileloc "cp/cp-tree.h" 1114)
nil )
(!pair "cp_global_trees"
(!type array 1564 nil gc_used "CPTI_MAX"
(!type already_seen 5)
)
- (!srcfileloc "cp/cp-tree.h" 902)
+ (!srcfileloc "cp/cp-tree.h" 904)
nil )
(!pair "lambda_cnt"
(!type already_seen 9)
@@ -25971,11 +25979,11 @@
nil )
(!pair "temp_count"
(!type already_seen 9)
- (!srcfileloc "cp/mangle.c" 3858)
+ (!srcfileloc "cp/mangle.c" 3862)
nil )
(!pair "conv_type_names"
(!type already_seen 53)
- (!srcfileloc "cp/mangle.c" 3730)
+ (!srcfileloc "cp/mangle.c" 3734)
(!options
(!option param_is type
(!type already_seen 7)
@@ -26036,7 +26044,7 @@
nil )
(!pair "pending_lang_change"
(!type already_seen 9)
- (!srcfileloc "c-family/c-common.h" 943)
+ (!srcfileloc "c-family/c-common.h" 944)
nil )
(!pair "c_global_trees"
(!type array 1567 nil gc_used "CTI_MAX"
@@ -26053,24 +26061,24 @@
)
(!pair "tree_vector_cache"
(!type already_seen 789)
- (!srcfileloc "c-family/c-common.c" 11540)
+ (!srcfileloc "c-family/c-common.c" 11546)
(!options
(!option deletable string "")
)
)
(!pair "optimize_args"
(!type already_seen 792)
- (!srcfileloc "c-family/c-common.c" 9005)
+ (!srcfileloc "c-family/c-common.c" 9011)
nil )
(!pair "compound_literal_number"
(!type already_seen 9)
- (!srcfileloc "c-family/c-common.c" 5726)
+ (!srcfileloc "c-family/c-common.c" 5732)
nil )
(!pair "built_in_attributes"
(!type array 1568 nil gc_used "(int) ATTR_LAST"
(!type already_seen 5)
)
- (!srcfileloc "c-family/c-common.c" 5101)
+ (!srcfileloc "c-family/c-common.c" 5107)
nil )
(!pair "type_hash_table"
(!type already_seen 53)
@@ -26387,7 +26395,7 @@
nil )
(!pair "clone_fn_id_num"
(!type already_seen 9)
- (!srcfileloc "cgraphclones.c" 479)
+ (!srcfileloc "cgraphclones.c" 491)
nil )
(!pair "vtable_entry_type"
(!type already_seen 5)
@@ -26406,15 +26414,15 @@
)
(!pair "va_list_type"
(!type already_seen 5)
- (!srcfileloc "config/aarch64/aarch64.c" 5553)
+ (!srcfileloc "config/aarch64/aarch64.c" 6473)
nil )
(!pair "tls_get_addr_libfunc"
(!type already_seen 136)
- (!srcfileloc "config/aarch64/aarch64.c" 492)
+ (!srcfileloc "config/aarch64/aarch64.c" 562)
nil )
(!pair "gty_dummy"
(!type already_seen 9)
- (!srcfileloc "config/aarch64/aarch64.c" 310)
+ (!srcfileloc "config/aarch64/aarch64.c" 362)
nil )
(!pair "stack_chk_fail_decl"
(!type already_seen 5)
@@ -26426,7 +26434,7 @@
nil )
(!pair "critical_name_mutexes"
(!type already_seen 1488)
- (!srcfileloc "omp-low.c" 8832)
+ (!srcfileloc "omp-low.c" 8883)
(!options
(!option param2_is type
(!type already_seen 5)
@@ -26607,7 +26615,7 @@
nil )
(!pair "gcc_eh_personality_decl"
(!type already_seen 5)
- (!srcfileloc "tree.c" 11746)
+ (!srcfileloc "tree.c" 11751)
nil )
(!pair "nonstandard_integer_type_cache"
(!type array 1579 nil gc_used "2 * MAX_INT_CACHED_PREC + 2"
@@ -26795,7 +26803,7 @@
nil )
(!pair "next_block_index"
(!type already_seen 9)
- (!srcfileloc "function.c" 4327)
+ (!srcfileloc "function.c" 4332)
nil )
(!pair "temp_slot_address_table"
(!type already_seen 1508)
@@ -27551,41 +27559,41 @@
nil )
(!pair "current_function_func_begin_label"
(!type already_seen 36)
- (!srcfileloc "tree-core.h" 1841)
+ (!srcfileloc "tree-core.h" 1846)
nil )
(!pair "current_function_decl"
(!type already_seen 5)
- (!srcfileloc "tree-core.h" 1838)
+ (!srcfileloc "tree-core.h" 1843)
nil )
(!pair "builtin_info"
(!type already_seen 1387)
- (!srcfileloc "tree-core.h" 1828)
+ (!srcfileloc "tree-core.h" 1833)
nil )
(!pair "sizetype_tab"
(!type array 1582 nil gc_used "(int) stk_type_kind_last"
(!type already_seen 5)
)
- (!srcfileloc "tree-core.h" 1816)
+ (!srcfileloc "tree-core.h" 1821)
nil )
(!pair "integer_types"
(!type array 1583 nil gc_used "itk_none"
(!type already_seen 5)
)
- (!srcfileloc "tree-core.h" 1813)
+ (!srcfileloc "tree-core.h" 1818)
nil )
(!pair "global_trees"
(!type array 1584 nil gc_used "TI_MAX"
(!type already_seen 5)
)
- (!srcfileloc "tree-core.h" 1809)
+ (!srcfileloc "tree-core.h" 1814)
nil )
(!pair "all_translation_units"
(!type already_seen 45)
- (!srcfileloc "tree-core.h" 1806)
+ (!srcfileloc "tree-core.h" 1811)
nil )
(!pair "alias_pairs"
(!type already_seen 1385)
- (!srcfileloc "tree-core.h" 1793)
+ (!srcfileloc "tree-core.h" 1798)
nil )
(!pair "stack_limit_rtx"
(!type already_seen 136)
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/auto-host.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/auto-host.h
index b14aea0..c2d78a8 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/auto-host.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/auto-host.h
@@ -1739,7 +1739,7 @@
/* Define to 1 if HOST_WIDE_INT must be 64 bits wide (see hwint.h). */
#ifndef USED_FOR_TARGET
-#define NEED_64BIT_HOST_WIDE_INT 1
+/* #undef NEED_64BIT_HOST_WIDE_INT */
#endif
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/b-header-vars b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/b-header-vars
index 458b1d5..6183b6f 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/b-header-vars
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/b-header-vars
@@ -1,4 +1,4 @@
-USER_H=arm_neon.h
+USER_H=arm_acle.h
HASHTAB_H=hashtab.h
OBSTACK_H=obstack.h
SPLAY_TREE_H=splay-tree.h
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-family/c-common.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-family/c-common.h
index 24959d8..06dcb7c 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-family/c-common.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-family/c-common.h
@@ -758,6 +758,7 @@ extern tree c_wrap_maybe_const (tree, bool);
extern tree c_save_expr (tree);
extern tree c_common_truthvalue_conversion (location_t, tree);
extern void c_apply_type_quals_to_decl (int, tree);
+extern unsigned int min_align_of_type (tree);
extern tree c_sizeof_or_alignof_type (location_t, tree, bool, bool, int);
extern tree c_alignof_expr (location_t, tree);
/* Print an error message for invalid operands to arith operation CODE.
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-tree.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-tree.h
index 85df885..53768d6 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-tree.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/c-tree.h
@@ -612,7 +612,8 @@ extern void push_init_level (int, struct obstack *);
extern struct c_expr pop_init_level (int, struct obstack *);
extern void set_init_index (tree, tree, struct obstack *);
extern void set_init_label (tree, struct obstack *);
-extern void process_init_element (struct c_expr, bool, struct obstack *);
+extern void process_init_element (location_t, struct c_expr, bool,
+ struct obstack *);
extern tree build_compound_literal (location_t, tree, tree, bool);
extern void check_compound_literal_type (location_t, struct c_type_name *);
extern tree c_start_case (location_t, location_t, tree);
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-linux.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-linux.h
index dae41b0..f40a03e 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-linux.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-linux.h
@@ -21,7 +21,7 @@
#ifndef GCC_AARCH64_LINUX_H
#define GCC_AARCH64_LINUX_H
-#define GLIBC_DYNAMIC_LINKER "/lib/ld-linux-aarch64%{mbig-endian:_be}.so.1"
+#define GLIBC_DYNAMIC_LINKER "/lib/ld-linux-aarch64%{mbig-endian:_be}%{mabi=ilp32:_ilp32}.so.1"
#undef BIONIC_DYNAMIC_LINKER
#define BIONIC_DYNAMIC_LINKER "/system/bin/linker64"
@@ -35,7 +35,7 @@
-dynamic-linker " GNU_USER_DYNAMIC_LINKER " \
-X \
%{mbig-endian:-EB} %{mlittle-endian:-EL} \
- -maarch64linux%{mbig-endian:b}"
+ -maarch64linux%{mabi=ilp32:32}%{mbig-endian:b}"
#define TARGET_OS_CPP_BUILTINS() \
do \
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-protos.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-protos.h
index 04cbc78..c4f75b3 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-protos.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64-protos.h
@@ -108,9 +108,22 @@ enum aarch64_symbol_type
cost models and vectors for address cost calculations, register
move costs and memory move costs. */
+/* Scaled addressing modes can vary cost depending on the mode of the
+ value to be loaded/stored. QImode values cannot use scaled
+ addressing modes. */
+
+struct scale_addr_mode_cost
+{
+ const int hi;
+ const int si;
+ const int di;
+ const int ti;
+};
+
/* Additional cost for addresses. */
struct cpu_addrcost_table
{
+ const struct scale_addr_mode_cost addr_scale_costs;
const int pre_modify;
const int post_modify;
const int register_offset;
@@ -167,6 +180,7 @@ bool aarch64_cannot_change_mode_class (enum machine_mode,
enum aarch64_symbol_type
aarch64_classify_symbolic_expression (rtx, enum aarch64_symbol_context);
bool aarch64_constant_address_p (rtx);
+bool aarch64_expand_movmem (rtx *);
bool aarch64_float_const_zero_rtx_p (rtx);
bool aarch64_function_arg_regno_p (unsigned);
bool aarch64_gen_movmemqi (rtx *);
@@ -202,6 +216,8 @@ enum aarch64_symbol_type aarch64_classify_symbol (rtx,
enum aarch64_symbol_type aarch64_classify_tls_symbol (rtx);
enum reg_class aarch64_regno_regclass (unsigned);
int aarch64_asm_preferred_eh_data_format (int, int);
+enum machine_mode aarch64_hard_regno_caller_save_mode (unsigned, unsigned,
+ enum machine_mode);
int aarch64_hard_regno_mode_ok (unsigned, enum machine_mode);
int aarch64_hard_regno_nregs (unsigned, enum machine_mode);
int aarch64_simd_attr_length_move (rtx);
@@ -291,4 +307,5 @@ extern void aarch64_split_combinev16qi (rtx operands[3]);
extern void aarch64_expand_vec_perm (rtx target, rtx op0, rtx op1, rtx sel);
extern bool
aarch64_expand_vec_perm_const (rtx target, rtx op0, rtx op1, rtx sel);
+void aarch64_atomic_assign_expand_fenv (tree *, tree *, tree *);
#endif /* GCC_AARCH64_PROTOS_H */
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64.h
index c9b30d0..e0384b3 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/config/aarch64/aarch64.h
@@ -35,6 +35,9 @@
if (TARGET_SIMD) \
builtin_define ("__ARM_NEON"); \
\
+ if (TARGET_CRC32) \
+ builtin_define ("__ARM_FEATURE_CRC32"); \
+ \
switch (aarch64_cmodel) \
{ \
case AARCH64_CMODEL_TINY: \
@@ -188,6 +191,9 @@ extern unsigned long aarch64_tune_flags;
/* Crypto is an optional extension to AdvSIMD. */
#define TARGET_CRYPTO (TARGET_SIMD && AARCH64_ISA_CRYPTO)
+/* CRC instructions that can be enabled through +crc arch extension. */
+#define TARGET_CRC32 (AARCH64_ISA_CRC)
+
/* Standard register usage. */
/* 31 64-bit general purpose registers R0-R30:
@@ -408,6 +414,7 @@ extern unsigned long aarch64_tune_flags;
enum reg_class
{
NO_REGS,
+ CALLER_SAVE_REGS,
CORE_REGS,
GENERAL_REGS,
STACK_REG,
@@ -423,6 +430,7 @@ enum reg_class
#define REG_CLASS_NAMES \
{ \
"NO_REGS", \
+ "CALLER_SAVE_REGS", \
"CORE_REGS", \
"GENERAL_REGS", \
"STACK_REG", \
@@ -435,6 +443,7 @@ enum reg_class
#define REG_CLASS_CONTENTS \
{ \
{ 0x00000000, 0x00000000, 0x00000000 }, /* NO_REGS */ \
+ { 0x0007ffff, 0x00000000, 0x00000000 }, /* CALLER_SAVE_REGS */ \
{ 0x7fffffff, 0x00000000, 0x00000003 }, /* CORE_REGS */ \
{ 0x7fffffff, 0x00000000, 0x00000003 }, /* GENERAL_REGS */ \
{ 0x80000000, 0x00000000, 0x00000000 }, /* STACK_REG */ \
@@ -659,12 +668,14 @@ do { \
/* The base cost overhead of a memcpy call, for MOVE_RATIO and friends. */
#define AARCH64_CALL_RATIO 8
-/* When optimizing for size, give a better estimate of the length of a memcpy
- call, but use the default otherwise. But move_by_pieces_ninsns() counts
- memory-to-memory moves, and we'll have to generate a load & store for each,
- so halve the value to take that into account. */
+/* MOVE_RATIO dictates when we will use the move_by_pieces infrastructure.
+ move_by_pieces will continually copy the largest safe chunks. So a
+ 7-byte copy is a 4-byte + 2-byte + byte copy. This proves inefficient
+ for both size and speed of copy, so we will instead use the "movmem"
+ standard name to implement the copy. This logic does not apply when
+ targeting -mstrict-align, so keep a sensible default in that case. */
#define MOVE_RATIO(speed) \
- (((speed) ? 15 : AARCH64_CALL_RATIO) / 2)
+ (!STRICT_ALIGNMENT ? 2 : (((speed) ? 15 : AARCH64_CALL_RATIO) / 2))
/* For CLEAR_RATIO, when optimizing for size, give a better estimate
of the length of a memset call, but use the default otherwise. */
@@ -824,6 +835,11 @@ do { \
#define SHIFT_COUNT_TRUNCATED !TARGET_SIMD
+/* Choose appropriate mode for caller saves, so we do the minimum
+ required size of load/store. */
+#define HARD_REGNO_CALLER_SAVE_MODE(REGNO, NREGS, MODE) \
+ aarch64_hard_regno_caller_save_mode ((REGNO), (NREGS), (MODE))
+
/* Callee only saves lower 64-bits of a 128-bit register. Tell the
compiler the callee clobbers the top 64-bits when restoring the
bottom 64-bits. */
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/configargs.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/configargs.h
index dbe4a51..1b088a2 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/configargs.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/configargs.h
@@ -1,5 +1,5 @@
/* Generated automatically. */
-static const char configuration_arguments[] = "/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/configure --prefix=/tmp/android-toolchain-eabi --host=x86_64-linux-gnu --build=x86_64-linux-gnu --with-gnu-as --with-gnu-ld --enable-languages=c,c++ --with-gmp=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/temp-install --with-mpfr=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/temp-install --with-mpc=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/temp-install --disable-libssp --disable-nls --disable-libmudflap --disable-shared --disable-sjlj-exceptions --disable-libquadmath --disable-libitm --disable-libsanitizer --prefix=/tmp/android-toolchain-eabi --disable-docs --disable-nls --with-gcc-version=linaro-4.9-2014.06 --with-binutils-version=current --with-gmp-version=current --with-mpfr-version=current --with-mpc-version=current --disable-graphite --target=aarch64-linux-android --enable-threads --enable-tls --enable-libgomp --disable-libitm --disable-libsanitizer --enable-libatomic --disable-libstdc__-v3 : (reconfigured) /mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/configure --prefix=/tmp/android-toolchain-eabi --host=x86_64-linux-gnu --build=x86_64-linux-gnu --with-gnu-as --with-gnu-ld --enable-languages=c,c++ --with-gmp=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/temp-install --with-mpfr=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/temp-install --with-mpc=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/temp-install --disable-libssp --disable-nls --disable-libmudflap --disable-shared --disable-sjlj-exceptions --disable-libquadmath --disable-libitm --disable-libsanitizer --prefix=/tmp/android-toolchain-eabi --disable-docs --disable-nls --with-gcc-version=linaro-4.9-2014.06 --with-binutils-version=current --with-gmp-version=current --with-mpfr-version=current --with-mpc-version=current --disable-graphite --target=aarch64-linux-android --enable-threads --enable-tls --enable-libgomp --disable-libitm --disable-libsanitizer --enable-libatomic --disable-libstdc__-v3";
+static const char configuration_arguments[] = "/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/configure --prefix=/tmp/android-toolchain-eabi --host=x86_64-linux-gnu --build=x86_64-linux-gnu --with-gnu-as --with-gnu-ld --enable-languages=c,c++ --with-gmp=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/temp-install --with-mpfr=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/temp-install --with-mpc=/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/temp-install --disable-libssp --disable-nls --disable-libmudflap --disable-shared --disable-sjlj-exceptions --disable-libquadmath --disable-libitm --disable-libsanitizer --prefix=/tmp/android-toolchain-eabi --disable-docs --disable-nls --with-gcc-version=linaro-4.9-2014.07 --with-binutils-version=current --with-gmp-version=current --with-mpfr-version=current --with-mpc-version=current --disable-graphite --target=aarch64-linux-android --enable-threads --enable-tls --disable-libgomp --disable-libstdc__-v3 --disable-libitm --disable-libsanitizer --disable-libatomic";
static const char thread_model[] = "posix";
static const struct {
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/cp/cp-tree.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/cp/cp-tree.h
index e9fe86e..26a63d0 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/cp/cp-tree.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/cp/cp-tree.h
@@ -101,12 +101,14 @@ c-common.h, not after.
FNDECL_USED_AUTO (in FUNCTION_DECL)
DECLTYPE_FOR_LAMBDA_PROXY (in DECLTYPE_TYPE)
REF_PARENTHESIZED_P (in COMPONENT_REF, SCOPE_REF)
+ AGGR_INIT_ZERO_FIRST (in AGGR_INIT_EXPR)
3: (TREE_REFERENCE_EXPR) (in NON_LVALUE_EXPR) (commented-out).
ICS_BAD_FLAG (in _CONV)
FN_TRY_BLOCK_P (in TRY_BLOCK)
IDENTIFIER_CTOR_OR_DTOR_P (in IDENTIFIER_NODE)
BIND_EXPR_BODY_BLOCK (in BIND_EXPR)
DECL_NON_TRIVIALLY_INITIALIZED_P (in VAR_DECL)
+ CALL_EXPR_LIST_INIT_P (in CALL_EXPR, AGGR_INIT_EXPR)
4: TREE_HAS_CONSTRUCTOR (in INDIRECT_REF, SAVE_EXPR, CONSTRUCTOR,
or FIELD_DECL).
IDENTIFIER_TYPENAME_P (in IDENTIFIER_NODE)
@@ -3026,6 +3028,10 @@ extern void decl_shadowed_for_var_insert (tree, tree);
should be performed at instantiation time. */
#define KOENIG_LOOKUP_P(NODE) TREE_LANG_FLAG_0 (CALL_EXPR_CHECK (NODE))
+/* True if CALL_EXPR expresses list-initialization of an object. */
+#define CALL_EXPR_LIST_INIT_P(NODE) \
+ TREE_LANG_FLAG_3 (TREE_CHECK2 ((NODE),CALL_EXPR,AGGR_INIT_EXPR))
+
/* Indicates whether a string literal has been parenthesized. Such
usages are disallowed in certain circumstances. */
@@ -3430,6 +3436,9 @@ more_aggr_init_expr_args_p (const aggr_init_expr_arg_iterator *iter)
B b{1,2}, not B b({1,2}) or B b = {1,2}. */
#define CONSTRUCTOR_IS_DIRECT_INIT(NODE) (TREE_LANG_FLAG_0 (CONSTRUCTOR_CHECK (NODE)))
+#define DIRECT_LIST_INIT_P(NODE) \
+ (BRACE_ENCLOSED_INITIALIZER_P (NODE) && CONSTRUCTOR_IS_DIRECT_INIT (NODE))
+
/* True if NODE represents a conversion for direct-initialization in a
template. Set by perform_implicit_conversion_flags. */
#define IMPLICIT_CONV_EXPR_DIRECT_INIT(NODE) \
@@ -4347,6 +4356,11 @@ extern int function_depth;
PARM_DECLs in cp_tree_equal. */
extern int comparing_specializations;
+/* A type-qualifier, or bitmask therefore, using the TYPE_QUAL
+ constants. */
+
+typedef int cp_cv_quals;
+
/* In parser.c. */
/* Nonzero if we are parsing an unevaluated operand: an operand to
@@ -4356,6 +4370,7 @@ extern int comparing_specializations;
extern int cp_unevaluated_operand;
extern tree cp_convert_range_for (tree, tree, tree, bool);
extern bool parsing_nsdmi (void);
+extern void inject_this_parameter (tree, cp_cv_quals);
/* in pt.c */
@@ -4735,11 +4750,6 @@ extern GTY(()) operator_name_info_t operator_name_info
extern GTY(()) operator_name_info_t assignment_operator_name_info
[(int) MAX_TREE_CODES];
-/* A type-qualifier, or bitmask therefore, using the TYPE_QUAL
- constants. */
-
-typedef int cp_cv_quals;
-
/* Non-static member functions have an optional virt-specifier-seq.
There is a VIRT_SPEC value for each virt-specifier.
They can be combined by bitwise-or to form the complete set of
@@ -5415,6 +5425,7 @@ extern tree get_type_value (tree);
extern tree build_zero_init (tree, tree, bool);
extern tree build_value_init (tree, tsubst_flags_t);
extern tree build_value_init_noctor (tree, tsubst_flags_t);
+extern tree get_nsdmi (tree, bool);
extern tree build_offset_ref (tree, tree, bool,
tsubst_flags_t);
extern tree throw_bad_array_new_length (void);
@@ -6151,6 +6162,7 @@ extern tree store_init_value (tree, tree, vec<tree, va_gc>**, int);
extern void check_narrowing (tree, tree);
extern tree digest_init (tree, tree, tsubst_flags_t);
extern tree digest_init_flags (tree, tree, int);
+extern tree digest_nsdmi_init (tree, tree);
extern tree build_scoped_ref (tree, tree, tree *);
extern tree build_x_arrow (location_t, tree,
tsubst_flags_t);
@@ -6201,7 +6213,7 @@ extern tree cxx_omp_clause_default_ctor (tree, tree, tree);
extern tree cxx_omp_clause_copy_ctor (tree, tree, tree);
extern tree cxx_omp_clause_assign_op (tree, tree, tree);
extern tree cxx_omp_clause_dtor (tree, tree);
-extern void cxx_omp_finish_clause (tree);
+extern void cxx_omp_finish_clause (tree, gimple_seq *);
extern bool cxx_omp_privatize_by_reference (const_tree);
/* in name-lookup.c */
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/gimplify.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/gimplify.h
index 47e7213..5085ccf 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/gimplify.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/gimplify.h
@@ -77,6 +77,7 @@ extern enum gimplify_status gimplify_expr (tree *, gimple_seq *, gimple_seq *,
extern void gimplify_type_sizes (tree, gimple_seq *);
extern void gimplify_one_sizepos (tree *, gimple_seq *);
extern gimple gimplify_body (tree, bool);
+extern enum gimplify_status gimplify_arg (tree *, gimple_seq *, location_t);
extern void gimplify_function_tree (tree);
extern enum gimplify_status gimplify_va_arg_expr (tree *, gimple_seq *,
gimple_seq *);
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-codes.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-codes.h
index 3385748..9eca9bb 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-codes.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-codes.h
@@ -64,2224 +64,2340 @@ enum insn_code {
CODE_FOR_cstorehi_neg = 359,
CODE_FOR_cstoresi_neg = 360,
CODE_FOR_cstoredi_neg = 361,
- CODE_FOR_csinc3si_insn = 372,
- CODE_FOR_csinc3di_insn = 373,
- CODE_FOR_andsi3 = 378,
- CODE_FOR_iorsi3 = 379,
- CODE_FOR_xorsi3 = 380,
- CODE_FOR_anddi3 = 381,
- CODE_FOR_iordi3 = 382,
- CODE_FOR_xordi3 = 383,
- CODE_FOR_one_cmplsi2 = 447,
- CODE_FOR_one_cmpldi2 = 448,
- CODE_FOR_clzsi2 = 502,
- CODE_FOR_clzdi2 = 503,
- CODE_FOR_clrsbsi2 = 504,
- CODE_FOR_clrsbdi2 = 505,
- CODE_FOR_rbitsi2 = 506,
- CODE_FOR_rbitdi2 = 507,
- CODE_FOR_bswapsi2 = 587,
- CODE_FOR_bswapdi2 = 588,
- CODE_FOR_bswaphi2 = 589,
- CODE_FOR_rev16si2 = 590,
- CODE_FOR_rev16di2 = 591,
- CODE_FOR_rev16si2_alt = 592,
- CODE_FOR_rev16di2_alt = 593,
- CODE_FOR_btruncsf2 = 595,
- CODE_FOR_ceilsf2 = 596,
- CODE_FOR_floorsf2 = 597,
- CODE_FOR_frintnsf2 = 598,
- CODE_FOR_nearbyintsf2 = 599,
- CODE_FOR_rintsf2 = 600,
- CODE_FOR_roundsf2 = 601,
- CODE_FOR_btruncdf2 = 602,
- CODE_FOR_ceildf2 = 603,
- CODE_FOR_floordf2 = 604,
- CODE_FOR_frintndf2 = 605,
- CODE_FOR_nearbyintdf2 = 606,
- CODE_FOR_rintdf2 = 607,
- CODE_FOR_rounddf2 = 608,
- CODE_FOR_lbtruncsfsi2 = 609,
- CODE_FOR_lceilsfsi2 = 610,
- CODE_FOR_lfloorsfsi2 = 611,
- CODE_FOR_lroundsfsi2 = 612,
- CODE_FOR_lfrintnsfsi2 = 613,
- CODE_FOR_lbtruncusfsi2 = 614,
- CODE_FOR_lceilusfsi2 = 615,
- CODE_FOR_lfloorusfsi2 = 616,
- CODE_FOR_lroundusfsi2 = 617,
- CODE_FOR_lfrintnusfsi2 = 618,
- CODE_FOR_lbtruncdfsi2 = 619,
- CODE_FOR_lceildfsi2 = 620,
- CODE_FOR_lfloordfsi2 = 621,
- CODE_FOR_lrounddfsi2 = 622,
- CODE_FOR_lfrintndfsi2 = 623,
- CODE_FOR_lbtruncudfsi2 = 624,
- CODE_FOR_lceiludfsi2 = 625,
- CODE_FOR_lfloorudfsi2 = 626,
- CODE_FOR_lroundudfsi2 = 627,
- CODE_FOR_lfrintnudfsi2 = 628,
- CODE_FOR_lbtruncsfdi2 = 629,
- CODE_FOR_lceilsfdi2 = 630,
- CODE_FOR_lfloorsfdi2 = 631,
- CODE_FOR_lroundsfdi2 = 632,
- CODE_FOR_lfrintnsfdi2 = 633,
- CODE_FOR_lbtruncusfdi2 = 634,
- CODE_FOR_lceilusfdi2 = 635,
- CODE_FOR_lfloorusfdi2 = 636,
- CODE_FOR_lroundusfdi2 = 637,
- CODE_FOR_lfrintnusfdi2 = 638,
- CODE_FOR_lbtruncdfdi2 = 639,
- CODE_FOR_lceildfdi2 = 640,
- CODE_FOR_lfloordfdi2 = 641,
- CODE_FOR_lrounddfdi2 = 642,
- CODE_FOR_lfrintndfdi2 = 643,
- CODE_FOR_lbtruncudfdi2 = 644,
- CODE_FOR_lceiludfdi2 = 645,
- CODE_FOR_lfloorudfdi2 = 646,
- CODE_FOR_lroundudfdi2 = 647,
- CODE_FOR_lfrintnudfdi2 = 648,
- CODE_FOR_fmasf4 = 649,
- CODE_FOR_fmadf4 = 650,
- CODE_FOR_fnmasf4 = 651,
- CODE_FOR_fnmadf4 = 652,
- CODE_FOR_fmssf4 = 653,
- CODE_FOR_fmsdf4 = 654,
- CODE_FOR_fnmssf4 = 655,
- CODE_FOR_fnmsdf4 = 656,
- CODE_FOR_extendsfdf2 = 659,
- CODE_FOR_truncdfsf2 = 660,
- CODE_FOR_fix_truncsfsi2 = 661,
- CODE_FOR_fix_truncdfsi2 = 662,
- CODE_FOR_fix_truncsfdi2 = 663,
- CODE_FOR_fix_truncdfdi2 = 664,
- CODE_FOR_fixuns_truncsfsi2 = 665,
- CODE_FOR_fixuns_truncdfsi2 = 666,
- CODE_FOR_fixuns_truncsfdi2 = 667,
- CODE_FOR_fixuns_truncdfdi2 = 668,
- CODE_FOR_floatsisf2 = 669,
- CODE_FOR_floatsidf2 = 670,
- CODE_FOR_floatdisf2 = 671,
- CODE_FOR_floatdidf2 = 672,
- CODE_FOR_floatunssisf2 = 673,
- CODE_FOR_floatunssidf2 = 674,
- CODE_FOR_floatunsdisf2 = 675,
- CODE_FOR_floatunsdidf2 = 676,
- CODE_FOR_addsf3 = 677,
- CODE_FOR_adddf3 = 678,
- CODE_FOR_subsf3 = 679,
- CODE_FOR_subdf3 = 680,
- CODE_FOR_mulsf3 = 681,
- CODE_FOR_muldf3 = 682,
- CODE_FOR_divsf3 = 685,
- CODE_FOR_divdf3 = 686,
- CODE_FOR_negsf2 = 687,
- CODE_FOR_negdf2 = 688,
- CODE_FOR_sqrtsf2 = 689,
- CODE_FOR_sqrtdf2 = 690,
- CODE_FOR_abssf2 = 691,
- CODE_FOR_absdf2 = 692,
- CODE_FOR_smaxsf3 = 693,
- CODE_FOR_smaxdf3 = 694,
- CODE_FOR_sminsf3 = 695,
- CODE_FOR_smindf3 = 696,
- CODE_FOR_aarch64_movdi_tilow = 697,
- CODE_FOR_aarch64_movdi_tflow = 698,
- CODE_FOR_aarch64_movdi_tihigh = 699,
- CODE_FOR_aarch64_movdi_tfhigh = 700,
- CODE_FOR_aarch64_movtihigh_di = 701,
- CODE_FOR_aarch64_movtfhigh_di = 702,
- CODE_FOR_aarch64_movtilow_di = 703,
- CODE_FOR_aarch64_movtflow_di = 704,
- CODE_FOR_aarch64_movtilow_tilow = 705,
- CODE_FOR_add_losym_si = 706,
- CODE_FOR_add_losym_di = 707,
- CODE_FOR_ldr_got_small_si = 708,
- CODE_FOR_ldr_got_small_di = 709,
- CODE_FOR_ldr_got_small_sidi = 710,
- CODE_FOR_ldr_got_tiny = 711,
- CODE_FOR_aarch64_load_tp_hard = 712,
- CODE_FOR_tlsie_small_si = 714,
- CODE_FOR_tlsie_small_di = 715,
- CODE_FOR_tlsie_small_sidi = 716,
- CODE_FOR_tlsle_small_si = 717,
- CODE_FOR_tlsle_small_di = 718,
- CODE_FOR_tlsdesc_small_si = 719,
- CODE_FOR_tlsdesc_small_di = 720,
- CODE_FOR_stack_tie = 721,
- CODE_FOR_stack_protect_set_si = 722,
- CODE_FOR_stack_protect_set_di = 723,
- CODE_FOR_stack_protect_test_si = 724,
- CODE_FOR_stack_protect_test_di = 725,
- CODE_FOR_aarch64_simd_dupv8qi = 726,
- CODE_FOR_aarch64_simd_dupv16qi = 727,
- CODE_FOR_aarch64_simd_dupv4hi = 728,
- CODE_FOR_aarch64_simd_dupv8hi = 729,
- CODE_FOR_aarch64_simd_dupv2si = 730,
- CODE_FOR_aarch64_simd_dupv4si = 731,
- CODE_FOR_aarch64_simd_dupv2di = 732,
- CODE_FOR_aarch64_simd_dupv2sf = 733,
- CODE_FOR_aarch64_simd_dupv4sf = 734,
- CODE_FOR_aarch64_simd_dupv2df = 735,
- CODE_FOR_aarch64_dup_lanev8qi = 736,
- CODE_FOR_aarch64_dup_lanev16qi = 737,
- CODE_FOR_aarch64_dup_lanev4hi = 738,
- CODE_FOR_aarch64_dup_lanev8hi = 739,
- CODE_FOR_aarch64_dup_lanev2si = 740,
- CODE_FOR_aarch64_dup_lanev4si = 741,
- CODE_FOR_aarch64_dup_lanev2di = 742,
- CODE_FOR_aarch64_dup_lanev2sf = 743,
- CODE_FOR_aarch64_dup_lanev4sf = 744,
- CODE_FOR_aarch64_dup_lanev2df = 745,
- CODE_FOR_aarch64_dup_lane_to_128v8qi = 746,
- CODE_FOR_aarch64_dup_lane_to_64v16qi = 747,
- CODE_FOR_aarch64_dup_lane_to_128v4hi = 748,
- CODE_FOR_aarch64_dup_lane_to_64v8hi = 749,
- CODE_FOR_aarch64_dup_lane_to_128v2si = 750,
- CODE_FOR_aarch64_dup_lane_to_64v4si = 751,
- CODE_FOR_aarch64_dup_lane_to_64v2di = 752,
- CODE_FOR_aarch64_dup_lane_to_128v2sf = 753,
- CODE_FOR_aarch64_dup_lane_to_64v4sf = 754,
- CODE_FOR_aarch64_dup_lane_to_64v2df = 755,
- CODE_FOR_aarch64_simd_mov_from_v16qilow = 766,
- CODE_FOR_aarch64_simd_mov_from_v8hilow = 767,
- CODE_FOR_aarch64_simd_mov_from_v4silow = 768,
- CODE_FOR_aarch64_simd_mov_from_v2dilow = 769,
- CODE_FOR_aarch64_simd_mov_from_v4sflow = 770,
- CODE_FOR_aarch64_simd_mov_from_v2dflow = 771,
- CODE_FOR_aarch64_simd_mov_from_v16qihigh = 772,
- CODE_FOR_aarch64_simd_mov_from_v8hihigh = 773,
- CODE_FOR_aarch64_simd_mov_from_v4sihigh = 774,
- CODE_FOR_aarch64_simd_mov_from_v2dihigh = 775,
- CODE_FOR_aarch64_simd_mov_from_v4sfhigh = 776,
- CODE_FOR_aarch64_simd_mov_from_v2dfhigh = 777,
- CODE_FOR_ornv8qi3 = 778,
- CODE_FOR_ornv16qi3 = 779,
- CODE_FOR_ornv4hi3 = 780,
- CODE_FOR_ornv8hi3 = 781,
- CODE_FOR_ornv2si3 = 782,
- CODE_FOR_ornv4si3 = 783,
- CODE_FOR_ornv2di3 = 784,
- CODE_FOR_bicv8qi3 = 785,
- CODE_FOR_bicv16qi3 = 786,
- CODE_FOR_bicv4hi3 = 787,
- CODE_FOR_bicv8hi3 = 788,
- CODE_FOR_bicv2si3 = 789,
- CODE_FOR_bicv4si3 = 790,
- CODE_FOR_bicv2di3 = 791,
- CODE_FOR_addv8qi3 = 792,
- CODE_FOR_addv16qi3 = 793,
- CODE_FOR_addv4hi3 = 794,
- CODE_FOR_addv8hi3 = 795,
- CODE_FOR_addv2si3 = 796,
- CODE_FOR_addv4si3 = 797,
- CODE_FOR_addv2di3 = 798,
- CODE_FOR_subv8qi3 = 799,
- CODE_FOR_subv16qi3 = 800,
- CODE_FOR_subv4hi3 = 801,
- CODE_FOR_subv8hi3 = 802,
- CODE_FOR_subv2si3 = 803,
- CODE_FOR_subv4si3 = 804,
- CODE_FOR_subv2di3 = 805,
- CODE_FOR_mulv8qi3 = 806,
- CODE_FOR_mulv16qi3 = 807,
- CODE_FOR_mulv4hi3 = 808,
- CODE_FOR_mulv8hi3 = 809,
- CODE_FOR_mulv2si3 = 810,
- CODE_FOR_mulv4si3 = 811,
- CODE_FOR_bswapv4hi = 812,
- CODE_FOR_bswapv8hi = 813,
- CODE_FOR_bswapv2si = 814,
- CODE_FOR_bswapv4si = 815,
- CODE_FOR_bswapv2di = 816,
- CODE_FOR_negv8qi2 = 832,
- CODE_FOR_negv16qi2 = 833,
- CODE_FOR_negv4hi2 = 834,
- CODE_FOR_negv8hi2 = 835,
- CODE_FOR_negv2si2 = 836,
- CODE_FOR_negv4si2 = 837,
- CODE_FOR_negv2di2 = 838,
- CODE_FOR_absv8qi2 = 839,
- CODE_FOR_absv16qi2 = 840,
- CODE_FOR_absv4hi2 = 841,
- CODE_FOR_absv8hi2 = 842,
- CODE_FOR_absv2si2 = 843,
- CODE_FOR_absv4si2 = 844,
- CODE_FOR_absv2di2 = 845,
- CODE_FOR_abdv8qi_3 = 846,
- CODE_FOR_abdv16qi_3 = 847,
- CODE_FOR_abdv4hi_3 = 848,
- CODE_FOR_abdv8hi_3 = 849,
- CODE_FOR_abdv2si_3 = 850,
- CODE_FOR_abdv4si_3 = 851,
- CODE_FOR_abav8qi_3 = 852,
- CODE_FOR_abav16qi_3 = 853,
- CODE_FOR_abav4hi_3 = 854,
- CODE_FOR_abav8hi_3 = 855,
- CODE_FOR_abav2si_3 = 856,
- CODE_FOR_abav4si_3 = 857,
- CODE_FOR_fabdv2sf_3 = 858,
- CODE_FOR_fabdv4sf_3 = 859,
- CODE_FOR_fabdv2df_3 = 860,
- CODE_FOR_andv8qi3 = 863,
- CODE_FOR_andv16qi3 = 864,
- CODE_FOR_andv4hi3 = 865,
- CODE_FOR_andv8hi3 = 866,
- CODE_FOR_andv2si3 = 867,
- CODE_FOR_andv4si3 = 868,
- CODE_FOR_andv2di3 = 869,
- CODE_FOR_iorv8qi3 = 870,
- CODE_FOR_iorv16qi3 = 871,
- CODE_FOR_iorv4hi3 = 872,
- CODE_FOR_iorv8hi3 = 873,
- CODE_FOR_iorv2si3 = 874,
- CODE_FOR_iorv4si3 = 875,
- CODE_FOR_iorv2di3 = 876,
- CODE_FOR_xorv8qi3 = 877,
- CODE_FOR_xorv16qi3 = 878,
- CODE_FOR_xorv4hi3 = 879,
- CODE_FOR_xorv8hi3 = 880,
- CODE_FOR_xorv2si3 = 881,
- CODE_FOR_xorv4si3 = 882,
- CODE_FOR_xorv2di3 = 883,
- CODE_FOR_one_cmplv8qi2 = 884,
- CODE_FOR_one_cmplv16qi2 = 885,
- CODE_FOR_one_cmplv4hi2 = 886,
- CODE_FOR_one_cmplv8hi2 = 887,
- CODE_FOR_one_cmplv2si2 = 888,
- CODE_FOR_one_cmplv4si2 = 889,
- CODE_FOR_one_cmplv2di2 = 890,
- CODE_FOR_aarch64_simd_vec_setv8qi = 891,
- CODE_FOR_aarch64_simd_vec_setv16qi = 892,
- CODE_FOR_aarch64_simd_vec_setv4hi = 893,
- CODE_FOR_aarch64_simd_vec_setv8hi = 894,
- CODE_FOR_aarch64_simd_vec_setv2si = 895,
- CODE_FOR_aarch64_simd_vec_setv4si = 896,
- CODE_FOR_aarch64_simd_lshrv8qi = 897,
- CODE_FOR_aarch64_simd_lshrv16qi = 898,
- CODE_FOR_aarch64_simd_lshrv4hi = 899,
- CODE_FOR_aarch64_simd_lshrv8hi = 900,
- CODE_FOR_aarch64_simd_lshrv2si = 901,
- CODE_FOR_aarch64_simd_lshrv4si = 902,
- CODE_FOR_aarch64_simd_lshrv2di = 903,
- CODE_FOR_aarch64_simd_ashrv8qi = 904,
- CODE_FOR_aarch64_simd_ashrv16qi = 905,
- CODE_FOR_aarch64_simd_ashrv4hi = 906,
- CODE_FOR_aarch64_simd_ashrv8hi = 907,
- CODE_FOR_aarch64_simd_ashrv2si = 908,
- CODE_FOR_aarch64_simd_ashrv4si = 909,
- CODE_FOR_aarch64_simd_ashrv2di = 910,
- CODE_FOR_aarch64_simd_imm_shlv8qi = 911,
- CODE_FOR_aarch64_simd_imm_shlv16qi = 912,
- CODE_FOR_aarch64_simd_imm_shlv4hi = 913,
- CODE_FOR_aarch64_simd_imm_shlv8hi = 914,
- CODE_FOR_aarch64_simd_imm_shlv2si = 915,
- CODE_FOR_aarch64_simd_imm_shlv4si = 916,
- CODE_FOR_aarch64_simd_imm_shlv2di = 917,
- CODE_FOR_aarch64_simd_reg_sshlv8qi = 918,
- CODE_FOR_aarch64_simd_reg_sshlv16qi = 919,
- CODE_FOR_aarch64_simd_reg_sshlv4hi = 920,
- CODE_FOR_aarch64_simd_reg_sshlv8hi = 921,
- CODE_FOR_aarch64_simd_reg_sshlv2si = 922,
- CODE_FOR_aarch64_simd_reg_sshlv4si = 923,
- CODE_FOR_aarch64_simd_reg_sshlv2di = 924,
- CODE_FOR_aarch64_simd_reg_shlv8qi_unsigned = 925,
- CODE_FOR_aarch64_simd_reg_shlv16qi_unsigned = 926,
- CODE_FOR_aarch64_simd_reg_shlv4hi_unsigned = 927,
- CODE_FOR_aarch64_simd_reg_shlv8hi_unsigned = 928,
- CODE_FOR_aarch64_simd_reg_shlv2si_unsigned = 929,
- CODE_FOR_aarch64_simd_reg_shlv4si_unsigned = 930,
- CODE_FOR_aarch64_simd_reg_shlv2di_unsigned = 931,
- CODE_FOR_aarch64_simd_reg_shlv8qi_signed = 932,
- CODE_FOR_aarch64_simd_reg_shlv16qi_signed = 933,
- CODE_FOR_aarch64_simd_reg_shlv4hi_signed = 934,
- CODE_FOR_aarch64_simd_reg_shlv8hi_signed = 935,
- CODE_FOR_aarch64_simd_reg_shlv2si_signed = 936,
- CODE_FOR_aarch64_simd_reg_shlv4si_signed = 937,
- CODE_FOR_aarch64_simd_reg_shlv2di_signed = 938,
- CODE_FOR_aarch64_sshr_simddi = 939,
- CODE_FOR_aarch64_ushr_simddi = 940,
- CODE_FOR_aarch64_simd_vec_setv2di = 941,
- CODE_FOR_aarch64_simd_vec_setv2sf = 942,
- CODE_FOR_aarch64_simd_vec_setv4sf = 943,
- CODE_FOR_aarch64_simd_vec_setv2df = 944,
- CODE_FOR_aarch64_mlav8qi = 945,
- CODE_FOR_aarch64_mlav16qi = 946,
- CODE_FOR_aarch64_mlav4hi = 947,
- CODE_FOR_aarch64_mlav8hi = 948,
- CODE_FOR_aarch64_mlav2si = 949,
- CODE_FOR_aarch64_mlav4si = 950,
- CODE_FOR_aarch64_mlsv8qi = 959,
- CODE_FOR_aarch64_mlsv16qi = 960,
- CODE_FOR_aarch64_mlsv4hi = 961,
- CODE_FOR_aarch64_mlsv8hi = 962,
- CODE_FOR_aarch64_mlsv2si = 963,
- CODE_FOR_aarch64_mlsv4si = 964,
- CODE_FOR_smaxv8qi3 = 973,
- CODE_FOR_sminv8qi3 = 974,
- CODE_FOR_umaxv8qi3 = 975,
- CODE_FOR_uminv8qi3 = 976,
- CODE_FOR_smaxv16qi3 = 977,
- CODE_FOR_sminv16qi3 = 978,
- CODE_FOR_umaxv16qi3 = 979,
- CODE_FOR_uminv16qi3 = 980,
- CODE_FOR_smaxv4hi3 = 981,
- CODE_FOR_sminv4hi3 = 982,
- CODE_FOR_umaxv4hi3 = 983,
- CODE_FOR_uminv4hi3 = 984,
- CODE_FOR_smaxv8hi3 = 985,
- CODE_FOR_sminv8hi3 = 986,
- CODE_FOR_umaxv8hi3 = 987,
- CODE_FOR_uminv8hi3 = 988,
- CODE_FOR_smaxv2si3 = 989,
- CODE_FOR_sminv2si3 = 990,
- CODE_FOR_umaxv2si3 = 991,
- CODE_FOR_uminv2si3 = 992,
- CODE_FOR_smaxv4si3 = 993,
- CODE_FOR_sminv4si3 = 994,
- CODE_FOR_umaxv4si3 = 995,
- CODE_FOR_uminv4si3 = 996,
- CODE_FOR_move_lo_quad_v16qi = 997,
- CODE_FOR_move_lo_quad_v8hi = 998,
- CODE_FOR_move_lo_quad_v4si = 999,
- CODE_FOR_move_lo_quad_v2di = 1000,
- CODE_FOR_move_lo_quad_v4sf = 1001,
- CODE_FOR_move_lo_quad_v2df = 1002,
- CODE_FOR_aarch64_simd_move_hi_quad_v16qi = 1003,
- CODE_FOR_aarch64_simd_move_hi_quad_v8hi = 1004,
- CODE_FOR_aarch64_simd_move_hi_quad_v4si = 1005,
- CODE_FOR_aarch64_simd_move_hi_quad_v2di = 1006,
- CODE_FOR_aarch64_simd_move_hi_quad_v4sf = 1007,
- CODE_FOR_aarch64_simd_move_hi_quad_v2df = 1008,
- CODE_FOR_aarch64_simd_vec_pack_trunc_v8hi = 1009,
- CODE_FOR_aarch64_simd_vec_pack_trunc_v4si = 1010,
- CODE_FOR_aarch64_simd_vec_pack_trunc_v2di = 1011,
- CODE_FOR_vec_pack_trunc_v8hi = 1012,
- CODE_FOR_vec_pack_trunc_v4si = 1013,
- CODE_FOR_vec_pack_trunc_v2di = 1014,
- CODE_FOR_aarch64_simd_vec_unpacks_lo_v16qi = 1015,
- CODE_FOR_aarch64_simd_vec_unpacku_lo_v16qi = 1016,
- CODE_FOR_aarch64_simd_vec_unpacks_lo_v8hi = 1017,
- CODE_FOR_aarch64_simd_vec_unpacku_lo_v8hi = 1018,
- CODE_FOR_aarch64_simd_vec_unpacks_lo_v4si = 1019,
- CODE_FOR_aarch64_simd_vec_unpacku_lo_v4si = 1020,
- CODE_FOR_aarch64_simd_vec_unpacks_hi_v16qi = 1021,
- CODE_FOR_aarch64_simd_vec_unpacku_hi_v16qi = 1022,
- CODE_FOR_aarch64_simd_vec_unpacks_hi_v8hi = 1023,
- CODE_FOR_aarch64_simd_vec_unpacku_hi_v8hi = 1024,
- CODE_FOR_aarch64_simd_vec_unpacks_hi_v4si = 1025,
- CODE_FOR_aarch64_simd_vec_unpacku_hi_v4si = 1026,
- CODE_FOR_aarch64_simd_vec_smult_lo_v16qi = 1063,
- CODE_FOR_aarch64_simd_vec_umult_lo_v16qi = 1064,
- CODE_FOR_aarch64_simd_vec_smult_lo_v8hi = 1065,
- CODE_FOR_aarch64_simd_vec_umult_lo_v8hi = 1066,
- CODE_FOR_aarch64_simd_vec_smult_lo_v4si = 1067,
- CODE_FOR_aarch64_simd_vec_umult_lo_v4si = 1068,
- CODE_FOR_aarch64_simd_vec_smult_hi_v16qi = 1069,
- CODE_FOR_aarch64_simd_vec_umult_hi_v16qi = 1070,
- CODE_FOR_aarch64_simd_vec_smult_hi_v8hi = 1071,
- CODE_FOR_aarch64_simd_vec_umult_hi_v8hi = 1072,
- CODE_FOR_aarch64_simd_vec_smult_hi_v4si = 1073,
- CODE_FOR_aarch64_simd_vec_umult_hi_v4si = 1074,
- CODE_FOR_addv2sf3 = 1075,
- CODE_FOR_addv4sf3 = 1076,
- CODE_FOR_addv2df3 = 1077,
- CODE_FOR_subv2sf3 = 1078,
- CODE_FOR_subv4sf3 = 1079,
- CODE_FOR_subv2df3 = 1080,
- CODE_FOR_mulv2sf3 = 1081,
- CODE_FOR_mulv4sf3 = 1082,
- CODE_FOR_mulv2df3 = 1083,
- CODE_FOR_divv2sf3 = 1084,
- CODE_FOR_divv4sf3 = 1085,
- CODE_FOR_divv2df3 = 1086,
- CODE_FOR_negv2sf2 = 1087,
- CODE_FOR_negv4sf2 = 1088,
- CODE_FOR_negv2df2 = 1089,
- CODE_FOR_absv2sf2 = 1090,
- CODE_FOR_absv4sf2 = 1091,
- CODE_FOR_absv2df2 = 1092,
- CODE_FOR_fmav2sf4 = 1093,
- CODE_FOR_fmav4sf4 = 1094,
- CODE_FOR_fmav2df4 = 1095,
- CODE_FOR_fnmav2sf4 = 1103,
- CODE_FOR_fnmav4sf4 = 1104,
- CODE_FOR_fnmav2df4 = 1105,
- CODE_FOR_btruncv2sf2 = 1113,
- CODE_FOR_ceilv2sf2 = 1114,
- CODE_FOR_floorv2sf2 = 1115,
- CODE_FOR_frintnv2sf2 = 1116,
- CODE_FOR_nearbyintv2sf2 = 1117,
- CODE_FOR_rintv2sf2 = 1118,
- CODE_FOR_roundv2sf2 = 1119,
- CODE_FOR_btruncv4sf2 = 1120,
- CODE_FOR_ceilv4sf2 = 1121,
- CODE_FOR_floorv4sf2 = 1122,
- CODE_FOR_frintnv4sf2 = 1123,
- CODE_FOR_nearbyintv4sf2 = 1124,
- CODE_FOR_rintv4sf2 = 1125,
- CODE_FOR_roundv4sf2 = 1126,
- CODE_FOR_btruncv2df2 = 1127,
- CODE_FOR_ceilv2df2 = 1128,
- CODE_FOR_floorv2df2 = 1129,
- CODE_FOR_frintnv2df2 = 1130,
- CODE_FOR_nearbyintv2df2 = 1131,
- CODE_FOR_rintv2df2 = 1132,
- CODE_FOR_roundv2df2 = 1133,
- CODE_FOR_lbtruncv2sfv2si2 = 1134,
- CODE_FOR_lceilv2sfv2si2 = 1135,
- CODE_FOR_lfloorv2sfv2si2 = 1136,
- CODE_FOR_lroundv2sfv2si2 = 1137,
- CODE_FOR_lfrintnv2sfv2si2 = 1138,
- CODE_FOR_lbtruncuv2sfv2si2 = 1139,
- CODE_FOR_lceiluv2sfv2si2 = 1140,
- CODE_FOR_lflooruv2sfv2si2 = 1141,
- CODE_FOR_lrounduv2sfv2si2 = 1142,
- CODE_FOR_lfrintnuv2sfv2si2 = 1143,
- CODE_FOR_lbtruncv4sfv4si2 = 1144,
- CODE_FOR_lceilv4sfv4si2 = 1145,
- CODE_FOR_lfloorv4sfv4si2 = 1146,
- CODE_FOR_lroundv4sfv4si2 = 1147,
- CODE_FOR_lfrintnv4sfv4si2 = 1148,
- CODE_FOR_lbtruncuv4sfv4si2 = 1149,
- CODE_FOR_lceiluv4sfv4si2 = 1150,
- CODE_FOR_lflooruv4sfv4si2 = 1151,
- CODE_FOR_lrounduv4sfv4si2 = 1152,
- CODE_FOR_lfrintnuv4sfv4si2 = 1153,
- CODE_FOR_lbtruncv2dfv2di2 = 1154,
- CODE_FOR_lceilv2dfv2di2 = 1155,
- CODE_FOR_lfloorv2dfv2di2 = 1156,
- CODE_FOR_lroundv2dfv2di2 = 1157,
- CODE_FOR_lfrintnv2dfv2di2 = 1158,
- CODE_FOR_lbtruncuv2dfv2di2 = 1159,
- CODE_FOR_lceiluv2dfv2di2 = 1160,
- CODE_FOR_lflooruv2dfv2di2 = 1161,
- CODE_FOR_lrounduv2dfv2di2 = 1162,
- CODE_FOR_lfrintnuv2dfv2di2 = 1163,
- CODE_FOR_floatv2siv2sf2 = 1164,
- CODE_FOR_floatunsv2siv2sf2 = 1165,
- CODE_FOR_floatv4siv4sf2 = 1166,
- CODE_FOR_floatunsv4siv4sf2 = 1167,
- CODE_FOR_floatv2div2df2 = 1168,
- CODE_FOR_floatunsv2div2df2 = 1169,
- CODE_FOR_vec_unpacks_lo_v4sf = 1170,
- CODE_FOR_aarch64_float_extend_lo_v2df = 1171,
- CODE_FOR_vec_unpacks_hi_v4sf = 1172,
- CODE_FOR_aarch64_float_truncate_lo_v2sf = 1173,
- CODE_FOR_aarch64_float_truncate_hi_v4sf = 1174,
- CODE_FOR_aarch64_vmlsv2sf = 1175,
- CODE_FOR_aarch64_vmlsv4sf = 1176,
- CODE_FOR_aarch64_vmlsv2df = 1177,
- CODE_FOR_smaxv2sf3 = 1178,
- CODE_FOR_sminv2sf3 = 1179,
- CODE_FOR_smaxv4sf3 = 1180,
- CODE_FOR_sminv4sf3 = 1181,
- CODE_FOR_smaxv2df3 = 1182,
- CODE_FOR_sminv2df3 = 1183,
- CODE_FOR_smax_nanv2sf3 = 1184,
- CODE_FOR_smin_nanv2sf3 = 1185,
- CODE_FOR_smax_nanv4sf3 = 1186,
- CODE_FOR_smin_nanv4sf3 = 1187,
- CODE_FOR_smax_nanv2df3 = 1188,
- CODE_FOR_smin_nanv2df3 = 1189,
- CODE_FOR_reduc_splus_v8qi = 1190,
- CODE_FOR_reduc_uplus_v8qi = 1191,
- CODE_FOR_reduc_splus_v16qi = 1192,
- CODE_FOR_reduc_uplus_v16qi = 1193,
- CODE_FOR_reduc_splus_v4hi = 1194,
- CODE_FOR_reduc_uplus_v4hi = 1195,
- CODE_FOR_reduc_splus_v8hi = 1196,
- CODE_FOR_reduc_uplus_v8hi = 1197,
- CODE_FOR_reduc_splus_v4si = 1198,
- CODE_FOR_reduc_uplus_v4si = 1199,
- CODE_FOR_reduc_splus_v2di = 1200,
- CODE_FOR_reduc_uplus_v2di = 1201,
- CODE_FOR_reduc_splus_v2si = 1202,
- CODE_FOR_reduc_uplus_v2si = 1203,
- CODE_FOR_reduc_splus_v2sf = 1204,
- CODE_FOR_reduc_splus_v2df = 1205,
- CODE_FOR_aarch64_addpv4sf = 1206,
- CODE_FOR_clzv8qi2 = 1207,
- CODE_FOR_clzv16qi2 = 1208,
- CODE_FOR_clzv4hi2 = 1209,
- CODE_FOR_clzv8hi2 = 1210,
- CODE_FOR_clzv2si2 = 1211,
- CODE_FOR_clzv4si2 = 1212,
- CODE_FOR_reduc_umax_v8qi = 1213,
- CODE_FOR_reduc_umin_v8qi = 1214,
- CODE_FOR_reduc_smax_v8qi = 1215,
- CODE_FOR_reduc_smin_v8qi = 1216,
- CODE_FOR_reduc_umax_v16qi = 1217,
- CODE_FOR_reduc_umin_v16qi = 1218,
- CODE_FOR_reduc_smax_v16qi = 1219,
- CODE_FOR_reduc_smin_v16qi = 1220,
- CODE_FOR_reduc_umax_v4hi = 1221,
- CODE_FOR_reduc_umin_v4hi = 1222,
- CODE_FOR_reduc_smax_v4hi = 1223,
- CODE_FOR_reduc_smin_v4hi = 1224,
- CODE_FOR_reduc_umax_v8hi = 1225,
- CODE_FOR_reduc_umin_v8hi = 1226,
- CODE_FOR_reduc_smax_v8hi = 1227,
- CODE_FOR_reduc_smin_v8hi = 1228,
- CODE_FOR_reduc_umax_v4si = 1229,
- CODE_FOR_reduc_umin_v4si = 1230,
- CODE_FOR_reduc_smax_v4si = 1231,
- CODE_FOR_reduc_smin_v4si = 1232,
- CODE_FOR_reduc_umax_v2si = 1233,
- CODE_FOR_reduc_umin_v2si = 1234,
- CODE_FOR_reduc_smax_v2si = 1235,
- CODE_FOR_reduc_smin_v2si = 1236,
- CODE_FOR_reduc_smax_nan_v2sf = 1237,
- CODE_FOR_reduc_smin_nan_v2sf = 1238,
- CODE_FOR_reduc_smax_v2sf = 1239,
- CODE_FOR_reduc_smin_v2sf = 1240,
- CODE_FOR_reduc_smax_nan_v2df = 1241,
- CODE_FOR_reduc_smin_nan_v2df = 1242,
- CODE_FOR_reduc_smax_v2df = 1243,
- CODE_FOR_reduc_smin_v2df = 1244,
- CODE_FOR_reduc_smax_nan_v4sf = 1245,
- CODE_FOR_reduc_smin_nan_v4sf = 1246,
- CODE_FOR_reduc_smax_v4sf = 1247,
- CODE_FOR_reduc_smin_v4sf = 1248,
- CODE_FOR_aarch64_simd_bslv8qi_internal = 1249,
- CODE_FOR_aarch64_simd_bslv16qi_internal = 1250,
- CODE_FOR_aarch64_simd_bslv4hi_internal = 1251,
- CODE_FOR_aarch64_simd_bslv8hi_internal = 1252,
- CODE_FOR_aarch64_simd_bslv2si_internal = 1253,
- CODE_FOR_aarch64_simd_bslv4si_internal = 1254,
- CODE_FOR_aarch64_simd_bslv2di_internal = 1255,
- CODE_FOR_aarch64_simd_bslv2sf_internal = 1256,
- CODE_FOR_aarch64_simd_bslv4sf_internal = 1257,
- CODE_FOR_aarch64_simd_bslv2df_internal = 1258,
- CODE_FOR_aarch64_simd_bsldi_internal = 1259,
- CODE_FOR_aarch64_simd_bsldf_internal = 1260,
- CODE_FOR_aarch64_get_lanev8qi = 1273,
- CODE_FOR_aarch64_get_lanev16qi = 1274,
- CODE_FOR_aarch64_get_lanev4hi = 1275,
- CODE_FOR_aarch64_get_lanev8hi = 1276,
- CODE_FOR_aarch64_get_lanev2si = 1277,
- CODE_FOR_aarch64_get_lanev4si = 1278,
- CODE_FOR_aarch64_get_lanev2di = 1279,
- CODE_FOR_aarch64_get_lanev2sf = 1280,
- CODE_FOR_aarch64_get_lanev4sf = 1281,
- CODE_FOR_aarch64_get_lanev2df = 1282,
- CODE_FOR_aarch64_combinev8qi = 1286,
- CODE_FOR_aarch64_combinev4hi = 1287,
- CODE_FOR_aarch64_combinev2si = 1288,
- CODE_FOR_aarch64_combinev2sf = 1289,
- CODE_FOR_aarch64_combinedi = 1290,
- CODE_FOR_aarch64_combinedf = 1291,
- CODE_FOR_aarch64_saddlv16qi_hi_internal = 1292,
- CODE_FOR_aarch64_ssublv16qi_hi_internal = 1293,
- CODE_FOR_aarch64_uaddlv16qi_hi_internal = 1294,
- CODE_FOR_aarch64_usublv16qi_hi_internal = 1295,
- CODE_FOR_aarch64_saddlv8hi_hi_internal = 1296,
- CODE_FOR_aarch64_ssublv8hi_hi_internal = 1297,
- CODE_FOR_aarch64_uaddlv8hi_hi_internal = 1298,
- CODE_FOR_aarch64_usublv8hi_hi_internal = 1299,
- CODE_FOR_aarch64_saddlv4si_hi_internal = 1300,
- CODE_FOR_aarch64_ssublv4si_hi_internal = 1301,
- CODE_FOR_aarch64_uaddlv4si_hi_internal = 1302,
- CODE_FOR_aarch64_usublv4si_hi_internal = 1303,
- CODE_FOR_aarch64_saddlv16qi_lo_internal = 1304,
- CODE_FOR_aarch64_ssublv16qi_lo_internal = 1305,
- CODE_FOR_aarch64_uaddlv16qi_lo_internal = 1306,
- CODE_FOR_aarch64_usublv16qi_lo_internal = 1307,
- CODE_FOR_aarch64_saddlv8hi_lo_internal = 1308,
- CODE_FOR_aarch64_ssublv8hi_lo_internal = 1309,
- CODE_FOR_aarch64_uaddlv8hi_lo_internal = 1310,
- CODE_FOR_aarch64_usublv8hi_lo_internal = 1311,
- CODE_FOR_aarch64_saddlv4si_lo_internal = 1312,
- CODE_FOR_aarch64_ssublv4si_lo_internal = 1313,
- CODE_FOR_aarch64_uaddlv4si_lo_internal = 1314,
- CODE_FOR_aarch64_usublv4si_lo_internal = 1315,
- CODE_FOR_aarch64_saddlv8qi = 1316,
- CODE_FOR_aarch64_ssublv8qi = 1317,
- CODE_FOR_aarch64_uaddlv8qi = 1318,
- CODE_FOR_aarch64_usublv8qi = 1319,
- CODE_FOR_aarch64_saddlv4hi = 1320,
- CODE_FOR_aarch64_ssublv4hi = 1321,
- CODE_FOR_aarch64_uaddlv4hi = 1322,
- CODE_FOR_aarch64_usublv4hi = 1323,
- CODE_FOR_aarch64_saddlv2si = 1324,
- CODE_FOR_aarch64_ssublv2si = 1325,
- CODE_FOR_aarch64_uaddlv2si = 1326,
- CODE_FOR_aarch64_usublv2si = 1327,
- CODE_FOR_aarch64_saddwv8qi = 1328,
- CODE_FOR_aarch64_ssubwv8qi = 1329,
- CODE_FOR_aarch64_uaddwv8qi = 1330,
- CODE_FOR_aarch64_usubwv8qi = 1331,
- CODE_FOR_aarch64_saddwv4hi = 1332,
- CODE_FOR_aarch64_ssubwv4hi = 1333,
- CODE_FOR_aarch64_uaddwv4hi = 1334,
- CODE_FOR_aarch64_usubwv4hi = 1335,
- CODE_FOR_aarch64_saddwv2si = 1336,
- CODE_FOR_aarch64_ssubwv2si = 1337,
- CODE_FOR_aarch64_uaddwv2si = 1338,
- CODE_FOR_aarch64_usubwv2si = 1339,
- CODE_FOR_aarch64_saddw2v16qi_internal = 1340,
- CODE_FOR_aarch64_ssubw2v16qi_internal = 1341,
- CODE_FOR_aarch64_uaddw2v16qi_internal = 1342,
- CODE_FOR_aarch64_usubw2v16qi_internal = 1343,
- CODE_FOR_aarch64_saddw2v8hi_internal = 1344,
- CODE_FOR_aarch64_ssubw2v8hi_internal = 1345,
- CODE_FOR_aarch64_uaddw2v8hi_internal = 1346,
- CODE_FOR_aarch64_usubw2v8hi_internal = 1347,
- CODE_FOR_aarch64_saddw2v4si_internal = 1348,
- CODE_FOR_aarch64_ssubw2v4si_internal = 1349,
- CODE_FOR_aarch64_uaddw2v4si_internal = 1350,
- CODE_FOR_aarch64_usubw2v4si_internal = 1351,
- CODE_FOR_aarch64_shaddv8qi = 1352,
- CODE_FOR_aarch64_uhaddv8qi = 1353,
- CODE_FOR_aarch64_srhaddv8qi = 1354,
- CODE_FOR_aarch64_urhaddv8qi = 1355,
- CODE_FOR_aarch64_shsubv8qi = 1356,
- CODE_FOR_aarch64_uhsubv8qi = 1357,
- CODE_FOR_aarch64_srhsubv8qi = 1358,
- CODE_FOR_aarch64_urhsubv8qi = 1359,
- CODE_FOR_aarch64_shaddv16qi = 1360,
- CODE_FOR_aarch64_uhaddv16qi = 1361,
- CODE_FOR_aarch64_srhaddv16qi = 1362,
- CODE_FOR_aarch64_urhaddv16qi = 1363,
- CODE_FOR_aarch64_shsubv16qi = 1364,
- CODE_FOR_aarch64_uhsubv16qi = 1365,
- CODE_FOR_aarch64_srhsubv16qi = 1366,
- CODE_FOR_aarch64_urhsubv16qi = 1367,
- CODE_FOR_aarch64_shaddv4hi = 1368,
- CODE_FOR_aarch64_uhaddv4hi = 1369,
- CODE_FOR_aarch64_srhaddv4hi = 1370,
- CODE_FOR_aarch64_urhaddv4hi = 1371,
- CODE_FOR_aarch64_shsubv4hi = 1372,
- CODE_FOR_aarch64_uhsubv4hi = 1373,
- CODE_FOR_aarch64_srhsubv4hi = 1374,
- CODE_FOR_aarch64_urhsubv4hi = 1375,
- CODE_FOR_aarch64_shaddv8hi = 1376,
- CODE_FOR_aarch64_uhaddv8hi = 1377,
- CODE_FOR_aarch64_srhaddv8hi = 1378,
- CODE_FOR_aarch64_urhaddv8hi = 1379,
- CODE_FOR_aarch64_shsubv8hi = 1380,
- CODE_FOR_aarch64_uhsubv8hi = 1381,
- CODE_FOR_aarch64_srhsubv8hi = 1382,
- CODE_FOR_aarch64_urhsubv8hi = 1383,
- CODE_FOR_aarch64_shaddv2si = 1384,
- CODE_FOR_aarch64_uhaddv2si = 1385,
- CODE_FOR_aarch64_srhaddv2si = 1386,
- CODE_FOR_aarch64_urhaddv2si = 1387,
- CODE_FOR_aarch64_shsubv2si = 1388,
- CODE_FOR_aarch64_uhsubv2si = 1389,
- CODE_FOR_aarch64_srhsubv2si = 1390,
- CODE_FOR_aarch64_urhsubv2si = 1391,
- CODE_FOR_aarch64_shaddv4si = 1392,
- CODE_FOR_aarch64_uhaddv4si = 1393,
- CODE_FOR_aarch64_srhaddv4si = 1394,
- CODE_FOR_aarch64_urhaddv4si = 1395,
- CODE_FOR_aarch64_shsubv4si = 1396,
- CODE_FOR_aarch64_uhsubv4si = 1397,
- CODE_FOR_aarch64_srhsubv4si = 1398,
- CODE_FOR_aarch64_urhsubv4si = 1399,
- CODE_FOR_aarch64_addhnv8hi = 1400,
- CODE_FOR_aarch64_raddhnv8hi = 1401,
- CODE_FOR_aarch64_subhnv8hi = 1402,
- CODE_FOR_aarch64_rsubhnv8hi = 1403,
- CODE_FOR_aarch64_addhnv4si = 1404,
- CODE_FOR_aarch64_raddhnv4si = 1405,
- CODE_FOR_aarch64_subhnv4si = 1406,
- CODE_FOR_aarch64_rsubhnv4si = 1407,
- CODE_FOR_aarch64_addhnv2di = 1408,
- CODE_FOR_aarch64_raddhnv2di = 1409,
- CODE_FOR_aarch64_subhnv2di = 1410,
- CODE_FOR_aarch64_rsubhnv2di = 1411,
- CODE_FOR_aarch64_addhn2v8hi = 1412,
- CODE_FOR_aarch64_raddhn2v8hi = 1413,
- CODE_FOR_aarch64_subhn2v8hi = 1414,
- CODE_FOR_aarch64_rsubhn2v8hi = 1415,
- CODE_FOR_aarch64_addhn2v4si = 1416,
- CODE_FOR_aarch64_raddhn2v4si = 1417,
- CODE_FOR_aarch64_subhn2v4si = 1418,
- CODE_FOR_aarch64_rsubhn2v4si = 1419,
- CODE_FOR_aarch64_addhn2v2di = 1420,
- CODE_FOR_aarch64_raddhn2v2di = 1421,
- CODE_FOR_aarch64_subhn2v2di = 1422,
- CODE_FOR_aarch64_rsubhn2v2di = 1423,
- CODE_FOR_aarch64_pmulv8qi = 1424,
- CODE_FOR_aarch64_pmulv16qi = 1425,
- CODE_FOR_aarch64_sqaddv8qi = 1426,
- CODE_FOR_aarch64_uqaddv8qi = 1427,
- CODE_FOR_aarch64_sqsubv8qi = 1428,
- CODE_FOR_aarch64_uqsubv8qi = 1429,
- CODE_FOR_aarch64_sqaddv16qi = 1430,
- CODE_FOR_aarch64_uqaddv16qi = 1431,
- CODE_FOR_aarch64_sqsubv16qi = 1432,
- CODE_FOR_aarch64_uqsubv16qi = 1433,
- CODE_FOR_aarch64_sqaddv4hi = 1434,
- CODE_FOR_aarch64_uqaddv4hi = 1435,
- CODE_FOR_aarch64_sqsubv4hi = 1436,
- CODE_FOR_aarch64_uqsubv4hi = 1437,
- CODE_FOR_aarch64_sqaddv8hi = 1438,
- CODE_FOR_aarch64_uqaddv8hi = 1439,
- CODE_FOR_aarch64_sqsubv8hi = 1440,
- CODE_FOR_aarch64_uqsubv8hi = 1441,
- CODE_FOR_aarch64_sqaddv2si = 1442,
- CODE_FOR_aarch64_uqaddv2si = 1443,
- CODE_FOR_aarch64_sqsubv2si = 1444,
- CODE_FOR_aarch64_uqsubv2si = 1445,
- CODE_FOR_aarch64_sqaddv4si = 1446,
- CODE_FOR_aarch64_uqaddv4si = 1447,
- CODE_FOR_aarch64_sqsubv4si = 1448,
- CODE_FOR_aarch64_uqsubv4si = 1449,
- CODE_FOR_aarch64_sqaddv2di = 1450,
- CODE_FOR_aarch64_uqaddv2di = 1451,
- CODE_FOR_aarch64_sqsubv2di = 1452,
- CODE_FOR_aarch64_uqsubv2di = 1453,
- CODE_FOR_aarch64_sqaddqi = 1454,
- CODE_FOR_aarch64_uqaddqi = 1455,
- CODE_FOR_aarch64_sqsubqi = 1456,
- CODE_FOR_aarch64_uqsubqi = 1457,
- CODE_FOR_aarch64_sqaddhi = 1458,
- CODE_FOR_aarch64_uqaddhi = 1459,
- CODE_FOR_aarch64_sqsubhi = 1460,
- CODE_FOR_aarch64_uqsubhi = 1461,
- CODE_FOR_aarch64_sqaddsi = 1462,
- CODE_FOR_aarch64_uqaddsi = 1463,
- CODE_FOR_aarch64_sqsubsi = 1464,
- CODE_FOR_aarch64_uqsubsi = 1465,
- CODE_FOR_aarch64_sqadddi = 1466,
- CODE_FOR_aarch64_uqadddi = 1467,
- CODE_FOR_aarch64_sqsubdi = 1468,
- CODE_FOR_aarch64_uqsubdi = 1469,
- CODE_FOR_aarch64_suqaddv8qi = 1470,
- CODE_FOR_aarch64_usqaddv8qi = 1471,
- CODE_FOR_aarch64_suqaddv16qi = 1472,
- CODE_FOR_aarch64_usqaddv16qi = 1473,
- CODE_FOR_aarch64_suqaddv4hi = 1474,
- CODE_FOR_aarch64_usqaddv4hi = 1475,
- CODE_FOR_aarch64_suqaddv8hi = 1476,
- CODE_FOR_aarch64_usqaddv8hi = 1477,
- CODE_FOR_aarch64_suqaddv2si = 1478,
- CODE_FOR_aarch64_usqaddv2si = 1479,
- CODE_FOR_aarch64_suqaddv4si = 1480,
- CODE_FOR_aarch64_usqaddv4si = 1481,
- CODE_FOR_aarch64_suqaddv2di = 1482,
- CODE_FOR_aarch64_usqaddv2di = 1483,
- CODE_FOR_aarch64_suqaddqi = 1484,
- CODE_FOR_aarch64_usqaddqi = 1485,
- CODE_FOR_aarch64_suqaddhi = 1486,
- CODE_FOR_aarch64_usqaddhi = 1487,
- CODE_FOR_aarch64_suqaddsi = 1488,
- CODE_FOR_aarch64_usqaddsi = 1489,
- CODE_FOR_aarch64_suqadddi = 1490,
- CODE_FOR_aarch64_usqadddi = 1491,
- CODE_FOR_aarch64_sqmovunv8hi = 1492,
- CODE_FOR_aarch64_sqmovunv4si = 1493,
- CODE_FOR_aarch64_sqmovunv2di = 1494,
- CODE_FOR_aarch64_sqmovunhi = 1495,
- CODE_FOR_aarch64_sqmovunsi = 1496,
- CODE_FOR_aarch64_sqmovundi = 1497,
- CODE_FOR_aarch64_sqmovnv8hi = 1498,
- CODE_FOR_aarch64_uqmovnv8hi = 1499,
- CODE_FOR_aarch64_sqmovnv4si = 1500,
- CODE_FOR_aarch64_uqmovnv4si = 1501,
- CODE_FOR_aarch64_sqmovnv2di = 1502,
- CODE_FOR_aarch64_uqmovnv2di = 1503,
- CODE_FOR_aarch64_sqmovnhi = 1504,
- CODE_FOR_aarch64_uqmovnhi = 1505,
- CODE_FOR_aarch64_sqmovnsi = 1506,
- CODE_FOR_aarch64_uqmovnsi = 1507,
- CODE_FOR_aarch64_sqmovndi = 1508,
- CODE_FOR_aarch64_uqmovndi = 1509,
- CODE_FOR_aarch64_sqnegv8qi = 1510,
- CODE_FOR_aarch64_sqabsv8qi = 1511,
- CODE_FOR_aarch64_sqnegv16qi = 1512,
- CODE_FOR_aarch64_sqabsv16qi = 1513,
- CODE_FOR_aarch64_sqnegv4hi = 1514,
- CODE_FOR_aarch64_sqabsv4hi = 1515,
- CODE_FOR_aarch64_sqnegv8hi = 1516,
- CODE_FOR_aarch64_sqabsv8hi = 1517,
- CODE_FOR_aarch64_sqnegv2si = 1518,
- CODE_FOR_aarch64_sqabsv2si = 1519,
- CODE_FOR_aarch64_sqnegv4si = 1520,
- CODE_FOR_aarch64_sqabsv4si = 1521,
- CODE_FOR_aarch64_sqnegv2di = 1522,
- CODE_FOR_aarch64_sqabsv2di = 1523,
- CODE_FOR_aarch64_sqnegqi = 1524,
- CODE_FOR_aarch64_sqabsqi = 1525,
- CODE_FOR_aarch64_sqneghi = 1526,
- CODE_FOR_aarch64_sqabshi = 1527,
- CODE_FOR_aarch64_sqnegsi = 1528,
- CODE_FOR_aarch64_sqabssi = 1529,
- CODE_FOR_aarch64_sqnegdi = 1530,
- CODE_FOR_aarch64_sqabsdi = 1531,
- CODE_FOR_aarch64_sqdmulhv4hi = 1532,
- CODE_FOR_aarch64_sqrdmulhv4hi = 1533,
- CODE_FOR_aarch64_sqdmulhv8hi = 1534,
- CODE_FOR_aarch64_sqrdmulhv8hi = 1535,
- CODE_FOR_aarch64_sqdmulhv2si = 1536,
- CODE_FOR_aarch64_sqrdmulhv2si = 1537,
- CODE_FOR_aarch64_sqdmulhv4si = 1538,
- CODE_FOR_aarch64_sqrdmulhv4si = 1539,
- CODE_FOR_aarch64_sqdmulhhi = 1540,
- CODE_FOR_aarch64_sqrdmulhhi = 1541,
- CODE_FOR_aarch64_sqdmulhsi = 1542,
- CODE_FOR_aarch64_sqrdmulhsi = 1543,
- CODE_FOR_aarch64_sqdmulh_lanev4hi = 1544,
- CODE_FOR_aarch64_sqrdmulh_lanev4hi = 1545,
- CODE_FOR_aarch64_sqdmulh_lanev8hi = 1546,
- CODE_FOR_aarch64_sqrdmulh_lanev8hi = 1547,
- CODE_FOR_aarch64_sqdmulh_lanev2si = 1548,
- CODE_FOR_aarch64_sqrdmulh_lanev2si = 1549,
- CODE_FOR_aarch64_sqdmulh_lanev4si = 1550,
- CODE_FOR_aarch64_sqrdmulh_lanev4si = 1551,
- CODE_FOR_aarch64_sqdmulh_laneqv4hi = 1552,
- CODE_FOR_aarch64_sqrdmulh_laneqv4hi = 1553,
- CODE_FOR_aarch64_sqdmulh_laneqv8hi = 1554,
- CODE_FOR_aarch64_sqrdmulh_laneqv8hi = 1555,
- CODE_FOR_aarch64_sqdmulh_laneqv2si = 1556,
- CODE_FOR_aarch64_sqrdmulh_laneqv2si = 1557,
- CODE_FOR_aarch64_sqdmulh_laneqv4si = 1558,
- CODE_FOR_aarch64_sqrdmulh_laneqv4si = 1559,
- CODE_FOR_aarch64_sqdmulh_lanehi = 1560,
- CODE_FOR_aarch64_sqrdmulh_lanehi = 1561,
- CODE_FOR_aarch64_sqdmulh_lanesi = 1562,
- CODE_FOR_aarch64_sqrdmulh_lanesi = 1563,
- CODE_FOR_aarch64_sqdmlalv4hi = 1564,
- CODE_FOR_aarch64_sqdmlslv4hi = 1565,
- CODE_FOR_aarch64_sqdmlalv2si = 1566,
- CODE_FOR_aarch64_sqdmlslv2si = 1567,
- CODE_FOR_aarch64_sqdmlalhi = 1568,
- CODE_FOR_aarch64_sqdmlslhi = 1569,
- CODE_FOR_aarch64_sqdmlalsi = 1570,
- CODE_FOR_aarch64_sqdmlslsi = 1571,
- CODE_FOR_aarch64_sqdmlal_lanev4hi_internal = 1572,
- CODE_FOR_aarch64_sqdmlsl_lanev4hi_internal = 1573,
- CODE_FOR_aarch64_sqdmlal_lanev2si_internal = 1574,
- CODE_FOR_aarch64_sqdmlsl_lanev2si_internal = 1575,
- CODE_FOR_aarch64_sqdmlal_lanehi_internal = 1576,
- CODE_FOR_aarch64_sqdmlsl_lanehi_internal = 1577,
- CODE_FOR_aarch64_sqdmlal_lanesi_internal = 1578,
- CODE_FOR_aarch64_sqdmlsl_lanesi_internal = 1579,
- CODE_FOR_aarch64_sqdmlal_nv4hi = 1580,
- CODE_FOR_aarch64_sqdmlsl_nv4hi = 1581,
- CODE_FOR_aarch64_sqdmlal_nv2si = 1582,
- CODE_FOR_aarch64_sqdmlsl_nv2si = 1583,
- CODE_FOR_aarch64_sqdmlal2v8hi_internal = 1584,
- CODE_FOR_aarch64_sqdmlsl2v8hi_internal = 1585,
- CODE_FOR_aarch64_sqdmlal2v4si_internal = 1586,
- CODE_FOR_aarch64_sqdmlsl2v4si_internal = 1587,
- CODE_FOR_aarch64_sqdmlal2_lanev8hi_internal = 1588,
- CODE_FOR_aarch64_sqdmlsl2_lanev8hi_internal = 1589,
- CODE_FOR_aarch64_sqdmlal2_lanev4si_internal = 1590,
- CODE_FOR_aarch64_sqdmlsl2_lanev4si_internal = 1591,
- CODE_FOR_aarch64_sqdmlal2_nv8hi_internal = 1592,
- CODE_FOR_aarch64_sqdmlsl2_nv8hi_internal = 1593,
- CODE_FOR_aarch64_sqdmlal2_nv4si_internal = 1594,
- CODE_FOR_aarch64_sqdmlsl2_nv4si_internal = 1595,
- CODE_FOR_aarch64_sqdmullv4hi = 1596,
- CODE_FOR_aarch64_sqdmullv2si = 1597,
- CODE_FOR_aarch64_sqdmullhi = 1598,
- CODE_FOR_aarch64_sqdmullsi = 1599,
- CODE_FOR_aarch64_sqdmull_lanev4hi_internal = 1600,
- CODE_FOR_aarch64_sqdmull_lanev2si_internal = 1601,
- CODE_FOR_aarch64_sqdmull_lanehi_internal = 1602,
- CODE_FOR_aarch64_sqdmull_lanesi_internal = 1603,
- CODE_FOR_aarch64_sqdmull_nv4hi = 1604,
- CODE_FOR_aarch64_sqdmull_nv2si = 1605,
- CODE_FOR_aarch64_sqdmull2v8hi_internal = 1606,
- CODE_FOR_aarch64_sqdmull2v4si_internal = 1607,
- CODE_FOR_aarch64_sqdmull2_lanev8hi_internal = 1608,
- CODE_FOR_aarch64_sqdmull2_lanev4si_internal = 1609,
- CODE_FOR_aarch64_sqdmull2_nv8hi_internal = 1610,
- CODE_FOR_aarch64_sqdmull2_nv4si_internal = 1611,
- CODE_FOR_aarch64_sshlv8qi = 1612,
- CODE_FOR_aarch64_ushlv8qi = 1613,
- CODE_FOR_aarch64_srshlv8qi = 1614,
- CODE_FOR_aarch64_urshlv8qi = 1615,
- CODE_FOR_aarch64_sshlv16qi = 1616,
- CODE_FOR_aarch64_ushlv16qi = 1617,
- CODE_FOR_aarch64_srshlv16qi = 1618,
- CODE_FOR_aarch64_urshlv16qi = 1619,
- CODE_FOR_aarch64_sshlv4hi = 1620,
- CODE_FOR_aarch64_ushlv4hi = 1621,
- CODE_FOR_aarch64_srshlv4hi = 1622,
- CODE_FOR_aarch64_urshlv4hi = 1623,
- CODE_FOR_aarch64_sshlv8hi = 1624,
- CODE_FOR_aarch64_ushlv8hi = 1625,
- CODE_FOR_aarch64_srshlv8hi = 1626,
- CODE_FOR_aarch64_urshlv8hi = 1627,
- CODE_FOR_aarch64_sshlv2si = 1628,
- CODE_FOR_aarch64_ushlv2si = 1629,
- CODE_FOR_aarch64_srshlv2si = 1630,
- CODE_FOR_aarch64_urshlv2si = 1631,
- CODE_FOR_aarch64_sshlv4si = 1632,
- CODE_FOR_aarch64_ushlv4si = 1633,
- CODE_FOR_aarch64_srshlv4si = 1634,
- CODE_FOR_aarch64_urshlv4si = 1635,
- CODE_FOR_aarch64_sshlv2di = 1636,
- CODE_FOR_aarch64_ushlv2di = 1637,
- CODE_FOR_aarch64_srshlv2di = 1638,
- CODE_FOR_aarch64_urshlv2di = 1639,
- CODE_FOR_aarch64_sshldi = 1640,
- CODE_FOR_aarch64_ushldi = 1641,
- CODE_FOR_aarch64_srshldi = 1642,
- CODE_FOR_aarch64_urshldi = 1643,
- CODE_FOR_aarch64_sqshlv8qi = 1644,
- CODE_FOR_aarch64_uqshlv8qi = 1645,
- CODE_FOR_aarch64_sqrshlv8qi = 1646,
- CODE_FOR_aarch64_uqrshlv8qi = 1647,
- CODE_FOR_aarch64_sqshlv16qi = 1648,
- CODE_FOR_aarch64_uqshlv16qi = 1649,
- CODE_FOR_aarch64_sqrshlv16qi = 1650,
- CODE_FOR_aarch64_uqrshlv16qi = 1651,
- CODE_FOR_aarch64_sqshlv4hi = 1652,
- CODE_FOR_aarch64_uqshlv4hi = 1653,
- CODE_FOR_aarch64_sqrshlv4hi = 1654,
- CODE_FOR_aarch64_uqrshlv4hi = 1655,
- CODE_FOR_aarch64_sqshlv8hi = 1656,
- CODE_FOR_aarch64_uqshlv8hi = 1657,
- CODE_FOR_aarch64_sqrshlv8hi = 1658,
- CODE_FOR_aarch64_uqrshlv8hi = 1659,
- CODE_FOR_aarch64_sqshlv2si = 1660,
- CODE_FOR_aarch64_uqshlv2si = 1661,
- CODE_FOR_aarch64_sqrshlv2si = 1662,
- CODE_FOR_aarch64_uqrshlv2si = 1663,
- CODE_FOR_aarch64_sqshlv4si = 1664,
- CODE_FOR_aarch64_uqshlv4si = 1665,
- CODE_FOR_aarch64_sqrshlv4si = 1666,
- CODE_FOR_aarch64_uqrshlv4si = 1667,
- CODE_FOR_aarch64_sqshlv2di = 1668,
- CODE_FOR_aarch64_uqshlv2di = 1669,
- CODE_FOR_aarch64_sqrshlv2di = 1670,
- CODE_FOR_aarch64_uqrshlv2di = 1671,
- CODE_FOR_aarch64_sqshlqi = 1672,
- CODE_FOR_aarch64_uqshlqi = 1673,
- CODE_FOR_aarch64_sqrshlqi = 1674,
- CODE_FOR_aarch64_uqrshlqi = 1675,
- CODE_FOR_aarch64_sqshlhi = 1676,
- CODE_FOR_aarch64_uqshlhi = 1677,
- CODE_FOR_aarch64_sqrshlhi = 1678,
- CODE_FOR_aarch64_uqrshlhi = 1679,
- CODE_FOR_aarch64_sqshlsi = 1680,
- CODE_FOR_aarch64_uqshlsi = 1681,
- CODE_FOR_aarch64_sqrshlsi = 1682,
- CODE_FOR_aarch64_uqrshlsi = 1683,
- CODE_FOR_aarch64_sqshldi = 1684,
- CODE_FOR_aarch64_uqshldi = 1685,
- CODE_FOR_aarch64_sqrshldi = 1686,
- CODE_FOR_aarch64_uqrshldi = 1687,
- CODE_FOR_aarch64_sshll_nv8qi = 1688,
- CODE_FOR_aarch64_ushll_nv8qi = 1689,
- CODE_FOR_aarch64_sshll_nv4hi = 1690,
- CODE_FOR_aarch64_ushll_nv4hi = 1691,
- CODE_FOR_aarch64_sshll_nv2si = 1692,
- CODE_FOR_aarch64_ushll_nv2si = 1693,
- CODE_FOR_aarch64_sshll2_nv16qi = 1694,
- CODE_FOR_aarch64_ushll2_nv16qi = 1695,
- CODE_FOR_aarch64_sshll2_nv8hi = 1696,
- CODE_FOR_aarch64_ushll2_nv8hi = 1697,
- CODE_FOR_aarch64_sshll2_nv4si = 1698,
- CODE_FOR_aarch64_ushll2_nv4si = 1699,
- CODE_FOR_aarch64_srshr_nv8qi = 1700,
- CODE_FOR_aarch64_urshr_nv8qi = 1701,
- CODE_FOR_aarch64_srshr_nv16qi = 1702,
- CODE_FOR_aarch64_urshr_nv16qi = 1703,
- CODE_FOR_aarch64_srshr_nv4hi = 1704,
- CODE_FOR_aarch64_urshr_nv4hi = 1705,
- CODE_FOR_aarch64_srshr_nv8hi = 1706,
- CODE_FOR_aarch64_urshr_nv8hi = 1707,
- CODE_FOR_aarch64_srshr_nv2si = 1708,
- CODE_FOR_aarch64_urshr_nv2si = 1709,
- CODE_FOR_aarch64_srshr_nv4si = 1710,
- CODE_FOR_aarch64_urshr_nv4si = 1711,
- CODE_FOR_aarch64_srshr_nv2di = 1712,
- CODE_FOR_aarch64_urshr_nv2di = 1713,
- CODE_FOR_aarch64_srshr_ndi = 1714,
- CODE_FOR_aarch64_urshr_ndi = 1715,
- CODE_FOR_aarch64_ssra_nv8qi = 1716,
- CODE_FOR_aarch64_usra_nv8qi = 1717,
- CODE_FOR_aarch64_srsra_nv8qi = 1718,
- CODE_FOR_aarch64_ursra_nv8qi = 1719,
- CODE_FOR_aarch64_ssra_nv16qi = 1720,
- CODE_FOR_aarch64_usra_nv16qi = 1721,
- CODE_FOR_aarch64_srsra_nv16qi = 1722,
- CODE_FOR_aarch64_ursra_nv16qi = 1723,
- CODE_FOR_aarch64_ssra_nv4hi = 1724,
- CODE_FOR_aarch64_usra_nv4hi = 1725,
- CODE_FOR_aarch64_srsra_nv4hi = 1726,
- CODE_FOR_aarch64_ursra_nv4hi = 1727,
- CODE_FOR_aarch64_ssra_nv8hi = 1728,
- CODE_FOR_aarch64_usra_nv8hi = 1729,
- CODE_FOR_aarch64_srsra_nv8hi = 1730,
- CODE_FOR_aarch64_ursra_nv8hi = 1731,
- CODE_FOR_aarch64_ssra_nv2si = 1732,
- CODE_FOR_aarch64_usra_nv2si = 1733,
- CODE_FOR_aarch64_srsra_nv2si = 1734,
- CODE_FOR_aarch64_ursra_nv2si = 1735,
- CODE_FOR_aarch64_ssra_nv4si = 1736,
- CODE_FOR_aarch64_usra_nv4si = 1737,
- CODE_FOR_aarch64_srsra_nv4si = 1738,
- CODE_FOR_aarch64_ursra_nv4si = 1739,
- CODE_FOR_aarch64_ssra_nv2di = 1740,
- CODE_FOR_aarch64_usra_nv2di = 1741,
- CODE_FOR_aarch64_srsra_nv2di = 1742,
- CODE_FOR_aarch64_ursra_nv2di = 1743,
- CODE_FOR_aarch64_ssra_ndi = 1744,
- CODE_FOR_aarch64_usra_ndi = 1745,
- CODE_FOR_aarch64_srsra_ndi = 1746,
- CODE_FOR_aarch64_ursra_ndi = 1747,
- CODE_FOR_aarch64_ssli_nv8qi = 1748,
- CODE_FOR_aarch64_usli_nv8qi = 1749,
- CODE_FOR_aarch64_ssri_nv8qi = 1750,
- CODE_FOR_aarch64_usri_nv8qi = 1751,
- CODE_FOR_aarch64_ssli_nv16qi = 1752,
- CODE_FOR_aarch64_usli_nv16qi = 1753,
- CODE_FOR_aarch64_ssri_nv16qi = 1754,
- CODE_FOR_aarch64_usri_nv16qi = 1755,
- CODE_FOR_aarch64_ssli_nv4hi = 1756,
- CODE_FOR_aarch64_usli_nv4hi = 1757,
- CODE_FOR_aarch64_ssri_nv4hi = 1758,
- CODE_FOR_aarch64_usri_nv4hi = 1759,
- CODE_FOR_aarch64_ssli_nv8hi = 1760,
- CODE_FOR_aarch64_usli_nv8hi = 1761,
- CODE_FOR_aarch64_ssri_nv8hi = 1762,
- CODE_FOR_aarch64_usri_nv8hi = 1763,
- CODE_FOR_aarch64_ssli_nv2si = 1764,
- CODE_FOR_aarch64_usli_nv2si = 1765,
- CODE_FOR_aarch64_ssri_nv2si = 1766,
- CODE_FOR_aarch64_usri_nv2si = 1767,
- CODE_FOR_aarch64_ssli_nv4si = 1768,
- CODE_FOR_aarch64_usli_nv4si = 1769,
- CODE_FOR_aarch64_ssri_nv4si = 1770,
- CODE_FOR_aarch64_usri_nv4si = 1771,
- CODE_FOR_aarch64_ssli_nv2di = 1772,
- CODE_FOR_aarch64_usli_nv2di = 1773,
- CODE_FOR_aarch64_ssri_nv2di = 1774,
- CODE_FOR_aarch64_usri_nv2di = 1775,
- CODE_FOR_aarch64_ssli_ndi = 1776,
- CODE_FOR_aarch64_usli_ndi = 1777,
- CODE_FOR_aarch64_ssri_ndi = 1778,
- CODE_FOR_aarch64_usri_ndi = 1779,
- CODE_FOR_aarch64_sqshlu_nv8qi = 1780,
- CODE_FOR_aarch64_sqshl_nv8qi = 1781,
- CODE_FOR_aarch64_uqshl_nv8qi = 1782,
- CODE_FOR_aarch64_sqshlu_nv16qi = 1783,
- CODE_FOR_aarch64_sqshl_nv16qi = 1784,
- CODE_FOR_aarch64_uqshl_nv16qi = 1785,
- CODE_FOR_aarch64_sqshlu_nv4hi = 1786,
- CODE_FOR_aarch64_sqshl_nv4hi = 1787,
- CODE_FOR_aarch64_uqshl_nv4hi = 1788,
- CODE_FOR_aarch64_sqshlu_nv8hi = 1789,
- CODE_FOR_aarch64_sqshl_nv8hi = 1790,
- CODE_FOR_aarch64_uqshl_nv8hi = 1791,
- CODE_FOR_aarch64_sqshlu_nv2si = 1792,
- CODE_FOR_aarch64_sqshl_nv2si = 1793,
- CODE_FOR_aarch64_uqshl_nv2si = 1794,
- CODE_FOR_aarch64_sqshlu_nv4si = 1795,
- CODE_FOR_aarch64_sqshl_nv4si = 1796,
- CODE_FOR_aarch64_uqshl_nv4si = 1797,
- CODE_FOR_aarch64_sqshlu_nv2di = 1798,
- CODE_FOR_aarch64_sqshl_nv2di = 1799,
- CODE_FOR_aarch64_uqshl_nv2di = 1800,
- CODE_FOR_aarch64_sqshlu_nqi = 1801,
- CODE_FOR_aarch64_sqshl_nqi = 1802,
- CODE_FOR_aarch64_uqshl_nqi = 1803,
- CODE_FOR_aarch64_sqshlu_nhi = 1804,
- CODE_FOR_aarch64_sqshl_nhi = 1805,
- CODE_FOR_aarch64_uqshl_nhi = 1806,
- CODE_FOR_aarch64_sqshlu_nsi = 1807,
- CODE_FOR_aarch64_sqshl_nsi = 1808,
- CODE_FOR_aarch64_uqshl_nsi = 1809,
- CODE_FOR_aarch64_sqshlu_ndi = 1810,
- CODE_FOR_aarch64_sqshl_ndi = 1811,
- CODE_FOR_aarch64_uqshl_ndi = 1812,
- CODE_FOR_aarch64_sqshrun_nv8hi = 1813,
- CODE_FOR_aarch64_sqrshrun_nv8hi = 1814,
- CODE_FOR_aarch64_sqshrn_nv8hi = 1815,
- CODE_FOR_aarch64_uqshrn_nv8hi = 1816,
- CODE_FOR_aarch64_sqrshrn_nv8hi = 1817,
- CODE_FOR_aarch64_uqrshrn_nv8hi = 1818,
- CODE_FOR_aarch64_sqshrun_nv4si = 1819,
- CODE_FOR_aarch64_sqrshrun_nv4si = 1820,
- CODE_FOR_aarch64_sqshrn_nv4si = 1821,
- CODE_FOR_aarch64_uqshrn_nv4si = 1822,
- CODE_FOR_aarch64_sqrshrn_nv4si = 1823,
- CODE_FOR_aarch64_uqrshrn_nv4si = 1824,
- CODE_FOR_aarch64_sqshrun_nv2di = 1825,
- CODE_FOR_aarch64_sqrshrun_nv2di = 1826,
- CODE_FOR_aarch64_sqshrn_nv2di = 1827,
- CODE_FOR_aarch64_uqshrn_nv2di = 1828,
- CODE_FOR_aarch64_sqrshrn_nv2di = 1829,
- CODE_FOR_aarch64_uqrshrn_nv2di = 1830,
- CODE_FOR_aarch64_sqshrun_nhi = 1831,
- CODE_FOR_aarch64_sqrshrun_nhi = 1832,
- CODE_FOR_aarch64_sqshrn_nhi = 1833,
- CODE_FOR_aarch64_uqshrn_nhi = 1834,
- CODE_FOR_aarch64_sqrshrn_nhi = 1835,
- CODE_FOR_aarch64_uqrshrn_nhi = 1836,
- CODE_FOR_aarch64_sqshrun_nsi = 1837,
- CODE_FOR_aarch64_sqrshrun_nsi = 1838,
- CODE_FOR_aarch64_sqshrn_nsi = 1839,
- CODE_FOR_aarch64_uqshrn_nsi = 1840,
- CODE_FOR_aarch64_sqrshrn_nsi = 1841,
- CODE_FOR_aarch64_uqrshrn_nsi = 1842,
- CODE_FOR_aarch64_sqshrun_ndi = 1843,
- CODE_FOR_aarch64_sqrshrun_ndi = 1844,
- CODE_FOR_aarch64_sqshrn_ndi = 1845,
- CODE_FOR_aarch64_uqshrn_ndi = 1846,
- CODE_FOR_aarch64_sqrshrn_ndi = 1847,
- CODE_FOR_aarch64_uqrshrn_ndi = 1848,
- CODE_FOR_aarch64_cmltv8qi = 1849,
- CODE_FOR_aarch64_cmlev8qi = 1850,
- CODE_FOR_aarch64_cmeqv8qi = 1851,
- CODE_FOR_aarch64_cmgev8qi = 1852,
- CODE_FOR_aarch64_cmgtv8qi = 1853,
- CODE_FOR_aarch64_cmltv16qi = 1854,
- CODE_FOR_aarch64_cmlev16qi = 1855,
- CODE_FOR_aarch64_cmeqv16qi = 1856,
- CODE_FOR_aarch64_cmgev16qi = 1857,
- CODE_FOR_aarch64_cmgtv16qi = 1858,
- CODE_FOR_aarch64_cmltv4hi = 1859,
- CODE_FOR_aarch64_cmlev4hi = 1860,
- CODE_FOR_aarch64_cmeqv4hi = 1861,
- CODE_FOR_aarch64_cmgev4hi = 1862,
- CODE_FOR_aarch64_cmgtv4hi = 1863,
- CODE_FOR_aarch64_cmltv8hi = 1864,
- CODE_FOR_aarch64_cmlev8hi = 1865,
- CODE_FOR_aarch64_cmeqv8hi = 1866,
- CODE_FOR_aarch64_cmgev8hi = 1867,
- CODE_FOR_aarch64_cmgtv8hi = 1868,
- CODE_FOR_aarch64_cmltv2si = 1869,
- CODE_FOR_aarch64_cmlev2si = 1870,
- CODE_FOR_aarch64_cmeqv2si = 1871,
- CODE_FOR_aarch64_cmgev2si = 1872,
- CODE_FOR_aarch64_cmgtv2si = 1873,
- CODE_FOR_aarch64_cmltv4si = 1874,
- CODE_FOR_aarch64_cmlev4si = 1875,
- CODE_FOR_aarch64_cmeqv4si = 1876,
- CODE_FOR_aarch64_cmgev4si = 1877,
- CODE_FOR_aarch64_cmgtv4si = 1878,
- CODE_FOR_aarch64_cmltv2di = 1879,
- CODE_FOR_aarch64_cmlev2di = 1880,
- CODE_FOR_aarch64_cmeqv2di = 1881,
- CODE_FOR_aarch64_cmgev2di = 1882,
- CODE_FOR_aarch64_cmgtv2di = 1883,
- CODE_FOR_aarch64_cmltdi = 1884,
- CODE_FOR_aarch64_cmledi = 1885,
- CODE_FOR_aarch64_cmeqdi = 1886,
- CODE_FOR_aarch64_cmgedi = 1887,
- CODE_FOR_aarch64_cmgtdi = 1888,
- CODE_FOR_aarch64_cmltuv8qi = 1894,
- CODE_FOR_aarch64_cmleuv8qi = 1895,
- CODE_FOR_aarch64_cmgeuv8qi = 1896,
- CODE_FOR_aarch64_cmgtuv8qi = 1897,
- CODE_FOR_aarch64_cmltuv16qi = 1898,
- CODE_FOR_aarch64_cmleuv16qi = 1899,
- CODE_FOR_aarch64_cmgeuv16qi = 1900,
- CODE_FOR_aarch64_cmgtuv16qi = 1901,
- CODE_FOR_aarch64_cmltuv4hi = 1902,
- CODE_FOR_aarch64_cmleuv4hi = 1903,
- CODE_FOR_aarch64_cmgeuv4hi = 1904,
- CODE_FOR_aarch64_cmgtuv4hi = 1905,
- CODE_FOR_aarch64_cmltuv8hi = 1906,
- CODE_FOR_aarch64_cmleuv8hi = 1907,
- CODE_FOR_aarch64_cmgeuv8hi = 1908,
- CODE_FOR_aarch64_cmgtuv8hi = 1909,
- CODE_FOR_aarch64_cmltuv2si = 1910,
- CODE_FOR_aarch64_cmleuv2si = 1911,
- CODE_FOR_aarch64_cmgeuv2si = 1912,
- CODE_FOR_aarch64_cmgtuv2si = 1913,
- CODE_FOR_aarch64_cmltuv4si = 1914,
- CODE_FOR_aarch64_cmleuv4si = 1915,
- CODE_FOR_aarch64_cmgeuv4si = 1916,
- CODE_FOR_aarch64_cmgtuv4si = 1917,
- CODE_FOR_aarch64_cmltuv2di = 1918,
- CODE_FOR_aarch64_cmleuv2di = 1919,
- CODE_FOR_aarch64_cmgeuv2di = 1920,
- CODE_FOR_aarch64_cmgtuv2di = 1921,
- CODE_FOR_aarch64_cmltudi = 1922,
- CODE_FOR_aarch64_cmleudi = 1923,
- CODE_FOR_aarch64_cmgeudi = 1924,
- CODE_FOR_aarch64_cmgtudi = 1925,
- CODE_FOR_aarch64_cmtstv8qi = 1930,
- CODE_FOR_aarch64_cmtstv16qi = 1931,
- CODE_FOR_aarch64_cmtstv4hi = 1932,
- CODE_FOR_aarch64_cmtstv8hi = 1933,
- CODE_FOR_aarch64_cmtstv2si = 1934,
- CODE_FOR_aarch64_cmtstv4si = 1935,
- CODE_FOR_aarch64_cmtstv2di = 1936,
- CODE_FOR_aarch64_cmtstdi = 1937,
- CODE_FOR_aarch64_cmltv2sf = 1939,
- CODE_FOR_aarch64_cmlev2sf = 1940,
- CODE_FOR_aarch64_cmeqv2sf = 1941,
- CODE_FOR_aarch64_cmgev2sf = 1942,
- CODE_FOR_aarch64_cmgtv2sf = 1943,
- CODE_FOR_aarch64_cmltv4sf = 1944,
- CODE_FOR_aarch64_cmlev4sf = 1945,
- CODE_FOR_aarch64_cmeqv4sf = 1946,
- CODE_FOR_aarch64_cmgev4sf = 1947,
- CODE_FOR_aarch64_cmgtv4sf = 1948,
- CODE_FOR_aarch64_cmltv2df = 1949,
- CODE_FOR_aarch64_cmlev2df = 1950,
- CODE_FOR_aarch64_cmeqv2df = 1951,
- CODE_FOR_aarch64_cmgev2df = 1952,
- CODE_FOR_aarch64_cmgtv2df = 1953,
- CODE_FOR_aarch64_cmltsf = 1954,
- CODE_FOR_aarch64_cmlesf = 1955,
- CODE_FOR_aarch64_cmeqsf = 1956,
- CODE_FOR_aarch64_cmgesf = 1957,
- CODE_FOR_aarch64_cmgtsf = 1958,
- CODE_FOR_aarch64_cmltdf = 1959,
- CODE_FOR_aarch64_cmledf = 1960,
- CODE_FOR_aarch64_cmeqdf = 1961,
- CODE_FOR_aarch64_cmgedf = 1962,
- CODE_FOR_aarch64_cmgtdf = 1963,
- CODE_FOR_aarch64_addpv8qi = 1984,
- CODE_FOR_aarch64_addpv4hi = 1985,
- CODE_FOR_aarch64_addpv2si = 1986,
- CODE_FOR_aarch64_addpdi = 1987,
- CODE_FOR_sqrtv2sf2 = 1988,
- CODE_FOR_sqrtv4sf2 = 1989,
- CODE_FOR_sqrtv2df2 = 1990,
- CODE_FOR_vec_load_lanesoiv16qi = 1991,
- CODE_FOR_vec_load_lanesoiv8hi = 1992,
- CODE_FOR_vec_load_lanesoiv4si = 1993,
- CODE_FOR_vec_load_lanesoiv2di = 1994,
- CODE_FOR_vec_load_lanesoiv4sf = 1995,
- CODE_FOR_vec_load_lanesoiv2df = 1996,
- CODE_FOR_vec_store_lanesoiv16qi = 1997,
- CODE_FOR_vec_store_lanesoiv8hi = 1998,
- CODE_FOR_vec_store_lanesoiv4si = 1999,
- CODE_FOR_vec_store_lanesoiv2di = 2000,
- CODE_FOR_vec_store_lanesoiv4sf = 2001,
- CODE_FOR_vec_store_lanesoiv2df = 2002,
- CODE_FOR_vec_store_lanesoi_lanev16qi = 2003,
- CODE_FOR_vec_store_lanesoi_lanev8hi = 2004,
- CODE_FOR_vec_store_lanesoi_lanev4si = 2005,
- CODE_FOR_vec_store_lanesoi_lanev2di = 2006,
- CODE_FOR_vec_store_lanesoi_lanev4sf = 2007,
- CODE_FOR_vec_store_lanesoi_lanev2df = 2008,
- CODE_FOR_vec_load_lanesciv16qi = 2009,
- CODE_FOR_vec_load_lanesciv8hi = 2010,
- CODE_FOR_vec_load_lanesciv4si = 2011,
- CODE_FOR_vec_load_lanesciv2di = 2012,
- CODE_FOR_vec_load_lanesciv4sf = 2013,
- CODE_FOR_vec_load_lanesciv2df = 2014,
- CODE_FOR_vec_store_lanesciv16qi = 2015,
- CODE_FOR_vec_store_lanesciv8hi = 2016,
- CODE_FOR_vec_store_lanesciv4si = 2017,
- CODE_FOR_vec_store_lanesciv2di = 2018,
- CODE_FOR_vec_store_lanesciv4sf = 2019,
- CODE_FOR_vec_store_lanesciv2df = 2020,
- CODE_FOR_vec_store_lanesci_lanev16qi = 2021,
- CODE_FOR_vec_store_lanesci_lanev8hi = 2022,
- CODE_FOR_vec_store_lanesci_lanev4si = 2023,
- CODE_FOR_vec_store_lanesci_lanev2di = 2024,
- CODE_FOR_vec_store_lanesci_lanev4sf = 2025,
- CODE_FOR_vec_store_lanesci_lanev2df = 2026,
- CODE_FOR_vec_load_lanesxiv16qi = 2027,
- CODE_FOR_vec_load_lanesxiv8hi = 2028,
- CODE_FOR_vec_load_lanesxiv4si = 2029,
- CODE_FOR_vec_load_lanesxiv2di = 2030,
- CODE_FOR_vec_load_lanesxiv4sf = 2031,
- CODE_FOR_vec_load_lanesxiv2df = 2032,
- CODE_FOR_vec_store_lanesxiv16qi = 2033,
- CODE_FOR_vec_store_lanesxiv8hi = 2034,
- CODE_FOR_vec_store_lanesxiv4si = 2035,
- CODE_FOR_vec_store_lanesxiv2di = 2036,
- CODE_FOR_vec_store_lanesxiv4sf = 2037,
- CODE_FOR_vec_store_lanesxiv2df = 2038,
- CODE_FOR_vec_store_lanesxi_lanev16qi = 2039,
- CODE_FOR_vec_store_lanesxi_lanev8hi = 2040,
- CODE_FOR_vec_store_lanesxi_lanev4si = 2041,
- CODE_FOR_vec_store_lanesxi_lanev2di = 2042,
- CODE_FOR_vec_store_lanesxi_lanev4sf = 2043,
- CODE_FOR_vec_store_lanesxi_lanev2df = 2044,
- CODE_FOR_aarch64_be_ld1v8qi = 2048,
- CODE_FOR_aarch64_be_ld1v16qi = 2049,
- CODE_FOR_aarch64_be_ld1v4hi = 2050,
- CODE_FOR_aarch64_be_ld1v8hi = 2051,
- CODE_FOR_aarch64_be_ld1v2si = 2052,
- CODE_FOR_aarch64_be_ld1v4si = 2053,
- CODE_FOR_aarch64_be_ld1v2di = 2054,
- CODE_FOR_aarch64_be_ld1v2sf = 2055,
- CODE_FOR_aarch64_be_ld1v4sf = 2056,
- CODE_FOR_aarch64_be_ld1v2df = 2057,
- CODE_FOR_aarch64_be_ld1di = 2058,
- CODE_FOR_aarch64_be_st1v8qi = 2059,
- CODE_FOR_aarch64_be_st1v16qi = 2060,
- CODE_FOR_aarch64_be_st1v4hi = 2061,
- CODE_FOR_aarch64_be_st1v8hi = 2062,
- CODE_FOR_aarch64_be_st1v2si = 2063,
- CODE_FOR_aarch64_be_st1v4si = 2064,
- CODE_FOR_aarch64_be_st1v2di = 2065,
- CODE_FOR_aarch64_be_st1v2sf = 2066,
- CODE_FOR_aarch64_be_st1v4sf = 2067,
- CODE_FOR_aarch64_be_st1v2df = 2068,
- CODE_FOR_aarch64_be_st1di = 2069,
- CODE_FOR_aarch64_ld2v8qi_dreg = 2070,
- CODE_FOR_aarch64_ld2v4hi_dreg = 2071,
- CODE_FOR_aarch64_ld2v2si_dreg = 2072,
- CODE_FOR_aarch64_ld2v2sf_dreg = 2073,
- CODE_FOR_aarch64_ld2di_dreg = 2074,
- CODE_FOR_aarch64_ld2df_dreg = 2075,
- CODE_FOR_aarch64_ld3v8qi_dreg = 2076,
- CODE_FOR_aarch64_ld3v4hi_dreg = 2077,
- CODE_FOR_aarch64_ld3v2si_dreg = 2078,
- CODE_FOR_aarch64_ld3v2sf_dreg = 2079,
- CODE_FOR_aarch64_ld3di_dreg = 2080,
- CODE_FOR_aarch64_ld3df_dreg = 2081,
- CODE_FOR_aarch64_ld4v8qi_dreg = 2082,
- CODE_FOR_aarch64_ld4v4hi_dreg = 2083,
- CODE_FOR_aarch64_ld4v2si_dreg = 2084,
- CODE_FOR_aarch64_ld4v2sf_dreg = 2085,
- CODE_FOR_aarch64_ld4di_dreg = 2086,
- CODE_FOR_aarch64_ld4df_dreg = 2087,
- CODE_FOR_aarch64_tbl1v8qi = 2088,
- CODE_FOR_aarch64_tbl1v16qi = 2089,
- CODE_FOR_aarch64_tbl2v16qi = 2090,
- CODE_FOR_aarch64_combinev16qi = 2091,
- CODE_FOR_aarch64_zip1v8qi = 2092,
- CODE_FOR_aarch64_zip2v8qi = 2093,
- CODE_FOR_aarch64_trn1v8qi = 2094,
- CODE_FOR_aarch64_trn2v8qi = 2095,
- CODE_FOR_aarch64_uzp1v8qi = 2096,
- CODE_FOR_aarch64_uzp2v8qi = 2097,
- CODE_FOR_aarch64_zip1v16qi = 2098,
- CODE_FOR_aarch64_zip2v16qi = 2099,
- CODE_FOR_aarch64_trn1v16qi = 2100,
- CODE_FOR_aarch64_trn2v16qi = 2101,
- CODE_FOR_aarch64_uzp1v16qi = 2102,
- CODE_FOR_aarch64_uzp2v16qi = 2103,
- CODE_FOR_aarch64_zip1v4hi = 2104,
- CODE_FOR_aarch64_zip2v4hi = 2105,
- CODE_FOR_aarch64_trn1v4hi = 2106,
- CODE_FOR_aarch64_trn2v4hi = 2107,
- CODE_FOR_aarch64_uzp1v4hi = 2108,
- CODE_FOR_aarch64_uzp2v4hi = 2109,
- CODE_FOR_aarch64_zip1v8hi = 2110,
- CODE_FOR_aarch64_zip2v8hi = 2111,
- CODE_FOR_aarch64_trn1v8hi = 2112,
- CODE_FOR_aarch64_trn2v8hi = 2113,
- CODE_FOR_aarch64_uzp1v8hi = 2114,
- CODE_FOR_aarch64_uzp2v8hi = 2115,
- CODE_FOR_aarch64_zip1v2si = 2116,
- CODE_FOR_aarch64_zip2v2si = 2117,
- CODE_FOR_aarch64_trn1v2si = 2118,
- CODE_FOR_aarch64_trn2v2si = 2119,
- CODE_FOR_aarch64_uzp1v2si = 2120,
- CODE_FOR_aarch64_uzp2v2si = 2121,
- CODE_FOR_aarch64_zip1v4si = 2122,
- CODE_FOR_aarch64_zip2v4si = 2123,
- CODE_FOR_aarch64_trn1v4si = 2124,
- CODE_FOR_aarch64_trn2v4si = 2125,
- CODE_FOR_aarch64_uzp1v4si = 2126,
- CODE_FOR_aarch64_uzp2v4si = 2127,
- CODE_FOR_aarch64_zip1v2di = 2128,
- CODE_FOR_aarch64_zip2v2di = 2129,
- CODE_FOR_aarch64_trn1v2di = 2130,
- CODE_FOR_aarch64_trn2v2di = 2131,
- CODE_FOR_aarch64_uzp1v2di = 2132,
- CODE_FOR_aarch64_uzp2v2di = 2133,
- CODE_FOR_aarch64_zip1v2sf = 2134,
- CODE_FOR_aarch64_zip2v2sf = 2135,
- CODE_FOR_aarch64_trn1v2sf = 2136,
- CODE_FOR_aarch64_trn2v2sf = 2137,
- CODE_FOR_aarch64_uzp1v2sf = 2138,
- CODE_FOR_aarch64_uzp2v2sf = 2139,
- CODE_FOR_aarch64_zip1v4sf = 2140,
- CODE_FOR_aarch64_zip2v4sf = 2141,
- CODE_FOR_aarch64_trn1v4sf = 2142,
- CODE_FOR_aarch64_trn2v4sf = 2143,
- CODE_FOR_aarch64_uzp1v4sf = 2144,
- CODE_FOR_aarch64_uzp2v4sf = 2145,
- CODE_FOR_aarch64_zip1v2df = 2146,
- CODE_FOR_aarch64_zip2v2df = 2147,
- CODE_FOR_aarch64_trn1v2df = 2148,
- CODE_FOR_aarch64_trn2v2df = 2149,
- CODE_FOR_aarch64_uzp1v2df = 2150,
- CODE_FOR_aarch64_uzp2v2df = 2151,
- CODE_FOR_aarch64_st2v8qi_dreg = 2152,
- CODE_FOR_aarch64_st2v4hi_dreg = 2153,
- CODE_FOR_aarch64_st2v2si_dreg = 2154,
- CODE_FOR_aarch64_st2v2sf_dreg = 2155,
- CODE_FOR_aarch64_st2di_dreg = 2156,
- CODE_FOR_aarch64_st2df_dreg = 2157,
- CODE_FOR_aarch64_st3v8qi_dreg = 2158,
- CODE_FOR_aarch64_st3v4hi_dreg = 2159,
- CODE_FOR_aarch64_st3v2si_dreg = 2160,
- CODE_FOR_aarch64_st3v2sf_dreg = 2161,
- CODE_FOR_aarch64_st3di_dreg = 2162,
- CODE_FOR_aarch64_st3df_dreg = 2163,
- CODE_FOR_aarch64_st4v8qi_dreg = 2164,
- CODE_FOR_aarch64_st4v4hi_dreg = 2165,
- CODE_FOR_aarch64_st4v2si_dreg = 2166,
- CODE_FOR_aarch64_st4v2sf_dreg = 2167,
- CODE_FOR_aarch64_st4di_dreg = 2168,
- CODE_FOR_aarch64_st4df_dreg = 2169,
- CODE_FOR_aarch64_frecpev2sf = 2181,
- CODE_FOR_aarch64_frecpev4sf = 2182,
- CODE_FOR_aarch64_frecpev2df = 2183,
- CODE_FOR_aarch64_frecpesf = 2184,
- CODE_FOR_aarch64_frecpxsf = 2185,
- CODE_FOR_aarch64_frecpedf = 2186,
- CODE_FOR_aarch64_frecpxdf = 2187,
- CODE_FOR_aarch64_frecpsv2sf = 2188,
- CODE_FOR_aarch64_frecpsv4sf = 2189,
- CODE_FOR_aarch64_frecpsv2df = 2190,
- CODE_FOR_aarch64_frecpssf = 2191,
- CODE_FOR_aarch64_frecpsdf = 2192,
- CODE_FOR_aarch64_crypto_aesev16qi = 2193,
- CODE_FOR_aarch64_crypto_aesdv16qi = 2194,
- CODE_FOR_aarch64_crypto_aesmcv16qi = 2195,
- CODE_FOR_aarch64_crypto_aesimcv16qi = 2196,
- CODE_FOR_aarch64_crypto_sha1hsi = 2197,
- CODE_FOR_aarch64_crypto_sha1su1v4si = 2198,
- CODE_FOR_aarch64_crypto_sha1cv4si = 2199,
- CODE_FOR_aarch64_crypto_sha1mv4si = 2200,
- CODE_FOR_aarch64_crypto_sha1pv4si = 2201,
- CODE_FOR_aarch64_crypto_sha1su0v4si = 2202,
- CODE_FOR_aarch64_crypto_sha256hv4si = 2203,
- CODE_FOR_aarch64_crypto_sha256h2v4si = 2204,
- CODE_FOR_aarch64_crypto_sha256su0v4si = 2205,
- CODE_FOR_aarch64_crypto_sha256su1v4si = 2206,
- CODE_FOR_aarch64_crypto_pmulldi = 2207,
- CODE_FOR_aarch64_crypto_pmullv2di = 2208,
- CODE_FOR_atomic_compare_and_swapqi_1 = 2209,
- CODE_FOR_atomic_compare_and_swaphi_1 = 2210,
- CODE_FOR_atomic_compare_and_swapsi_1 = 2211,
- CODE_FOR_atomic_compare_and_swapdi_1 = 2212,
- CODE_FOR_atomic_exchangeqi = 2213,
- CODE_FOR_atomic_exchangehi = 2214,
- CODE_FOR_atomic_exchangesi = 2215,
- CODE_FOR_atomic_exchangedi = 2216,
- CODE_FOR_atomic_addqi = 2217,
- CODE_FOR_atomic_subqi = 2218,
- CODE_FOR_atomic_orqi = 2219,
- CODE_FOR_atomic_xorqi = 2220,
- CODE_FOR_atomic_andqi = 2221,
- CODE_FOR_atomic_addhi = 2222,
- CODE_FOR_atomic_subhi = 2223,
- CODE_FOR_atomic_orhi = 2224,
- CODE_FOR_atomic_xorhi = 2225,
- CODE_FOR_atomic_andhi = 2226,
- CODE_FOR_atomic_addsi = 2227,
- CODE_FOR_atomic_subsi = 2228,
- CODE_FOR_atomic_orsi = 2229,
- CODE_FOR_atomic_xorsi = 2230,
- CODE_FOR_atomic_andsi = 2231,
- CODE_FOR_atomic_adddi = 2232,
- CODE_FOR_atomic_subdi = 2233,
- CODE_FOR_atomic_ordi = 2234,
- CODE_FOR_atomic_xordi = 2235,
- CODE_FOR_atomic_anddi = 2236,
- CODE_FOR_atomic_nandqi = 2237,
- CODE_FOR_atomic_nandhi = 2238,
- CODE_FOR_atomic_nandsi = 2239,
- CODE_FOR_atomic_nanddi = 2240,
- CODE_FOR_atomic_fetch_addqi = 2241,
- CODE_FOR_atomic_fetch_subqi = 2242,
- CODE_FOR_atomic_fetch_orqi = 2243,
- CODE_FOR_atomic_fetch_xorqi = 2244,
- CODE_FOR_atomic_fetch_andqi = 2245,
- CODE_FOR_atomic_fetch_addhi = 2246,
- CODE_FOR_atomic_fetch_subhi = 2247,
- CODE_FOR_atomic_fetch_orhi = 2248,
- CODE_FOR_atomic_fetch_xorhi = 2249,
- CODE_FOR_atomic_fetch_andhi = 2250,
- CODE_FOR_atomic_fetch_addsi = 2251,
- CODE_FOR_atomic_fetch_subsi = 2252,
- CODE_FOR_atomic_fetch_orsi = 2253,
- CODE_FOR_atomic_fetch_xorsi = 2254,
- CODE_FOR_atomic_fetch_andsi = 2255,
- CODE_FOR_atomic_fetch_adddi = 2256,
- CODE_FOR_atomic_fetch_subdi = 2257,
- CODE_FOR_atomic_fetch_ordi = 2258,
- CODE_FOR_atomic_fetch_xordi = 2259,
- CODE_FOR_atomic_fetch_anddi = 2260,
- CODE_FOR_atomic_fetch_nandqi = 2261,
- CODE_FOR_atomic_fetch_nandhi = 2262,
- CODE_FOR_atomic_fetch_nandsi = 2263,
- CODE_FOR_atomic_fetch_nanddi = 2264,
- CODE_FOR_atomic_add_fetchqi = 2265,
- CODE_FOR_atomic_sub_fetchqi = 2266,
- CODE_FOR_atomic_or_fetchqi = 2267,
- CODE_FOR_atomic_xor_fetchqi = 2268,
- CODE_FOR_atomic_and_fetchqi = 2269,
- CODE_FOR_atomic_add_fetchhi = 2270,
- CODE_FOR_atomic_sub_fetchhi = 2271,
- CODE_FOR_atomic_or_fetchhi = 2272,
- CODE_FOR_atomic_xor_fetchhi = 2273,
- CODE_FOR_atomic_and_fetchhi = 2274,
- CODE_FOR_atomic_add_fetchsi = 2275,
- CODE_FOR_atomic_sub_fetchsi = 2276,
- CODE_FOR_atomic_or_fetchsi = 2277,
- CODE_FOR_atomic_xor_fetchsi = 2278,
- CODE_FOR_atomic_and_fetchsi = 2279,
- CODE_FOR_atomic_add_fetchdi = 2280,
- CODE_FOR_atomic_sub_fetchdi = 2281,
- CODE_FOR_atomic_or_fetchdi = 2282,
- CODE_FOR_atomic_xor_fetchdi = 2283,
- CODE_FOR_atomic_and_fetchdi = 2284,
- CODE_FOR_atomic_nand_fetchqi = 2285,
- CODE_FOR_atomic_nand_fetchhi = 2286,
- CODE_FOR_atomic_nand_fetchsi = 2287,
- CODE_FOR_atomic_nand_fetchdi = 2288,
- CODE_FOR_atomic_loadqi = 2289,
- CODE_FOR_atomic_loadhi = 2290,
- CODE_FOR_atomic_loadsi = 2291,
- CODE_FOR_atomic_loaddi = 2292,
- CODE_FOR_atomic_storeqi = 2293,
- CODE_FOR_atomic_storehi = 2294,
- CODE_FOR_atomic_storesi = 2295,
- CODE_FOR_atomic_storedi = 2296,
- CODE_FOR_aarch64_load_exclusiveqi = 2297,
- CODE_FOR_aarch64_load_exclusivehi = 2298,
- CODE_FOR_aarch64_load_exclusivesi = 2299,
- CODE_FOR_aarch64_load_exclusivedi = 2300,
- CODE_FOR_aarch64_store_exclusiveqi = 2301,
- CODE_FOR_aarch64_store_exclusivehi = 2302,
- CODE_FOR_aarch64_store_exclusivesi = 2303,
- CODE_FOR_aarch64_store_exclusivedi = 2304,
- CODE_FOR_cbranchsi4 = 2306,
- CODE_FOR_cbranchdi4 = 2307,
- CODE_FOR_cbranchsf4 = 2308,
- CODE_FOR_cbranchdf4 = 2309,
- CODE_FOR_casesi = 2310,
- CODE_FOR_prologue = 2311,
- CODE_FOR_epilogue = 2312,
- CODE_FOR_sibcall_epilogue = 2313,
- CODE_FOR_call = 2315,
- CODE_FOR_call_value = 2316,
- CODE_FOR_sibcall = 2317,
- CODE_FOR_sibcall_value = 2318,
- CODE_FOR_untyped_call = 2319,
- CODE_FOR_movqi = 2320,
- CODE_FOR_movhi = 2321,
- CODE_FOR_movsi = 2322,
- CODE_FOR_movdi = 2323,
- CODE_FOR_movti = 2324,
- CODE_FOR_movsf = 2326,
- CODE_FOR_movdf = 2327,
- CODE_FOR_movtf = 2328,
- CODE_FOR_extendsidi2 = 2330,
- CODE_FOR_zero_extendsidi2 = 2331,
- CODE_FOR_extendqisi2 = 2332,
- CODE_FOR_zero_extendqisi2 = 2333,
- CODE_FOR_extendqidi2 = 2334,
- CODE_FOR_zero_extendqidi2 = 2335,
- CODE_FOR_extendhisi2 = 2336,
- CODE_FOR_zero_extendhisi2 = 2337,
- CODE_FOR_extendhidi2 = 2338,
- CODE_FOR_zero_extendhidi2 = 2339,
- CODE_FOR_extendqihi2 = 2340,
- CODE_FOR_zero_extendqihi2 = 2341,
- CODE_FOR_addsi3 = 2342,
- CODE_FOR_adddi3 = 2343,
- CODE_FOR_addti3 = 2344,
- CODE_FOR_subti3 = 2345,
- CODE_FOR_mulditi3 = 2347,
- CODE_FOR_umulditi3 = 2348,
- CODE_FOR_multi3 = 2349,
- CODE_FOR_cstoresi4 = 2350,
- CODE_FOR_cstoredi4 = 2351,
- CODE_FOR_cstoresf4 = 2352,
- CODE_FOR_cstoredf4 = 2353,
- CODE_FOR_cmovsi6 = 2354,
- CODE_FOR_cmovdi6 = 2355,
- CODE_FOR_cmovsf6 = 2356,
- CODE_FOR_cmovdf6 = 2357,
- CODE_FOR_movqicc = 2358,
- CODE_FOR_movhicc = 2359,
- CODE_FOR_movsicc = 2360,
- CODE_FOR_movdicc = 2361,
- CODE_FOR_movsfsicc = 2362,
- CODE_FOR_movdfsicc = 2363,
- CODE_FOR_movsfdicc = 2364,
- CODE_FOR_movdfdicc = 2365,
- CODE_FOR_movsfcc = 2366,
- CODE_FOR_movdfcc = 2367,
- CODE_FOR_ffssi2 = 2368,
- CODE_FOR_ffsdi2 = 2369,
- CODE_FOR_ctzsi2 = 2370,
- CODE_FOR_ctzdi2 = 2371,
- CODE_FOR_ashlsi3 = 2372,
- CODE_FOR_ashrsi3 = 2373,
- CODE_FOR_lshrsi3 = 2374,
- CODE_FOR_ashldi3 = 2375,
- CODE_FOR_ashrdi3 = 2376,
- CODE_FOR_lshrdi3 = 2377,
- CODE_FOR_ashlqi3 = 2378,
- CODE_FOR_ashlhi3 = 2379,
- CODE_FOR_rotrsi3 = 2380,
- CODE_FOR_rotrdi3 = 2381,
- CODE_FOR_rotlsi3 = 2382,
- CODE_FOR_rotldi3 = 2383,
- CODE_FOR_extv = 2388,
- CODE_FOR_extzv = 2389,
- CODE_FOR_insvsi = 2390,
- CODE_FOR_insvdi = 2391,
- CODE_FOR_aarch64_reload_movti = 2392,
- CODE_FOR_aarch64_reload_movtf = 2393,
- CODE_FOR_add_losym = 2394,
- CODE_FOR_tlsgd_small = 2395,
- CODE_FOR_tlsle_small = 2396,
- CODE_FOR_get_thread_pointerdi = 2397,
- CODE_FOR_stack_protect_set = 2398,
- CODE_FOR_stack_protect_test = 2399,
- CODE_FOR_movv8qi = 2400,
- CODE_FOR_movv16qi = 2401,
- CODE_FOR_movv4hi = 2402,
- CODE_FOR_movv8hi = 2403,
- CODE_FOR_movv2si = 2404,
- CODE_FOR_movv4si = 2405,
- CODE_FOR_movv2di = 2406,
- CODE_FOR_movv2sf = 2407,
- CODE_FOR_movv4sf = 2408,
- CODE_FOR_movv2df = 2409,
- CODE_FOR_movmisalignv8qi = 2410,
- CODE_FOR_movmisalignv16qi = 2411,
- CODE_FOR_movmisalignv4hi = 2412,
- CODE_FOR_movmisalignv8hi = 2413,
- CODE_FOR_movmisalignv2si = 2414,
- CODE_FOR_movmisalignv4si = 2415,
- CODE_FOR_movmisalignv2di = 2416,
- CODE_FOR_movmisalignv2sf = 2417,
- CODE_FOR_movmisalignv4sf = 2418,
- CODE_FOR_movmisalignv2df = 2419,
- CODE_FOR_aarch64_split_simd_movv16qi = 2432,
- CODE_FOR_aarch64_split_simd_movv8hi = 2433,
- CODE_FOR_aarch64_split_simd_movv4si = 2434,
- CODE_FOR_aarch64_split_simd_movv2di = 2435,
- CODE_FOR_aarch64_split_simd_movv4sf = 2436,
- CODE_FOR_aarch64_split_simd_movv2df = 2437,
- CODE_FOR_ashlv8qi3 = 2438,
- CODE_FOR_ashlv16qi3 = 2439,
- CODE_FOR_ashlv4hi3 = 2440,
- CODE_FOR_ashlv8hi3 = 2441,
- CODE_FOR_ashlv2si3 = 2442,
- CODE_FOR_ashlv4si3 = 2443,
- CODE_FOR_ashlv2di3 = 2444,
- CODE_FOR_lshrv8qi3 = 2445,
- CODE_FOR_lshrv16qi3 = 2446,
- CODE_FOR_lshrv4hi3 = 2447,
- CODE_FOR_lshrv8hi3 = 2448,
- CODE_FOR_lshrv2si3 = 2449,
- CODE_FOR_lshrv4si3 = 2450,
- CODE_FOR_lshrv2di3 = 2451,
- CODE_FOR_ashrv8qi3 = 2452,
- CODE_FOR_ashrv16qi3 = 2453,
- CODE_FOR_ashrv4hi3 = 2454,
- CODE_FOR_ashrv8hi3 = 2455,
- CODE_FOR_ashrv2si3 = 2456,
- CODE_FOR_ashrv4si3 = 2457,
- CODE_FOR_ashrv2di3 = 2458,
- CODE_FOR_vashlv8qi3 = 2459,
- CODE_FOR_vashlv16qi3 = 2460,
- CODE_FOR_vashlv4hi3 = 2461,
- CODE_FOR_vashlv8hi3 = 2462,
- CODE_FOR_vashlv2si3 = 2463,
- CODE_FOR_vashlv4si3 = 2464,
- CODE_FOR_vashlv2di3 = 2465,
- CODE_FOR_vashrv8qi3 = 2466,
- CODE_FOR_vashrv16qi3 = 2467,
- CODE_FOR_vashrv4hi3 = 2468,
- CODE_FOR_vashrv8hi3 = 2469,
- CODE_FOR_vashrv2si3 = 2470,
- CODE_FOR_vashrv4si3 = 2471,
- CODE_FOR_aarch64_ashr_simddi = 2472,
- CODE_FOR_vlshrv8qi3 = 2473,
- CODE_FOR_vlshrv16qi3 = 2474,
- CODE_FOR_vlshrv4hi3 = 2475,
- CODE_FOR_vlshrv8hi3 = 2476,
- CODE_FOR_vlshrv2si3 = 2477,
- CODE_FOR_vlshrv4si3 = 2478,
- CODE_FOR_aarch64_lshr_simddi = 2479,
- CODE_FOR_vec_setv8qi = 2480,
- CODE_FOR_vec_setv16qi = 2481,
- CODE_FOR_vec_setv4hi = 2482,
- CODE_FOR_vec_setv8hi = 2483,
- CODE_FOR_vec_setv2si = 2484,
- CODE_FOR_vec_setv4si = 2485,
- CODE_FOR_vec_setv2di = 2486,
- CODE_FOR_vec_setv2sf = 2487,
- CODE_FOR_vec_setv4sf = 2488,
- CODE_FOR_vec_setv2df = 2489,
- CODE_FOR_move_hi_quad_v16qi = 2490,
- CODE_FOR_move_hi_quad_v8hi = 2491,
- CODE_FOR_move_hi_quad_v4si = 2492,
- CODE_FOR_move_hi_quad_v2di = 2493,
- CODE_FOR_move_hi_quad_v4sf = 2494,
- CODE_FOR_move_hi_quad_v2df = 2495,
- CODE_FOR_vec_pack_trunc_v4hi = 2496,
- CODE_FOR_vec_pack_trunc_v2si = 2497,
- CODE_FOR_vec_pack_trunc_di = 2498,
- CODE_FOR_vec_unpacks_hi_v16qi = 2499,
- CODE_FOR_vec_unpacku_hi_v16qi = 2500,
- CODE_FOR_vec_unpacks_hi_v8hi = 2501,
- CODE_FOR_vec_unpacku_hi_v8hi = 2502,
- CODE_FOR_vec_unpacks_hi_v4si = 2503,
- CODE_FOR_vec_unpacku_hi_v4si = 2504,
- CODE_FOR_vec_unpacks_lo_v16qi = 2505,
- CODE_FOR_vec_unpacku_lo_v16qi = 2506,
- CODE_FOR_vec_unpacks_lo_v8hi = 2507,
- CODE_FOR_vec_unpacku_lo_v8hi = 2508,
- CODE_FOR_vec_unpacks_lo_v4si = 2509,
- CODE_FOR_vec_unpacku_lo_v4si = 2510,
- CODE_FOR_vec_widen_smult_lo_v16qi = 2511,
- CODE_FOR_vec_widen_umult_lo_v16qi = 2512,
- CODE_FOR_vec_widen_smult_lo_v8hi = 2513,
- CODE_FOR_vec_widen_umult_lo_v8hi = 2514,
- CODE_FOR_vec_widen_smult_lo_v4si = 2515,
- CODE_FOR_vec_widen_umult_lo_v4si = 2516,
- CODE_FOR_vec_widen_smult_hi_v16qi = 2517,
- CODE_FOR_vec_widen_umult_hi_v16qi = 2518,
- CODE_FOR_vec_widen_smult_hi_v8hi = 2519,
- CODE_FOR_vec_widen_umult_hi_v8hi = 2520,
- CODE_FOR_vec_widen_smult_hi_v4si = 2521,
- CODE_FOR_vec_widen_umult_hi_v4si = 2522,
- CODE_FOR_fixv2sfv2si2 = 2523,
- CODE_FOR_fixunsv2sfv2si2 = 2524,
- CODE_FOR_fixv4sfv4si2 = 2525,
- CODE_FOR_fixunsv4sfv4si2 = 2526,
- CODE_FOR_fixv2dfv2di2 = 2527,
- CODE_FOR_fixunsv2dfv2di2 = 2528,
- CODE_FOR_fix_truncv2sfv2si2 = 2529,
- CODE_FOR_fixuns_truncv2sfv2si2 = 2530,
- CODE_FOR_fix_truncv4sfv4si2 = 2531,
- CODE_FOR_fixuns_truncv4sfv4si2 = 2532,
- CODE_FOR_fix_truncv2dfv2di2 = 2533,
- CODE_FOR_fixuns_truncv2dfv2di2 = 2534,
- CODE_FOR_ftruncv2sf2 = 2535,
- CODE_FOR_ftruncv4sf2 = 2536,
- CODE_FOR_ftruncv2df2 = 2537,
- CODE_FOR_vec_pack_trunc_v2df = 2538,
- CODE_FOR_vec_pack_trunc_df = 2539,
- CODE_FOR_reduc_splus_v4sf = 2540,
- CODE_FOR_aarch64_simd_bslv8qi = 2541,
- CODE_FOR_aarch64_simd_bslv16qi = 2542,
- CODE_FOR_aarch64_simd_bslv4hi = 2543,
- CODE_FOR_aarch64_simd_bslv8hi = 2544,
- CODE_FOR_aarch64_simd_bslv2si = 2545,
- CODE_FOR_aarch64_simd_bslv4si = 2546,
- CODE_FOR_aarch64_simd_bslv2di = 2547,
- CODE_FOR_aarch64_simd_bslv2sf = 2548,
- CODE_FOR_aarch64_simd_bslv4sf = 2549,
- CODE_FOR_aarch64_simd_bslv2df = 2550,
- CODE_FOR_aarch64_simd_bsldi = 2551,
- CODE_FOR_aarch64_simd_bsldf = 2552,
- CODE_FOR_aarch64_vcond_internalv8qiv8qi = 2553,
- CODE_FOR_aarch64_vcond_internalv16qiv16qi = 2554,
- CODE_FOR_aarch64_vcond_internalv4hiv4hi = 2555,
- CODE_FOR_aarch64_vcond_internalv8hiv8hi = 2556,
- CODE_FOR_aarch64_vcond_internalv2siv2si = 2557,
- CODE_FOR_aarch64_vcond_internalv4siv4si = 2558,
- CODE_FOR_aarch64_vcond_internalv2div2di = 2559,
- CODE_FOR_aarch64_vcond_internalv2sfv2sf = 2560,
- CODE_FOR_aarch64_vcond_internalv2siv2sf = 2561,
- CODE_FOR_aarch64_vcond_internalv4sfv2sf = 2562,
- CODE_FOR_aarch64_vcond_internalv4siv2sf = 2563,
- CODE_FOR_aarch64_vcond_internalv2dfv2sf = 2564,
- CODE_FOR_aarch64_vcond_internalv2div2sf = 2565,
- CODE_FOR_aarch64_vcond_internalv2sfv4sf = 2566,
- CODE_FOR_aarch64_vcond_internalv2siv4sf = 2567,
- CODE_FOR_aarch64_vcond_internalv4sfv4sf = 2568,
- CODE_FOR_aarch64_vcond_internalv4siv4sf = 2569,
- CODE_FOR_aarch64_vcond_internalv2dfv4sf = 2570,
- CODE_FOR_aarch64_vcond_internalv2div4sf = 2571,
- CODE_FOR_aarch64_vcond_internalv2sfv2df = 2572,
- CODE_FOR_aarch64_vcond_internalv2siv2df = 2573,
- CODE_FOR_aarch64_vcond_internalv4sfv2df = 2574,
- CODE_FOR_aarch64_vcond_internalv4siv2df = 2575,
- CODE_FOR_aarch64_vcond_internalv2dfv2df = 2576,
- CODE_FOR_aarch64_vcond_internalv2div2df = 2577,
- CODE_FOR_vcondv8qiv8qi = 2578,
- CODE_FOR_vcondv16qiv16qi = 2579,
- CODE_FOR_vcondv4hiv4hi = 2580,
- CODE_FOR_vcondv8hiv8hi = 2581,
- CODE_FOR_vcondv2siv2si = 2582,
- CODE_FOR_vcondv4siv4si = 2583,
- CODE_FOR_vcondv2div2di = 2584,
- CODE_FOR_vcondv2sfv2sf = 2585,
- CODE_FOR_vcondv4sfv4sf = 2586,
- CODE_FOR_vcondv2dfv2df = 2587,
- CODE_FOR_vcondv2siv2sf = 2588,
- CODE_FOR_vcondv4siv4sf = 2589,
- CODE_FOR_vcondv2div2df = 2590,
- CODE_FOR_vconduv8qiv8qi = 2591,
- CODE_FOR_vconduv16qiv16qi = 2592,
- CODE_FOR_vconduv4hiv4hi = 2593,
- CODE_FOR_vconduv8hiv8hi = 2594,
- CODE_FOR_vconduv2siv2si = 2595,
- CODE_FOR_vconduv4siv4si = 2596,
- CODE_FOR_vconduv2div2di = 2597,
- CODE_FOR_aarch64_createv8qi = 2598,
- CODE_FOR_aarch64_createv4hi = 2599,
- CODE_FOR_aarch64_createv2si = 2600,
- CODE_FOR_aarch64_createdi = 2601,
- CODE_FOR_aarch64_createdf = 2602,
- CODE_FOR_aarch64_createv2sf = 2603,
- CODE_FOR_aarch64_be_checked_get_lanev8qi = 2604,
- CODE_FOR_aarch64_be_checked_get_lanev16qi = 2605,
- CODE_FOR_aarch64_be_checked_get_lanev4hi = 2606,
- CODE_FOR_aarch64_be_checked_get_lanev8hi = 2607,
- CODE_FOR_aarch64_be_checked_get_lanev2si = 2608,
- CODE_FOR_aarch64_be_checked_get_lanev4si = 2609,
- CODE_FOR_aarch64_be_checked_get_lanev2di = 2610,
- CODE_FOR_aarch64_be_checked_get_lanev2sf = 2611,
- CODE_FOR_aarch64_be_checked_get_lanev4sf = 2612,
- CODE_FOR_aarch64_be_checked_get_lanev2df = 2613,
- CODE_FOR_aarch64_get_lanedi = 2614,
- CODE_FOR_aarch64_reinterpretv8qiv8qi = 2615,
- CODE_FOR_aarch64_reinterpretv8qiv4hi = 2616,
- CODE_FOR_aarch64_reinterpretv8qiv2si = 2617,
- CODE_FOR_aarch64_reinterpretv8qiv2sf = 2618,
- CODE_FOR_aarch64_reinterpretv8qidi = 2619,
- CODE_FOR_aarch64_reinterpretv8qidf = 2620,
- CODE_FOR_aarch64_reinterpretv4hiv8qi = 2621,
- CODE_FOR_aarch64_reinterpretv4hiv4hi = 2622,
- CODE_FOR_aarch64_reinterpretv4hiv2si = 2623,
- CODE_FOR_aarch64_reinterpretv4hiv2sf = 2624,
- CODE_FOR_aarch64_reinterpretv4hidi = 2625,
- CODE_FOR_aarch64_reinterpretv4hidf = 2626,
- CODE_FOR_aarch64_reinterpretv2siv8qi = 2627,
- CODE_FOR_aarch64_reinterpretv2siv4hi = 2628,
- CODE_FOR_aarch64_reinterpretv2siv2si = 2629,
- CODE_FOR_aarch64_reinterpretv2siv2sf = 2630,
- CODE_FOR_aarch64_reinterpretv2sidi = 2631,
- CODE_FOR_aarch64_reinterpretv2sidf = 2632,
- CODE_FOR_aarch64_reinterpretv2sfv8qi = 2633,
- CODE_FOR_aarch64_reinterpretv2sfv4hi = 2634,
- CODE_FOR_aarch64_reinterpretv2sfv2si = 2635,
- CODE_FOR_aarch64_reinterpretv2sfv2sf = 2636,
- CODE_FOR_aarch64_reinterpretv2sfdi = 2637,
- CODE_FOR_aarch64_reinterpretv2sfdf = 2638,
- CODE_FOR_aarch64_reinterpretdiv8qi = 2639,
- CODE_FOR_aarch64_reinterpretdiv4hi = 2640,
- CODE_FOR_aarch64_reinterpretdiv2si = 2641,
- CODE_FOR_aarch64_reinterpretdidi = 2642,
- CODE_FOR_aarch64_reinterpretdidf = 2643,
- CODE_FOR_aarch64_reinterpretdiv2sf = 2644,
- CODE_FOR_aarch64_reinterpretdfv8qi = 2645,
- CODE_FOR_aarch64_reinterpretdfv4hi = 2646,
- CODE_FOR_aarch64_reinterpretdfv2si = 2647,
- CODE_FOR_aarch64_reinterpretdfdi = 2648,
- CODE_FOR_aarch64_reinterpretdfdf = 2649,
- CODE_FOR_aarch64_reinterpretdfv2sf = 2650,
- CODE_FOR_aarch64_reinterpretv16qiv16qi = 2651,
- CODE_FOR_aarch64_reinterpretv16qiv8hi = 2652,
- CODE_FOR_aarch64_reinterpretv16qiv4si = 2653,
- CODE_FOR_aarch64_reinterpretv16qiv2di = 2654,
- CODE_FOR_aarch64_reinterpretv16qiv4sf = 2655,
- CODE_FOR_aarch64_reinterpretv16qiv2df = 2656,
- CODE_FOR_aarch64_reinterpretv8hiv16qi = 2657,
- CODE_FOR_aarch64_reinterpretv8hiv8hi = 2658,
- CODE_FOR_aarch64_reinterpretv8hiv4si = 2659,
- CODE_FOR_aarch64_reinterpretv8hiv2di = 2660,
- CODE_FOR_aarch64_reinterpretv8hiv4sf = 2661,
- CODE_FOR_aarch64_reinterpretv8hiv2df = 2662,
- CODE_FOR_aarch64_reinterpretv4siv16qi = 2663,
- CODE_FOR_aarch64_reinterpretv4siv8hi = 2664,
- CODE_FOR_aarch64_reinterpretv4siv4si = 2665,
- CODE_FOR_aarch64_reinterpretv4siv2di = 2666,
- CODE_FOR_aarch64_reinterpretv4siv4sf = 2667,
- CODE_FOR_aarch64_reinterpretv4siv2df = 2668,
- CODE_FOR_aarch64_reinterpretv4sfv16qi = 2669,
- CODE_FOR_aarch64_reinterpretv4sfv8hi = 2670,
- CODE_FOR_aarch64_reinterpretv4sfv4si = 2671,
- CODE_FOR_aarch64_reinterpretv4sfv2di = 2672,
- CODE_FOR_aarch64_reinterpretv4sfv4sf = 2673,
- CODE_FOR_aarch64_reinterpretv4sfv2df = 2674,
- CODE_FOR_aarch64_reinterpretv2div16qi = 2675,
- CODE_FOR_aarch64_reinterpretv2div8hi = 2676,
- CODE_FOR_aarch64_reinterpretv2div4si = 2677,
- CODE_FOR_aarch64_reinterpretv2div2di = 2678,
- CODE_FOR_aarch64_reinterpretv2div4sf = 2679,
- CODE_FOR_aarch64_reinterpretv2div2df = 2680,
- CODE_FOR_aarch64_reinterpretv2dfv16qi = 2681,
- CODE_FOR_aarch64_reinterpretv2dfv8hi = 2682,
- CODE_FOR_aarch64_reinterpretv2dfv4si = 2683,
- CODE_FOR_aarch64_reinterpretv2dfv2di = 2684,
- CODE_FOR_aarch64_reinterpretv2dfv4sf = 2685,
- CODE_FOR_aarch64_reinterpretv2dfv2df = 2686,
- CODE_FOR_aarch64_simd_combinev8qi = 2693,
- CODE_FOR_aarch64_simd_combinev4hi = 2694,
- CODE_FOR_aarch64_simd_combinev2si = 2695,
- CODE_FOR_aarch64_simd_combinev2sf = 2696,
- CODE_FOR_aarch64_simd_combinedi = 2697,
- CODE_FOR_aarch64_simd_combinedf = 2698,
- CODE_FOR_aarch64_saddl2v16qi = 2699,
- CODE_FOR_aarch64_saddl2v8hi = 2700,
- CODE_FOR_aarch64_saddl2v4si = 2701,
- CODE_FOR_aarch64_uaddl2v16qi = 2702,
- CODE_FOR_aarch64_uaddl2v8hi = 2703,
- CODE_FOR_aarch64_uaddl2v4si = 2704,
- CODE_FOR_aarch64_ssubl2v16qi = 2705,
- CODE_FOR_aarch64_ssubl2v8hi = 2706,
- CODE_FOR_aarch64_ssubl2v4si = 2707,
- CODE_FOR_aarch64_usubl2v16qi = 2708,
- CODE_FOR_aarch64_usubl2v8hi = 2709,
- CODE_FOR_aarch64_usubl2v4si = 2710,
- CODE_FOR_aarch64_saddw2v16qi = 2711,
- CODE_FOR_aarch64_saddw2v8hi = 2712,
- CODE_FOR_aarch64_saddw2v4si = 2713,
- CODE_FOR_aarch64_uaddw2v16qi = 2714,
- CODE_FOR_aarch64_uaddw2v8hi = 2715,
- CODE_FOR_aarch64_uaddw2v4si = 2716,
- CODE_FOR_aarch64_ssubw2v16qi = 2717,
- CODE_FOR_aarch64_ssubw2v8hi = 2718,
- CODE_FOR_aarch64_ssubw2v4si = 2719,
- CODE_FOR_aarch64_usubw2v16qi = 2720,
- CODE_FOR_aarch64_usubw2v8hi = 2721,
- CODE_FOR_aarch64_usubw2v4si = 2722,
- CODE_FOR_aarch64_sqdmlal_lanev4hi = 2723,
- CODE_FOR_aarch64_sqdmlal_lanev2si = 2724,
- CODE_FOR_aarch64_sqdmlal_lanehi = 2725,
- CODE_FOR_aarch64_sqdmlal_lanesi = 2726,
- CODE_FOR_aarch64_sqdmlal_laneqv4hi = 2727,
- CODE_FOR_aarch64_sqdmlal_laneqv2si = 2728,
- CODE_FOR_aarch64_sqdmlal_laneqhi = 2729,
- CODE_FOR_aarch64_sqdmlal_laneqsi = 2730,
- CODE_FOR_aarch64_sqdmlsl_lanev4hi = 2731,
- CODE_FOR_aarch64_sqdmlsl_lanev2si = 2732,
- CODE_FOR_aarch64_sqdmlsl_lanehi = 2733,
- CODE_FOR_aarch64_sqdmlsl_lanesi = 2734,
- CODE_FOR_aarch64_sqdmlsl_laneqv4hi = 2735,
- CODE_FOR_aarch64_sqdmlsl_laneqv2si = 2736,
- CODE_FOR_aarch64_sqdmlsl_laneqhi = 2737,
- CODE_FOR_aarch64_sqdmlsl_laneqsi = 2738,
- CODE_FOR_aarch64_sqdmlal2v8hi = 2739,
- CODE_FOR_aarch64_sqdmlal2v4si = 2740,
- CODE_FOR_aarch64_sqdmlsl2v8hi = 2741,
- CODE_FOR_aarch64_sqdmlsl2v4si = 2742,
- CODE_FOR_aarch64_sqdmlal2_lanev8hi = 2743,
- CODE_FOR_aarch64_sqdmlal2_lanev4si = 2744,
- CODE_FOR_aarch64_sqdmlal2_laneqv8hi = 2745,
- CODE_FOR_aarch64_sqdmlal2_laneqv4si = 2746,
- CODE_FOR_aarch64_sqdmlsl2_lanev8hi = 2747,
- CODE_FOR_aarch64_sqdmlsl2_lanev4si = 2748,
- CODE_FOR_aarch64_sqdmlsl2_laneqv8hi = 2749,
- CODE_FOR_aarch64_sqdmlsl2_laneqv4si = 2750,
- CODE_FOR_aarch64_sqdmlal2_nv8hi = 2751,
- CODE_FOR_aarch64_sqdmlal2_nv4si = 2752,
- CODE_FOR_aarch64_sqdmlsl2_nv8hi = 2753,
- CODE_FOR_aarch64_sqdmlsl2_nv4si = 2754,
- CODE_FOR_aarch64_sqdmull_lanev4hi = 2755,
- CODE_FOR_aarch64_sqdmull_lanev2si = 2756,
- CODE_FOR_aarch64_sqdmull_lanehi = 2757,
- CODE_FOR_aarch64_sqdmull_lanesi = 2758,
- CODE_FOR_aarch64_sqdmull_laneqv4hi = 2759,
- CODE_FOR_aarch64_sqdmull_laneqv2si = 2760,
- CODE_FOR_aarch64_sqdmull2v8hi = 2761,
- CODE_FOR_aarch64_sqdmull2v4si = 2762,
- CODE_FOR_aarch64_sqdmull2_lanev8hi = 2763,
- CODE_FOR_aarch64_sqdmull2_lanev4si = 2764,
- CODE_FOR_aarch64_sqdmull2_laneqv8hi = 2765,
- CODE_FOR_aarch64_sqdmull2_laneqv4si = 2766,
- CODE_FOR_aarch64_sqdmull2_nv8hi = 2767,
- CODE_FOR_aarch64_sqdmull2_nv4si = 2768,
- CODE_FOR_movoi = 2779,
- CODE_FOR_movci = 2780,
- CODE_FOR_movxi = 2781,
- CODE_FOR_aarch64_ld2v8qi = 2785,
- CODE_FOR_aarch64_ld2v4hi = 2786,
- CODE_FOR_aarch64_ld2v2si = 2787,
- CODE_FOR_aarch64_ld2v2sf = 2788,
- CODE_FOR_aarch64_ld2di = 2789,
- CODE_FOR_aarch64_ld2df = 2790,
- CODE_FOR_aarch64_ld3v8qi = 2791,
- CODE_FOR_aarch64_ld3v4hi = 2792,
- CODE_FOR_aarch64_ld3v2si = 2793,
- CODE_FOR_aarch64_ld3v2sf = 2794,
- CODE_FOR_aarch64_ld3di = 2795,
- CODE_FOR_aarch64_ld3df = 2796,
- CODE_FOR_aarch64_ld4v8qi = 2797,
- CODE_FOR_aarch64_ld4v4hi = 2798,
- CODE_FOR_aarch64_ld4v2si = 2799,
- CODE_FOR_aarch64_ld4v2sf = 2800,
- CODE_FOR_aarch64_ld4di = 2801,
- CODE_FOR_aarch64_ld4df = 2802,
- CODE_FOR_aarch64_ld1v8qi = 2803,
- CODE_FOR_aarch64_ld1v16qi = 2804,
- CODE_FOR_aarch64_ld1v4hi = 2805,
- CODE_FOR_aarch64_ld1v8hi = 2806,
- CODE_FOR_aarch64_ld1v2si = 2807,
- CODE_FOR_aarch64_ld1v4si = 2808,
- CODE_FOR_aarch64_ld1v2di = 2809,
- CODE_FOR_aarch64_ld1v2sf = 2810,
- CODE_FOR_aarch64_ld1v4sf = 2811,
- CODE_FOR_aarch64_ld1v2df = 2812,
- CODE_FOR_aarch64_ld2v16qi = 2813,
- CODE_FOR_aarch64_ld3v16qi = 2814,
- CODE_FOR_aarch64_ld4v16qi = 2815,
- CODE_FOR_aarch64_ld2v8hi = 2816,
- CODE_FOR_aarch64_ld3v8hi = 2817,
- CODE_FOR_aarch64_ld4v8hi = 2818,
- CODE_FOR_aarch64_ld2v4si = 2819,
- CODE_FOR_aarch64_ld3v4si = 2820,
- CODE_FOR_aarch64_ld4v4si = 2821,
- CODE_FOR_aarch64_ld2v2di = 2822,
- CODE_FOR_aarch64_ld3v2di = 2823,
- CODE_FOR_aarch64_ld4v2di = 2824,
- CODE_FOR_aarch64_ld2v4sf = 2825,
- CODE_FOR_aarch64_ld3v4sf = 2826,
- CODE_FOR_aarch64_ld4v4sf = 2827,
- CODE_FOR_aarch64_ld2v2df = 2828,
- CODE_FOR_aarch64_ld3v2df = 2829,
- CODE_FOR_aarch64_ld4v2df = 2830,
- CODE_FOR_aarch64_get_dregoiv8qi = 2831,
- CODE_FOR_aarch64_get_dregoiv4hi = 2832,
- CODE_FOR_aarch64_get_dregoiv2si = 2833,
- CODE_FOR_aarch64_get_dregoiv2sf = 2834,
- CODE_FOR_aarch64_get_dregoidi = 2835,
- CODE_FOR_aarch64_get_dregoidf = 2836,
- CODE_FOR_aarch64_get_dregciv8qi = 2837,
- CODE_FOR_aarch64_get_dregciv4hi = 2838,
- CODE_FOR_aarch64_get_dregciv2si = 2839,
- CODE_FOR_aarch64_get_dregciv2sf = 2840,
- CODE_FOR_aarch64_get_dregcidi = 2841,
- CODE_FOR_aarch64_get_dregcidf = 2842,
- CODE_FOR_aarch64_get_dregxiv8qi = 2843,
- CODE_FOR_aarch64_get_dregxiv4hi = 2844,
- CODE_FOR_aarch64_get_dregxiv2si = 2845,
- CODE_FOR_aarch64_get_dregxiv2sf = 2846,
- CODE_FOR_aarch64_get_dregxidi = 2847,
- CODE_FOR_aarch64_get_dregxidf = 2848,
- CODE_FOR_aarch64_get_qregoiv16qi = 2849,
- CODE_FOR_aarch64_get_qregciv16qi = 2850,
- CODE_FOR_aarch64_get_qregxiv16qi = 2851,
- CODE_FOR_aarch64_get_qregoiv8hi = 2852,
- CODE_FOR_aarch64_get_qregciv8hi = 2853,
- CODE_FOR_aarch64_get_qregxiv8hi = 2854,
- CODE_FOR_aarch64_get_qregoiv4si = 2855,
- CODE_FOR_aarch64_get_qregciv4si = 2856,
- CODE_FOR_aarch64_get_qregxiv4si = 2857,
- CODE_FOR_aarch64_get_qregoiv2di = 2858,
- CODE_FOR_aarch64_get_qregciv2di = 2859,
- CODE_FOR_aarch64_get_qregxiv2di = 2860,
- CODE_FOR_aarch64_get_qregoiv4sf = 2861,
- CODE_FOR_aarch64_get_qregciv4sf = 2862,
- CODE_FOR_aarch64_get_qregxiv4sf = 2863,
- CODE_FOR_aarch64_get_qregoiv2df = 2864,
- CODE_FOR_aarch64_get_qregciv2df = 2865,
- CODE_FOR_aarch64_get_qregxiv2df = 2866,
- CODE_FOR_vec_perm_constv8qi = 2867,
- CODE_FOR_vec_perm_constv16qi = 2868,
- CODE_FOR_vec_perm_constv4hi = 2869,
- CODE_FOR_vec_perm_constv8hi = 2870,
- CODE_FOR_vec_perm_constv2si = 2871,
- CODE_FOR_vec_perm_constv4si = 2872,
- CODE_FOR_vec_perm_constv2di = 2873,
- CODE_FOR_vec_perm_constv2sf = 2874,
- CODE_FOR_vec_perm_constv4sf = 2875,
- CODE_FOR_vec_perm_constv2df = 2876,
- CODE_FOR_vec_permv8qi = 2877,
- CODE_FOR_vec_permv16qi = 2878,
- CODE_FOR_aarch64_st2v8qi = 2880,
- CODE_FOR_aarch64_st2v4hi = 2881,
- CODE_FOR_aarch64_st2v2si = 2882,
- CODE_FOR_aarch64_st2v2sf = 2883,
- CODE_FOR_aarch64_st2di = 2884,
- CODE_FOR_aarch64_st2df = 2885,
- CODE_FOR_aarch64_st3v8qi = 2886,
- CODE_FOR_aarch64_st3v4hi = 2887,
- CODE_FOR_aarch64_st3v2si = 2888,
- CODE_FOR_aarch64_st3v2sf = 2889,
- CODE_FOR_aarch64_st3di = 2890,
- CODE_FOR_aarch64_st3df = 2891,
- CODE_FOR_aarch64_st4v8qi = 2892,
- CODE_FOR_aarch64_st4v4hi = 2893,
- CODE_FOR_aarch64_st4v2si = 2894,
- CODE_FOR_aarch64_st4v2sf = 2895,
- CODE_FOR_aarch64_st4di = 2896,
- CODE_FOR_aarch64_st4df = 2897,
- CODE_FOR_aarch64_st2v16qi = 2898,
- CODE_FOR_aarch64_st3v16qi = 2899,
- CODE_FOR_aarch64_st4v16qi = 2900,
- CODE_FOR_aarch64_st2v8hi = 2901,
- CODE_FOR_aarch64_st3v8hi = 2902,
- CODE_FOR_aarch64_st4v8hi = 2903,
- CODE_FOR_aarch64_st2v4si = 2904,
- CODE_FOR_aarch64_st3v4si = 2905,
- CODE_FOR_aarch64_st4v4si = 2906,
- CODE_FOR_aarch64_st2v2di = 2907,
- CODE_FOR_aarch64_st3v2di = 2908,
- CODE_FOR_aarch64_st4v2di = 2909,
- CODE_FOR_aarch64_st2v4sf = 2910,
- CODE_FOR_aarch64_st3v4sf = 2911,
- CODE_FOR_aarch64_st4v4sf = 2912,
- CODE_FOR_aarch64_st2v2df = 2913,
- CODE_FOR_aarch64_st3v2df = 2914,
- CODE_FOR_aarch64_st4v2df = 2915,
- CODE_FOR_aarch64_st2_lanev16qi = 2916,
- CODE_FOR_aarch64_st2_lanev8hi = 2917,
- CODE_FOR_aarch64_st2_lanev4si = 2918,
- CODE_FOR_aarch64_st2_lanev2di = 2919,
- CODE_FOR_aarch64_st2_lanev4sf = 2920,
- CODE_FOR_aarch64_st2_lanev2df = 2921,
- CODE_FOR_aarch64_st3_lanev16qi = 2922,
- CODE_FOR_aarch64_st3_lanev8hi = 2923,
- CODE_FOR_aarch64_st3_lanev4si = 2924,
- CODE_FOR_aarch64_st3_lanev2di = 2925,
- CODE_FOR_aarch64_st3_lanev4sf = 2926,
- CODE_FOR_aarch64_st3_lanev2df = 2927,
- CODE_FOR_aarch64_st4_lanev16qi = 2928,
- CODE_FOR_aarch64_st4_lanev8hi = 2929,
- CODE_FOR_aarch64_st4_lanev4si = 2930,
- CODE_FOR_aarch64_st4_lanev2di = 2931,
- CODE_FOR_aarch64_st4_lanev4sf = 2932,
- CODE_FOR_aarch64_st4_lanev2df = 2933,
- CODE_FOR_aarch64_st1v8qi = 2934,
- CODE_FOR_aarch64_st1v16qi = 2935,
- CODE_FOR_aarch64_st1v4hi = 2936,
- CODE_FOR_aarch64_st1v8hi = 2937,
- CODE_FOR_aarch64_st1v2si = 2938,
- CODE_FOR_aarch64_st1v4si = 2939,
- CODE_FOR_aarch64_st1v2di = 2940,
- CODE_FOR_aarch64_st1v2sf = 2941,
- CODE_FOR_aarch64_st1v4sf = 2942,
- CODE_FOR_aarch64_st1v2df = 2943,
- CODE_FOR_aarch64_set_qregoiv16qi = 2944,
- CODE_FOR_aarch64_set_qregciv16qi = 2945,
- CODE_FOR_aarch64_set_qregxiv16qi = 2946,
- CODE_FOR_aarch64_set_qregoiv8hi = 2947,
- CODE_FOR_aarch64_set_qregciv8hi = 2948,
- CODE_FOR_aarch64_set_qregxiv8hi = 2949,
- CODE_FOR_aarch64_set_qregoiv4si = 2950,
- CODE_FOR_aarch64_set_qregciv4si = 2951,
- CODE_FOR_aarch64_set_qregxiv4si = 2952,
- CODE_FOR_aarch64_set_qregoiv2di = 2953,
- CODE_FOR_aarch64_set_qregciv2di = 2954,
- CODE_FOR_aarch64_set_qregxiv2di = 2955,
- CODE_FOR_aarch64_set_qregoiv4sf = 2956,
- CODE_FOR_aarch64_set_qregciv4sf = 2957,
- CODE_FOR_aarch64_set_qregxiv4sf = 2958,
- CODE_FOR_aarch64_set_qregoiv2df = 2959,
- CODE_FOR_aarch64_set_qregciv2df = 2960,
- CODE_FOR_aarch64_set_qregxiv2df = 2961,
- CODE_FOR_vec_initv8qi = 2962,
- CODE_FOR_vec_initv16qi = 2963,
- CODE_FOR_vec_initv4hi = 2964,
- CODE_FOR_vec_initv8hi = 2965,
- CODE_FOR_vec_initv2si = 2966,
- CODE_FOR_vec_initv4si = 2967,
- CODE_FOR_vec_initv2di = 2968,
- CODE_FOR_vec_initv2sf = 2969,
- CODE_FOR_vec_initv4sf = 2970,
- CODE_FOR_vec_initv2df = 2971,
- CODE_FOR_vec_extractv8qi = 2972,
- CODE_FOR_vec_extractv16qi = 2973,
- CODE_FOR_vec_extractv4hi = 2974,
- CODE_FOR_vec_extractv8hi = 2975,
- CODE_FOR_vec_extractv2si = 2976,
- CODE_FOR_vec_extractv4si = 2977,
- CODE_FOR_vec_extractv2di = 2978,
- CODE_FOR_vec_extractv2sf = 2979,
- CODE_FOR_vec_extractv4sf = 2980,
- CODE_FOR_vec_extractv2df = 2981,
- CODE_FOR_atomic_compare_and_swapqi = 2982,
- CODE_FOR_atomic_compare_and_swaphi = 2983,
- CODE_FOR_atomic_compare_and_swapsi = 2984,
- CODE_FOR_atomic_compare_and_swapdi = 2985,
- CODE_FOR_mem_thread_fence = 3066,
- CODE_FOR_dmb = 3067,
+ CODE_FOR_aarch64_crc32b = 370,
+ CODE_FOR_aarch64_crc32h = 371,
+ CODE_FOR_aarch64_crc32w = 372,
+ CODE_FOR_aarch64_crc32x = 373,
+ CODE_FOR_aarch64_crc32cb = 374,
+ CODE_FOR_aarch64_crc32ch = 375,
+ CODE_FOR_aarch64_crc32cw = 376,
+ CODE_FOR_aarch64_crc32cx = 377,
+ CODE_FOR_csinc3si_insn = 380,
+ CODE_FOR_csinc3di_insn = 381,
+ CODE_FOR_andsi3 = 386,
+ CODE_FOR_iorsi3 = 387,
+ CODE_FOR_xorsi3 = 388,
+ CODE_FOR_anddi3 = 389,
+ CODE_FOR_iordi3 = 390,
+ CODE_FOR_xordi3 = 391,
+ CODE_FOR_one_cmplsi2 = 455,
+ CODE_FOR_one_cmpldi2 = 456,
+ CODE_FOR_clzsi2 = 510,
+ CODE_FOR_clzdi2 = 511,
+ CODE_FOR_clrsbsi2 = 512,
+ CODE_FOR_clrsbdi2 = 513,
+ CODE_FOR_rbitsi2 = 514,
+ CODE_FOR_rbitdi2 = 515,
+ CODE_FOR_bswapsi2 = 595,
+ CODE_FOR_bswapdi2 = 596,
+ CODE_FOR_bswaphi2 = 597,
+ CODE_FOR_rev16si2 = 598,
+ CODE_FOR_rev16di2 = 599,
+ CODE_FOR_rev16si2_alt = 600,
+ CODE_FOR_rev16di2_alt = 601,
+ CODE_FOR_btruncsf2 = 603,
+ CODE_FOR_ceilsf2 = 604,
+ CODE_FOR_floorsf2 = 605,
+ CODE_FOR_frintnsf2 = 606,
+ CODE_FOR_nearbyintsf2 = 607,
+ CODE_FOR_rintsf2 = 608,
+ CODE_FOR_roundsf2 = 609,
+ CODE_FOR_btruncdf2 = 610,
+ CODE_FOR_ceildf2 = 611,
+ CODE_FOR_floordf2 = 612,
+ CODE_FOR_frintndf2 = 613,
+ CODE_FOR_nearbyintdf2 = 614,
+ CODE_FOR_rintdf2 = 615,
+ CODE_FOR_rounddf2 = 616,
+ CODE_FOR_lbtruncsfsi2 = 617,
+ CODE_FOR_lceilsfsi2 = 618,
+ CODE_FOR_lfloorsfsi2 = 619,
+ CODE_FOR_lroundsfsi2 = 620,
+ CODE_FOR_lfrintnsfsi2 = 621,
+ CODE_FOR_lbtruncusfsi2 = 622,
+ CODE_FOR_lceilusfsi2 = 623,
+ CODE_FOR_lfloorusfsi2 = 624,
+ CODE_FOR_lroundusfsi2 = 625,
+ CODE_FOR_lfrintnusfsi2 = 626,
+ CODE_FOR_lbtruncdfsi2 = 627,
+ CODE_FOR_lceildfsi2 = 628,
+ CODE_FOR_lfloordfsi2 = 629,
+ CODE_FOR_lrounddfsi2 = 630,
+ CODE_FOR_lfrintndfsi2 = 631,
+ CODE_FOR_lbtruncudfsi2 = 632,
+ CODE_FOR_lceiludfsi2 = 633,
+ CODE_FOR_lfloorudfsi2 = 634,
+ CODE_FOR_lroundudfsi2 = 635,
+ CODE_FOR_lfrintnudfsi2 = 636,
+ CODE_FOR_lbtruncsfdi2 = 637,
+ CODE_FOR_lceilsfdi2 = 638,
+ CODE_FOR_lfloorsfdi2 = 639,
+ CODE_FOR_lroundsfdi2 = 640,
+ CODE_FOR_lfrintnsfdi2 = 641,
+ CODE_FOR_lbtruncusfdi2 = 642,
+ CODE_FOR_lceilusfdi2 = 643,
+ CODE_FOR_lfloorusfdi2 = 644,
+ CODE_FOR_lroundusfdi2 = 645,
+ CODE_FOR_lfrintnusfdi2 = 646,
+ CODE_FOR_lbtruncdfdi2 = 647,
+ CODE_FOR_lceildfdi2 = 648,
+ CODE_FOR_lfloordfdi2 = 649,
+ CODE_FOR_lrounddfdi2 = 650,
+ CODE_FOR_lfrintndfdi2 = 651,
+ CODE_FOR_lbtruncudfdi2 = 652,
+ CODE_FOR_lceiludfdi2 = 653,
+ CODE_FOR_lfloorudfdi2 = 654,
+ CODE_FOR_lroundudfdi2 = 655,
+ CODE_FOR_lfrintnudfdi2 = 656,
+ CODE_FOR_fmasf4 = 657,
+ CODE_FOR_fmadf4 = 658,
+ CODE_FOR_fnmasf4 = 659,
+ CODE_FOR_fnmadf4 = 660,
+ CODE_FOR_fmssf4 = 661,
+ CODE_FOR_fmsdf4 = 662,
+ CODE_FOR_fnmssf4 = 663,
+ CODE_FOR_fnmsdf4 = 664,
+ CODE_FOR_extendsfdf2 = 667,
+ CODE_FOR_truncdfsf2 = 668,
+ CODE_FOR_fix_truncsfsi2 = 669,
+ CODE_FOR_fix_truncdfsi2 = 670,
+ CODE_FOR_fix_truncsfdi2 = 671,
+ CODE_FOR_fix_truncdfdi2 = 672,
+ CODE_FOR_fixuns_truncsfsi2 = 673,
+ CODE_FOR_fixuns_truncdfsi2 = 674,
+ CODE_FOR_fixuns_truncsfdi2 = 675,
+ CODE_FOR_fixuns_truncdfdi2 = 676,
+ CODE_FOR_floatsisf2 = 677,
+ CODE_FOR_floatsidf2 = 678,
+ CODE_FOR_floatdisf2 = 679,
+ CODE_FOR_floatdidf2 = 680,
+ CODE_FOR_floatunssisf2 = 681,
+ CODE_FOR_floatunssidf2 = 682,
+ CODE_FOR_floatunsdisf2 = 683,
+ CODE_FOR_floatunsdidf2 = 684,
+ CODE_FOR_addsf3 = 685,
+ CODE_FOR_adddf3 = 686,
+ CODE_FOR_subsf3 = 687,
+ CODE_FOR_subdf3 = 688,
+ CODE_FOR_mulsf3 = 689,
+ CODE_FOR_muldf3 = 690,
+ CODE_FOR_divsf3 = 693,
+ CODE_FOR_divdf3 = 694,
+ CODE_FOR_negsf2 = 695,
+ CODE_FOR_negdf2 = 696,
+ CODE_FOR_sqrtsf2 = 697,
+ CODE_FOR_sqrtdf2 = 698,
+ CODE_FOR_abssf2 = 699,
+ CODE_FOR_absdf2 = 700,
+ CODE_FOR_smaxsf3 = 701,
+ CODE_FOR_smaxdf3 = 702,
+ CODE_FOR_sminsf3 = 703,
+ CODE_FOR_smindf3 = 704,
+ CODE_FOR_aarch64_movdi_tilow = 705,
+ CODE_FOR_aarch64_movdi_tflow = 706,
+ CODE_FOR_aarch64_movdi_tihigh = 707,
+ CODE_FOR_aarch64_movdi_tfhigh = 708,
+ CODE_FOR_aarch64_movtihigh_di = 709,
+ CODE_FOR_aarch64_movtfhigh_di = 710,
+ CODE_FOR_aarch64_movtilow_di = 711,
+ CODE_FOR_aarch64_movtflow_di = 712,
+ CODE_FOR_aarch64_movtilow_tilow = 713,
+ CODE_FOR_add_losym_si = 714,
+ CODE_FOR_add_losym_di = 715,
+ CODE_FOR_ldr_got_small_si = 716,
+ CODE_FOR_ldr_got_small_di = 717,
+ CODE_FOR_ldr_got_small_sidi = 718,
+ CODE_FOR_ldr_got_tiny = 719,
+ CODE_FOR_aarch64_load_tp_hard = 720,
+ CODE_FOR_tlsie_small_si = 722,
+ CODE_FOR_tlsie_small_di = 723,
+ CODE_FOR_tlsie_small_sidi = 724,
+ CODE_FOR_tlsle_small_si = 725,
+ CODE_FOR_tlsle_small_di = 726,
+ CODE_FOR_tlsdesc_small_si = 727,
+ CODE_FOR_tlsdesc_small_di = 728,
+ CODE_FOR_stack_tie = 729,
+ CODE_FOR_stack_protect_set_si = 730,
+ CODE_FOR_stack_protect_set_di = 731,
+ CODE_FOR_stack_protect_test_si = 732,
+ CODE_FOR_stack_protect_test_di = 733,
+ CODE_FOR_set_fpcr = 734,
+ CODE_FOR_get_fpcr = 735,
+ CODE_FOR_set_fpsr = 736,
+ CODE_FOR_get_fpsr = 737,
+ CODE_FOR_aarch64_simd_dupv8qi = 738,
+ CODE_FOR_aarch64_simd_dupv16qi = 739,
+ CODE_FOR_aarch64_simd_dupv4hi = 740,
+ CODE_FOR_aarch64_simd_dupv8hi = 741,
+ CODE_FOR_aarch64_simd_dupv2si = 742,
+ CODE_FOR_aarch64_simd_dupv4si = 743,
+ CODE_FOR_aarch64_simd_dupv2di = 744,
+ CODE_FOR_aarch64_simd_dupv2sf = 745,
+ CODE_FOR_aarch64_simd_dupv4sf = 746,
+ CODE_FOR_aarch64_simd_dupv2df = 747,
+ CODE_FOR_aarch64_dup_lanev8qi = 748,
+ CODE_FOR_aarch64_dup_lanev16qi = 749,
+ CODE_FOR_aarch64_dup_lanev4hi = 750,
+ CODE_FOR_aarch64_dup_lanev8hi = 751,
+ CODE_FOR_aarch64_dup_lanev2si = 752,
+ CODE_FOR_aarch64_dup_lanev4si = 753,
+ CODE_FOR_aarch64_dup_lanev2di = 754,
+ CODE_FOR_aarch64_dup_lanev2sf = 755,
+ CODE_FOR_aarch64_dup_lanev4sf = 756,
+ CODE_FOR_aarch64_dup_lanev2df = 757,
+ CODE_FOR_aarch64_dup_lane_to_128v8qi = 758,
+ CODE_FOR_aarch64_dup_lane_to_64v16qi = 759,
+ CODE_FOR_aarch64_dup_lane_to_128v4hi = 760,
+ CODE_FOR_aarch64_dup_lane_to_64v8hi = 761,
+ CODE_FOR_aarch64_dup_lane_to_128v2si = 762,
+ CODE_FOR_aarch64_dup_lane_to_64v4si = 763,
+ CODE_FOR_aarch64_dup_lane_to_64v2di = 764,
+ CODE_FOR_aarch64_dup_lane_to_128v2sf = 765,
+ CODE_FOR_aarch64_dup_lane_to_64v4sf = 766,
+ CODE_FOR_aarch64_dup_lane_to_64v2df = 767,
+ CODE_FOR_aarch64_simd_mov_from_v16qilow = 778,
+ CODE_FOR_aarch64_simd_mov_from_v8hilow = 779,
+ CODE_FOR_aarch64_simd_mov_from_v4silow = 780,
+ CODE_FOR_aarch64_simd_mov_from_v2dilow = 781,
+ CODE_FOR_aarch64_simd_mov_from_v4sflow = 782,
+ CODE_FOR_aarch64_simd_mov_from_v2dflow = 783,
+ CODE_FOR_aarch64_simd_mov_from_v16qihigh = 784,
+ CODE_FOR_aarch64_simd_mov_from_v8hihigh = 785,
+ CODE_FOR_aarch64_simd_mov_from_v4sihigh = 786,
+ CODE_FOR_aarch64_simd_mov_from_v2dihigh = 787,
+ CODE_FOR_aarch64_simd_mov_from_v4sfhigh = 788,
+ CODE_FOR_aarch64_simd_mov_from_v2dfhigh = 789,
+ CODE_FOR_ornv8qi3 = 790,
+ CODE_FOR_ornv16qi3 = 791,
+ CODE_FOR_ornv4hi3 = 792,
+ CODE_FOR_ornv8hi3 = 793,
+ CODE_FOR_ornv2si3 = 794,
+ CODE_FOR_ornv4si3 = 795,
+ CODE_FOR_ornv2di3 = 796,
+ CODE_FOR_bicv8qi3 = 797,
+ CODE_FOR_bicv16qi3 = 798,
+ CODE_FOR_bicv4hi3 = 799,
+ CODE_FOR_bicv8hi3 = 800,
+ CODE_FOR_bicv2si3 = 801,
+ CODE_FOR_bicv4si3 = 802,
+ CODE_FOR_bicv2di3 = 803,
+ CODE_FOR_addv8qi3 = 804,
+ CODE_FOR_addv16qi3 = 805,
+ CODE_FOR_addv4hi3 = 806,
+ CODE_FOR_addv8hi3 = 807,
+ CODE_FOR_addv2si3 = 808,
+ CODE_FOR_addv4si3 = 809,
+ CODE_FOR_addv2di3 = 810,
+ CODE_FOR_subv8qi3 = 811,
+ CODE_FOR_subv16qi3 = 812,
+ CODE_FOR_subv4hi3 = 813,
+ CODE_FOR_subv8hi3 = 814,
+ CODE_FOR_subv2si3 = 815,
+ CODE_FOR_subv4si3 = 816,
+ CODE_FOR_subv2di3 = 817,
+ CODE_FOR_mulv8qi3 = 818,
+ CODE_FOR_mulv16qi3 = 819,
+ CODE_FOR_mulv4hi3 = 820,
+ CODE_FOR_mulv8hi3 = 821,
+ CODE_FOR_mulv2si3 = 822,
+ CODE_FOR_mulv4si3 = 823,
+ CODE_FOR_bswapv4hi = 824,
+ CODE_FOR_bswapv8hi = 825,
+ CODE_FOR_bswapv2si = 826,
+ CODE_FOR_bswapv4si = 827,
+ CODE_FOR_bswapv2di = 828,
+ CODE_FOR_negv8qi2 = 844,
+ CODE_FOR_negv16qi2 = 845,
+ CODE_FOR_negv4hi2 = 846,
+ CODE_FOR_negv8hi2 = 847,
+ CODE_FOR_negv2si2 = 848,
+ CODE_FOR_negv4si2 = 849,
+ CODE_FOR_negv2di2 = 850,
+ CODE_FOR_absv8qi2 = 851,
+ CODE_FOR_absv16qi2 = 852,
+ CODE_FOR_absv4hi2 = 853,
+ CODE_FOR_absv8hi2 = 854,
+ CODE_FOR_absv2si2 = 855,
+ CODE_FOR_absv4si2 = 856,
+ CODE_FOR_absv2di2 = 857,
+ CODE_FOR_abdv8qi_3 = 858,
+ CODE_FOR_abdv16qi_3 = 859,
+ CODE_FOR_abdv4hi_3 = 860,
+ CODE_FOR_abdv8hi_3 = 861,
+ CODE_FOR_abdv2si_3 = 862,
+ CODE_FOR_abdv4si_3 = 863,
+ CODE_FOR_abav8qi_3 = 864,
+ CODE_FOR_abav16qi_3 = 865,
+ CODE_FOR_abav4hi_3 = 866,
+ CODE_FOR_abav8hi_3 = 867,
+ CODE_FOR_abav2si_3 = 868,
+ CODE_FOR_abav4si_3 = 869,
+ CODE_FOR_fabdv2sf_3 = 870,
+ CODE_FOR_fabdv4sf_3 = 871,
+ CODE_FOR_fabdv2df_3 = 872,
+ CODE_FOR_andv8qi3 = 875,
+ CODE_FOR_andv16qi3 = 876,
+ CODE_FOR_andv4hi3 = 877,
+ CODE_FOR_andv8hi3 = 878,
+ CODE_FOR_andv2si3 = 879,
+ CODE_FOR_andv4si3 = 880,
+ CODE_FOR_andv2di3 = 881,
+ CODE_FOR_iorv8qi3 = 882,
+ CODE_FOR_iorv16qi3 = 883,
+ CODE_FOR_iorv4hi3 = 884,
+ CODE_FOR_iorv8hi3 = 885,
+ CODE_FOR_iorv2si3 = 886,
+ CODE_FOR_iorv4si3 = 887,
+ CODE_FOR_iorv2di3 = 888,
+ CODE_FOR_xorv8qi3 = 889,
+ CODE_FOR_xorv16qi3 = 890,
+ CODE_FOR_xorv4hi3 = 891,
+ CODE_FOR_xorv8hi3 = 892,
+ CODE_FOR_xorv2si3 = 893,
+ CODE_FOR_xorv4si3 = 894,
+ CODE_FOR_xorv2di3 = 895,
+ CODE_FOR_one_cmplv8qi2 = 896,
+ CODE_FOR_one_cmplv16qi2 = 897,
+ CODE_FOR_one_cmplv4hi2 = 898,
+ CODE_FOR_one_cmplv8hi2 = 899,
+ CODE_FOR_one_cmplv2si2 = 900,
+ CODE_FOR_one_cmplv4si2 = 901,
+ CODE_FOR_one_cmplv2di2 = 902,
+ CODE_FOR_aarch64_simd_vec_setv8qi = 903,
+ CODE_FOR_aarch64_simd_vec_setv16qi = 904,
+ CODE_FOR_aarch64_simd_vec_setv4hi = 905,
+ CODE_FOR_aarch64_simd_vec_setv8hi = 906,
+ CODE_FOR_aarch64_simd_vec_setv2si = 907,
+ CODE_FOR_aarch64_simd_vec_setv4si = 908,
+ CODE_FOR_aarch64_simd_lshrv8qi = 909,
+ CODE_FOR_aarch64_simd_lshrv16qi = 910,
+ CODE_FOR_aarch64_simd_lshrv4hi = 911,
+ CODE_FOR_aarch64_simd_lshrv8hi = 912,
+ CODE_FOR_aarch64_simd_lshrv2si = 913,
+ CODE_FOR_aarch64_simd_lshrv4si = 914,
+ CODE_FOR_aarch64_simd_lshrv2di = 915,
+ CODE_FOR_aarch64_simd_ashrv8qi = 916,
+ CODE_FOR_aarch64_simd_ashrv16qi = 917,
+ CODE_FOR_aarch64_simd_ashrv4hi = 918,
+ CODE_FOR_aarch64_simd_ashrv8hi = 919,
+ CODE_FOR_aarch64_simd_ashrv2si = 920,
+ CODE_FOR_aarch64_simd_ashrv4si = 921,
+ CODE_FOR_aarch64_simd_ashrv2di = 922,
+ CODE_FOR_aarch64_simd_imm_shlv8qi = 923,
+ CODE_FOR_aarch64_simd_imm_shlv16qi = 924,
+ CODE_FOR_aarch64_simd_imm_shlv4hi = 925,
+ CODE_FOR_aarch64_simd_imm_shlv8hi = 926,
+ CODE_FOR_aarch64_simd_imm_shlv2si = 927,
+ CODE_FOR_aarch64_simd_imm_shlv4si = 928,
+ CODE_FOR_aarch64_simd_imm_shlv2di = 929,
+ CODE_FOR_aarch64_simd_reg_sshlv8qi = 930,
+ CODE_FOR_aarch64_simd_reg_sshlv16qi = 931,
+ CODE_FOR_aarch64_simd_reg_sshlv4hi = 932,
+ CODE_FOR_aarch64_simd_reg_sshlv8hi = 933,
+ CODE_FOR_aarch64_simd_reg_sshlv2si = 934,
+ CODE_FOR_aarch64_simd_reg_sshlv4si = 935,
+ CODE_FOR_aarch64_simd_reg_sshlv2di = 936,
+ CODE_FOR_aarch64_simd_reg_shlv8qi_unsigned = 937,
+ CODE_FOR_aarch64_simd_reg_shlv16qi_unsigned = 938,
+ CODE_FOR_aarch64_simd_reg_shlv4hi_unsigned = 939,
+ CODE_FOR_aarch64_simd_reg_shlv8hi_unsigned = 940,
+ CODE_FOR_aarch64_simd_reg_shlv2si_unsigned = 941,
+ CODE_FOR_aarch64_simd_reg_shlv4si_unsigned = 942,
+ CODE_FOR_aarch64_simd_reg_shlv2di_unsigned = 943,
+ CODE_FOR_aarch64_simd_reg_shlv8qi_signed = 944,
+ CODE_FOR_aarch64_simd_reg_shlv16qi_signed = 945,
+ CODE_FOR_aarch64_simd_reg_shlv4hi_signed = 946,
+ CODE_FOR_aarch64_simd_reg_shlv8hi_signed = 947,
+ CODE_FOR_aarch64_simd_reg_shlv2si_signed = 948,
+ CODE_FOR_aarch64_simd_reg_shlv4si_signed = 949,
+ CODE_FOR_aarch64_simd_reg_shlv2di_signed = 950,
+ CODE_FOR_aarch64_sshr_simddi = 951,
+ CODE_FOR_aarch64_ushr_simddi = 952,
+ CODE_FOR_aarch64_simd_vec_setv2di = 953,
+ CODE_FOR_aarch64_simd_vec_setv2sf = 954,
+ CODE_FOR_aarch64_simd_vec_setv4sf = 955,
+ CODE_FOR_aarch64_simd_vec_setv2df = 956,
+ CODE_FOR_aarch64_mlav8qi = 957,
+ CODE_FOR_aarch64_mlav16qi = 958,
+ CODE_FOR_aarch64_mlav4hi = 959,
+ CODE_FOR_aarch64_mlav8hi = 960,
+ CODE_FOR_aarch64_mlav2si = 961,
+ CODE_FOR_aarch64_mlav4si = 962,
+ CODE_FOR_aarch64_mlsv8qi = 971,
+ CODE_FOR_aarch64_mlsv16qi = 972,
+ CODE_FOR_aarch64_mlsv4hi = 973,
+ CODE_FOR_aarch64_mlsv8hi = 974,
+ CODE_FOR_aarch64_mlsv2si = 975,
+ CODE_FOR_aarch64_mlsv4si = 976,
+ CODE_FOR_smaxv8qi3 = 985,
+ CODE_FOR_sminv8qi3 = 986,
+ CODE_FOR_umaxv8qi3 = 987,
+ CODE_FOR_uminv8qi3 = 988,
+ CODE_FOR_smaxv16qi3 = 989,
+ CODE_FOR_sminv16qi3 = 990,
+ CODE_FOR_umaxv16qi3 = 991,
+ CODE_FOR_uminv16qi3 = 992,
+ CODE_FOR_smaxv4hi3 = 993,
+ CODE_FOR_sminv4hi3 = 994,
+ CODE_FOR_umaxv4hi3 = 995,
+ CODE_FOR_uminv4hi3 = 996,
+ CODE_FOR_smaxv8hi3 = 997,
+ CODE_FOR_sminv8hi3 = 998,
+ CODE_FOR_umaxv8hi3 = 999,
+ CODE_FOR_uminv8hi3 = 1000,
+ CODE_FOR_smaxv2si3 = 1001,
+ CODE_FOR_sminv2si3 = 1002,
+ CODE_FOR_umaxv2si3 = 1003,
+ CODE_FOR_uminv2si3 = 1004,
+ CODE_FOR_smaxv4si3 = 1005,
+ CODE_FOR_sminv4si3 = 1006,
+ CODE_FOR_umaxv4si3 = 1007,
+ CODE_FOR_uminv4si3 = 1008,
+ CODE_FOR_move_lo_quad_internal_v16qi = 1009,
+ CODE_FOR_move_lo_quad_internal_v8hi = 1010,
+ CODE_FOR_move_lo_quad_internal_v4si = 1011,
+ CODE_FOR_move_lo_quad_internal_v2di = 1012,
+ CODE_FOR_move_lo_quad_internal_v4sf = 1013,
+ CODE_FOR_move_lo_quad_internal_v2df = 1014,
+ CODE_FOR_move_lo_quad_internal_be_v16qi = 1015,
+ CODE_FOR_move_lo_quad_internal_be_v8hi = 1016,
+ CODE_FOR_move_lo_quad_internal_be_v4si = 1017,
+ CODE_FOR_move_lo_quad_internal_be_v2di = 1018,
+ CODE_FOR_move_lo_quad_internal_be_v4sf = 1019,
+ CODE_FOR_move_lo_quad_internal_be_v2df = 1020,
+ CODE_FOR_aarch64_simd_move_hi_quad_v16qi = 1021,
+ CODE_FOR_aarch64_simd_move_hi_quad_v8hi = 1022,
+ CODE_FOR_aarch64_simd_move_hi_quad_v4si = 1023,
+ CODE_FOR_aarch64_simd_move_hi_quad_v2di = 1024,
+ CODE_FOR_aarch64_simd_move_hi_quad_v4sf = 1025,
+ CODE_FOR_aarch64_simd_move_hi_quad_v2df = 1026,
+ CODE_FOR_aarch64_simd_move_hi_quad_be_v16qi = 1027,
+ CODE_FOR_aarch64_simd_move_hi_quad_be_v8hi = 1028,
+ CODE_FOR_aarch64_simd_move_hi_quad_be_v4si = 1029,
+ CODE_FOR_aarch64_simd_move_hi_quad_be_v2di = 1030,
+ CODE_FOR_aarch64_simd_move_hi_quad_be_v4sf = 1031,
+ CODE_FOR_aarch64_simd_move_hi_quad_be_v2df = 1032,
+ CODE_FOR_aarch64_simd_vec_pack_trunc_v8hi = 1033,
+ CODE_FOR_aarch64_simd_vec_pack_trunc_v4si = 1034,
+ CODE_FOR_aarch64_simd_vec_pack_trunc_v2di = 1035,
+ CODE_FOR_vec_pack_trunc_v8hi = 1036,
+ CODE_FOR_vec_pack_trunc_v4si = 1037,
+ CODE_FOR_vec_pack_trunc_v2di = 1038,
+ CODE_FOR_aarch64_simd_vec_unpacks_lo_v16qi = 1039,
+ CODE_FOR_aarch64_simd_vec_unpacku_lo_v16qi = 1040,
+ CODE_FOR_aarch64_simd_vec_unpacks_lo_v8hi = 1041,
+ CODE_FOR_aarch64_simd_vec_unpacku_lo_v8hi = 1042,
+ CODE_FOR_aarch64_simd_vec_unpacks_lo_v4si = 1043,
+ CODE_FOR_aarch64_simd_vec_unpacku_lo_v4si = 1044,
+ CODE_FOR_aarch64_simd_vec_unpacks_hi_v16qi = 1045,
+ CODE_FOR_aarch64_simd_vec_unpacku_hi_v16qi = 1046,
+ CODE_FOR_aarch64_simd_vec_unpacks_hi_v8hi = 1047,
+ CODE_FOR_aarch64_simd_vec_unpacku_hi_v8hi = 1048,
+ CODE_FOR_aarch64_simd_vec_unpacks_hi_v4si = 1049,
+ CODE_FOR_aarch64_simd_vec_unpacku_hi_v4si = 1050,
+ CODE_FOR_aarch64_simd_vec_smult_lo_v16qi = 1087,
+ CODE_FOR_aarch64_simd_vec_umult_lo_v16qi = 1088,
+ CODE_FOR_aarch64_simd_vec_smult_lo_v8hi = 1089,
+ CODE_FOR_aarch64_simd_vec_umult_lo_v8hi = 1090,
+ CODE_FOR_aarch64_simd_vec_smult_lo_v4si = 1091,
+ CODE_FOR_aarch64_simd_vec_umult_lo_v4si = 1092,
+ CODE_FOR_aarch64_simd_vec_smult_hi_v16qi = 1093,
+ CODE_FOR_aarch64_simd_vec_umult_hi_v16qi = 1094,
+ CODE_FOR_aarch64_simd_vec_smult_hi_v8hi = 1095,
+ CODE_FOR_aarch64_simd_vec_umult_hi_v8hi = 1096,
+ CODE_FOR_aarch64_simd_vec_smult_hi_v4si = 1097,
+ CODE_FOR_aarch64_simd_vec_umult_hi_v4si = 1098,
+ CODE_FOR_addv2sf3 = 1099,
+ CODE_FOR_addv4sf3 = 1100,
+ CODE_FOR_addv2df3 = 1101,
+ CODE_FOR_subv2sf3 = 1102,
+ CODE_FOR_subv4sf3 = 1103,
+ CODE_FOR_subv2df3 = 1104,
+ CODE_FOR_mulv2sf3 = 1105,
+ CODE_FOR_mulv4sf3 = 1106,
+ CODE_FOR_mulv2df3 = 1107,
+ CODE_FOR_divv2sf3 = 1108,
+ CODE_FOR_divv4sf3 = 1109,
+ CODE_FOR_divv2df3 = 1110,
+ CODE_FOR_negv2sf2 = 1111,
+ CODE_FOR_negv4sf2 = 1112,
+ CODE_FOR_negv2df2 = 1113,
+ CODE_FOR_absv2sf2 = 1114,
+ CODE_FOR_absv4sf2 = 1115,
+ CODE_FOR_absv2df2 = 1116,
+ CODE_FOR_fmav2sf4 = 1117,
+ CODE_FOR_fmav4sf4 = 1118,
+ CODE_FOR_fmav2df4 = 1119,
+ CODE_FOR_fnmav2sf4 = 1127,
+ CODE_FOR_fnmav4sf4 = 1128,
+ CODE_FOR_fnmav2df4 = 1129,
+ CODE_FOR_btruncv2sf2 = 1137,
+ CODE_FOR_ceilv2sf2 = 1138,
+ CODE_FOR_floorv2sf2 = 1139,
+ CODE_FOR_frintnv2sf2 = 1140,
+ CODE_FOR_nearbyintv2sf2 = 1141,
+ CODE_FOR_rintv2sf2 = 1142,
+ CODE_FOR_roundv2sf2 = 1143,
+ CODE_FOR_btruncv4sf2 = 1144,
+ CODE_FOR_ceilv4sf2 = 1145,
+ CODE_FOR_floorv4sf2 = 1146,
+ CODE_FOR_frintnv4sf2 = 1147,
+ CODE_FOR_nearbyintv4sf2 = 1148,
+ CODE_FOR_rintv4sf2 = 1149,
+ CODE_FOR_roundv4sf2 = 1150,
+ CODE_FOR_btruncv2df2 = 1151,
+ CODE_FOR_ceilv2df2 = 1152,
+ CODE_FOR_floorv2df2 = 1153,
+ CODE_FOR_frintnv2df2 = 1154,
+ CODE_FOR_nearbyintv2df2 = 1155,
+ CODE_FOR_rintv2df2 = 1156,
+ CODE_FOR_roundv2df2 = 1157,
+ CODE_FOR_lbtruncv2sfv2si2 = 1158,
+ CODE_FOR_lceilv2sfv2si2 = 1159,
+ CODE_FOR_lfloorv2sfv2si2 = 1160,
+ CODE_FOR_lroundv2sfv2si2 = 1161,
+ CODE_FOR_lfrintnv2sfv2si2 = 1162,
+ CODE_FOR_lbtruncuv2sfv2si2 = 1163,
+ CODE_FOR_lceiluv2sfv2si2 = 1164,
+ CODE_FOR_lflooruv2sfv2si2 = 1165,
+ CODE_FOR_lrounduv2sfv2si2 = 1166,
+ CODE_FOR_lfrintnuv2sfv2si2 = 1167,
+ CODE_FOR_lbtruncv4sfv4si2 = 1168,
+ CODE_FOR_lceilv4sfv4si2 = 1169,
+ CODE_FOR_lfloorv4sfv4si2 = 1170,
+ CODE_FOR_lroundv4sfv4si2 = 1171,
+ CODE_FOR_lfrintnv4sfv4si2 = 1172,
+ CODE_FOR_lbtruncuv4sfv4si2 = 1173,
+ CODE_FOR_lceiluv4sfv4si2 = 1174,
+ CODE_FOR_lflooruv4sfv4si2 = 1175,
+ CODE_FOR_lrounduv4sfv4si2 = 1176,
+ CODE_FOR_lfrintnuv4sfv4si2 = 1177,
+ CODE_FOR_lbtruncv2dfv2di2 = 1178,
+ CODE_FOR_lceilv2dfv2di2 = 1179,
+ CODE_FOR_lfloorv2dfv2di2 = 1180,
+ CODE_FOR_lroundv2dfv2di2 = 1181,
+ CODE_FOR_lfrintnv2dfv2di2 = 1182,
+ CODE_FOR_lbtruncuv2dfv2di2 = 1183,
+ CODE_FOR_lceiluv2dfv2di2 = 1184,
+ CODE_FOR_lflooruv2dfv2di2 = 1185,
+ CODE_FOR_lrounduv2dfv2di2 = 1186,
+ CODE_FOR_lfrintnuv2dfv2di2 = 1187,
+ CODE_FOR_floatv2siv2sf2 = 1188,
+ CODE_FOR_floatunsv2siv2sf2 = 1189,
+ CODE_FOR_floatv4siv4sf2 = 1190,
+ CODE_FOR_floatunsv4siv4sf2 = 1191,
+ CODE_FOR_floatv2div2df2 = 1192,
+ CODE_FOR_floatunsv2div2df2 = 1193,
+ CODE_FOR_vec_unpacks_lo_v4sf = 1194,
+ CODE_FOR_aarch64_float_extend_lo_v2df = 1195,
+ CODE_FOR_vec_unpacks_hi_v4sf = 1196,
+ CODE_FOR_aarch64_float_truncate_lo_v2sf = 1197,
+ CODE_FOR_aarch64_float_truncate_hi_v4sf = 1198,
+ CODE_FOR_aarch64_vmlsv2sf = 1199,
+ CODE_FOR_aarch64_vmlsv4sf = 1200,
+ CODE_FOR_aarch64_vmlsv2df = 1201,
+ CODE_FOR_smaxv2sf3 = 1202,
+ CODE_FOR_sminv2sf3 = 1203,
+ CODE_FOR_smaxv4sf3 = 1204,
+ CODE_FOR_sminv4sf3 = 1205,
+ CODE_FOR_smaxv2df3 = 1206,
+ CODE_FOR_sminv2df3 = 1207,
+ CODE_FOR_smax_nanv2sf3 = 1208,
+ CODE_FOR_smin_nanv2sf3 = 1209,
+ CODE_FOR_smax_nanv4sf3 = 1210,
+ CODE_FOR_smin_nanv4sf3 = 1211,
+ CODE_FOR_smax_nanv2df3 = 1212,
+ CODE_FOR_smin_nanv2df3 = 1213,
+ CODE_FOR_reduc_splus_v8qi = 1214,
+ CODE_FOR_reduc_uplus_v8qi = 1215,
+ CODE_FOR_reduc_splus_v16qi = 1216,
+ CODE_FOR_reduc_uplus_v16qi = 1217,
+ CODE_FOR_reduc_splus_v4hi = 1218,
+ CODE_FOR_reduc_uplus_v4hi = 1219,
+ CODE_FOR_reduc_splus_v8hi = 1220,
+ CODE_FOR_reduc_uplus_v8hi = 1221,
+ CODE_FOR_reduc_splus_v4si = 1222,
+ CODE_FOR_reduc_uplus_v4si = 1223,
+ CODE_FOR_reduc_splus_v2di = 1224,
+ CODE_FOR_reduc_uplus_v2di = 1225,
+ CODE_FOR_reduc_splus_v2si = 1226,
+ CODE_FOR_reduc_uplus_v2si = 1227,
+ CODE_FOR_reduc_splus_v2sf = 1228,
+ CODE_FOR_reduc_splus_v2df = 1229,
+ CODE_FOR_aarch64_addpv4sf = 1230,
+ CODE_FOR_clzv8qi2 = 1231,
+ CODE_FOR_clzv16qi2 = 1232,
+ CODE_FOR_clzv4hi2 = 1233,
+ CODE_FOR_clzv8hi2 = 1234,
+ CODE_FOR_clzv2si2 = 1235,
+ CODE_FOR_clzv4si2 = 1236,
+ CODE_FOR_reduc_umax_v8qi = 1237,
+ CODE_FOR_reduc_umin_v8qi = 1238,
+ CODE_FOR_reduc_smax_v8qi = 1239,
+ CODE_FOR_reduc_smin_v8qi = 1240,
+ CODE_FOR_reduc_umax_v16qi = 1241,
+ CODE_FOR_reduc_umin_v16qi = 1242,
+ CODE_FOR_reduc_smax_v16qi = 1243,
+ CODE_FOR_reduc_smin_v16qi = 1244,
+ CODE_FOR_reduc_umax_v4hi = 1245,
+ CODE_FOR_reduc_umin_v4hi = 1246,
+ CODE_FOR_reduc_smax_v4hi = 1247,
+ CODE_FOR_reduc_smin_v4hi = 1248,
+ CODE_FOR_reduc_umax_v8hi = 1249,
+ CODE_FOR_reduc_umin_v8hi = 1250,
+ CODE_FOR_reduc_smax_v8hi = 1251,
+ CODE_FOR_reduc_smin_v8hi = 1252,
+ CODE_FOR_reduc_umax_v4si = 1253,
+ CODE_FOR_reduc_umin_v4si = 1254,
+ CODE_FOR_reduc_smax_v4si = 1255,
+ CODE_FOR_reduc_smin_v4si = 1256,
+ CODE_FOR_reduc_umax_v2si = 1257,
+ CODE_FOR_reduc_umin_v2si = 1258,
+ CODE_FOR_reduc_smax_v2si = 1259,
+ CODE_FOR_reduc_smin_v2si = 1260,
+ CODE_FOR_reduc_smax_nan_v2sf = 1261,
+ CODE_FOR_reduc_smin_nan_v2sf = 1262,
+ CODE_FOR_reduc_smax_v2sf = 1263,
+ CODE_FOR_reduc_smin_v2sf = 1264,
+ CODE_FOR_reduc_smax_nan_v2df = 1265,
+ CODE_FOR_reduc_smin_nan_v2df = 1266,
+ CODE_FOR_reduc_smax_v2df = 1267,
+ CODE_FOR_reduc_smin_v2df = 1268,
+ CODE_FOR_reduc_smax_nan_v4sf = 1269,
+ CODE_FOR_reduc_smin_nan_v4sf = 1270,
+ CODE_FOR_reduc_smax_v4sf = 1271,
+ CODE_FOR_reduc_smin_v4sf = 1272,
+ CODE_FOR_aarch64_simd_bslv8qi_internal = 1273,
+ CODE_FOR_aarch64_simd_bslv16qi_internal = 1274,
+ CODE_FOR_aarch64_simd_bslv4hi_internal = 1275,
+ CODE_FOR_aarch64_simd_bslv8hi_internal = 1276,
+ CODE_FOR_aarch64_simd_bslv2si_internal = 1277,
+ CODE_FOR_aarch64_simd_bslv4si_internal = 1278,
+ CODE_FOR_aarch64_simd_bslv2di_internal = 1279,
+ CODE_FOR_aarch64_simd_bslv2sf_internal = 1280,
+ CODE_FOR_aarch64_simd_bslv4sf_internal = 1281,
+ CODE_FOR_aarch64_simd_bslv2df_internal = 1282,
+ CODE_FOR_aarch64_simd_bsldi_internal = 1283,
+ CODE_FOR_aarch64_simd_bsldf_internal = 1284,
+ CODE_FOR_aarch64_get_lanev8qi = 1297,
+ CODE_FOR_aarch64_get_lanev16qi = 1298,
+ CODE_FOR_aarch64_get_lanev4hi = 1299,
+ CODE_FOR_aarch64_get_lanev8hi = 1300,
+ CODE_FOR_aarch64_get_lanev2si = 1301,
+ CODE_FOR_aarch64_get_lanev4si = 1302,
+ CODE_FOR_aarch64_get_lanev2di = 1303,
+ CODE_FOR_aarch64_get_lanev2sf = 1304,
+ CODE_FOR_aarch64_get_lanev4sf = 1305,
+ CODE_FOR_aarch64_get_lanev2df = 1306,
+ CODE_FOR_aarch64_combine_internalv8qi = 1313,
+ CODE_FOR_aarch64_combine_internalv4hi = 1314,
+ CODE_FOR_aarch64_combine_internalv2si = 1315,
+ CODE_FOR_aarch64_combine_internalv2sf = 1316,
+ CODE_FOR_aarch64_combine_internaldi = 1317,
+ CODE_FOR_aarch64_combine_internaldf = 1318,
+ CODE_FOR_aarch64_saddlv16qi_hi_internal = 1319,
+ CODE_FOR_aarch64_ssublv16qi_hi_internal = 1320,
+ CODE_FOR_aarch64_uaddlv16qi_hi_internal = 1321,
+ CODE_FOR_aarch64_usublv16qi_hi_internal = 1322,
+ CODE_FOR_aarch64_saddlv8hi_hi_internal = 1323,
+ CODE_FOR_aarch64_ssublv8hi_hi_internal = 1324,
+ CODE_FOR_aarch64_uaddlv8hi_hi_internal = 1325,
+ CODE_FOR_aarch64_usublv8hi_hi_internal = 1326,
+ CODE_FOR_aarch64_saddlv4si_hi_internal = 1327,
+ CODE_FOR_aarch64_ssublv4si_hi_internal = 1328,
+ CODE_FOR_aarch64_uaddlv4si_hi_internal = 1329,
+ CODE_FOR_aarch64_usublv4si_hi_internal = 1330,
+ CODE_FOR_aarch64_saddlv16qi_lo_internal = 1331,
+ CODE_FOR_aarch64_ssublv16qi_lo_internal = 1332,
+ CODE_FOR_aarch64_uaddlv16qi_lo_internal = 1333,
+ CODE_FOR_aarch64_usublv16qi_lo_internal = 1334,
+ CODE_FOR_aarch64_saddlv8hi_lo_internal = 1335,
+ CODE_FOR_aarch64_ssublv8hi_lo_internal = 1336,
+ CODE_FOR_aarch64_uaddlv8hi_lo_internal = 1337,
+ CODE_FOR_aarch64_usublv8hi_lo_internal = 1338,
+ CODE_FOR_aarch64_saddlv4si_lo_internal = 1339,
+ CODE_FOR_aarch64_ssublv4si_lo_internal = 1340,
+ CODE_FOR_aarch64_uaddlv4si_lo_internal = 1341,
+ CODE_FOR_aarch64_usublv4si_lo_internal = 1342,
+ CODE_FOR_aarch64_saddlv8qi = 1343,
+ CODE_FOR_aarch64_ssublv8qi = 1344,
+ CODE_FOR_aarch64_uaddlv8qi = 1345,
+ CODE_FOR_aarch64_usublv8qi = 1346,
+ CODE_FOR_aarch64_saddlv4hi = 1347,
+ CODE_FOR_aarch64_ssublv4hi = 1348,
+ CODE_FOR_aarch64_uaddlv4hi = 1349,
+ CODE_FOR_aarch64_usublv4hi = 1350,
+ CODE_FOR_aarch64_saddlv2si = 1351,
+ CODE_FOR_aarch64_ssublv2si = 1352,
+ CODE_FOR_aarch64_uaddlv2si = 1353,
+ CODE_FOR_aarch64_usublv2si = 1354,
+ CODE_FOR_aarch64_saddwv8qi = 1355,
+ CODE_FOR_aarch64_ssubwv8qi = 1356,
+ CODE_FOR_aarch64_uaddwv8qi = 1357,
+ CODE_FOR_aarch64_usubwv8qi = 1358,
+ CODE_FOR_aarch64_saddwv4hi = 1359,
+ CODE_FOR_aarch64_ssubwv4hi = 1360,
+ CODE_FOR_aarch64_uaddwv4hi = 1361,
+ CODE_FOR_aarch64_usubwv4hi = 1362,
+ CODE_FOR_aarch64_saddwv2si = 1363,
+ CODE_FOR_aarch64_ssubwv2si = 1364,
+ CODE_FOR_aarch64_uaddwv2si = 1365,
+ CODE_FOR_aarch64_usubwv2si = 1366,
+ CODE_FOR_aarch64_saddw2v16qi_internal = 1367,
+ CODE_FOR_aarch64_ssubw2v16qi_internal = 1368,
+ CODE_FOR_aarch64_uaddw2v16qi_internal = 1369,
+ CODE_FOR_aarch64_usubw2v16qi_internal = 1370,
+ CODE_FOR_aarch64_saddw2v8hi_internal = 1371,
+ CODE_FOR_aarch64_ssubw2v8hi_internal = 1372,
+ CODE_FOR_aarch64_uaddw2v8hi_internal = 1373,
+ CODE_FOR_aarch64_usubw2v8hi_internal = 1374,
+ CODE_FOR_aarch64_saddw2v4si_internal = 1375,
+ CODE_FOR_aarch64_ssubw2v4si_internal = 1376,
+ CODE_FOR_aarch64_uaddw2v4si_internal = 1377,
+ CODE_FOR_aarch64_usubw2v4si_internal = 1378,
+ CODE_FOR_aarch64_shaddv8qi = 1379,
+ CODE_FOR_aarch64_uhaddv8qi = 1380,
+ CODE_FOR_aarch64_srhaddv8qi = 1381,
+ CODE_FOR_aarch64_urhaddv8qi = 1382,
+ CODE_FOR_aarch64_shsubv8qi = 1383,
+ CODE_FOR_aarch64_uhsubv8qi = 1384,
+ CODE_FOR_aarch64_srhsubv8qi = 1385,
+ CODE_FOR_aarch64_urhsubv8qi = 1386,
+ CODE_FOR_aarch64_shaddv16qi = 1387,
+ CODE_FOR_aarch64_uhaddv16qi = 1388,
+ CODE_FOR_aarch64_srhaddv16qi = 1389,
+ CODE_FOR_aarch64_urhaddv16qi = 1390,
+ CODE_FOR_aarch64_shsubv16qi = 1391,
+ CODE_FOR_aarch64_uhsubv16qi = 1392,
+ CODE_FOR_aarch64_srhsubv16qi = 1393,
+ CODE_FOR_aarch64_urhsubv16qi = 1394,
+ CODE_FOR_aarch64_shaddv4hi = 1395,
+ CODE_FOR_aarch64_uhaddv4hi = 1396,
+ CODE_FOR_aarch64_srhaddv4hi = 1397,
+ CODE_FOR_aarch64_urhaddv4hi = 1398,
+ CODE_FOR_aarch64_shsubv4hi = 1399,
+ CODE_FOR_aarch64_uhsubv4hi = 1400,
+ CODE_FOR_aarch64_srhsubv4hi = 1401,
+ CODE_FOR_aarch64_urhsubv4hi = 1402,
+ CODE_FOR_aarch64_shaddv8hi = 1403,
+ CODE_FOR_aarch64_uhaddv8hi = 1404,
+ CODE_FOR_aarch64_srhaddv8hi = 1405,
+ CODE_FOR_aarch64_urhaddv8hi = 1406,
+ CODE_FOR_aarch64_shsubv8hi = 1407,
+ CODE_FOR_aarch64_uhsubv8hi = 1408,
+ CODE_FOR_aarch64_srhsubv8hi = 1409,
+ CODE_FOR_aarch64_urhsubv8hi = 1410,
+ CODE_FOR_aarch64_shaddv2si = 1411,
+ CODE_FOR_aarch64_uhaddv2si = 1412,
+ CODE_FOR_aarch64_srhaddv2si = 1413,
+ CODE_FOR_aarch64_urhaddv2si = 1414,
+ CODE_FOR_aarch64_shsubv2si = 1415,
+ CODE_FOR_aarch64_uhsubv2si = 1416,
+ CODE_FOR_aarch64_srhsubv2si = 1417,
+ CODE_FOR_aarch64_urhsubv2si = 1418,
+ CODE_FOR_aarch64_shaddv4si = 1419,
+ CODE_FOR_aarch64_uhaddv4si = 1420,
+ CODE_FOR_aarch64_srhaddv4si = 1421,
+ CODE_FOR_aarch64_urhaddv4si = 1422,
+ CODE_FOR_aarch64_shsubv4si = 1423,
+ CODE_FOR_aarch64_uhsubv4si = 1424,
+ CODE_FOR_aarch64_srhsubv4si = 1425,
+ CODE_FOR_aarch64_urhsubv4si = 1426,
+ CODE_FOR_aarch64_addhnv8hi = 1427,
+ CODE_FOR_aarch64_raddhnv8hi = 1428,
+ CODE_FOR_aarch64_subhnv8hi = 1429,
+ CODE_FOR_aarch64_rsubhnv8hi = 1430,
+ CODE_FOR_aarch64_addhnv4si = 1431,
+ CODE_FOR_aarch64_raddhnv4si = 1432,
+ CODE_FOR_aarch64_subhnv4si = 1433,
+ CODE_FOR_aarch64_rsubhnv4si = 1434,
+ CODE_FOR_aarch64_addhnv2di = 1435,
+ CODE_FOR_aarch64_raddhnv2di = 1436,
+ CODE_FOR_aarch64_subhnv2di = 1437,
+ CODE_FOR_aarch64_rsubhnv2di = 1438,
+ CODE_FOR_aarch64_addhn2v8hi = 1439,
+ CODE_FOR_aarch64_raddhn2v8hi = 1440,
+ CODE_FOR_aarch64_subhn2v8hi = 1441,
+ CODE_FOR_aarch64_rsubhn2v8hi = 1442,
+ CODE_FOR_aarch64_addhn2v4si = 1443,
+ CODE_FOR_aarch64_raddhn2v4si = 1444,
+ CODE_FOR_aarch64_subhn2v4si = 1445,
+ CODE_FOR_aarch64_rsubhn2v4si = 1446,
+ CODE_FOR_aarch64_addhn2v2di = 1447,
+ CODE_FOR_aarch64_raddhn2v2di = 1448,
+ CODE_FOR_aarch64_subhn2v2di = 1449,
+ CODE_FOR_aarch64_rsubhn2v2di = 1450,
+ CODE_FOR_aarch64_pmulv8qi = 1451,
+ CODE_FOR_aarch64_pmulv16qi = 1452,
+ CODE_FOR_aarch64_sqaddv8qi = 1453,
+ CODE_FOR_aarch64_uqaddv8qi = 1454,
+ CODE_FOR_aarch64_sqsubv8qi = 1455,
+ CODE_FOR_aarch64_uqsubv8qi = 1456,
+ CODE_FOR_aarch64_sqaddv16qi = 1457,
+ CODE_FOR_aarch64_uqaddv16qi = 1458,
+ CODE_FOR_aarch64_sqsubv16qi = 1459,
+ CODE_FOR_aarch64_uqsubv16qi = 1460,
+ CODE_FOR_aarch64_sqaddv4hi = 1461,
+ CODE_FOR_aarch64_uqaddv4hi = 1462,
+ CODE_FOR_aarch64_sqsubv4hi = 1463,
+ CODE_FOR_aarch64_uqsubv4hi = 1464,
+ CODE_FOR_aarch64_sqaddv8hi = 1465,
+ CODE_FOR_aarch64_uqaddv8hi = 1466,
+ CODE_FOR_aarch64_sqsubv8hi = 1467,
+ CODE_FOR_aarch64_uqsubv8hi = 1468,
+ CODE_FOR_aarch64_sqaddv2si = 1469,
+ CODE_FOR_aarch64_uqaddv2si = 1470,
+ CODE_FOR_aarch64_sqsubv2si = 1471,
+ CODE_FOR_aarch64_uqsubv2si = 1472,
+ CODE_FOR_aarch64_sqaddv4si = 1473,
+ CODE_FOR_aarch64_uqaddv4si = 1474,
+ CODE_FOR_aarch64_sqsubv4si = 1475,
+ CODE_FOR_aarch64_uqsubv4si = 1476,
+ CODE_FOR_aarch64_sqaddv2di = 1477,
+ CODE_FOR_aarch64_uqaddv2di = 1478,
+ CODE_FOR_aarch64_sqsubv2di = 1479,
+ CODE_FOR_aarch64_uqsubv2di = 1480,
+ CODE_FOR_aarch64_sqaddqi = 1481,
+ CODE_FOR_aarch64_uqaddqi = 1482,
+ CODE_FOR_aarch64_sqsubqi = 1483,
+ CODE_FOR_aarch64_uqsubqi = 1484,
+ CODE_FOR_aarch64_sqaddhi = 1485,
+ CODE_FOR_aarch64_uqaddhi = 1486,
+ CODE_FOR_aarch64_sqsubhi = 1487,
+ CODE_FOR_aarch64_uqsubhi = 1488,
+ CODE_FOR_aarch64_sqaddsi = 1489,
+ CODE_FOR_aarch64_uqaddsi = 1490,
+ CODE_FOR_aarch64_sqsubsi = 1491,
+ CODE_FOR_aarch64_uqsubsi = 1492,
+ CODE_FOR_aarch64_sqadddi = 1493,
+ CODE_FOR_aarch64_uqadddi = 1494,
+ CODE_FOR_aarch64_sqsubdi = 1495,
+ CODE_FOR_aarch64_uqsubdi = 1496,
+ CODE_FOR_aarch64_suqaddv8qi = 1497,
+ CODE_FOR_aarch64_usqaddv8qi = 1498,
+ CODE_FOR_aarch64_suqaddv16qi = 1499,
+ CODE_FOR_aarch64_usqaddv16qi = 1500,
+ CODE_FOR_aarch64_suqaddv4hi = 1501,
+ CODE_FOR_aarch64_usqaddv4hi = 1502,
+ CODE_FOR_aarch64_suqaddv8hi = 1503,
+ CODE_FOR_aarch64_usqaddv8hi = 1504,
+ CODE_FOR_aarch64_suqaddv2si = 1505,
+ CODE_FOR_aarch64_usqaddv2si = 1506,
+ CODE_FOR_aarch64_suqaddv4si = 1507,
+ CODE_FOR_aarch64_usqaddv4si = 1508,
+ CODE_FOR_aarch64_suqaddv2di = 1509,
+ CODE_FOR_aarch64_usqaddv2di = 1510,
+ CODE_FOR_aarch64_suqaddqi = 1511,
+ CODE_FOR_aarch64_usqaddqi = 1512,
+ CODE_FOR_aarch64_suqaddhi = 1513,
+ CODE_FOR_aarch64_usqaddhi = 1514,
+ CODE_FOR_aarch64_suqaddsi = 1515,
+ CODE_FOR_aarch64_usqaddsi = 1516,
+ CODE_FOR_aarch64_suqadddi = 1517,
+ CODE_FOR_aarch64_usqadddi = 1518,
+ CODE_FOR_aarch64_sqmovunv8hi = 1519,
+ CODE_FOR_aarch64_sqmovunv4si = 1520,
+ CODE_FOR_aarch64_sqmovunv2di = 1521,
+ CODE_FOR_aarch64_sqmovunhi = 1522,
+ CODE_FOR_aarch64_sqmovunsi = 1523,
+ CODE_FOR_aarch64_sqmovundi = 1524,
+ CODE_FOR_aarch64_sqmovnv8hi = 1525,
+ CODE_FOR_aarch64_uqmovnv8hi = 1526,
+ CODE_FOR_aarch64_sqmovnv4si = 1527,
+ CODE_FOR_aarch64_uqmovnv4si = 1528,
+ CODE_FOR_aarch64_sqmovnv2di = 1529,
+ CODE_FOR_aarch64_uqmovnv2di = 1530,
+ CODE_FOR_aarch64_sqmovnhi = 1531,
+ CODE_FOR_aarch64_uqmovnhi = 1532,
+ CODE_FOR_aarch64_sqmovnsi = 1533,
+ CODE_FOR_aarch64_uqmovnsi = 1534,
+ CODE_FOR_aarch64_sqmovndi = 1535,
+ CODE_FOR_aarch64_uqmovndi = 1536,
+ CODE_FOR_aarch64_sqnegv8qi = 1537,
+ CODE_FOR_aarch64_sqabsv8qi = 1538,
+ CODE_FOR_aarch64_sqnegv16qi = 1539,
+ CODE_FOR_aarch64_sqabsv16qi = 1540,
+ CODE_FOR_aarch64_sqnegv4hi = 1541,
+ CODE_FOR_aarch64_sqabsv4hi = 1542,
+ CODE_FOR_aarch64_sqnegv8hi = 1543,
+ CODE_FOR_aarch64_sqabsv8hi = 1544,
+ CODE_FOR_aarch64_sqnegv2si = 1545,
+ CODE_FOR_aarch64_sqabsv2si = 1546,
+ CODE_FOR_aarch64_sqnegv4si = 1547,
+ CODE_FOR_aarch64_sqabsv4si = 1548,
+ CODE_FOR_aarch64_sqnegv2di = 1549,
+ CODE_FOR_aarch64_sqabsv2di = 1550,
+ CODE_FOR_aarch64_sqnegqi = 1551,
+ CODE_FOR_aarch64_sqabsqi = 1552,
+ CODE_FOR_aarch64_sqneghi = 1553,
+ CODE_FOR_aarch64_sqabshi = 1554,
+ CODE_FOR_aarch64_sqnegsi = 1555,
+ CODE_FOR_aarch64_sqabssi = 1556,
+ CODE_FOR_aarch64_sqnegdi = 1557,
+ CODE_FOR_aarch64_sqabsdi = 1558,
+ CODE_FOR_aarch64_sqdmulhv4hi = 1559,
+ CODE_FOR_aarch64_sqrdmulhv4hi = 1560,
+ CODE_FOR_aarch64_sqdmulhv8hi = 1561,
+ CODE_FOR_aarch64_sqrdmulhv8hi = 1562,
+ CODE_FOR_aarch64_sqdmulhv2si = 1563,
+ CODE_FOR_aarch64_sqrdmulhv2si = 1564,
+ CODE_FOR_aarch64_sqdmulhv4si = 1565,
+ CODE_FOR_aarch64_sqrdmulhv4si = 1566,
+ CODE_FOR_aarch64_sqdmulhhi = 1567,
+ CODE_FOR_aarch64_sqrdmulhhi = 1568,
+ CODE_FOR_aarch64_sqdmulhsi = 1569,
+ CODE_FOR_aarch64_sqrdmulhsi = 1570,
+ CODE_FOR_aarch64_sqdmulh_lanev4hi_internal = 1571,
+ CODE_FOR_aarch64_sqrdmulh_lanev4hi_internal = 1572,
+ CODE_FOR_aarch64_sqdmulh_lanev8hi_internal = 1573,
+ CODE_FOR_aarch64_sqrdmulh_lanev8hi_internal = 1574,
+ CODE_FOR_aarch64_sqdmulh_lanev2si_internal = 1575,
+ CODE_FOR_aarch64_sqrdmulh_lanev2si_internal = 1576,
+ CODE_FOR_aarch64_sqdmulh_lanev4si_internal = 1577,
+ CODE_FOR_aarch64_sqrdmulh_lanev4si_internal = 1578,
+ CODE_FOR_aarch64_sqdmulh_laneqv4hi_internal = 1579,
+ CODE_FOR_aarch64_sqrdmulh_laneqv4hi_internal = 1580,
+ CODE_FOR_aarch64_sqdmulh_laneqv8hi_internal = 1581,
+ CODE_FOR_aarch64_sqrdmulh_laneqv8hi_internal = 1582,
+ CODE_FOR_aarch64_sqdmulh_laneqv2si_internal = 1583,
+ CODE_FOR_aarch64_sqrdmulh_laneqv2si_internal = 1584,
+ CODE_FOR_aarch64_sqdmulh_laneqv4si_internal = 1585,
+ CODE_FOR_aarch64_sqrdmulh_laneqv4si_internal = 1586,
+ CODE_FOR_aarch64_sqdmulh_lanehi_internal = 1587,
+ CODE_FOR_aarch64_sqrdmulh_lanehi_internal = 1588,
+ CODE_FOR_aarch64_sqdmulh_lanesi_internal = 1589,
+ CODE_FOR_aarch64_sqrdmulh_lanesi_internal = 1590,
+ CODE_FOR_aarch64_sqdmlalv4hi = 1591,
+ CODE_FOR_aarch64_sqdmlslv4hi = 1592,
+ CODE_FOR_aarch64_sqdmlalv2si = 1593,
+ CODE_FOR_aarch64_sqdmlslv2si = 1594,
+ CODE_FOR_aarch64_sqdmlalhi = 1595,
+ CODE_FOR_aarch64_sqdmlslhi = 1596,
+ CODE_FOR_aarch64_sqdmlalsi = 1597,
+ CODE_FOR_aarch64_sqdmlslsi = 1598,
+ CODE_FOR_aarch64_sqdmlal_lanev4hi_internal = 1599,
+ CODE_FOR_aarch64_sqdmlsl_lanev4hi_internal = 1600,
+ CODE_FOR_aarch64_sqdmlal_lanev2si_internal = 1601,
+ CODE_FOR_aarch64_sqdmlsl_lanev2si_internal = 1602,
+ CODE_FOR_aarch64_sqdmlal_laneqv4hi_internal = 1603,
+ CODE_FOR_aarch64_sqdmlsl_laneqv4hi_internal = 1604,
+ CODE_FOR_aarch64_sqdmlal_laneqv2si_internal = 1605,
+ CODE_FOR_aarch64_sqdmlsl_laneqv2si_internal = 1606,
+ CODE_FOR_aarch64_sqdmlal_lanehi_internal = 1607,
+ CODE_FOR_aarch64_sqdmlsl_lanehi_internal = 1608,
+ CODE_FOR_aarch64_sqdmlal_lanesi_internal = 1609,
+ CODE_FOR_aarch64_sqdmlsl_lanesi_internal = 1610,
+ CODE_FOR_aarch64_sqdmlal_laneqhi_internal = 1611,
+ CODE_FOR_aarch64_sqdmlsl_laneqhi_internal = 1612,
+ CODE_FOR_aarch64_sqdmlal_laneqsi_internal = 1613,
+ CODE_FOR_aarch64_sqdmlsl_laneqsi_internal = 1614,
+ CODE_FOR_aarch64_sqdmlal_nv4hi = 1615,
+ CODE_FOR_aarch64_sqdmlsl_nv4hi = 1616,
+ CODE_FOR_aarch64_sqdmlal_nv2si = 1617,
+ CODE_FOR_aarch64_sqdmlsl_nv2si = 1618,
+ CODE_FOR_aarch64_sqdmlal2v8hi_internal = 1619,
+ CODE_FOR_aarch64_sqdmlsl2v8hi_internal = 1620,
+ CODE_FOR_aarch64_sqdmlal2v4si_internal = 1621,
+ CODE_FOR_aarch64_sqdmlsl2v4si_internal = 1622,
+ CODE_FOR_aarch64_sqdmlal2_lanev8hi_internal = 1623,
+ CODE_FOR_aarch64_sqdmlsl2_lanev8hi_internal = 1624,
+ CODE_FOR_aarch64_sqdmlal2_lanev4si_internal = 1625,
+ CODE_FOR_aarch64_sqdmlsl2_lanev4si_internal = 1626,
+ CODE_FOR_aarch64_sqdmlal2_laneqv8hi_internal = 1627,
+ CODE_FOR_aarch64_sqdmlsl2_laneqv8hi_internal = 1628,
+ CODE_FOR_aarch64_sqdmlal2_laneqv4si_internal = 1629,
+ CODE_FOR_aarch64_sqdmlsl2_laneqv4si_internal = 1630,
+ CODE_FOR_aarch64_sqdmlal2_nv8hi_internal = 1631,
+ CODE_FOR_aarch64_sqdmlsl2_nv8hi_internal = 1632,
+ CODE_FOR_aarch64_sqdmlal2_nv4si_internal = 1633,
+ CODE_FOR_aarch64_sqdmlsl2_nv4si_internal = 1634,
+ CODE_FOR_aarch64_sqdmullv4hi = 1635,
+ CODE_FOR_aarch64_sqdmullv2si = 1636,
+ CODE_FOR_aarch64_sqdmullhi = 1637,
+ CODE_FOR_aarch64_sqdmullsi = 1638,
+ CODE_FOR_aarch64_sqdmull_lanev4hi_internal = 1639,
+ CODE_FOR_aarch64_sqdmull_lanev2si_internal = 1640,
+ CODE_FOR_aarch64_sqdmull_laneqv4hi_internal = 1641,
+ CODE_FOR_aarch64_sqdmull_laneqv2si_internal = 1642,
+ CODE_FOR_aarch64_sqdmull_lanehi_internal = 1643,
+ CODE_FOR_aarch64_sqdmull_lanesi_internal = 1644,
+ CODE_FOR_aarch64_sqdmull_laneqhi_internal = 1645,
+ CODE_FOR_aarch64_sqdmull_laneqsi_internal = 1646,
+ CODE_FOR_aarch64_sqdmull_nv4hi = 1647,
+ CODE_FOR_aarch64_sqdmull_nv2si = 1648,
+ CODE_FOR_aarch64_sqdmull2v8hi_internal = 1649,
+ CODE_FOR_aarch64_sqdmull2v4si_internal = 1650,
+ CODE_FOR_aarch64_sqdmull2_lanev8hi_internal = 1651,
+ CODE_FOR_aarch64_sqdmull2_lanev4si_internal = 1652,
+ CODE_FOR_aarch64_sqdmull2_laneqv8hi_internal = 1653,
+ CODE_FOR_aarch64_sqdmull2_laneqv4si_internal = 1654,
+ CODE_FOR_aarch64_sqdmull2_nv8hi_internal = 1655,
+ CODE_FOR_aarch64_sqdmull2_nv4si_internal = 1656,
+ CODE_FOR_aarch64_sshlv8qi = 1657,
+ CODE_FOR_aarch64_ushlv8qi = 1658,
+ CODE_FOR_aarch64_srshlv8qi = 1659,
+ CODE_FOR_aarch64_urshlv8qi = 1660,
+ CODE_FOR_aarch64_sshlv16qi = 1661,
+ CODE_FOR_aarch64_ushlv16qi = 1662,
+ CODE_FOR_aarch64_srshlv16qi = 1663,
+ CODE_FOR_aarch64_urshlv16qi = 1664,
+ CODE_FOR_aarch64_sshlv4hi = 1665,
+ CODE_FOR_aarch64_ushlv4hi = 1666,
+ CODE_FOR_aarch64_srshlv4hi = 1667,
+ CODE_FOR_aarch64_urshlv4hi = 1668,
+ CODE_FOR_aarch64_sshlv8hi = 1669,
+ CODE_FOR_aarch64_ushlv8hi = 1670,
+ CODE_FOR_aarch64_srshlv8hi = 1671,
+ CODE_FOR_aarch64_urshlv8hi = 1672,
+ CODE_FOR_aarch64_sshlv2si = 1673,
+ CODE_FOR_aarch64_ushlv2si = 1674,
+ CODE_FOR_aarch64_srshlv2si = 1675,
+ CODE_FOR_aarch64_urshlv2si = 1676,
+ CODE_FOR_aarch64_sshlv4si = 1677,
+ CODE_FOR_aarch64_ushlv4si = 1678,
+ CODE_FOR_aarch64_srshlv4si = 1679,
+ CODE_FOR_aarch64_urshlv4si = 1680,
+ CODE_FOR_aarch64_sshlv2di = 1681,
+ CODE_FOR_aarch64_ushlv2di = 1682,
+ CODE_FOR_aarch64_srshlv2di = 1683,
+ CODE_FOR_aarch64_urshlv2di = 1684,
+ CODE_FOR_aarch64_sshldi = 1685,
+ CODE_FOR_aarch64_ushldi = 1686,
+ CODE_FOR_aarch64_srshldi = 1687,
+ CODE_FOR_aarch64_urshldi = 1688,
+ CODE_FOR_aarch64_sqshlv8qi = 1689,
+ CODE_FOR_aarch64_uqshlv8qi = 1690,
+ CODE_FOR_aarch64_sqrshlv8qi = 1691,
+ CODE_FOR_aarch64_uqrshlv8qi = 1692,
+ CODE_FOR_aarch64_sqshlv16qi = 1693,
+ CODE_FOR_aarch64_uqshlv16qi = 1694,
+ CODE_FOR_aarch64_sqrshlv16qi = 1695,
+ CODE_FOR_aarch64_uqrshlv16qi = 1696,
+ CODE_FOR_aarch64_sqshlv4hi = 1697,
+ CODE_FOR_aarch64_uqshlv4hi = 1698,
+ CODE_FOR_aarch64_sqrshlv4hi = 1699,
+ CODE_FOR_aarch64_uqrshlv4hi = 1700,
+ CODE_FOR_aarch64_sqshlv8hi = 1701,
+ CODE_FOR_aarch64_uqshlv8hi = 1702,
+ CODE_FOR_aarch64_sqrshlv8hi = 1703,
+ CODE_FOR_aarch64_uqrshlv8hi = 1704,
+ CODE_FOR_aarch64_sqshlv2si = 1705,
+ CODE_FOR_aarch64_uqshlv2si = 1706,
+ CODE_FOR_aarch64_sqrshlv2si = 1707,
+ CODE_FOR_aarch64_uqrshlv2si = 1708,
+ CODE_FOR_aarch64_sqshlv4si = 1709,
+ CODE_FOR_aarch64_uqshlv4si = 1710,
+ CODE_FOR_aarch64_sqrshlv4si = 1711,
+ CODE_FOR_aarch64_uqrshlv4si = 1712,
+ CODE_FOR_aarch64_sqshlv2di = 1713,
+ CODE_FOR_aarch64_uqshlv2di = 1714,
+ CODE_FOR_aarch64_sqrshlv2di = 1715,
+ CODE_FOR_aarch64_uqrshlv2di = 1716,
+ CODE_FOR_aarch64_sqshlqi = 1717,
+ CODE_FOR_aarch64_uqshlqi = 1718,
+ CODE_FOR_aarch64_sqrshlqi = 1719,
+ CODE_FOR_aarch64_uqrshlqi = 1720,
+ CODE_FOR_aarch64_sqshlhi = 1721,
+ CODE_FOR_aarch64_uqshlhi = 1722,
+ CODE_FOR_aarch64_sqrshlhi = 1723,
+ CODE_FOR_aarch64_uqrshlhi = 1724,
+ CODE_FOR_aarch64_sqshlsi = 1725,
+ CODE_FOR_aarch64_uqshlsi = 1726,
+ CODE_FOR_aarch64_sqrshlsi = 1727,
+ CODE_FOR_aarch64_uqrshlsi = 1728,
+ CODE_FOR_aarch64_sqshldi = 1729,
+ CODE_FOR_aarch64_uqshldi = 1730,
+ CODE_FOR_aarch64_sqrshldi = 1731,
+ CODE_FOR_aarch64_uqrshldi = 1732,
+ CODE_FOR_aarch64_sshll_nv8qi = 1733,
+ CODE_FOR_aarch64_ushll_nv8qi = 1734,
+ CODE_FOR_aarch64_sshll_nv4hi = 1735,
+ CODE_FOR_aarch64_ushll_nv4hi = 1736,
+ CODE_FOR_aarch64_sshll_nv2si = 1737,
+ CODE_FOR_aarch64_ushll_nv2si = 1738,
+ CODE_FOR_aarch64_sshll2_nv16qi = 1739,
+ CODE_FOR_aarch64_ushll2_nv16qi = 1740,
+ CODE_FOR_aarch64_sshll2_nv8hi = 1741,
+ CODE_FOR_aarch64_ushll2_nv8hi = 1742,
+ CODE_FOR_aarch64_sshll2_nv4si = 1743,
+ CODE_FOR_aarch64_ushll2_nv4si = 1744,
+ CODE_FOR_aarch64_srshr_nv8qi = 1745,
+ CODE_FOR_aarch64_urshr_nv8qi = 1746,
+ CODE_FOR_aarch64_srshr_nv16qi = 1747,
+ CODE_FOR_aarch64_urshr_nv16qi = 1748,
+ CODE_FOR_aarch64_srshr_nv4hi = 1749,
+ CODE_FOR_aarch64_urshr_nv4hi = 1750,
+ CODE_FOR_aarch64_srshr_nv8hi = 1751,
+ CODE_FOR_aarch64_urshr_nv8hi = 1752,
+ CODE_FOR_aarch64_srshr_nv2si = 1753,
+ CODE_FOR_aarch64_urshr_nv2si = 1754,
+ CODE_FOR_aarch64_srshr_nv4si = 1755,
+ CODE_FOR_aarch64_urshr_nv4si = 1756,
+ CODE_FOR_aarch64_srshr_nv2di = 1757,
+ CODE_FOR_aarch64_urshr_nv2di = 1758,
+ CODE_FOR_aarch64_srshr_ndi = 1759,
+ CODE_FOR_aarch64_urshr_ndi = 1760,
+ CODE_FOR_aarch64_ssra_nv8qi = 1761,
+ CODE_FOR_aarch64_usra_nv8qi = 1762,
+ CODE_FOR_aarch64_srsra_nv8qi = 1763,
+ CODE_FOR_aarch64_ursra_nv8qi = 1764,
+ CODE_FOR_aarch64_ssra_nv16qi = 1765,
+ CODE_FOR_aarch64_usra_nv16qi = 1766,
+ CODE_FOR_aarch64_srsra_nv16qi = 1767,
+ CODE_FOR_aarch64_ursra_nv16qi = 1768,
+ CODE_FOR_aarch64_ssra_nv4hi = 1769,
+ CODE_FOR_aarch64_usra_nv4hi = 1770,
+ CODE_FOR_aarch64_srsra_nv4hi = 1771,
+ CODE_FOR_aarch64_ursra_nv4hi = 1772,
+ CODE_FOR_aarch64_ssra_nv8hi = 1773,
+ CODE_FOR_aarch64_usra_nv8hi = 1774,
+ CODE_FOR_aarch64_srsra_nv8hi = 1775,
+ CODE_FOR_aarch64_ursra_nv8hi = 1776,
+ CODE_FOR_aarch64_ssra_nv2si = 1777,
+ CODE_FOR_aarch64_usra_nv2si = 1778,
+ CODE_FOR_aarch64_srsra_nv2si = 1779,
+ CODE_FOR_aarch64_ursra_nv2si = 1780,
+ CODE_FOR_aarch64_ssra_nv4si = 1781,
+ CODE_FOR_aarch64_usra_nv4si = 1782,
+ CODE_FOR_aarch64_srsra_nv4si = 1783,
+ CODE_FOR_aarch64_ursra_nv4si = 1784,
+ CODE_FOR_aarch64_ssra_nv2di = 1785,
+ CODE_FOR_aarch64_usra_nv2di = 1786,
+ CODE_FOR_aarch64_srsra_nv2di = 1787,
+ CODE_FOR_aarch64_ursra_nv2di = 1788,
+ CODE_FOR_aarch64_ssra_ndi = 1789,
+ CODE_FOR_aarch64_usra_ndi = 1790,
+ CODE_FOR_aarch64_srsra_ndi = 1791,
+ CODE_FOR_aarch64_ursra_ndi = 1792,
+ CODE_FOR_aarch64_ssli_nv8qi = 1793,
+ CODE_FOR_aarch64_usli_nv8qi = 1794,
+ CODE_FOR_aarch64_ssri_nv8qi = 1795,
+ CODE_FOR_aarch64_usri_nv8qi = 1796,
+ CODE_FOR_aarch64_ssli_nv16qi = 1797,
+ CODE_FOR_aarch64_usli_nv16qi = 1798,
+ CODE_FOR_aarch64_ssri_nv16qi = 1799,
+ CODE_FOR_aarch64_usri_nv16qi = 1800,
+ CODE_FOR_aarch64_ssli_nv4hi = 1801,
+ CODE_FOR_aarch64_usli_nv4hi = 1802,
+ CODE_FOR_aarch64_ssri_nv4hi = 1803,
+ CODE_FOR_aarch64_usri_nv4hi = 1804,
+ CODE_FOR_aarch64_ssli_nv8hi = 1805,
+ CODE_FOR_aarch64_usli_nv8hi = 1806,
+ CODE_FOR_aarch64_ssri_nv8hi = 1807,
+ CODE_FOR_aarch64_usri_nv8hi = 1808,
+ CODE_FOR_aarch64_ssli_nv2si = 1809,
+ CODE_FOR_aarch64_usli_nv2si = 1810,
+ CODE_FOR_aarch64_ssri_nv2si = 1811,
+ CODE_FOR_aarch64_usri_nv2si = 1812,
+ CODE_FOR_aarch64_ssli_nv4si = 1813,
+ CODE_FOR_aarch64_usli_nv4si = 1814,
+ CODE_FOR_aarch64_ssri_nv4si = 1815,
+ CODE_FOR_aarch64_usri_nv4si = 1816,
+ CODE_FOR_aarch64_ssli_nv2di = 1817,
+ CODE_FOR_aarch64_usli_nv2di = 1818,
+ CODE_FOR_aarch64_ssri_nv2di = 1819,
+ CODE_FOR_aarch64_usri_nv2di = 1820,
+ CODE_FOR_aarch64_ssli_ndi = 1821,
+ CODE_FOR_aarch64_usli_ndi = 1822,
+ CODE_FOR_aarch64_ssri_ndi = 1823,
+ CODE_FOR_aarch64_usri_ndi = 1824,
+ CODE_FOR_aarch64_sqshlu_nv8qi = 1825,
+ CODE_FOR_aarch64_sqshl_nv8qi = 1826,
+ CODE_FOR_aarch64_uqshl_nv8qi = 1827,
+ CODE_FOR_aarch64_sqshlu_nv16qi = 1828,
+ CODE_FOR_aarch64_sqshl_nv16qi = 1829,
+ CODE_FOR_aarch64_uqshl_nv16qi = 1830,
+ CODE_FOR_aarch64_sqshlu_nv4hi = 1831,
+ CODE_FOR_aarch64_sqshl_nv4hi = 1832,
+ CODE_FOR_aarch64_uqshl_nv4hi = 1833,
+ CODE_FOR_aarch64_sqshlu_nv8hi = 1834,
+ CODE_FOR_aarch64_sqshl_nv8hi = 1835,
+ CODE_FOR_aarch64_uqshl_nv8hi = 1836,
+ CODE_FOR_aarch64_sqshlu_nv2si = 1837,
+ CODE_FOR_aarch64_sqshl_nv2si = 1838,
+ CODE_FOR_aarch64_uqshl_nv2si = 1839,
+ CODE_FOR_aarch64_sqshlu_nv4si = 1840,
+ CODE_FOR_aarch64_sqshl_nv4si = 1841,
+ CODE_FOR_aarch64_uqshl_nv4si = 1842,
+ CODE_FOR_aarch64_sqshlu_nv2di = 1843,
+ CODE_FOR_aarch64_sqshl_nv2di = 1844,
+ CODE_FOR_aarch64_uqshl_nv2di = 1845,
+ CODE_FOR_aarch64_sqshlu_nqi = 1846,
+ CODE_FOR_aarch64_sqshl_nqi = 1847,
+ CODE_FOR_aarch64_uqshl_nqi = 1848,
+ CODE_FOR_aarch64_sqshlu_nhi = 1849,
+ CODE_FOR_aarch64_sqshl_nhi = 1850,
+ CODE_FOR_aarch64_uqshl_nhi = 1851,
+ CODE_FOR_aarch64_sqshlu_nsi = 1852,
+ CODE_FOR_aarch64_sqshl_nsi = 1853,
+ CODE_FOR_aarch64_uqshl_nsi = 1854,
+ CODE_FOR_aarch64_sqshlu_ndi = 1855,
+ CODE_FOR_aarch64_sqshl_ndi = 1856,
+ CODE_FOR_aarch64_uqshl_ndi = 1857,
+ CODE_FOR_aarch64_sqshrun_nv8hi = 1858,
+ CODE_FOR_aarch64_sqrshrun_nv8hi = 1859,
+ CODE_FOR_aarch64_sqshrn_nv8hi = 1860,
+ CODE_FOR_aarch64_uqshrn_nv8hi = 1861,
+ CODE_FOR_aarch64_sqrshrn_nv8hi = 1862,
+ CODE_FOR_aarch64_uqrshrn_nv8hi = 1863,
+ CODE_FOR_aarch64_sqshrun_nv4si = 1864,
+ CODE_FOR_aarch64_sqrshrun_nv4si = 1865,
+ CODE_FOR_aarch64_sqshrn_nv4si = 1866,
+ CODE_FOR_aarch64_uqshrn_nv4si = 1867,
+ CODE_FOR_aarch64_sqrshrn_nv4si = 1868,
+ CODE_FOR_aarch64_uqrshrn_nv4si = 1869,
+ CODE_FOR_aarch64_sqshrun_nv2di = 1870,
+ CODE_FOR_aarch64_sqrshrun_nv2di = 1871,
+ CODE_FOR_aarch64_sqshrn_nv2di = 1872,
+ CODE_FOR_aarch64_uqshrn_nv2di = 1873,
+ CODE_FOR_aarch64_sqrshrn_nv2di = 1874,
+ CODE_FOR_aarch64_uqrshrn_nv2di = 1875,
+ CODE_FOR_aarch64_sqshrun_nhi = 1876,
+ CODE_FOR_aarch64_sqrshrun_nhi = 1877,
+ CODE_FOR_aarch64_sqshrn_nhi = 1878,
+ CODE_FOR_aarch64_uqshrn_nhi = 1879,
+ CODE_FOR_aarch64_sqrshrn_nhi = 1880,
+ CODE_FOR_aarch64_uqrshrn_nhi = 1881,
+ CODE_FOR_aarch64_sqshrun_nsi = 1882,
+ CODE_FOR_aarch64_sqrshrun_nsi = 1883,
+ CODE_FOR_aarch64_sqshrn_nsi = 1884,
+ CODE_FOR_aarch64_uqshrn_nsi = 1885,
+ CODE_FOR_aarch64_sqrshrn_nsi = 1886,
+ CODE_FOR_aarch64_uqrshrn_nsi = 1887,
+ CODE_FOR_aarch64_sqshrun_ndi = 1888,
+ CODE_FOR_aarch64_sqrshrun_ndi = 1889,
+ CODE_FOR_aarch64_sqshrn_ndi = 1890,
+ CODE_FOR_aarch64_uqshrn_ndi = 1891,
+ CODE_FOR_aarch64_sqrshrn_ndi = 1892,
+ CODE_FOR_aarch64_uqrshrn_ndi = 1893,
+ CODE_FOR_aarch64_cmltv8qi = 1894,
+ CODE_FOR_aarch64_cmlev8qi = 1895,
+ CODE_FOR_aarch64_cmeqv8qi = 1896,
+ CODE_FOR_aarch64_cmgev8qi = 1897,
+ CODE_FOR_aarch64_cmgtv8qi = 1898,
+ CODE_FOR_aarch64_cmltv16qi = 1899,
+ CODE_FOR_aarch64_cmlev16qi = 1900,
+ CODE_FOR_aarch64_cmeqv16qi = 1901,
+ CODE_FOR_aarch64_cmgev16qi = 1902,
+ CODE_FOR_aarch64_cmgtv16qi = 1903,
+ CODE_FOR_aarch64_cmltv4hi = 1904,
+ CODE_FOR_aarch64_cmlev4hi = 1905,
+ CODE_FOR_aarch64_cmeqv4hi = 1906,
+ CODE_FOR_aarch64_cmgev4hi = 1907,
+ CODE_FOR_aarch64_cmgtv4hi = 1908,
+ CODE_FOR_aarch64_cmltv8hi = 1909,
+ CODE_FOR_aarch64_cmlev8hi = 1910,
+ CODE_FOR_aarch64_cmeqv8hi = 1911,
+ CODE_FOR_aarch64_cmgev8hi = 1912,
+ CODE_FOR_aarch64_cmgtv8hi = 1913,
+ CODE_FOR_aarch64_cmltv2si = 1914,
+ CODE_FOR_aarch64_cmlev2si = 1915,
+ CODE_FOR_aarch64_cmeqv2si = 1916,
+ CODE_FOR_aarch64_cmgev2si = 1917,
+ CODE_FOR_aarch64_cmgtv2si = 1918,
+ CODE_FOR_aarch64_cmltv4si = 1919,
+ CODE_FOR_aarch64_cmlev4si = 1920,
+ CODE_FOR_aarch64_cmeqv4si = 1921,
+ CODE_FOR_aarch64_cmgev4si = 1922,
+ CODE_FOR_aarch64_cmgtv4si = 1923,
+ CODE_FOR_aarch64_cmltv2di = 1924,
+ CODE_FOR_aarch64_cmlev2di = 1925,
+ CODE_FOR_aarch64_cmeqv2di = 1926,
+ CODE_FOR_aarch64_cmgev2di = 1927,
+ CODE_FOR_aarch64_cmgtv2di = 1928,
+ CODE_FOR_aarch64_cmltdi = 1929,
+ CODE_FOR_aarch64_cmledi = 1930,
+ CODE_FOR_aarch64_cmeqdi = 1931,
+ CODE_FOR_aarch64_cmgedi = 1932,
+ CODE_FOR_aarch64_cmgtdi = 1933,
+ CODE_FOR_aarch64_cmltuv8qi = 1939,
+ CODE_FOR_aarch64_cmleuv8qi = 1940,
+ CODE_FOR_aarch64_cmgeuv8qi = 1941,
+ CODE_FOR_aarch64_cmgtuv8qi = 1942,
+ CODE_FOR_aarch64_cmltuv16qi = 1943,
+ CODE_FOR_aarch64_cmleuv16qi = 1944,
+ CODE_FOR_aarch64_cmgeuv16qi = 1945,
+ CODE_FOR_aarch64_cmgtuv16qi = 1946,
+ CODE_FOR_aarch64_cmltuv4hi = 1947,
+ CODE_FOR_aarch64_cmleuv4hi = 1948,
+ CODE_FOR_aarch64_cmgeuv4hi = 1949,
+ CODE_FOR_aarch64_cmgtuv4hi = 1950,
+ CODE_FOR_aarch64_cmltuv8hi = 1951,
+ CODE_FOR_aarch64_cmleuv8hi = 1952,
+ CODE_FOR_aarch64_cmgeuv8hi = 1953,
+ CODE_FOR_aarch64_cmgtuv8hi = 1954,
+ CODE_FOR_aarch64_cmltuv2si = 1955,
+ CODE_FOR_aarch64_cmleuv2si = 1956,
+ CODE_FOR_aarch64_cmgeuv2si = 1957,
+ CODE_FOR_aarch64_cmgtuv2si = 1958,
+ CODE_FOR_aarch64_cmltuv4si = 1959,
+ CODE_FOR_aarch64_cmleuv4si = 1960,
+ CODE_FOR_aarch64_cmgeuv4si = 1961,
+ CODE_FOR_aarch64_cmgtuv4si = 1962,
+ CODE_FOR_aarch64_cmltuv2di = 1963,
+ CODE_FOR_aarch64_cmleuv2di = 1964,
+ CODE_FOR_aarch64_cmgeuv2di = 1965,
+ CODE_FOR_aarch64_cmgtuv2di = 1966,
+ CODE_FOR_aarch64_cmltudi = 1967,
+ CODE_FOR_aarch64_cmleudi = 1968,
+ CODE_FOR_aarch64_cmgeudi = 1969,
+ CODE_FOR_aarch64_cmgtudi = 1970,
+ CODE_FOR_aarch64_cmtstv8qi = 1975,
+ CODE_FOR_aarch64_cmtstv16qi = 1976,
+ CODE_FOR_aarch64_cmtstv4hi = 1977,
+ CODE_FOR_aarch64_cmtstv8hi = 1978,
+ CODE_FOR_aarch64_cmtstv2si = 1979,
+ CODE_FOR_aarch64_cmtstv4si = 1980,
+ CODE_FOR_aarch64_cmtstv2di = 1981,
+ CODE_FOR_aarch64_cmtstdi = 1982,
+ CODE_FOR_aarch64_cmltv2sf = 1984,
+ CODE_FOR_aarch64_cmlev2sf = 1985,
+ CODE_FOR_aarch64_cmeqv2sf = 1986,
+ CODE_FOR_aarch64_cmgev2sf = 1987,
+ CODE_FOR_aarch64_cmgtv2sf = 1988,
+ CODE_FOR_aarch64_cmltv4sf = 1989,
+ CODE_FOR_aarch64_cmlev4sf = 1990,
+ CODE_FOR_aarch64_cmeqv4sf = 1991,
+ CODE_FOR_aarch64_cmgev4sf = 1992,
+ CODE_FOR_aarch64_cmgtv4sf = 1993,
+ CODE_FOR_aarch64_cmltv2df = 1994,
+ CODE_FOR_aarch64_cmlev2df = 1995,
+ CODE_FOR_aarch64_cmeqv2df = 1996,
+ CODE_FOR_aarch64_cmgev2df = 1997,
+ CODE_FOR_aarch64_cmgtv2df = 1998,
+ CODE_FOR_aarch64_cmltsf = 1999,
+ CODE_FOR_aarch64_cmlesf = 2000,
+ CODE_FOR_aarch64_cmeqsf = 2001,
+ CODE_FOR_aarch64_cmgesf = 2002,
+ CODE_FOR_aarch64_cmgtsf = 2003,
+ CODE_FOR_aarch64_cmltdf = 2004,
+ CODE_FOR_aarch64_cmledf = 2005,
+ CODE_FOR_aarch64_cmeqdf = 2006,
+ CODE_FOR_aarch64_cmgedf = 2007,
+ CODE_FOR_aarch64_cmgtdf = 2008,
+ CODE_FOR_aarch64_addpv8qi = 2029,
+ CODE_FOR_aarch64_addpv4hi = 2030,
+ CODE_FOR_aarch64_addpv2si = 2031,
+ CODE_FOR_aarch64_addpdi = 2032,
+ CODE_FOR_sqrtv2sf2 = 2033,
+ CODE_FOR_sqrtv4sf2 = 2034,
+ CODE_FOR_sqrtv2df2 = 2035,
+ CODE_FOR_vec_load_lanesoiv16qi = 2036,
+ CODE_FOR_vec_load_lanesoiv8hi = 2037,
+ CODE_FOR_vec_load_lanesoiv4si = 2038,
+ CODE_FOR_vec_load_lanesoiv2di = 2039,
+ CODE_FOR_vec_load_lanesoiv4sf = 2040,
+ CODE_FOR_vec_load_lanesoiv2df = 2041,
+ CODE_FOR_vec_store_lanesoiv16qi = 2042,
+ CODE_FOR_vec_store_lanesoiv8hi = 2043,
+ CODE_FOR_vec_store_lanesoiv4si = 2044,
+ CODE_FOR_vec_store_lanesoiv2di = 2045,
+ CODE_FOR_vec_store_lanesoiv4sf = 2046,
+ CODE_FOR_vec_store_lanesoiv2df = 2047,
+ CODE_FOR_vec_store_lanesoi_lanev16qi = 2048,
+ CODE_FOR_vec_store_lanesoi_lanev8hi = 2049,
+ CODE_FOR_vec_store_lanesoi_lanev4si = 2050,
+ CODE_FOR_vec_store_lanesoi_lanev2di = 2051,
+ CODE_FOR_vec_store_lanesoi_lanev4sf = 2052,
+ CODE_FOR_vec_store_lanesoi_lanev2df = 2053,
+ CODE_FOR_vec_load_lanesciv16qi = 2054,
+ CODE_FOR_vec_load_lanesciv8hi = 2055,
+ CODE_FOR_vec_load_lanesciv4si = 2056,
+ CODE_FOR_vec_load_lanesciv2di = 2057,
+ CODE_FOR_vec_load_lanesciv4sf = 2058,
+ CODE_FOR_vec_load_lanesciv2df = 2059,
+ CODE_FOR_vec_store_lanesciv16qi = 2060,
+ CODE_FOR_vec_store_lanesciv8hi = 2061,
+ CODE_FOR_vec_store_lanesciv4si = 2062,
+ CODE_FOR_vec_store_lanesciv2di = 2063,
+ CODE_FOR_vec_store_lanesciv4sf = 2064,
+ CODE_FOR_vec_store_lanesciv2df = 2065,
+ CODE_FOR_vec_store_lanesci_lanev16qi = 2066,
+ CODE_FOR_vec_store_lanesci_lanev8hi = 2067,
+ CODE_FOR_vec_store_lanesci_lanev4si = 2068,
+ CODE_FOR_vec_store_lanesci_lanev2di = 2069,
+ CODE_FOR_vec_store_lanesci_lanev4sf = 2070,
+ CODE_FOR_vec_store_lanesci_lanev2df = 2071,
+ CODE_FOR_vec_load_lanesxiv16qi = 2072,
+ CODE_FOR_vec_load_lanesxiv8hi = 2073,
+ CODE_FOR_vec_load_lanesxiv4si = 2074,
+ CODE_FOR_vec_load_lanesxiv2di = 2075,
+ CODE_FOR_vec_load_lanesxiv4sf = 2076,
+ CODE_FOR_vec_load_lanesxiv2df = 2077,
+ CODE_FOR_vec_store_lanesxiv16qi = 2078,
+ CODE_FOR_vec_store_lanesxiv8hi = 2079,
+ CODE_FOR_vec_store_lanesxiv4si = 2080,
+ CODE_FOR_vec_store_lanesxiv2di = 2081,
+ CODE_FOR_vec_store_lanesxiv4sf = 2082,
+ CODE_FOR_vec_store_lanesxiv2df = 2083,
+ CODE_FOR_vec_store_lanesxi_lanev16qi = 2084,
+ CODE_FOR_vec_store_lanesxi_lanev8hi = 2085,
+ CODE_FOR_vec_store_lanesxi_lanev4si = 2086,
+ CODE_FOR_vec_store_lanesxi_lanev2di = 2087,
+ CODE_FOR_vec_store_lanesxi_lanev4sf = 2088,
+ CODE_FOR_vec_store_lanesxi_lanev2df = 2089,
+ CODE_FOR_aarch64_be_ld1v8qi = 2093,
+ CODE_FOR_aarch64_be_ld1v16qi = 2094,
+ CODE_FOR_aarch64_be_ld1v4hi = 2095,
+ CODE_FOR_aarch64_be_ld1v8hi = 2096,
+ CODE_FOR_aarch64_be_ld1v2si = 2097,
+ CODE_FOR_aarch64_be_ld1v4si = 2098,
+ CODE_FOR_aarch64_be_ld1v2di = 2099,
+ CODE_FOR_aarch64_be_ld1v2sf = 2100,
+ CODE_FOR_aarch64_be_ld1v4sf = 2101,
+ CODE_FOR_aarch64_be_ld1v2df = 2102,
+ CODE_FOR_aarch64_be_ld1di = 2103,
+ CODE_FOR_aarch64_be_st1v8qi = 2104,
+ CODE_FOR_aarch64_be_st1v16qi = 2105,
+ CODE_FOR_aarch64_be_st1v4hi = 2106,
+ CODE_FOR_aarch64_be_st1v8hi = 2107,
+ CODE_FOR_aarch64_be_st1v2si = 2108,
+ CODE_FOR_aarch64_be_st1v4si = 2109,
+ CODE_FOR_aarch64_be_st1v2di = 2110,
+ CODE_FOR_aarch64_be_st1v2sf = 2111,
+ CODE_FOR_aarch64_be_st1v4sf = 2112,
+ CODE_FOR_aarch64_be_st1v2df = 2113,
+ CODE_FOR_aarch64_be_st1di = 2114,
+ CODE_FOR_aarch64_ld2v8qi_dreg = 2115,
+ CODE_FOR_aarch64_ld2v4hi_dreg = 2116,
+ CODE_FOR_aarch64_ld2v2si_dreg = 2117,
+ CODE_FOR_aarch64_ld2v2sf_dreg = 2118,
+ CODE_FOR_aarch64_ld2di_dreg = 2119,
+ CODE_FOR_aarch64_ld2df_dreg = 2120,
+ CODE_FOR_aarch64_ld3v8qi_dreg = 2121,
+ CODE_FOR_aarch64_ld3v4hi_dreg = 2122,
+ CODE_FOR_aarch64_ld3v2si_dreg = 2123,
+ CODE_FOR_aarch64_ld3v2sf_dreg = 2124,
+ CODE_FOR_aarch64_ld3di_dreg = 2125,
+ CODE_FOR_aarch64_ld3df_dreg = 2126,
+ CODE_FOR_aarch64_ld4v8qi_dreg = 2127,
+ CODE_FOR_aarch64_ld4v4hi_dreg = 2128,
+ CODE_FOR_aarch64_ld4v2si_dreg = 2129,
+ CODE_FOR_aarch64_ld4v2sf_dreg = 2130,
+ CODE_FOR_aarch64_ld4di_dreg = 2131,
+ CODE_FOR_aarch64_ld4df_dreg = 2132,
+ CODE_FOR_aarch64_tbl1v8qi = 2133,
+ CODE_FOR_aarch64_tbl1v16qi = 2134,
+ CODE_FOR_aarch64_tbl2v16qi = 2135,
+ CODE_FOR_aarch64_combinev16qi = 2136,
+ CODE_FOR_aarch64_zip1v8qi = 2137,
+ CODE_FOR_aarch64_zip2v8qi = 2138,
+ CODE_FOR_aarch64_trn1v8qi = 2139,
+ CODE_FOR_aarch64_trn2v8qi = 2140,
+ CODE_FOR_aarch64_uzp1v8qi = 2141,
+ CODE_FOR_aarch64_uzp2v8qi = 2142,
+ CODE_FOR_aarch64_zip1v16qi = 2143,
+ CODE_FOR_aarch64_zip2v16qi = 2144,
+ CODE_FOR_aarch64_trn1v16qi = 2145,
+ CODE_FOR_aarch64_trn2v16qi = 2146,
+ CODE_FOR_aarch64_uzp1v16qi = 2147,
+ CODE_FOR_aarch64_uzp2v16qi = 2148,
+ CODE_FOR_aarch64_zip1v4hi = 2149,
+ CODE_FOR_aarch64_zip2v4hi = 2150,
+ CODE_FOR_aarch64_trn1v4hi = 2151,
+ CODE_FOR_aarch64_trn2v4hi = 2152,
+ CODE_FOR_aarch64_uzp1v4hi = 2153,
+ CODE_FOR_aarch64_uzp2v4hi = 2154,
+ CODE_FOR_aarch64_zip1v8hi = 2155,
+ CODE_FOR_aarch64_zip2v8hi = 2156,
+ CODE_FOR_aarch64_trn1v8hi = 2157,
+ CODE_FOR_aarch64_trn2v8hi = 2158,
+ CODE_FOR_aarch64_uzp1v8hi = 2159,
+ CODE_FOR_aarch64_uzp2v8hi = 2160,
+ CODE_FOR_aarch64_zip1v2si = 2161,
+ CODE_FOR_aarch64_zip2v2si = 2162,
+ CODE_FOR_aarch64_trn1v2si = 2163,
+ CODE_FOR_aarch64_trn2v2si = 2164,
+ CODE_FOR_aarch64_uzp1v2si = 2165,
+ CODE_FOR_aarch64_uzp2v2si = 2166,
+ CODE_FOR_aarch64_zip1v4si = 2167,
+ CODE_FOR_aarch64_zip2v4si = 2168,
+ CODE_FOR_aarch64_trn1v4si = 2169,
+ CODE_FOR_aarch64_trn2v4si = 2170,
+ CODE_FOR_aarch64_uzp1v4si = 2171,
+ CODE_FOR_aarch64_uzp2v4si = 2172,
+ CODE_FOR_aarch64_zip1v2di = 2173,
+ CODE_FOR_aarch64_zip2v2di = 2174,
+ CODE_FOR_aarch64_trn1v2di = 2175,
+ CODE_FOR_aarch64_trn2v2di = 2176,
+ CODE_FOR_aarch64_uzp1v2di = 2177,
+ CODE_FOR_aarch64_uzp2v2di = 2178,
+ CODE_FOR_aarch64_zip1v2sf = 2179,
+ CODE_FOR_aarch64_zip2v2sf = 2180,
+ CODE_FOR_aarch64_trn1v2sf = 2181,
+ CODE_FOR_aarch64_trn2v2sf = 2182,
+ CODE_FOR_aarch64_uzp1v2sf = 2183,
+ CODE_FOR_aarch64_uzp2v2sf = 2184,
+ CODE_FOR_aarch64_zip1v4sf = 2185,
+ CODE_FOR_aarch64_zip2v4sf = 2186,
+ CODE_FOR_aarch64_trn1v4sf = 2187,
+ CODE_FOR_aarch64_trn2v4sf = 2188,
+ CODE_FOR_aarch64_uzp1v4sf = 2189,
+ CODE_FOR_aarch64_uzp2v4sf = 2190,
+ CODE_FOR_aarch64_zip1v2df = 2191,
+ CODE_FOR_aarch64_zip2v2df = 2192,
+ CODE_FOR_aarch64_trn1v2df = 2193,
+ CODE_FOR_aarch64_trn2v2df = 2194,
+ CODE_FOR_aarch64_uzp1v2df = 2195,
+ CODE_FOR_aarch64_uzp2v2df = 2196,
+ CODE_FOR_aarch64_extv8qi = 2197,
+ CODE_FOR_aarch64_extv16qi = 2198,
+ CODE_FOR_aarch64_extv4hi = 2199,
+ CODE_FOR_aarch64_extv8hi = 2200,
+ CODE_FOR_aarch64_extv2si = 2201,
+ CODE_FOR_aarch64_extv4si = 2202,
+ CODE_FOR_aarch64_extv2di = 2203,
+ CODE_FOR_aarch64_extv2sf = 2204,
+ CODE_FOR_aarch64_extv4sf = 2205,
+ CODE_FOR_aarch64_extv2df = 2206,
+ CODE_FOR_aarch64_rev64v8qi = 2207,
+ CODE_FOR_aarch64_rev32v8qi = 2208,
+ CODE_FOR_aarch64_rev16v8qi = 2209,
+ CODE_FOR_aarch64_rev64v16qi = 2210,
+ CODE_FOR_aarch64_rev32v16qi = 2211,
+ CODE_FOR_aarch64_rev16v16qi = 2212,
+ CODE_FOR_aarch64_rev64v4hi = 2213,
+ CODE_FOR_aarch64_rev32v4hi = 2214,
+ CODE_FOR_aarch64_rev16v4hi = 2215,
+ CODE_FOR_aarch64_rev64v8hi = 2216,
+ CODE_FOR_aarch64_rev32v8hi = 2217,
+ CODE_FOR_aarch64_rev16v8hi = 2218,
+ CODE_FOR_aarch64_rev64v2si = 2219,
+ CODE_FOR_aarch64_rev32v2si = 2220,
+ CODE_FOR_aarch64_rev16v2si = 2221,
+ CODE_FOR_aarch64_rev64v4si = 2222,
+ CODE_FOR_aarch64_rev32v4si = 2223,
+ CODE_FOR_aarch64_rev16v4si = 2224,
+ CODE_FOR_aarch64_rev64v2di = 2225,
+ CODE_FOR_aarch64_rev32v2di = 2226,
+ CODE_FOR_aarch64_rev16v2di = 2227,
+ CODE_FOR_aarch64_rev64v2sf = 2228,
+ CODE_FOR_aarch64_rev32v2sf = 2229,
+ CODE_FOR_aarch64_rev16v2sf = 2230,
+ CODE_FOR_aarch64_rev64v4sf = 2231,
+ CODE_FOR_aarch64_rev32v4sf = 2232,
+ CODE_FOR_aarch64_rev16v4sf = 2233,
+ CODE_FOR_aarch64_rev64v2df = 2234,
+ CODE_FOR_aarch64_rev32v2df = 2235,
+ CODE_FOR_aarch64_rev16v2df = 2236,
+ CODE_FOR_aarch64_st2v8qi_dreg = 2237,
+ CODE_FOR_aarch64_st2v4hi_dreg = 2238,
+ CODE_FOR_aarch64_st2v2si_dreg = 2239,
+ CODE_FOR_aarch64_st2v2sf_dreg = 2240,
+ CODE_FOR_aarch64_st2di_dreg = 2241,
+ CODE_FOR_aarch64_st2df_dreg = 2242,
+ CODE_FOR_aarch64_st3v8qi_dreg = 2243,
+ CODE_FOR_aarch64_st3v4hi_dreg = 2244,
+ CODE_FOR_aarch64_st3v2si_dreg = 2245,
+ CODE_FOR_aarch64_st3v2sf_dreg = 2246,
+ CODE_FOR_aarch64_st3di_dreg = 2247,
+ CODE_FOR_aarch64_st3df_dreg = 2248,
+ CODE_FOR_aarch64_st4v8qi_dreg = 2249,
+ CODE_FOR_aarch64_st4v4hi_dreg = 2250,
+ CODE_FOR_aarch64_st4v2si_dreg = 2251,
+ CODE_FOR_aarch64_st4v2sf_dreg = 2252,
+ CODE_FOR_aarch64_st4di_dreg = 2253,
+ CODE_FOR_aarch64_st4df_dreg = 2254,
+ CODE_FOR_aarch64_frecpev2sf = 2266,
+ CODE_FOR_aarch64_frecpev4sf = 2267,
+ CODE_FOR_aarch64_frecpev2df = 2268,
+ CODE_FOR_aarch64_frecpesf = 2269,
+ CODE_FOR_aarch64_frecpxsf = 2270,
+ CODE_FOR_aarch64_frecpedf = 2271,
+ CODE_FOR_aarch64_frecpxdf = 2272,
+ CODE_FOR_aarch64_frecpsv2sf = 2273,
+ CODE_FOR_aarch64_frecpsv4sf = 2274,
+ CODE_FOR_aarch64_frecpsv2df = 2275,
+ CODE_FOR_aarch64_frecpssf = 2276,
+ CODE_FOR_aarch64_frecpsdf = 2277,
+ CODE_FOR_aarch64_crypto_aesev16qi = 2278,
+ CODE_FOR_aarch64_crypto_aesdv16qi = 2279,
+ CODE_FOR_aarch64_crypto_aesmcv16qi = 2280,
+ CODE_FOR_aarch64_crypto_aesimcv16qi = 2281,
+ CODE_FOR_aarch64_crypto_sha1hsi = 2282,
+ CODE_FOR_aarch64_crypto_sha1su1v4si = 2283,
+ CODE_FOR_aarch64_crypto_sha1cv4si = 2284,
+ CODE_FOR_aarch64_crypto_sha1mv4si = 2285,
+ CODE_FOR_aarch64_crypto_sha1pv4si = 2286,
+ CODE_FOR_aarch64_crypto_sha1su0v4si = 2287,
+ CODE_FOR_aarch64_crypto_sha256hv4si = 2288,
+ CODE_FOR_aarch64_crypto_sha256h2v4si = 2289,
+ CODE_FOR_aarch64_crypto_sha256su0v4si = 2290,
+ CODE_FOR_aarch64_crypto_sha256su1v4si = 2291,
+ CODE_FOR_aarch64_crypto_pmulldi = 2292,
+ CODE_FOR_aarch64_crypto_pmullv2di = 2293,
+ CODE_FOR_atomic_compare_and_swapqi_1 = 2294,
+ CODE_FOR_atomic_compare_and_swaphi_1 = 2295,
+ CODE_FOR_atomic_compare_and_swapsi_1 = 2296,
+ CODE_FOR_atomic_compare_and_swapdi_1 = 2297,
+ CODE_FOR_atomic_exchangeqi = 2298,
+ CODE_FOR_atomic_exchangehi = 2299,
+ CODE_FOR_atomic_exchangesi = 2300,
+ CODE_FOR_atomic_exchangedi = 2301,
+ CODE_FOR_atomic_addqi = 2302,
+ CODE_FOR_atomic_subqi = 2303,
+ CODE_FOR_atomic_orqi = 2304,
+ CODE_FOR_atomic_xorqi = 2305,
+ CODE_FOR_atomic_andqi = 2306,
+ CODE_FOR_atomic_addhi = 2307,
+ CODE_FOR_atomic_subhi = 2308,
+ CODE_FOR_atomic_orhi = 2309,
+ CODE_FOR_atomic_xorhi = 2310,
+ CODE_FOR_atomic_andhi = 2311,
+ CODE_FOR_atomic_addsi = 2312,
+ CODE_FOR_atomic_subsi = 2313,
+ CODE_FOR_atomic_orsi = 2314,
+ CODE_FOR_atomic_xorsi = 2315,
+ CODE_FOR_atomic_andsi = 2316,
+ CODE_FOR_atomic_adddi = 2317,
+ CODE_FOR_atomic_subdi = 2318,
+ CODE_FOR_atomic_ordi = 2319,
+ CODE_FOR_atomic_xordi = 2320,
+ CODE_FOR_atomic_anddi = 2321,
+ CODE_FOR_atomic_nandqi = 2322,
+ CODE_FOR_atomic_nandhi = 2323,
+ CODE_FOR_atomic_nandsi = 2324,
+ CODE_FOR_atomic_nanddi = 2325,
+ CODE_FOR_atomic_fetch_addqi = 2326,
+ CODE_FOR_atomic_fetch_subqi = 2327,
+ CODE_FOR_atomic_fetch_orqi = 2328,
+ CODE_FOR_atomic_fetch_xorqi = 2329,
+ CODE_FOR_atomic_fetch_andqi = 2330,
+ CODE_FOR_atomic_fetch_addhi = 2331,
+ CODE_FOR_atomic_fetch_subhi = 2332,
+ CODE_FOR_atomic_fetch_orhi = 2333,
+ CODE_FOR_atomic_fetch_xorhi = 2334,
+ CODE_FOR_atomic_fetch_andhi = 2335,
+ CODE_FOR_atomic_fetch_addsi = 2336,
+ CODE_FOR_atomic_fetch_subsi = 2337,
+ CODE_FOR_atomic_fetch_orsi = 2338,
+ CODE_FOR_atomic_fetch_xorsi = 2339,
+ CODE_FOR_atomic_fetch_andsi = 2340,
+ CODE_FOR_atomic_fetch_adddi = 2341,
+ CODE_FOR_atomic_fetch_subdi = 2342,
+ CODE_FOR_atomic_fetch_ordi = 2343,
+ CODE_FOR_atomic_fetch_xordi = 2344,
+ CODE_FOR_atomic_fetch_anddi = 2345,
+ CODE_FOR_atomic_fetch_nandqi = 2346,
+ CODE_FOR_atomic_fetch_nandhi = 2347,
+ CODE_FOR_atomic_fetch_nandsi = 2348,
+ CODE_FOR_atomic_fetch_nanddi = 2349,
+ CODE_FOR_atomic_add_fetchqi = 2350,
+ CODE_FOR_atomic_sub_fetchqi = 2351,
+ CODE_FOR_atomic_or_fetchqi = 2352,
+ CODE_FOR_atomic_xor_fetchqi = 2353,
+ CODE_FOR_atomic_and_fetchqi = 2354,
+ CODE_FOR_atomic_add_fetchhi = 2355,
+ CODE_FOR_atomic_sub_fetchhi = 2356,
+ CODE_FOR_atomic_or_fetchhi = 2357,
+ CODE_FOR_atomic_xor_fetchhi = 2358,
+ CODE_FOR_atomic_and_fetchhi = 2359,
+ CODE_FOR_atomic_add_fetchsi = 2360,
+ CODE_FOR_atomic_sub_fetchsi = 2361,
+ CODE_FOR_atomic_or_fetchsi = 2362,
+ CODE_FOR_atomic_xor_fetchsi = 2363,
+ CODE_FOR_atomic_and_fetchsi = 2364,
+ CODE_FOR_atomic_add_fetchdi = 2365,
+ CODE_FOR_atomic_sub_fetchdi = 2366,
+ CODE_FOR_atomic_or_fetchdi = 2367,
+ CODE_FOR_atomic_xor_fetchdi = 2368,
+ CODE_FOR_atomic_and_fetchdi = 2369,
+ CODE_FOR_atomic_nand_fetchqi = 2370,
+ CODE_FOR_atomic_nand_fetchhi = 2371,
+ CODE_FOR_atomic_nand_fetchsi = 2372,
+ CODE_FOR_atomic_nand_fetchdi = 2373,
+ CODE_FOR_atomic_loadqi = 2374,
+ CODE_FOR_atomic_loadhi = 2375,
+ CODE_FOR_atomic_loadsi = 2376,
+ CODE_FOR_atomic_loaddi = 2377,
+ CODE_FOR_atomic_storeqi = 2378,
+ CODE_FOR_atomic_storehi = 2379,
+ CODE_FOR_atomic_storesi = 2380,
+ CODE_FOR_atomic_storedi = 2381,
+ CODE_FOR_aarch64_load_exclusiveqi = 2382,
+ CODE_FOR_aarch64_load_exclusivehi = 2383,
+ CODE_FOR_aarch64_load_exclusivesi = 2384,
+ CODE_FOR_aarch64_load_exclusivedi = 2385,
+ CODE_FOR_aarch64_store_exclusiveqi = 2386,
+ CODE_FOR_aarch64_store_exclusivehi = 2387,
+ CODE_FOR_aarch64_store_exclusivesi = 2388,
+ CODE_FOR_aarch64_store_exclusivedi = 2389,
+ CODE_FOR_cbranchsi4 = 2391,
+ CODE_FOR_cbranchdi4 = 2392,
+ CODE_FOR_cbranchsf4 = 2393,
+ CODE_FOR_cbranchdf4 = 2394,
+ CODE_FOR_casesi = 2395,
+ CODE_FOR_prologue = 2396,
+ CODE_FOR_epilogue = 2397,
+ CODE_FOR_sibcall_epilogue = 2398,
+ CODE_FOR_call = 2400,
+ CODE_FOR_call_value = 2401,
+ CODE_FOR_sibcall = 2402,
+ CODE_FOR_sibcall_value = 2403,
+ CODE_FOR_untyped_call = 2404,
+ CODE_FOR_movqi = 2405,
+ CODE_FOR_movhi = 2406,
+ CODE_FOR_movsi = 2407,
+ CODE_FOR_movdi = 2408,
+ CODE_FOR_movti = 2409,
+ CODE_FOR_movsf = 2411,
+ CODE_FOR_movdf = 2412,
+ CODE_FOR_movtf = 2413,
+ CODE_FOR_movmemdi = 2415,
+ CODE_FOR_extendsidi2 = 2416,
+ CODE_FOR_zero_extendsidi2 = 2417,
+ CODE_FOR_extendqisi2 = 2418,
+ CODE_FOR_zero_extendqisi2 = 2419,
+ CODE_FOR_extendqidi2 = 2420,
+ CODE_FOR_zero_extendqidi2 = 2421,
+ CODE_FOR_extendhisi2 = 2422,
+ CODE_FOR_zero_extendhisi2 = 2423,
+ CODE_FOR_extendhidi2 = 2424,
+ CODE_FOR_zero_extendhidi2 = 2425,
+ CODE_FOR_extendqihi2 = 2426,
+ CODE_FOR_zero_extendqihi2 = 2427,
+ CODE_FOR_addsi3 = 2428,
+ CODE_FOR_adddi3 = 2429,
+ CODE_FOR_addti3 = 2430,
+ CODE_FOR_subti3 = 2431,
+ CODE_FOR_mulditi3 = 2433,
+ CODE_FOR_umulditi3 = 2434,
+ CODE_FOR_multi3 = 2435,
+ CODE_FOR_cstoresi4 = 2436,
+ CODE_FOR_cstoredi4 = 2437,
+ CODE_FOR_cstoresf4 = 2438,
+ CODE_FOR_cstoredf4 = 2439,
+ CODE_FOR_cmovsi6 = 2440,
+ CODE_FOR_cmovdi6 = 2441,
+ CODE_FOR_cmovsf6 = 2442,
+ CODE_FOR_cmovdf6 = 2443,
+ CODE_FOR_movqicc = 2444,
+ CODE_FOR_movhicc = 2445,
+ CODE_FOR_movsicc = 2446,
+ CODE_FOR_movdicc = 2447,
+ CODE_FOR_movsfsicc = 2448,
+ CODE_FOR_movdfsicc = 2449,
+ CODE_FOR_movsfdicc = 2450,
+ CODE_FOR_movdfdicc = 2451,
+ CODE_FOR_movsfcc = 2452,
+ CODE_FOR_movdfcc = 2453,
+ CODE_FOR_ffssi2 = 2454,
+ CODE_FOR_ffsdi2 = 2455,
+ CODE_FOR_ctzsi2 = 2456,
+ CODE_FOR_ctzdi2 = 2457,
+ CODE_FOR_ashlsi3 = 2458,
+ CODE_FOR_ashrsi3 = 2459,
+ CODE_FOR_lshrsi3 = 2460,
+ CODE_FOR_ashldi3 = 2461,
+ CODE_FOR_ashrdi3 = 2462,
+ CODE_FOR_lshrdi3 = 2463,
+ CODE_FOR_ashlqi3 = 2464,
+ CODE_FOR_ashlhi3 = 2465,
+ CODE_FOR_rotrsi3 = 2466,
+ CODE_FOR_rotrdi3 = 2467,
+ CODE_FOR_rotlsi3 = 2468,
+ CODE_FOR_rotldi3 = 2469,
+ CODE_FOR_extv = 2474,
+ CODE_FOR_extzv = 2475,
+ CODE_FOR_insvsi = 2476,
+ CODE_FOR_insvdi = 2477,
+ CODE_FOR_aarch64_reload_movti = 2478,
+ CODE_FOR_aarch64_reload_movtf = 2479,
+ CODE_FOR_add_losym = 2480,
+ CODE_FOR_tlsgd_small = 2481,
+ CODE_FOR_tlsle_small = 2482,
+ CODE_FOR_get_thread_pointerdi = 2483,
+ CODE_FOR_stack_protect_set = 2484,
+ CODE_FOR_stack_protect_test = 2485,
+ CODE_FOR_movv8qi = 2486,
+ CODE_FOR_movv16qi = 2487,
+ CODE_FOR_movv4hi = 2488,
+ CODE_FOR_movv8hi = 2489,
+ CODE_FOR_movv2si = 2490,
+ CODE_FOR_movv4si = 2491,
+ CODE_FOR_movv2di = 2492,
+ CODE_FOR_movv2sf = 2493,
+ CODE_FOR_movv4sf = 2494,
+ CODE_FOR_movv2df = 2495,
+ CODE_FOR_movmisalignv8qi = 2496,
+ CODE_FOR_movmisalignv16qi = 2497,
+ CODE_FOR_movmisalignv4hi = 2498,
+ CODE_FOR_movmisalignv8hi = 2499,
+ CODE_FOR_movmisalignv2si = 2500,
+ CODE_FOR_movmisalignv4si = 2501,
+ CODE_FOR_movmisalignv2di = 2502,
+ CODE_FOR_movmisalignv2sf = 2503,
+ CODE_FOR_movmisalignv4sf = 2504,
+ CODE_FOR_movmisalignv2df = 2505,
+ CODE_FOR_aarch64_split_simd_movv16qi = 2518,
+ CODE_FOR_aarch64_split_simd_movv8hi = 2519,
+ CODE_FOR_aarch64_split_simd_movv4si = 2520,
+ CODE_FOR_aarch64_split_simd_movv2di = 2521,
+ CODE_FOR_aarch64_split_simd_movv4sf = 2522,
+ CODE_FOR_aarch64_split_simd_movv2df = 2523,
+ CODE_FOR_ashlv8qi3 = 2524,
+ CODE_FOR_ashlv16qi3 = 2525,
+ CODE_FOR_ashlv4hi3 = 2526,
+ CODE_FOR_ashlv8hi3 = 2527,
+ CODE_FOR_ashlv2si3 = 2528,
+ CODE_FOR_ashlv4si3 = 2529,
+ CODE_FOR_ashlv2di3 = 2530,
+ CODE_FOR_lshrv8qi3 = 2531,
+ CODE_FOR_lshrv16qi3 = 2532,
+ CODE_FOR_lshrv4hi3 = 2533,
+ CODE_FOR_lshrv8hi3 = 2534,
+ CODE_FOR_lshrv2si3 = 2535,
+ CODE_FOR_lshrv4si3 = 2536,
+ CODE_FOR_lshrv2di3 = 2537,
+ CODE_FOR_ashrv8qi3 = 2538,
+ CODE_FOR_ashrv16qi3 = 2539,
+ CODE_FOR_ashrv4hi3 = 2540,
+ CODE_FOR_ashrv8hi3 = 2541,
+ CODE_FOR_ashrv2si3 = 2542,
+ CODE_FOR_ashrv4si3 = 2543,
+ CODE_FOR_ashrv2di3 = 2544,
+ CODE_FOR_vashlv8qi3 = 2545,
+ CODE_FOR_vashlv16qi3 = 2546,
+ CODE_FOR_vashlv4hi3 = 2547,
+ CODE_FOR_vashlv8hi3 = 2548,
+ CODE_FOR_vashlv2si3 = 2549,
+ CODE_FOR_vashlv4si3 = 2550,
+ CODE_FOR_vashlv2di3 = 2551,
+ CODE_FOR_vashrv8qi3 = 2552,
+ CODE_FOR_vashrv16qi3 = 2553,
+ CODE_FOR_vashrv4hi3 = 2554,
+ CODE_FOR_vashrv8hi3 = 2555,
+ CODE_FOR_vashrv2si3 = 2556,
+ CODE_FOR_vashrv4si3 = 2557,
+ CODE_FOR_aarch64_ashr_simddi = 2558,
+ CODE_FOR_vlshrv8qi3 = 2559,
+ CODE_FOR_vlshrv16qi3 = 2560,
+ CODE_FOR_vlshrv4hi3 = 2561,
+ CODE_FOR_vlshrv8hi3 = 2562,
+ CODE_FOR_vlshrv2si3 = 2563,
+ CODE_FOR_vlshrv4si3 = 2564,
+ CODE_FOR_aarch64_lshr_simddi = 2565,
+ CODE_FOR_vec_setv8qi = 2566,
+ CODE_FOR_vec_setv16qi = 2567,
+ CODE_FOR_vec_setv4hi = 2568,
+ CODE_FOR_vec_setv8hi = 2569,
+ CODE_FOR_vec_setv2si = 2570,
+ CODE_FOR_vec_setv4si = 2571,
+ CODE_FOR_vec_setv2di = 2572,
+ CODE_FOR_vec_setv2sf = 2573,
+ CODE_FOR_vec_setv4sf = 2574,
+ CODE_FOR_vec_setv2df = 2575,
+ CODE_FOR_move_lo_quad_v16qi = 2576,
+ CODE_FOR_move_lo_quad_v8hi = 2577,
+ CODE_FOR_move_lo_quad_v4si = 2578,
+ CODE_FOR_move_lo_quad_v2di = 2579,
+ CODE_FOR_move_lo_quad_v4sf = 2580,
+ CODE_FOR_move_lo_quad_v2df = 2581,
+ CODE_FOR_move_hi_quad_v16qi = 2582,
+ CODE_FOR_move_hi_quad_v8hi = 2583,
+ CODE_FOR_move_hi_quad_v4si = 2584,
+ CODE_FOR_move_hi_quad_v2di = 2585,
+ CODE_FOR_move_hi_quad_v4sf = 2586,
+ CODE_FOR_move_hi_quad_v2df = 2587,
+ CODE_FOR_vec_pack_trunc_v4hi = 2588,
+ CODE_FOR_vec_pack_trunc_v2si = 2589,
+ CODE_FOR_vec_pack_trunc_di = 2590,
+ CODE_FOR_vec_unpacks_hi_v16qi = 2591,
+ CODE_FOR_vec_unpacku_hi_v16qi = 2592,
+ CODE_FOR_vec_unpacks_hi_v8hi = 2593,
+ CODE_FOR_vec_unpacku_hi_v8hi = 2594,
+ CODE_FOR_vec_unpacks_hi_v4si = 2595,
+ CODE_FOR_vec_unpacku_hi_v4si = 2596,
+ CODE_FOR_vec_unpacks_lo_v16qi = 2597,
+ CODE_FOR_vec_unpacku_lo_v16qi = 2598,
+ CODE_FOR_vec_unpacks_lo_v8hi = 2599,
+ CODE_FOR_vec_unpacku_lo_v8hi = 2600,
+ CODE_FOR_vec_unpacks_lo_v4si = 2601,
+ CODE_FOR_vec_unpacku_lo_v4si = 2602,
+ CODE_FOR_vec_widen_smult_lo_v16qi = 2603,
+ CODE_FOR_vec_widen_umult_lo_v16qi = 2604,
+ CODE_FOR_vec_widen_smult_lo_v8hi = 2605,
+ CODE_FOR_vec_widen_umult_lo_v8hi = 2606,
+ CODE_FOR_vec_widen_smult_lo_v4si = 2607,
+ CODE_FOR_vec_widen_umult_lo_v4si = 2608,
+ CODE_FOR_vec_widen_smult_hi_v16qi = 2609,
+ CODE_FOR_vec_widen_umult_hi_v16qi = 2610,
+ CODE_FOR_vec_widen_smult_hi_v8hi = 2611,
+ CODE_FOR_vec_widen_umult_hi_v8hi = 2612,
+ CODE_FOR_vec_widen_smult_hi_v4si = 2613,
+ CODE_FOR_vec_widen_umult_hi_v4si = 2614,
+ CODE_FOR_fixv2sfv2si2 = 2615,
+ CODE_FOR_fixunsv2sfv2si2 = 2616,
+ CODE_FOR_fixv4sfv4si2 = 2617,
+ CODE_FOR_fixunsv4sfv4si2 = 2618,
+ CODE_FOR_fixv2dfv2di2 = 2619,
+ CODE_FOR_fixunsv2dfv2di2 = 2620,
+ CODE_FOR_fix_truncv2sfv2si2 = 2621,
+ CODE_FOR_fixuns_truncv2sfv2si2 = 2622,
+ CODE_FOR_fix_truncv4sfv4si2 = 2623,
+ CODE_FOR_fixuns_truncv4sfv4si2 = 2624,
+ CODE_FOR_fix_truncv2dfv2di2 = 2625,
+ CODE_FOR_fixuns_truncv2dfv2di2 = 2626,
+ CODE_FOR_ftruncv2sf2 = 2627,
+ CODE_FOR_ftruncv4sf2 = 2628,
+ CODE_FOR_ftruncv2df2 = 2629,
+ CODE_FOR_vec_pack_trunc_v2df = 2630,
+ CODE_FOR_vec_pack_trunc_df = 2631,
+ CODE_FOR_reduc_splus_v4sf = 2632,
+ CODE_FOR_aarch64_simd_bslv8qi = 2633,
+ CODE_FOR_aarch64_simd_bslv16qi = 2634,
+ CODE_FOR_aarch64_simd_bslv4hi = 2635,
+ CODE_FOR_aarch64_simd_bslv8hi = 2636,
+ CODE_FOR_aarch64_simd_bslv2si = 2637,
+ CODE_FOR_aarch64_simd_bslv4si = 2638,
+ CODE_FOR_aarch64_simd_bslv2di = 2639,
+ CODE_FOR_aarch64_simd_bslv2sf = 2640,
+ CODE_FOR_aarch64_simd_bslv4sf = 2641,
+ CODE_FOR_aarch64_simd_bslv2df = 2642,
+ CODE_FOR_aarch64_simd_bsldi = 2643,
+ CODE_FOR_aarch64_simd_bsldf = 2644,
+ CODE_FOR_aarch64_vcond_internalv8qiv8qi = 2645,
+ CODE_FOR_aarch64_vcond_internalv16qiv16qi = 2646,
+ CODE_FOR_aarch64_vcond_internalv4hiv4hi = 2647,
+ CODE_FOR_aarch64_vcond_internalv8hiv8hi = 2648,
+ CODE_FOR_aarch64_vcond_internalv2siv2si = 2649,
+ CODE_FOR_aarch64_vcond_internalv4siv4si = 2650,
+ CODE_FOR_aarch64_vcond_internalv2div2di = 2651,
+ CODE_FOR_aarch64_vcond_internalv2sfv2sf = 2652,
+ CODE_FOR_aarch64_vcond_internalv2siv2sf = 2653,
+ CODE_FOR_aarch64_vcond_internalv4sfv2sf = 2654,
+ CODE_FOR_aarch64_vcond_internalv4siv2sf = 2655,
+ CODE_FOR_aarch64_vcond_internalv2dfv2sf = 2656,
+ CODE_FOR_aarch64_vcond_internalv2div2sf = 2657,
+ CODE_FOR_aarch64_vcond_internalv2sfv4sf = 2658,
+ CODE_FOR_aarch64_vcond_internalv2siv4sf = 2659,
+ CODE_FOR_aarch64_vcond_internalv4sfv4sf = 2660,
+ CODE_FOR_aarch64_vcond_internalv4siv4sf = 2661,
+ CODE_FOR_aarch64_vcond_internalv2dfv4sf = 2662,
+ CODE_FOR_aarch64_vcond_internalv2div4sf = 2663,
+ CODE_FOR_aarch64_vcond_internalv2sfv2df = 2664,
+ CODE_FOR_aarch64_vcond_internalv2siv2df = 2665,
+ CODE_FOR_aarch64_vcond_internalv4sfv2df = 2666,
+ CODE_FOR_aarch64_vcond_internalv4siv2df = 2667,
+ CODE_FOR_aarch64_vcond_internalv2dfv2df = 2668,
+ CODE_FOR_aarch64_vcond_internalv2div2df = 2669,
+ CODE_FOR_vcondv8qiv8qi = 2670,
+ CODE_FOR_vcondv16qiv16qi = 2671,
+ CODE_FOR_vcondv4hiv4hi = 2672,
+ CODE_FOR_vcondv8hiv8hi = 2673,
+ CODE_FOR_vcondv2siv2si = 2674,
+ CODE_FOR_vcondv4siv4si = 2675,
+ CODE_FOR_vcondv2div2di = 2676,
+ CODE_FOR_vcondv2sfv2sf = 2677,
+ CODE_FOR_vcondv4sfv4sf = 2678,
+ CODE_FOR_vcondv2dfv2df = 2679,
+ CODE_FOR_vcondv2siv2sf = 2680,
+ CODE_FOR_vcondv4siv4sf = 2681,
+ CODE_FOR_vcondv2div2df = 2682,
+ CODE_FOR_vconduv8qiv8qi = 2683,
+ CODE_FOR_vconduv16qiv16qi = 2684,
+ CODE_FOR_vconduv4hiv4hi = 2685,
+ CODE_FOR_vconduv8hiv8hi = 2686,
+ CODE_FOR_vconduv2siv2si = 2687,
+ CODE_FOR_vconduv4siv4si = 2688,
+ CODE_FOR_vconduv2div2di = 2689,
+ CODE_FOR_aarch64_createv8qi = 2690,
+ CODE_FOR_aarch64_createv4hi = 2691,
+ CODE_FOR_aarch64_createv2si = 2692,
+ CODE_FOR_aarch64_createdi = 2693,
+ CODE_FOR_aarch64_createdf = 2694,
+ CODE_FOR_aarch64_createv2sf = 2695,
+ CODE_FOR_aarch64_be_checked_get_lanev8qi = 2696,
+ CODE_FOR_aarch64_be_checked_get_lanev16qi = 2697,
+ CODE_FOR_aarch64_be_checked_get_lanev4hi = 2698,
+ CODE_FOR_aarch64_be_checked_get_lanev8hi = 2699,
+ CODE_FOR_aarch64_be_checked_get_lanev2si = 2700,
+ CODE_FOR_aarch64_be_checked_get_lanev4si = 2701,
+ CODE_FOR_aarch64_be_checked_get_lanev2di = 2702,
+ CODE_FOR_aarch64_be_checked_get_lanev2sf = 2703,
+ CODE_FOR_aarch64_be_checked_get_lanev4sf = 2704,
+ CODE_FOR_aarch64_be_checked_get_lanev2df = 2705,
+ CODE_FOR_aarch64_get_lanedi = 2706,
+ CODE_FOR_aarch64_reinterpretv8qiv8qi = 2707,
+ CODE_FOR_aarch64_reinterpretv8qiv4hi = 2708,
+ CODE_FOR_aarch64_reinterpretv8qiv2si = 2709,
+ CODE_FOR_aarch64_reinterpretv8qiv2sf = 2710,
+ CODE_FOR_aarch64_reinterpretv8qidi = 2711,
+ CODE_FOR_aarch64_reinterpretv8qidf = 2712,
+ CODE_FOR_aarch64_reinterpretv4hiv8qi = 2713,
+ CODE_FOR_aarch64_reinterpretv4hiv4hi = 2714,
+ CODE_FOR_aarch64_reinterpretv4hiv2si = 2715,
+ CODE_FOR_aarch64_reinterpretv4hiv2sf = 2716,
+ CODE_FOR_aarch64_reinterpretv4hidi = 2717,
+ CODE_FOR_aarch64_reinterpretv4hidf = 2718,
+ CODE_FOR_aarch64_reinterpretv2siv8qi = 2719,
+ CODE_FOR_aarch64_reinterpretv2siv4hi = 2720,
+ CODE_FOR_aarch64_reinterpretv2siv2si = 2721,
+ CODE_FOR_aarch64_reinterpretv2siv2sf = 2722,
+ CODE_FOR_aarch64_reinterpretv2sidi = 2723,
+ CODE_FOR_aarch64_reinterpretv2sidf = 2724,
+ CODE_FOR_aarch64_reinterpretv2sfv8qi = 2725,
+ CODE_FOR_aarch64_reinterpretv2sfv4hi = 2726,
+ CODE_FOR_aarch64_reinterpretv2sfv2si = 2727,
+ CODE_FOR_aarch64_reinterpretv2sfv2sf = 2728,
+ CODE_FOR_aarch64_reinterpretv2sfdi = 2729,
+ CODE_FOR_aarch64_reinterpretv2sfdf = 2730,
+ CODE_FOR_aarch64_reinterpretdiv8qi = 2731,
+ CODE_FOR_aarch64_reinterpretdiv4hi = 2732,
+ CODE_FOR_aarch64_reinterpretdiv2si = 2733,
+ CODE_FOR_aarch64_reinterpretdidi = 2734,
+ CODE_FOR_aarch64_reinterpretdidf = 2735,
+ CODE_FOR_aarch64_reinterpretdiv2sf = 2736,
+ CODE_FOR_aarch64_reinterpretdfv8qi = 2737,
+ CODE_FOR_aarch64_reinterpretdfv4hi = 2738,
+ CODE_FOR_aarch64_reinterpretdfv2si = 2739,
+ CODE_FOR_aarch64_reinterpretdfdi = 2740,
+ CODE_FOR_aarch64_reinterpretdfdf = 2741,
+ CODE_FOR_aarch64_reinterpretdfv2sf = 2742,
+ CODE_FOR_aarch64_reinterpretv16qiv16qi = 2743,
+ CODE_FOR_aarch64_reinterpretv16qiv8hi = 2744,
+ CODE_FOR_aarch64_reinterpretv16qiv4si = 2745,
+ CODE_FOR_aarch64_reinterpretv16qiv2di = 2746,
+ CODE_FOR_aarch64_reinterpretv16qiv4sf = 2747,
+ CODE_FOR_aarch64_reinterpretv16qiv2df = 2748,
+ CODE_FOR_aarch64_reinterpretv8hiv16qi = 2749,
+ CODE_FOR_aarch64_reinterpretv8hiv8hi = 2750,
+ CODE_FOR_aarch64_reinterpretv8hiv4si = 2751,
+ CODE_FOR_aarch64_reinterpretv8hiv2di = 2752,
+ CODE_FOR_aarch64_reinterpretv8hiv4sf = 2753,
+ CODE_FOR_aarch64_reinterpretv8hiv2df = 2754,
+ CODE_FOR_aarch64_reinterpretv4siv16qi = 2755,
+ CODE_FOR_aarch64_reinterpretv4siv8hi = 2756,
+ CODE_FOR_aarch64_reinterpretv4siv4si = 2757,
+ CODE_FOR_aarch64_reinterpretv4siv2di = 2758,
+ CODE_FOR_aarch64_reinterpretv4siv4sf = 2759,
+ CODE_FOR_aarch64_reinterpretv4siv2df = 2760,
+ CODE_FOR_aarch64_reinterpretv4sfv16qi = 2761,
+ CODE_FOR_aarch64_reinterpretv4sfv8hi = 2762,
+ CODE_FOR_aarch64_reinterpretv4sfv4si = 2763,
+ CODE_FOR_aarch64_reinterpretv4sfv2di = 2764,
+ CODE_FOR_aarch64_reinterpretv4sfv4sf = 2765,
+ CODE_FOR_aarch64_reinterpretv4sfv2df = 2766,
+ CODE_FOR_aarch64_reinterpretv2div16qi = 2767,
+ CODE_FOR_aarch64_reinterpretv2div8hi = 2768,
+ CODE_FOR_aarch64_reinterpretv2div4si = 2769,
+ CODE_FOR_aarch64_reinterpretv2div2di = 2770,
+ CODE_FOR_aarch64_reinterpretv2div4sf = 2771,
+ CODE_FOR_aarch64_reinterpretv2div2df = 2772,
+ CODE_FOR_aarch64_reinterpretv2dfv16qi = 2773,
+ CODE_FOR_aarch64_reinterpretv2dfv8hi = 2774,
+ CODE_FOR_aarch64_reinterpretv2dfv4si = 2775,
+ CODE_FOR_aarch64_reinterpretv2dfv2di = 2776,
+ CODE_FOR_aarch64_reinterpretv2dfv4sf = 2777,
+ CODE_FOR_aarch64_reinterpretv2dfv2df = 2778,
+ CODE_FOR_aarch64_combinev8qi = 2779,
+ CODE_FOR_aarch64_combinev4hi = 2780,
+ CODE_FOR_aarch64_combinev2si = 2781,
+ CODE_FOR_aarch64_combinev2sf = 2782,
+ CODE_FOR_aarch64_combinedi = 2783,
+ CODE_FOR_aarch64_combinedf = 2784,
+ CODE_FOR_aarch64_simd_combinev8qi = 2791,
+ CODE_FOR_aarch64_simd_combinev4hi = 2792,
+ CODE_FOR_aarch64_simd_combinev2si = 2793,
+ CODE_FOR_aarch64_simd_combinev2sf = 2794,
+ CODE_FOR_aarch64_simd_combinedi = 2795,
+ CODE_FOR_aarch64_simd_combinedf = 2796,
+ CODE_FOR_aarch64_saddl2v16qi = 2797,
+ CODE_FOR_aarch64_saddl2v8hi = 2798,
+ CODE_FOR_aarch64_saddl2v4si = 2799,
+ CODE_FOR_aarch64_uaddl2v16qi = 2800,
+ CODE_FOR_aarch64_uaddl2v8hi = 2801,
+ CODE_FOR_aarch64_uaddl2v4si = 2802,
+ CODE_FOR_aarch64_ssubl2v16qi = 2803,
+ CODE_FOR_aarch64_ssubl2v8hi = 2804,
+ CODE_FOR_aarch64_ssubl2v4si = 2805,
+ CODE_FOR_aarch64_usubl2v16qi = 2806,
+ CODE_FOR_aarch64_usubl2v8hi = 2807,
+ CODE_FOR_aarch64_usubl2v4si = 2808,
+ CODE_FOR_aarch64_saddw2v16qi = 2809,
+ CODE_FOR_aarch64_saddw2v8hi = 2810,
+ CODE_FOR_aarch64_saddw2v4si = 2811,
+ CODE_FOR_aarch64_uaddw2v16qi = 2812,
+ CODE_FOR_aarch64_uaddw2v8hi = 2813,
+ CODE_FOR_aarch64_uaddw2v4si = 2814,
+ CODE_FOR_aarch64_ssubw2v16qi = 2815,
+ CODE_FOR_aarch64_ssubw2v8hi = 2816,
+ CODE_FOR_aarch64_ssubw2v4si = 2817,
+ CODE_FOR_aarch64_usubw2v16qi = 2818,
+ CODE_FOR_aarch64_usubw2v8hi = 2819,
+ CODE_FOR_aarch64_usubw2v4si = 2820,
+ CODE_FOR_aarch64_sqdmulh_lanev4hi = 2821,
+ CODE_FOR_aarch64_sqdmulh_lanev8hi = 2822,
+ CODE_FOR_aarch64_sqdmulh_lanev2si = 2823,
+ CODE_FOR_aarch64_sqdmulh_lanev4si = 2824,
+ CODE_FOR_aarch64_sqrdmulh_lanev4hi = 2825,
+ CODE_FOR_aarch64_sqrdmulh_lanev8hi = 2826,
+ CODE_FOR_aarch64_sqrdmulh_lanev2si = 2827,
+ CODE_FOR_aarch64_sqrdmulh_lanev4si = 2828,
+ CODE_FOR_aarch64_sqdmulh_laneqv4hi = 2829,
+ CODE_FOR_aarch64_sqdmulh_laneqv8hi = 2830,
+ CODE_FOR_aarch64_sqdmulh_laneqv2si = 2831,
+ CODE_FOR_aarch64_sqdmulh_laneqv4si = 2832,
+ CODE_FOR_aarch64_sqrdmulh_laneqv4hi = 2833,
+ CODE_FOR_aarch64_sqrdmulh_laneqv8hi = 2834,
+ CODE_FOR_aarch64_sqrdmulh_laneqv2si = 2835,
+ CODE_FOR_aarch64_sqrdmulh_laneqv4si = 2836,
+ CODE_FOR_aarch64_sqdmulh_lanehi = 2837,
+ CODE_FOR_aarch64_sqdmulh_lanesi = 2838,
+ CODE_FOR_aarch64_sqrdmulh_lanehi = 2839,
+ CODE_FOR_aarch64_sqrdmulh_lanesi = 2840,
+ CODE_FOR_aarch64_sqdmlal_lanev4hi = 2841,
+ CODE_FOR_aarch64_sqdmlal_lanev2si = 2842,
+ CODE_FOR_aarch64_sqdmlal_lanehi = 2843,
+ CODE_FOR_aarch64_sqdmlal_lanesi = 2844,
+ CODE_FOR_aarch64_sqdmlal_laneqv4hi = 2845,
+ CODE_FOR_aarch64_sqdmlal_laneqv2si = 2846,
+ CODE_FOR_aarch64_sqdmlal_laneqhi = 2847,
+ CODE_FOR_aarch64_sqdmlal_laneqsi = 2848,
+ CODE_FOR_aarch64_sqdmlsl_lanev4hi = 2849,
+ CODE_FOR_aarch64_sqdmlsl_lanev2si = 2850,
+ CODE_FOR_aarch64_sqdmlsl_lanehi = 2851,
+ CODE_FOR_aarch64_sqdmlsl_lanesi = 2852,
+ CODE_FOR_aarch64_sqdmlsl_laneqv4hi = 2853,
+ CODE_FOR_aarch64_sqdmlsl_laneqv2si = 2854,
+ CODE_FOR_aarch64_sqdmlsl_laneqhi = 2855,
+ CODE_FOR_aarch64_sqdmlsl_laneqsi = 2856,
+ CODE_FOR_aarch64_sqdmlal2v8hi = 2857,
+ CODE_FOR_aarch64_sqdmlal2v4si = 2858,
+ CODE_FOR_aarch64_sqdmlsl2v8hi = 2859,
+ CODE_FOR_aarch64_sqdmlsl2v4si = 2860,
+ CODE_FOR_aarch64_sqdmlal2_lanev8hi = 2861,
+ CODE_FOR_aarch64_sqdmlal2_lanev4si = 2862,
+ CODE_FOR_aarch64_sqdmlal2_laneqv8hi = 2863,
+ CODE_FOR_aarch64_sqdmlal2_laneqv4si = 2864,
+ CODE_FOR_aarch64_sqdmlsl2_lanev8hi = 2865,
+ CODE_FOR_aarch64_sqdmlsl2_lanev4si = 2866,
+ CODE_FOR_aarch64_sqdmlsl2_laneqv8hi = 2867,
+ CODE_FOR_aarch64_sqdmlsl2_laneqv4si = 2868,
+ CODE_FOR_aarch64_sqdmlal2_nv8hi = 2869,
+ CODE_FOR_aarch64_sqdmlal2_nv4si = 2870,
+ CODE_FOR_aarch64_sqdmlsl2_nv8hi = 2871,
+ CODE_FOR_aarch64_sqdmlsl2_nv4si = 2872,
+ CODE_FOR_aarch64_sqdmull_lanev4hi = 2873,
+ CODE_FOR_aarch64_sqdmull_lanev2si = 2874,
+ CODE_FOR_aarch64_sqdmull_lanehi = 2875,
+ CODE_FOR_aarch64_sqdmull_lanesi = 2876,
+ CODE_FOR_aarch64_sqdmull_laneqv4hi = 2877,
+ CODE_FOR_aarch64_sqdmull_laneqv2si = 2878,
+ CODE_FOR_aarch64_sqdmull2v8hi = 2879,
+ CODE_FOR_aarch64_sqdmull2v4si = 2880,
+ CODE_FOR_aarch64_sqdmull2_lanev8hi = 2881,
+ CODE_FOR_aarch64_sqdmull2_lanev4si = 2882,
+ CODE_FOR_aarch64_sqdmull2_laneqv8hi = 2883,
+ CODE_FOR_aarch64_sqdmull2_laneqv4si = 2884,
+ CODE_FOR_aarch64_sqdmull2_nv8hi = 2885,
+ CODE_FOR_aarch64_sqdmull2_nv4si = 2886,
+ CODE_FOR_movoi = 2897,
+ CODE_FOR_movci = 2898,
+ CODE_FOR_movxi = 2899,
+ CODE_FOR_aarch64_ld2v8qi = 2903,
+ CODE_FOR_aarch64_ld2v4hi = 2904,
+ CODE_FOR_aarch64_ld2v2si = 2905,
+ CODE_FOR_aarch64_ld2v2sf = 2906,
+ CODE_FOR_aarch64_ld2di = 2907,
+ CODE_FOR_aarch64_ld2df = 2908,
+ CODE_FOR_aarch64_ld3v8qi = 2909,
+ CODE_FOR_aarch64_ld3v4hi = 2910,
+ CODE_FOR_aarch64_ld3v2si = 2911,
+ CODE_FOR_aarch64_ld3v2sf = 2912,
+ CODE_FOR_aarch64_ld3di = 2913,
+ CODE_FOR_aarch64_ld3df = 2914,
+ CODE_FOR_aarch64_ld4v8qi = 2915,
+ CODE_FOR_aarch64_ld4v4hi = 2916,
+ CODE_FOR_aarch64_ld4v2si = 2917,
+ CODE_FOR_aarch64_ld4v2sf = 2918,
+ CODE_FOR_aarch64_ld4di = 2919,
+ CODE_FOR_aarch64_ld4df = 2920,
+ CODE_FOR_aarch64_ld1v8qi = 2921,
+ CODE_FOR_aarch64_ld1v16qi = 2922,
+ CODE_FOR_aarch64_ld1v4hi = 2923,
+ CODE_FOR_aarch64_ld1v8hi = 2924,
+ CODE_FOR_aarch64_ld1v2si = 2925,
+ CODE_FOR_aarch64_ld1v4si = 2926,
+ CODE_FOR_aarch64_ld1v2di = 2927,
+ CODE_FOR_aarch64_ld1v2sf = 2928,
+ CODE_FOR_aarch64_ld1v4sf = 2929,
+ CODE_FOR_aarch64_ld1v2df = 2930,
+ CODE_FOR_aarch64_ld2v16qi = 2931,
+ CODE_FOR_aarch64_ld3v16qi = 2932,
+ CODE_FOR_aarch64_ld4v16qi = 2933,
+ CODE_FOR_aarch64_ld2v8hi = 2934,
+ CODE_FOR_aarch64_ld3v8hi = 2935,
+ CODE_FOR_aarch64_ld4v8hi = 2936,
+ CODE_FOR_aarch64_ld2v4si = 2937,
+ CODE_FOR_aarch64_ld3v4si = 2938,
+ CODE_FOR_aarch64_ld4v4si = 2939,
+ CODE_FOR_aarch64_ld2v2di = 2940,
+ CODE_FOR_aarch64_ld3v2di = 2941,
+ CODE_FOR_aarch64_ld4v2di = 2942,
+ CODE_FOR_aarch64_ld2v4sf = 2943,
+ CODE_FOR_aarch64_ld3v4sf = 2944,
+ CODE_FOR_aarch64_ld4v4sf = 2945,
+ CODE_FOR_aarch64_ld2v2df = 2946,
+ CODE_FOR_aarch64_ld3v2df = 2947,
+ CODE_FOR_aarch64_ld4v2df = 2948,
+ CODE_FOR_aarch64_get_dregoiv8qi = 2949,
+ CODE_FOR_aarch64_get_dregoiv4hi = 2950,
+ CODE_FOR_aarch64_get_dregoiv2si = 2951,
+ CODE_FOR_aarch64_get_dregoiv2sf = 2952,
+ CODE_FOR_aarch64_get_dregoidi = 2953,
+ CODE_FOR_aarch64_get_dregoidf = 2954,
+ CODE_FOR_aarch64_get_dregciv8qi = 2955,
+ CODE_FOR_aarch64_get_dregciv4hi = 2956,
+ CODE_FOR_aarch64_get_dregciv2si = 2957,
+ CODE_FOR_aarch64_get_dregciv2sf = 2958,
+ CODE_FOR_aarch64_get_dregcidi = 2959,
+ CODE_FOR_aarch64_get_dregcidf = 2960,
+ CODE_FOR_aarch64_get_dregxiv8qi = 2961,
+ CODE_FOR_aarch64_get_dregxiv4hi = 2962,
+ CODE_FOR_aarch64_get_dregxiv2si = 2963,
+ CODE_FOR_aarch64_get_dregxiv2sf = 2964,
+ CODE_FOR_aarch64_get_dregxidi = 2965,
+ CODE_FOR_aarch64_get_dregxidf = 2966,
+ CODE_FOR_aarch64_get_qregoiv16qi = 2967,
+ CODE_FOR_aarch64_get_qregciv16qi = 2968,
+ CODE_FOR_aarch64_get_qregxiv16qi = 2969,
+ CODE_FOR_aarch64_get_qregoiv8hi = 2970,
+ CODE_FOR_aarch64_get_qregciv8hi = 2971,
+ CODE_FOR_aarch64_get_qregxiv8hi = 2972,
+ CODE_FOR_aarch64_get_qregoiv4si = 2973,
+ CODE_FOR_aarch64_get_qregciv4si = 2974,
+ CODE_FOR_aarch64_get_qregxiv4si = 2975,
+ CODE_FOR_aarch64_get_qregoiv2di = 2976,
+ CODE_FOR_aarch64_get_qregciv2di = 2977,
+ CODE_FOR_aarch64_get_qregxiv2di = 2978,
+ CODE_FOR_aarch64_get_qregoiv4sf = 2979,
+ CODE_FOR_aarch64_get_qregciv4sf = 2980,
+ CODE_FOR_aarch64_get_qregxiv4sf = 2981,
+ CODE_FOR_aarch64_get_qregoiv2df = 2982,
+ CODE_FOR_aarch64_get_qregciv2df = 2983,
+ CODE_FOR_aarch64_get_qregxiv2df = 2984,
+ CODE_FOR_vec_perm_constv8qi = 2985,
+ CODE_FOR_vec_perm_constv16qi = 2986,
+ CODE_FOR_vec_perm_constv4hi = 2987,
+ CODE_FOR_vec_perm_constv8hi = 2988,
+ CODE_FOR_vec_perm_constv2si = 2989,
+ CODE_FOR_vec_perm_constv4si = 2990,
+ CODE_FOR_vec_perm_constv2di = 2991,
+ CODE_FOR_vec_perm_constv2sf = 2992,
+ CODE_FOR_vec_perm_constv4sf = 2993,
+ CODE_FOR_vec_perm_constv2df = 2994,
+ CODE_FOR_vec_permv8qi = 2995,
+ CODE_FOR_vec_permv16qi = 2996,
+ CODE_FOR_aarch64_im_lane_boundsi = 2998,
+ CODE_FOR_aarch64_st2v8qi = 2999,
+ CODE_FOR_aarch64_st2v4hi = 3000,
+ CODE_FOR_aarch64_st2v2si = 3001,
+ CODE_FOR_aarch64_st2v2sf = 3002,
+ CODE_FOR_aarch64_st2di = 3003,
+ CODE_FOR_aarch64_st2df = 3004,
+ CODE_FOR_aarch64_st3v8qi = 3005,
+ CODE_FOR_aarch64_st3v4hi = 3006,
+ CODE_FOR_aarch64_st3v2si = 3007,
+ CODE_FOR_aarch64_st3v2sf = 3008,
+ CODE_FOR_aarch64_st3di = 3009,
+ CODE_FOR_aarch64_st3df = 3010,
+ CODE_FOR_aarch64_st4v8qi = 3011,
+ CODE_FOR_aarch64_st4v4hi = 3012,
+ CODE_FOR_aarch64_st4v2si = 3013,
+ CODE_FOR_aarch64_st4v2sf = 3014,
+ CODE_FOR_aarch64_st4di = 3015,
+ CODE_FOR_aarch64_st4df = 3016,
+ CODE_FOR_aarch64_st2v16qi = 3017,
+ CODE_FOR_aarch64_st3v16qi = 3018,
+ CODE_FOR_aarch64_st4v16qi = 3019,
+ CODE_FOR_aarch64_st2v8hi = 3020,
+ CODE_FOR_aarch64_st3v8hi = 3021,
+ CODE_FOR_aarch64_st4v8hi = 3022,
+ CODE_FOR_aarch64_st2v4si = 3023,
+ CODE_FOR_aarch64_st3v4si = 3024,
+ CODE_FOR_aarch64_st4v4si = 3025,
+ CODE_FOR_aarch64_st2v2di = 3026,
+ CODE_FOR_aarch64_st3v2di = 3027,
+ CODE_FOR_aarch64_st4v2di = 3028,
+ CODE_FOR_aarch64_st2v4sf = 3029,
+ CODE_FOR_aarch64_st3v4sf = 3030,
+ CODE_FOR_aarch64_st4v4sf = 3031,
+ CODE_FOR_aarch64_st2v2df = 3032,
+ CODE_FOR_aarch64_st3v2df = 3033,
+ CODE_FOR_aarch64_st4v2df = 3034,
+ CODE_FOR_aarch64_st2_lanev16qi = 3035,
+ CODE_FOR_aarch64_st2_lanev8hi = 3036,
+ CODE_FOR_aarch64_st2_lanev4si = 3037,
+ CODE_FOR_aarch64_st2_lanev2di = 3038,
+ CODE_FOR_aarch64_st2_lanev4sf = 3039,
+ CODE_FOR_aarch64_st2_lanev2df = 3040,
+ CODE_FOR_aarch64_st3_lanev16qi = 3041,
+ CODE_FOR_aarch64_st3_lanev8hi = 3042,
+ CODE_FOR_aarch64_st3_lanev4si = 3043,
+ CODE_FOR_aarch64_st3_lanev2di = 3044,
+ CODE_FOR_aarch64_st3_lanev4sf = 3045,
+ CODE_FOR_aarch64_st3_lanev2df = 3046,
+ CODE_FOR_aarch64_st4_lanev16qi = 3047,
+ CODE_FOR_aarch64_st4_lanev8hi = 3048,
+ CODE_FOR_aarch64_st4_lanev4si = 3049,
+ CODE_FOR_aarch64_st4_lanev2di = 3050,
+ CODE_FOR_aarch64_st4_lanev4sf = 3051,
+ CODE_FOR_aarch64_st4_lanev2df = 3052,
+ CODE_FOR_aarch64_st1v8qi = 3053,
+ CODE_FOR_aarch64_st1v16qi = 3054,
+ CODE_FOR_aarch64_st1v4hi = 3055,
+ CODE_FOR_aarch64_st1v8hi = 3056,
+ CODE_FOR_aarch64_st1v2si = 3057,
+ CODE_FOR_aarch64_st1v4si = 3058,
+ CODE_FOR_aarch64_st1v2di = 3059,
+ CODE_FOR_aarch64_st1v2sf = 3060,
+ CODE_FOR_aarch64_st1v4sf = 3061,
+ CODE_FOR_aarch64_st1v2df = 3062,
+ CODE_FOR_aarch64_set_qregoiv16qi = 3063,
+ CODE_FOR_aarch64_set_qregciv16qi = 3064,
+ CODE_FOR_aarch64_set_qregxiv16qi = 3065,
+ CODE_FOR_aarch64_set_qregoiv8hi = 3066,
+ CODE_FOR_aarch64_set_qregciv8hi = 3067,
+ CODE_FOR_aarch64_set_qregxiv8hi = 3068,
+ CODE_FOR_aarch64_set_qregoiv4si = 3069,
+ CODE_FOR_aarch64_set_qregciv4si = 3070,
+ CODE_FOR_aarch64_set_qregxiv4si = 3071,
+ CODE_FOR_aarch64_set_qregoiv2di = 3072,
+ CODE_FOR_aarch64_set_qregciv2di = 3073,
+ CODE_FOR_aarch64_set_qregxiv2di = 3074,
+ CODE_FOR_aarch64_set_qregoiv4sf = 3075,
+ CODE_FOR_aarch64_set_qregciv4sf = 3076,
+ CODE_FOR_aarch64_set_qregxiv4sf = 3077,
+ CODE_FOR_aarch64_set_qregoiv2df = 3078,
+ CODE_FOR_aarch64_set_qregciv2df = 3079,
+ CODE_FOR_aarch64_set_qregxiv2df = 3080,
+ CODE_FOR_vec_initv8qi = 3081,
+ CODE_FOR_vec_initv16qi = 3082,
+ CODE_FOR_vec_initv4hi = 3083,
+ CODE_FOR_vec_initv8hi = 3084,
+ CODE_FOR_vec_initv2si = 3085,
+ CODE_FOR_vec_initv4si = 3086,
+ CODE_FOR_vec_initv2di = 3087,
+ CODE_FOR_vec_initv2sf = 3088,
+ CODE_FOR_vec_initv4sf = 3089,
+ CODE_FOR_vec_initv2df = 3090,
+ CODE_FOR_vec_extractv8qi = 3091,
+ CODE_FOR_vec_extractv16qi = 3092,
+ CODE_FOR_vec_extractv4hi = 3093,
+ CODE_FOR_vec_extractv8hi = 3094,
+ CODE_FOR_vec_extractv2si = 3095,
+ CODE_FOR_vec_extractv4si = 3096,
+ CODE_FOR_vec_extractv2di = 3097,
+ CODE_FOR_vec_extractv2sf = 3098,
+ CODE_FOR_vec_extractv4sf = 3099,
+ CODE_FOR_vec_extractv2df = 3100,
+ CODE_FOR_atomic_compare_and_swapqi = 3101,
+ CODE_FOR_atomic_compare_and_swaphi = 3102,
+ CODE_FOR_atomic_compare_and_swapsi = 3103,
+ CODE_FOR_atomic_compare_and_swapdi = 3104,
+ CODE_FOR_mem_thread_fence = 3185,
+ CODE_FOR_dmb = 3186,
LAST_INSN_CODE
};
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-constants.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-constants.h
index 6cd61f0..2cd05cb 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-constants.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-constants.h
@@ -4,197 +4,213 @@
#ifndef GCC_INSN_CONSTANTS_H
#define GCC_INSN_CONSTANTS_H
-#define R13_REGNUM 13
-#define R7_REGNUM 7
-#define R16_REGNUM 16
-#define R23_REGNUM 23
-#define R19_REGNUM 19
-#define R26_REGNUM 26
+#define SFP_REGNUM 64
+#define R12_REGNUM 12
+#define CC_REGNUM 66
#define R29_REGNUM 29
-#define R0_REGNUM 0
-#define R3_REGNUM 3
-#define SP_REGNUM 31
-#define R6_REGNUM 6
-#define R9_REGNUM 9
-#define V0_REGNUM 32
-#define R8_REGNUM 8
+#define R15_REGNUM 15
#define V15_REGNUM 47
-#define R11_REGNUM 11
+#define R2_REGNUM 2
+#define R18_REGNUM 18
+#define IP0_REGNUM 16
+#define R5_REGNUM 5
+#define R25_REGNUM 25
#define V31_REGNUM 63
+#define R9_REGNUM 9
+#define V0_REGNUM 32
+#define R10_REGNUM 10
#define R14_REGNUM 14
+#define SP_REGNUM 31
+#define R1_REGNUM 1
#define R21_REGNUM 21
#define R17_REGNUM 17
-#define R24_REGNUM 24
-#define R1_REGNUM 1
-#define R4_REGNUM 4
-#define AP_REGNUM 65
#define R20_REGNUM 20
+#define R4_REGNUM 4
+#define R24_REGNUM 24
+#define R22_REGNUM 22
#define R30_REGNUM 30
-#define LR_REGNUM 30
-#define SFP_REGNUM 64
-#define R12_REGNUM 12
+#define R8_REGNUM 8
#define R27_REGNUM 27
-#define R15_REGNUM 15
-#define R22_REGNUM 22
-#define CC_REGNUM 66
-#define R18_REGNUM 18
-#define R25_REGNUM 25
+#define R13_REGNUM 13
+#define R11_REGNUM 11
+#define AP_REGNUM 65
+#define R0_REGNUM 0
+#define R6_REGNUM 6
+#define LR_REGNUM 30
#define R28_REGNUM 28
-#define R2_REGNUM 2
+#define R16_REGNUM 16
+#define R3_REGNUM 3
+#define R23_REGNUM 23
+#define R19_REGNUM 19
#define IP1_REGNUM 17
-#define R10_REGNUM 10
-#define R5_REGNUM 5
-#define IP0_REGNUM 16
+#define R7_REGNUM 7
+#define R26_REGNUM 26
enum unspec {
UNSPEC_CASESI = 0,
UNSPEC_CLS = 1,
- UNSPEC_FRECPE = 2,
- UNSPEC_FRECPS = 3,
- UNSPEC_FRECPX = 4,
- UNSPEC_FRINTA = 5,
- UNSPEC_FRINTI = 6,
- UNSPEC_FRINTM = 7,
- UNSPEC_FRINTN = 8,
- UNSPEC_FRINTP = 9,
- UNSPEC_FRINTX = 10,
- UNSPEC_FRINTZ = 11,
- UNSPEC_GOTSMALLPIC = 12,
- UNSPEC_GOTSMALLTLS = 13,
- UNSPEC_GOTTINYPIC = 14,
- UNSPEC_LD1 = 15,
- UNSPEC_LD2 = 16,
- UNSPEC_LD3 = 17,
- UNSPEC_LD4 = 18,
- UNSPEC_MB = 19,
- UNSPEC_NOP = 20,
- UNSPEC_PRLG_STK = 21,
- UNSPEC_RBIT = 22,
- UNSPEC_SISD_NEG = 23,
- UNSPEC_SISD_SSHL = 24,
- UNSPEC_SISD_USHL = 25,
- UNSPEC_SSHL_2S = 26,
- UNSPEC_SSHR64 = 27,
- UNSPEC_ST1 = 28,
- UNSPEC_ST2 = 29,
- UNSPEC_ST3 = 30,
- UNSPEC_ST4 = 31,
- UNSPEC_ST2_LANE = 32,
- UNSPEC_ST3_LANE = 33,
- UNSPEC_ST4_LANE = 34,
- UNSPEC_TLS = 35,
- UNSPEC_TLSDESC = 36,
- UNSPEC_USHL_2S = 37,
- UNSPEC_USHR64 = 38,
- UNSPEC_VSTRUCTDUMMY = 39,
- UNSPEC_SP_SET = 40,
- UNSPEC_SP_TEST = 41,
- UNSPEC_ASHIFT_SIGNED = 42,
- UNSPEC_ASHIFT_UNSIGNED = 43,
- UNSPEC_FMAX = 44,
- UNSPEC_FMAXNMV = 45,
- UNSPEC_FMAXV = 46,
- UNSPEC_FMIN = 47,
- UNSPEC_FMINNMV = 48,
- UNSPEC_FMINV = 49,
- UNSPEC_FADDV = 50,
- UNSPEC_SADDV = 51,
- UNSPEC_UADDV = 52,
- UNSPEC_SMAXV = 53,
- UNSPEC_SMINV = 54,
- UNSPEC_UMAXV = 55,
- UNSPEC_UMINV = 56,
- UNSPEC_SHADD = 57,
- UNSPEC_UHADD = 58,
- UNSPEC_SRHADD = 59,
- UNSPEC_URHADD = 60,
- UNSPEC_SHSUB = 61,
- UNSPEC_UHSUB = 62,
- UNSPEC_SRHSUB = 63,
- UNSPEC_URHSUB = 64,
- UNSPEC_ADDHN = 65,
- UNSPEC_RADDHN = 66,
- UNSPEC_SUBHN = 67,
- UNSPEC_RSUBHN = 68,
- UNSPEC_ADDHN2 = 69,
- UNSPEC_RADDHN2 = 70,
- UNSPEC_SUBHN2 = 71,
- UNSPEC_RSUBHN2 = 72,
- UNSPEC_SQDMULH = 73,
- UNSPEC_SQRDMULH = 74,
- UNSPEC_PMUL = 75,
- UNSPEC_USQADD = 76,
- UNSPEC_SUQADD = 77,
- UNSPEC_SQXTUN = 78,
- UNSPEC_SQXTN = 79,
- UNSPEC_UQXTN = 80,
- UNSPEC_SSRA = 81,
- UNSPEC_USRA = 82,
- UNSPEC_SRSRA = 83,
- UNSPEC_URSRA = 84,
- UNSPEC_SRSHR = 85,
- UNSPEC_URSHR = 86,
- UNSPEC_SQSHLU = 87,
- UNSPEC_SQSHL = 88,
- UNSPEC_UQSHL = 89,
- UNSPEC_SQSHRUN = 90,
- UNSPEC_SQRSHRUN = 91,
- UNSPEC_SQSHRN = 92,
- UNSPEC_UQSHRN = 93,
- UNSPEC_SQRSHRN = 94,
- UNSPEC_UQRSHRN = 95,
- UNSPEC_SSHL = 96,
- UNSPEC_USHL = 97,
- UNSPEC_SRSHL = 98,
- UNSPEC_URSHL = 99,
- UNSPEC_SQRSHL = 100,
- UNSPEC_UQRSHL = 101,
- UNSPEC_SSLI = 102,
- UNSPEC_USLI = 103,
- UNSPEC_SSRI = 104,
- UNSPEC_USRI = 105,
- UNSPEC_SSHLL = 106,
- UNSPEC_USHLL = 107,
- UNSPEC_ADDP = 108,
- UNSPEC_TBL = 109,
- UNSPEC_CONCAT = 110,
- UNSPEC_ZIP1 = 111,
- UNSPEC_ZIP2 = 112,
- UNSPEC_UZP1 = 113,
- UNSPEC_UZP2 = 114,
- UNSPEC_TRN1 = 115,
- UNSPEC_TRN2 = 116,
- UNSPEC_AESE = 117,
- UNSPEC_AESD = 118,
- UNSPEC_AESMC = 119,
- UNSPEC_AESIMC = 120,
- UNSPEC_SHA1C = 121,
- UNSPEC_SHA1M = 122,
- UNSPEC_SHA1P = 123,
- UNSPEC_SHA1H = 124,
- UNSPEC_SHA1SU0 = 125,
- UNSPEC_SHA1SU1 = 126,
- UNSPEC_SHA256H = 127,
- UNSPEC_SHA256H2 = 128,
- UNSPEC_SHA256SU0 = 129,
- UNSPEC_SHA256SU1 = 130,
- UNSPEC_PMULL = 131,
- UNSPEC_PMULL2 = 132
+ UNSPEC_CRC32B = 2,
+ UNSPEC_CRC32CB = 3,
+ UNSPEC_CRC32CH = 4,
+ UNSPEC_CRC32CW = 5,
+ UNSPEC_CRC32CX = 6,
+ UNSPEC_CRC32H = 7,
+ UNSPEC_CRC32W = 8,
+ UNSPEC_CRC32X = 9,
+ UNSPEC_FRECPE = 10,
+ UNSPEC_FRECPS = 11,
+ UNSPEC_FRECPX = 12,
+ UNSPEC_FRINTA = 13,
+ UNSPEC_FRINTI = 14,
+ UNSPEC_FRINTM = 15,
+ UNSPEC_FRINTN = 16,
+ UNSPEC_FRINTP = 17,
+ UNSPEC_FRINTX = 18,
+ UNSPEC_FRINTZ = 19,
+ UNSPEC_GOTSMALLPIC = 20,
+ UNSPEC_GOTSMALLTLS = 21,
+ UNSPEC_GOTTINYPIC = 22,
+ UNSPEC_LD1 = 23,
+ UNSPEC_LD2 = 24,
+ UNSPEC_LD3 = 25,
+ UNSPEC_LD4 = 26,
+ UNSPEC_MB = 27,
+ UNSPEC_NOP = 28,
+ UNSPEC_PRLG_STK = 29,
+ UNSPEC_RBIT = 30,
+ UNSPEC_SISD_NEG = 31,
+ UNSPEC_SISD_SSHL = 32,
+ UNSPEC_SISD_USHL = 33,
+ UNSPEC_SSHL_2S = 34,
+ UNSPEC_SSHR64 = 35,
+ UNSPEC_ST1 = 36,
+ UNSPEC_ST2 = 37,
+ UNSPEC_ST3 = 38,
+ UNSPEC_ST4 = 39,
+ UNSPEC_ST2_LANE = 40,
+ UNSPEC_ST3_LANE = 41,
+ UNSPEC_ST4_LANE = 42,
+ UNSPEC_TLS = 43,
+ UNSPEC_TLSDESC = 44,
+ UNSPEC_USHL_2S = 45,
+ UNSPEC_USHR64 = 46,
+ UNSPEC_VSTRUCTDUMMY = 47,
+ UNSPEC_SP_SET = 48,
+ UNSPEC_SP_TEST = 49,
+ UNSPEC_ASHIFT_SIGNED = 50,
+ UNSPEC_ASHIFT_UNSIGNED = 51,
+ UNSPEC_FMAX = 52,
+ UNSPEC_FMAXNMV = 53,
+ UNSPEC_FMAXV = 54,
+ UNSPEC_FMIN = 55,
+ UNSPEC_FMINNMV = 56,
+ UNSPEC_FMINV = 57,
+ UNSPEC_FADDV = 58,
+ UNSPEC_SADDV = 59,
+ UNSPEC_UADDV = 60,
+ UNSPEC_SMAXV = 61,
+ UNSPEC_SMINV = 62,
+ UNSPEC_UMAXV = 63,
+ UNSPEC_UMINV = 64,
+ UNSPEC_SHADD = 65,
+ UNSPEC_UHADD = 66,
+ UNSPEC_SRHADD = 67,
+ UNSPEC_URHADD = 68,
+ UNSPEC_SHSUB = 69,
+ UNSPEC_UHSUB = 70,
+ UNSPEC_SRHSUB = 71,
+ UNSPEC_URHSUB = 72,
+ UNSPEC_ADDHN = 73,
+ UNSPEC_RADDHN = 74,
+ UNSPEC_SUBHN = 75,
+ UNSPEC_RSUBHN = 76,
+ UNSPEC_ADDHN2 = 77,
+ UNSPEC_RADDHN2 = 78,
+ UNSPEC_SUBHN2 = 79,
+ UNSPEC_RSUBHN2 = 80,
+ UNSPEC_SQDMULH = 81,
+ UNSPEC_SQRDMULH = 82,
+ UNSPEC_PMUL = 83,
+ UNSPEC_USQADD = 84,
+ UNSPEC_SUQADD = 85,
+ UNSPEC_SQXTUN = 86,
+ UNSPEC_SQXTN = 87,
+ UNSPEC_UQXTN = 88,
+ UNSPEC_SSRA = 89,
+ UNSPEC_USRA = 90,
+ UNSPEC_SRSRA = 91,
+ UNSPEC_URSRA = 92,
+ UNSPEC_SRSHR = 93,
+ UNSPEC_URSHR = 94,
+ UNSPEC_SQSHLU = 95,
+ UNSPEC_SQSHL = 96,
+ UNSPEC_UQSHL = 97,
+ UNSPEC_SQSHRUN = 98,
+ UNSPEC_SQRSHRUN = 99,
+ UNSPEC_SQSHRN = 100,
+ UNSPEC_UQSHRN = 101,
+ UNSPEC_SQRSHRN = 102,
+ UNSPEC_UQRSHRN = 103,
+ UNSPEC_SSHL = 104,
+ UNSPEC_USHL = 105,
+ UNSPEC_SRSHL = 106,
+ UNSPEC_URSHL = 107,
+ UNSPEC_SQRSHL = 108,
+ UNSPEC_UQRSHL = 109,
+ UNSPEC_SSLI = 110,
+ UNSPEC_USLI = 111,
+ UNSPEC_SSRI = 112,
+ UNSPEC_USRI = 113,
+ UNSPEC_SSHLL = 114,
+ UNSPEC_USHLL = 115,
+ UNSPEC_ADDP = 116,
+ UNSPEC_TBL = 117,
+ UNSPEC_CONCAT = 118,
+ UNSPEC_ZIP1 = 119,
+ UNSPEC_ZIP2 = 120,
+ UNSPEC_UZP1 = 121,
+ UNSPEC_UZP2 = 122,
+ UNSPEC_TRN1 = 123,
+ UNSPEC_TRN2 = 124,
+ UNSPEC_EXT = 125,
+ UNSPEC_REV64 = 126,
+ UNSPEC_REV32 = 127,
+ UNSPEC_REV16 = 128,
+ UNSPEC_AESE = 129,
+ UNSPEC_AESD = 130,
+ UNSPEC_AESMC = 131,
+ UNSPEC_AESIMC = 132,
+ UNSPEC_SHA1C = 133,
+ UNSPEC_SHA1M = 134,
+ UNSPEC_SHA1P = 135,
+ UNSPEC_SHA1H = 136,
+ UNSPEC_SHA1SU0 = 137,
+ UNSPEC_SHA1SU1 = 138,
+ UNSPEC_SHA256H = 139,
+ UNSPEC_SHA256H2 = 140,
+ UNSPEC_SHA256SU0 = 141,
+ UNSPEC_SHA256SU1 = 142,
+ UNSPEC_PMULL = 143,
+ UNSPEC_PMULL2 = 144
};
-#define NUM_UNSPEC_VALUES 133
+#define NUM_UNSPEC_VALUES 145
extern const char *const unspec_strings[];
enum unspecv {
UNSPECV_EH_RETURN = 0,
- UNSPECV_LX = 1,
- UNSPECV_SX = 2,
- UNSPECV_LDA = 3,
- UNSPECV_STL = 4,
- UNSPECV_ATOMIC_CMPSW = 5,
- UNSPECV_ATOMIC_EXCHG = 6,
- UNSPECV_ATOMIC_OP = 7
+ UNSPECV_GET_FPCR = 1,
+ UNSPECV_SET_FPCR = 2,
+ UNSPECV_GET_FPSR = 3,
+ UNSPECV_SET_FPSR = 4,
+ UNSPECV_LX = 5,
+ UNSPECV_SX = 6,
+ UNSPECV_LDA = 7,
+ UNSPECV_STL = 8,
+ UNSPECV_ATOMIC_CMPSW = 9,
+ UNSPECV_ATOMIC_EXCHG = 10,
+ UNSPECV_ATOMIC_OP = 11
};
-#define NUM_UNSPECV_VALUES 8
+#define NUM_UNSPECV_VALUES 12
extern const char *const unspecv_strings[];
#endif /* GCC_INSN_CONSTANTS_H */
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-flags.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-flags.h
index 469874b..28e9777 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-flags.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-flags.h
@@ -87,6 +87,14 @@
#define HAVE_cstorehi_neg 1
#define HAVE_cstoresi_neg 1
#define HAVE_cstoredi_neg 1
+#define HAVE_aarch64_crc32b (TARGET_CRC32)
+#define HAVE_aarch64_crc32h (TARGET_CRC32)
+#define HAVE_aarch64_crc32w (TARGET_CRC32)
+#define HAVE_aarch64_crc32x (TARGET_CRC32)
+#define HAVE_aarch64_crc32cb (TARGET_CRC32)
+#define HAVE_aarch64_crc32ch (TARGET_CRC32)
+#define HAVE_aarch64_crc32cw (TARGET_CRC32)
+#define HAVE_aarch64_crc32cx (TARGET_CRC32)
#define HAVE_csinc3si_insn 1
#define HAVE_csinc3di_insn 1
#define HAVE_andsi3 1
@@ -240,6 +248,10 @@
#define HAVE_stack_protect_set_di (ptr_mode == DImode)
#define HAVE_stack_protect_test_si (ptr_mode == SImode)
#define HAVE_stack_protect_test_di (ptr_mode == DImode)
+#define HAVE_set_fpcr 1
+#define HAVE_get_fpcr 1
+#define HAVE_set_fpsr 1
+#define HAVE_get_fpsr 1
#define HAVE_aarch64_simd_dupv8qi (TARGET_SIMD)
#define HAVE_aarch64_simd_dupv16qi (TARGET_SIMD)
#define HAVE_aarch64_simd_dupv4hi (TARGET_SIMD)
@@ -468,18 +480,30 @@
#define HAVE_sminv4si3 (TARGET_SIMD)
#define HAVE_umaxv4si3 (TARGET_SIMD)
#define HAVE_uminv4si3 (TARGET_SIMD)
-#define HAVE_move_lo_quad_v16qi (TARGET_SIMD)
-#define HAVE_move_lo_quad_v8hi (TARGET_SIMD)
-#define HAVE_move_lo_quad_v4si (TARGET_SIMD)
-#define HAVE_move_lo_quad_v2di (TARGET_SIMD)
-#define HAVE_move_lo_quad_v4sf (TARGET_SIMD)
-#define HAVE_move_lo_quad_v2df (TARGET_SIMD)
-#define HAVE_aarch64_simd_move_hi_quad_v16qi (TARGET_SIMD)
-#define HAVE_aarch64_simd_move_hi_quad_v8hi (TARGET_SIMD)
-#define HAVE_aarch64_simd_move_hi_quad_v4si (TARGET_SIMD)
-#define HAVE_aarch64_simd_move_hi_quad_v2di (TARGET_SIMD)
-#define HAVE_aarch64_simd_move_hi_quad_v4sf (TARGET_SIMD)
-#define HAVE_aarch64_simd_move_hi_quad_v2df (TARGET_SIMD)
+#define HAVE_move_lo_quad_internal_v16qi (TARGET_SIMD && !BYTES_BIG_ENDIAN)
+#define HAVE_move_lo_quad_internal_v8hi (TARGET_SIMD && !BYTES_BIG_ENDIAN)
+#define HAVE_move_lo_quad_internal_v4si (TARGET_SIMD && !BYTES_BIG_ENDIAN)
+#define HAVE_move_lo_quad_internal_v2di (TARGET_SIMD && !BYTES_BIG_ENDIAN)
+#define HAVE_move_lo_quad_internal_v4sf (TARGET_SIMD && !BYTES_BIG_ENDIAN)
+#define HAVE_move_lo_quad_internal_v2df (TARGET_SIMD && !BYTES_BIG_ENDIAN)
+#define HAVE_move_lo_quad_internal_be_v16qi (TARGET_SIMD && BYTES_BIG_ENDIAN)
+#define HAVE_move_lo_quad_internal_be_v8hi (TARGET_SIMD && BYTES_BIG_ENDIAN)
+#define HAVE_move_lo_quad_internal_be_v4si (TARGET_SIMD && BYTES_BIG_ENDIAN)
+#define HAVE_move_lo_quad_internal_be_v2di (TARGET_SIMD && BYTES_BIG_ENDIAN)
+#define HAVE_move_lo_quad_internal_be_v4sf (TARGET_SIMD && BYTES_BIG_ENDIAN)
+#define HAVE_move_lo_quad_internal_be_v2df (TARGET_SIMD && BYTES_BIG_ENDIAN)
+#define HAVE_aarch64_simd_move_hi_quad_v16qi (TARGET_SIMD && !BYTES_BIG_ENDIAN)
+#define HAVE_aarch64_simd_move_hi_quad_v8hi (TARGET_SIMD && !BYTES_BIG_ENDIAN)
+#define HAVE_aarch64_simd_move_hi_quad_v4si (TARGET_SIMD && !BYTES_BIG_ENDIAN)
+#define HAVE_aarch64_simd_move_hi_quad_v2di (TARGET_SIMD && !BYTES_BIG_ENDIAN)
+#define HAVE_aarch64_simd_move_hi_quad_v4sf (TARGET_SIMD && !BYTES_BIG_ENDIAN)
+#define HAVE_aarch64_simd_move_hi_quad_v2df (TARGET_SIMD && !BYTES_BIG_ENDIAN)
+#define HAVE_aarch64_simd_move_hi_quad_be_v16qi (TARGET_SIMD && BYTES_BIG_ENDIAN)
+#define HAVE_aarch64_simd_move_hi_quad_be_v8hi (TARGET_SIMD && BYTES_BIG_ENDIAN)
+#define HAVE_aarch64_simd_move_hi_quad_be_v4si (TARGET_SIMD && BYTES_BIG_ENDIAN)
+#define HAVE_aarch64_simd_move_hi_quad_be_v2di (TARGET_SIMD && BYTES_BIG_ENDIAN)
+#define HAVE_aarch64_simd_move_hi_quad_be_v4sf (TARGET_SIMD && BYTES_BIG_ENDIAN)
+#define HAVE_aarch64_simd_move_hi_quad_be_v2df (TARGET_SIMD && BYTES_BIG_ENDIAN)
#define HAVE_aarch64_simd_vec_pack_trunc_v8hi (TARGET_SIMD)
#define HAVE_aarch64_simd_vec_pack_trunc_v4si (TARGET_SIMD)
#define HAVE_aarch64_simd_vec_pack_trunc_v2di (TARGET_SIMD)
@@ -692,12 +716,12 @@
#define HAVE_aarch64_get_lanev2sf (TARGET_SIMD)
#define HAVE_aarch64_get_lanev4sf (TARGET_SIMD)
#define HAVE_aarch64_get_lanev2df (TARGET_SIMD)
-#define HAVE_aarch64_combinev8qi (TARGET_SIMD)
-#define HAVE_aarch64_combinev4hi (TARGET_SIMD)
-#define HAVE_aarch64_combinev2si (TARGET_SIMD)
-#define HAVE_aarch64_combinev2sf (TARGET_SIMD)
-#define HAVE_aarch64_combinedi (TARGET_SIMD)
-#define HAVE_aarch64_combinedf (TARGET_SIMD)
+#define HAVE_aarch64_combine_internalv8qi (TARGET_SIMD)
+#define HAVE_aarch64_combine_internalv4hi (TARGET_SIMD)
+#define HAVE_aarch64_combine_internalv2si (TARGET_SIMD)
+#define HAVE_aarch64_combine_internalv2sf (TARGET_SIMD)
+#define HAVE_aarch64_combine_internaldi (TARGET_SIMD)
+#define HAVE_aarch64_combine_internaldf (TARGET_SIMD)
#define HAVE_aarch64_saddlv16qi_hi_internal (TARGET_SIMD)
#define HAVE_aarch64_ssublv16qi_hi_internal (TARGET_SIMD)
#define HAVE_aarch64_uaddlv16qi_hi_internal (TARGET_SIMD)
@@ -950,26 +974,26 @@
#define HAVE_aarch64_sqrdmulhhi (TARGET_SIMD)
#define HAVE_aarch64_sqdmulhsi (TARGET_SIMD)
#define HAVE_aarch64_sqrdmulhsi (TARGET_SIMD)
-#define HAVE_aarch64_sqdmulh_lanev4hi (TARGET_SIMD)
-#define HAVE_aarch64_sqrdmulh_lanev4hi (TARGET_SIMD)
-#define HAVE_aarch64_sqdmulh_lanev8hi (TARGET_SIMD)
-#define HAVE_aarch64_sqrdmulh_lanev8hi (TARGET_SIMD)
-#define HAVE_aarch64_sqdmulh_lanev2si (TARGET_SIMD)
-#define HAVE_aarch64_sqrdmulh_lanev2si (TARGET_SIMD)
-#define HAVE_aarch64_sqdmulh_lanev4si (TARGET_SIMD)
-#define HAVE_aarch64_sqrdmulh_lanev4si (TARGET_SIMD)
-#define HAVE_aarch64_sqdmulh_laneqv4hi (TARGET_SIMD)
-#define HAVE_aarch64_sqrdmulh_laneqv4hi (TARGET_SIMD)
-#define HAVE_aarch64_sqdmulh_laneqv8hi (TARGET_SIMD)
-#define HAVE_aarch64_sqrdmulh_laneqv8hi (TARGET_SIMD)
-#define HAVE_aarch64_sqdmulh_laneqv2si (TARGET_SIMD)
-#define HAVE_aarch64_sqrdmulh_laneqv2si (TARGET_SIMD)
-#define HAVE_aarch64_sqdmulh_laneqv4si (TARGET_SIMD)
-#define HAVE_aarch64_sqrdmulh_laneqv4si (TARGET_SIMD)
-#define HAVE_aarch64_sqdmulh_lanehi (TARGET_SIMD)
-#define HAVE_aarch64_sqrdmulh_lanehi (TARGET_SIMD)
-#define HAVE_aarch64_sqdmulh_lanesi (TARGET_SIMD)
-#define HAVE_aarch64_sqrdmulh_lanesi (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_lanev4hi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_lanev4hi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_lanev8hi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_lanev8hi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_lanev2si_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_lanev2si_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_lanev4si_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_lanev4si_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_laneqv4hi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_laneqv4hi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_laneqv8hi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_laneqv8hi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_laneqv2si_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_laneqv2si_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_laneqv4si_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_laneqv4si_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_lanehi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_lanehi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_lanesi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_lanesi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmlalv4hi (TARGET_SIMD)
#define HAVE_aarch64_sqdmlslv4hi (TARGET_SIMD)
#define HAVE_aarch64_sqdmlalv2si (TARGET_SIMD)
@@ -982,10 +1006,18 @@
#define HAVE_aarch64_sqdmlsl_lanev4hi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmlal_lanev2si_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmlsl_lanev2si_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmlal_laneqv4hi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmlsl_laneqv4hi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmlal_laneqv2si_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmlsl_laneqv2si_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmlal_lanehi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmlsl_lanehi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmlal_lanesi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmlsl_lanesi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmlal_laneqhi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmlsl_laneqhi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmlal_laneqsi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmlsl_laneqsi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmlal_nv4hi (TARGET_SIMD)
#define HAVE_aarch64_sqdmlsl_nv4hi (TARGET_SIMD)
#define HAVE_aarch64_sqdmlal_nv2si (TARGET_SIMD)
@@ -998,6 +1030,10 @@
#define HAVE_aarch64_sqdmlsl2_lanev8hi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmlal2_lanev4si_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmlsl2_lanev4si_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmlal2_laneqv8hi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmlsl2_laneqv8hi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmlal2_laneqv4si_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmlsl2_laneqv4si_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmlal2_nv8hi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmlsl2_nv8hi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmlal2_nv4si_internal (TARGET_SIMD)
@@ -1008,14 +1044,20 @@
#define HAVE_aarch64_sqdmullsi (TARGET_SIMD)
#define HAVE_aarch64_sqdmull_lanev4hi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmull_lanev2si_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmull_laneqv4hi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmull_laneqv2si_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmull_lanehi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmull_lanesi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmull_laneqhi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmull_laneqsi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmull_nv4hi (TARGET_SIMD)
#define HAVE_aarch64_sqdmull_nv2si (TARGET_SIMD)
#define HAVE_aarch64_sqdmull2v8hi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmull2v4si_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmull2_lanev8hi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmull2_lanev4si_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmull2_laneqv8hi_internal (TARGET_SIMD)
+#define HAVE_aarch64_sqdmull2_laneqv4si_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmull2_nv8hi_internal (TARGET_SIMD)
#define HAVE_aarch64_sqdmull2_nv4si_internal (TARGET_SIMD)
#define HAVE_aarch64_sshlv8qi (TARGET_SIMD)
@@ -1525,6 +1567,46 @@
#define HAVE_aarch64_trn2v2df (TARGET_SIMD)
#define HAVE_aarch64_uzp1v2df (TARGET_SIMD)
#define HAVE_aarch64_uzp2v2df (TARGET_SIMD)
+#define HAVE_aarch64_extv8qi (TARGET_SIMD)
+#define HAVE_aarch64_extv16qi (TARGET_SIMD)
+#define HAVE_aarch64_extv4hi (TARGET_SIMD)
+#define HAVE_aarch64_extv8hi (TARGET_SIMD)
+#define HAVE_aarch64_extv2si (TARGET_SIMD)
+#define HAVE_aarch64_extv4si (TARGET_SIMD)
+#define HAVE_aarch64_extv2di (TARGET_SIMD)
+#define HAVE_aarch64_extv2sf (TARGET_SIMD)
+#define HAVE_aarch64_extv4sf (TARGET_SIMD)
+#define HAVE_aarch64_extv2df (TARGET_SIMD)
+#define HAVE_aarch64_rev64v8qi (TARGET_SIMD)
+#define HAVE_aarch64_rev32v8qi (TARGET_SIMD)
+#define HAVE_aarch64_rev16v8qi (TARGET_SIMD)
+#define HAVE_aarch64_rev64v16qi (TARGET_SIMD)
+#define HAVE_aarch64_rev32v16qi (TARGET_SIMD)
+#define HAVE_aarch64_rev16v16qi (TARGET_SIMD)
+#define HAVE_aarch64_rev64v4hi (TARGET_SIMD)
+#define HAVE_aarch64_rev32v4hi (TARGET_SIMD)
+#define HAVE_aarch64_rev16v4hi (TARGET_SIMD)
+#define HAVE_aarch64_rev64v8hi (TARGET_SIMD)
+#define HAVE_aarch64_rev32v8hi (TARGET_SIMD)
+#define HAVE_aarch64_rev16v8hi (TARGET_SIMD)
+#define HAVE_aarch64_rev64v2si (TARGET_SIMD)
+#define HAVE_aarch64_rev32v2si (TARGET_SIMD)
+#define HAVE_aarch64_rev16v2si (TARGET_SIMD)
+#define HAVE_aarch64_rev64v4si (TARGET_SIMD)
+#define HAVE_aarch64_rev32v4si (TARGET_SIMD)
+#define HAVE_aarch64_rev16v4si (TARGET_SIMD)
+#define HAVE_aarch64_rev64v2di (TARGET_SIMD)
+#define HAVE_aarch64_rev32v2di (TARGET_SIMD)
+#define HAVE_aarch64_rev16v2di (TARGET_SIMD)
+#define HAVE_aarch64_rev64v2sf (TARGET_SIMD)
+#define HAVE_aarch64_rev32v2sf (TARGET_SIMD)
+#define HAVE_aarch64_rev16v2sf (TARGET_SIMD)
+#define HAVE_aarch64_rev64v4sf (TARGET_SIMD)
+#define HAVE_aarch64_rev32v4sf (TARGET_SIMD)
+#define HAVE_aarch64_rev16v4sf (TARGET_SIMD)
+#define HAVE_aarch64_rev64v2df (TARGET_SIMD)
+#define HAVE_aarch64_rev32v2df (TARGET_SIMD)
+#define HAVE_aarch64_rev16v2df (TARGET_SIMD)
#define HAVE_aarch64_st2v8qi_dreg (TARGET_SIMD)
#define HAVE_aarch64_st2v4hi_dreg (TARGET_SIMD)
#define HAVE_aarch64_st2v2si_dreg (TARGET_SIMD)
@@ -1688,6 +1770,7 @@
#define HAVE_movsf 1
#define HAVE_movdf 1
#define HAVE_movtf 1
+#define HAVE_movmemdi (!STRICT_ALIGNMENT)
#define HAVE_extendsidi2 1
#define HAVE_zero_extendsidi2 1
#define HAVE_extendqisi2 1
@@ -1831,6 +1914,12 @@
#define HAVE_vec_setv2sf (TARGET_SIMD)
#define HAVE_vec_setv4sf (TARGET_SIMD)
#define HAVE_vec_setv2df (TARGET_SIMD)
+#define HAVE_move_lo_quad_v16qi (TARGET_SIMD)
+#define HAVE_move_lo_quad_v8hi (TARGET_SIMD)
+#define HAVE_move_lo_quad_v4si (TARGET_SIMD)
+#define HAVE_move_lo_quad_v2di (TARGET_SIMD)
+#define HAVE_move_lo_quad_v4sf (TARGET_SIMD)
+#define HAVE_move_lo_quad_v2df (TARGET_SIMD)
#define HAVE_move_hi_quad_v16qi (TARGET_SIMD)
#define HAVE_move_hi_quad_v8hi (TARGET_SIMD)
#define HAVE_move_hi_quad_v4si (TARGET_SIMD)
@@ -2028,6 +2117,12 @@
#define HAVE_aarch64_reinterpretv2dfv2di (TARGET_SIMD)
#define HAVE_aarch64_reinterpretv2dfv4sf (TARGET_SIMD)
#define HAVE_aarch64_reinterpretv2dfv2df (TARGET_SIMD)
+#define HAVE_aarch64_combinev8qi (TARGET_SIMD)
+#define HAVE_aarch64_combinev4hi (TARGET_SIMD)
+#define HAVE_aarch64_combinev2si (TARGET_SIMD)
+#define HAVE_aarch64_combinev2sf (TARGET_SIMD)
+#define HAVE_aarch64_combinedi (TARGET_SIMD)
+#define HAVE_aarch64_combinedf (TARGET_SIMD)
#define HAVE_aarch64_simd_combinev8qi (TARGET_SIMD)
#define HAVE_aarch64_simd_combinev4hi (TARGET_SIMD)
#define HAVE_aarch64_simd_combinev2si (TARGET_SIMD)
@@ -2058,6 +2153,26 @@
#define HAVE_aarch64_usubw2v16qi (TARGET_SIMD)
#define HAVE_aarch64_usubw2v8hi (TARGET_SIMD)
#define HAVE_aarch64_usubw2v4si (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_lanev4hi (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_lanev8hi (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_lanev2si (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_lanev4si (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_lanev4hi (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_lanev8hi (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_lanev2si (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_lanev4si (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_laneqv4hi (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_laneqv8hi (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_laneqv2si (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_laneqv4si (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_laneqv4hi (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_laneqv8hi (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_laneqv2si (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_laneqv4si (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_lanehi (TARGET_SIMD)
+#define HAVE_aarch64_sqdmulh_lanesi (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_lanehi (TARGET_SIMD)
+#define HAVE_aarch64_sqrdmulh_lanesi (TARGET_SIMD)
#define HAVE_aarch64_sqdmlal_lanev4hi (TARGET_SIMD)
#define HAVE_aarch64_sqdmlal_lanev2si (TARGET_SIMD)
#define HAVE_aarch64_sqdmlal_lanehi (TARGET_SIMD)
@@ -2201,6 +2316,7 @@
#define HAVE_vec_perm_constv2df (TARGET_SIMD)
#define HAVE_vec_permv8qi (TARGET_SIMD && !BYTES_BIG_ENDIAN)
#define HAVE_vec_permv16qi (TARGET_SIMD && !BYTES_BIG_ENDIAN)
+#define HAVE_aarch64_im_lane_boundsi (TARGET_SIMD)
#define HAVE_aarch64_st2v8qi (TARGET_SIMD)
#define HAVE_aarch64_st2v4hi (TARGET_SIMD)
#define HAVE_aarch64_st2v2si (TARGET_SIMD)
@@ -2309,2284 +2425,2400 @@
#define HAVE_atomic_compare_and_swapdi 1
#define HAVE_mem_thread_fence 1
#define HAVE_dmb 1
-extern rtx gen_indirect_jump (rtx);
-extern rtx gen_jump (rtx);
-extern rtx gen_casesi_dispatch (rtx, rtx, rtx);
-extern rtx gen_nop (void);
-extern rtx gen_trap (void);
-extern rtx gen_eh_return (rtx);
-extern rtx gen_insv_immsi (rtx, rtx, rtx);
-extern rtx gen_insv_immdi (rtx, rtx, rtx);
-extern rtx gen_load_pairsi (rtx, rtx, rtx, rtx);
-extern rtx gen_load_pairdi (rtx, rtx, rtx, rtx);
-extern rtx gen_store_pairsi (rtx, rtx, rtx, rtx);
-extern rtx gen_store_pairdi (rtx, rtx, rtx, rtx);
-extern rtx gen_load_pairsf (rtx, rtx, rtx, rtx);
-extern rtx gen_load_pairdf (rtx, rtx, rtx, rtx);
-extern rtx gen_store_pairsf (rtx, rtx, rtx, rtx);
-extern rtx gen_store_pairdf (rtx, rtx, rtx, rtx);
-extern rtx gen_loadwb_pairsi_si (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_loadwb_pairsi_di (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_loadwb_pairdi_si (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_loadwb_pairdi_di (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_storewb_pairsi_si (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_storewb_pairsi_di (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_storewb_pairdi_si (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_storewb_pairdi_di (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_addsi3_compare0 (rtx, rtx, rtx);
-extern rtx gen_adddi3_compare0 (rtx, rtx, rtx);
-extern rtx gen_addsi3_carryin (rtx, rtx, rtx);
-extern rtx gen_adddi3_carryin (rtx, rtx, rtx);
-extern rtx gen_subsi3 (rtx, rtx, rtx);
-extern rtx gen_subdi3 (rtx, rtx, rtx);
-extern rtx gen_subsi3_compare0 (rtx, rtx, rtx);
-extern rtx gen_subdi3_compare0 (rtx, rtx, rtx);
-extern rtx gen_subsi3_carryin (rtx, rtx, rtx);
-extern rtx gen_subdi3_carryin (rtx, rtx, rtx);
-extern rtx gen_absdi2 (rtx, rtx);
-extern rtx gen_negsi2 (rtx, rtx);
-extern rtx gen_negdi2 (rtx, rtx);
-extern rtx gen_mulsi3 (rtx, rtx, rtx);
-extern rtx gen_muldi3 (rtx, rtx, rtx);
-extern rtx gen_maddsi (rtx, rtx, rtx, rtx);
-extern rtx gen_madddi (rtx, rtx, rtx, rtx);
-extern rtx gen_mulsidi3 (rtx, rtx, rtx);
-extern rtx gen_umulsidi3 (rtx, rtx, rtx);
-extern rtx gen_maddsidi4 (rtx, rtx, rtx, rtx);
-extern rtx gen_umaddsidi4 (rtx, rtx, rtx, rtx);
-extern rtx gen_msubsidi4 (rtx, rtx, rtx, rtx);
-extern rtx gen_umsubsidi4 (rtx, rtx, rtx, rtx);
-extern rtx gen_smuldi3_highpart (rtx, rtx, rtx);
-extern rtx gen_umuldi3_highpart (rtx, rtx, rtx);
-extern rtx gen_divsi3 (rtx, rtx, rtx);
-extern rtx gen_udivsi3 (rtx, rtx, rtx);
-extern rtx gen_divdi3 (rtx, rtx, rtx);
-extern rtx gen_udivdi3 (rtx, rtx, rtx);
-extern rtx gen_cstoreqi_neg (rtx, rtx, rtx);
-extern rtx gen_cstorehi_neg (rtx, rtx, rtx);
-extern rtx gen_cstoresi_neg (rtx, rtx, rtx);
-extern rtx gen_cstoredi_neg (rtx, rtx, rtx);
-extern rtx gen_csinc3si_insn (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_csinc3di_insn (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_andsi3 (rtx, rtx, rtx);
-extern rtx gen_iorsi3 (rtx, rtx, rtx);
-extern rtx gen_xorsi3 (rtx, rtx, rtx);
-extern rtx gen_anddi3 (rtx, rtx, rtx);
-extern rtx gen_iordi3 (rtx, rtx, rtx);
-extern rtx gen_xordi3 (rtx, rtx, rtx);
-extern rtx gen_one_cmplsi2 (rtx, rtx);
-extern rtx gen_one_cmpldi2 (rtx, rtx);
-extern rtx gen_clzsi2 (rtx, rtx);
-extern rtx gen_clzdi2 (rtx, rtx);
-extern rtx gen_clrsbsi2 (rtx, rtx);
-extern rtx gen_clrsbdi2 (rtx, rtx);
-extern rtx gen_rbitsi2 (rtx, rtx);
-extern rtx gen_rbitdi2 (rtx, rtx);
-extern rtx gen_bswapsi2 (rtx, rtx);
-extern rtx gen_bswapdi2 (rtx, rtx);
-extern rtx gen_bswaphi2 (rtx, rtx);
-extern rtx gen_rev16si2 (rtx, rtx, rtx, rtx);
-extern rtx gen_rev16di2 (rtx, rtx, rtx, rtx);
-extern rtx gen_rev16si2_alt (rtx, rtx, rtx, rtx);
-extern rtx gen_rev16di2_alt (rtx, rtx, rtx, rtx);
-extern rtx gen_btruncsf2 (rtx, rtx);
-extern rtx gen_ceilsf2 (rtx, rtx);
-extern rtx gen_floorsf2 (rtx, rtx);
-extern rtx gen_frintnsf2 (rtx, rtx);
-extern rtx gen_nearbyintsf2 (rtx, rtx);
-extern rtx gen_rintsf2 (rtx, rtx);
-extern rtx gen_roundsf2 (rtx, rtx);
-extern rtx gen_btruncdf2 (rtx, rtx);
-extern rtx gen_ceildf2 (rtx, rtx);
-extern rtx gen_floordf2 (rtx, rtx);
-extern rtx gen_frintndf2 (rtx, rtx);
-extern rtx gen_nearbyintdf2 (rtx, rtx);
-extern rtx gen_rintdf2 (rtx, rtx);
-extern rtx gen_rounddf2 (rtx, rtx);
-extern rtx gen_lbtruncsfsi2 (rtx, rtx);
-extern rtx gen_lceilsfsi2 (rtx, rtx);
-extern rtx gen_lfloorsfsi2 (rtx, rtx);
-extern rtx gen_lroundsfsi2 (rtx, rtx);
-extern rtx gen_lfrintnsfsi2 (rtx, rtx);
-extern rtx gen_lbtruncusfsi2 (rtx, rtx);
-extern rtx gen_lceilusfsi2 (rtx, rtx);
-extern rtx gen_lfloorusfsi2 (rtx, rtx);
-extern rtx gen_lroundusfsi2 (rtx, rtx);
-extern rtx gen_lfrintnusfsi2 (rtx, rtx);
-extern rtx gen_lbtruncdfsi2 (rtx, rtx);
-extern rtx gen_lceildfsi2 (rtx, rtx);
-extern rtx gen_lfloordfsi2 (rtx, rtx);
-extern rtx gen_lrounddfsi2 (rtx, rtx);
-extern rtx gen_lfrintndfsi2 (rtx, rtx);
-extern rtx gen_lbtruncudfsi2 (rtx, rtx);
-extern rtx gen_lceiludfsi2 (rtx, rtx);
-extern rtx gen_lfloorudfsi2 (rtx, rtx);
-extern rtx gen_lroundudfsi2 (rtx, rtx);
-extern rtx gen_lfrintnudfsi2 (rtx, rtx);
-extern rtx gen_lbtruncsfdi2 (rtx, rtx);
-extern rtx gen_lceilsfdi2 (rtx, rtx);
-extern rtx gen_lfloorsfdi2 (rtx, rtx);
-extern rtx gen_lroundsfdi2 (rtx, rtx);
-extern rtx gen_lfrintnsfdi2 (rtx, rtx);
-extern rtx gen_lbtruncusfdi2 (rtx, rtx);
-extern rtx gen_lceilusfdi2 (rtx, rtx);
-extern rtx gen_lfloorusfdi2 (rtx, rtx);
-extern rtx gen_lroundusfdi2 (rtx, rtx);
-extern rtx gen_lfrintnusfdi2 (rtx, rtx);
-extern rtx gen_lbtruncdfdi2 (rtx, rtx);
-extern rtx gen_lceildfdi2 (rtx, rtx);
-extern rtx gen_lfloordfdi2 (rtx, rtx);
-extern rtx gen_lrounddfdi2 (rtx, rtx);
-extern rtx gen_lfrintndfdi2 (rtx, rtx);
-extern rtx gen_lbtruncudfdi2 (rtx, rtx);
-extern rtx gen_lceiludfdi2 (rtx, rtx);
-extern rtx gen_lfloorudfdi2 (rtx, rtx);
-extern rtx gen_lroundudfdi2 (rtx, rtx);
-extern rtx gen_lfrintnudfdi2 (rtx, rtx);
-extern rtx gen_fmasf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_fmadf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_fnmasf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_fnmadf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_fmssf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_fmsdf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_fnmssf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_fnmsdf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_extendsfdf2 (rtx, rtx);
-extern rtx gen_truncdfsf2 (rtx, rtx);
-extern rtx gen_fix_truncsfsi2 (rtx, rtx);
-extern rtx gen_fix_truncdfsi2 (rtx, rtx);
-extern rtx gen_fix_truncsfdi2 (rtx, rtx);
-extern rtx gen_fix_truncdfdi2 (rtx, rtx);
-extern rtx gen_fixuns_truncsfsi2 (rtx, rtx);
-extern rtx gen_fixuns_truncdfsi2 (rtx, rtx);
-extern rtx gen_fixuns_truncsfdi2 (rtx, rtx);
-extern rtx gen_fixuns_truncdfdi2 (rtx, rtx);
-extern rtx gen_floatsisf2 (rtx, rtx);
-extern rtx gen_floatsidf2 (rtx, rtx);
-extern rtx gen_floatdisf2 (rtx, rtx);
-extern rtx gen_floatdidf2 (rtx, rtx);
-extern rtx gen_floatunssisf2 (rtx, rtx);
-extern rtx gen_floatunssidf2 (rtx, rtx);
-extern rtx gen_floatunsdisf2 (rtx, rtx);
-extern rtx gen_floatunsdidf2 (rtx, rtx);
-extern rtx gen_addsf3 (rtx, rtx, rtx);
-extern rtx gen_adddf3 (rtx, rtx, rtx);
-extern rtx gen_subsf3 (rtx, rtx, rtx);
-extern rtx gen_subdf3 (rtx, rtx, rtx);
-extern rtx gen_mulsf3 (rtx, rtx, rtx);
-extern rtx gen_muldf3 (rtx, rtx, rtx);
-extern rtx gen_divsf3 (rtx, rtx, rtx);
-extern rtx gen_divdf3 (rtx, rtx, rtx);
-extern rtx gen_negsf2 (rtx, rtx);
-extern rtx gen_negdf2 (rtx, rtx);
-extern rtx gen_sqrtsf2 (rtx, rtx);
-extern rtx gen_sqrtdf2 (rtx, rtx);
-extern rtx gen_abssf2 (rtx, rtx);
-extern rtx gen_absdf2 (rtx, rtx);
-extern rtx gen_smaxsf3 (rtx, rtx, rtx);
-extern rtx gen_smaxdf3 (rtx, rtx, rtx);
-extern rtx gen_sminsf3 (rtx, rtx, rtx);
-extern rtx gen_smindf3 (rtx, rtx, rtx);
-extern rtx gen_aarch64_movdi_tilow (rtx, rtx);
-extern rtx gen_aarch64_movdi_tflow (rtx, rtx);
-extern rtx gen_aarch64_movdi_tihigh (rtx, rtx);
-extern rtx gen_aarch64_movdi_tfhigh (rtx, rtx);
-extern rtx gen_aarch64_movtihigh_di (rtx, rtx);
-extern rtx gen_aarch64_movtfhigh_di (rtx, rtx);
-extern rtx gen_aarch64_movtilow_di (rtx, rtx);
-extern rtx gen_aarch64_movtflow_di (rtx, rtx);
-extern rtx gen_aarch64_movtilow_tilow (rtx, rtx);
-extern rtx gen_add_losym_si (rtx, rtx, rtx);
-extern rtx gen_add_losym_di (rtx, rtx, rtx);
-extern rtx gen_ldr_got_small_si (rtx, rtx, rtx);
-extern rtx gen_ldr_got_small_di (rtx, rtx, rtx);
-extern rtx gen_ldr_got_small_sidi (rtx, rtx, rtx);
-extern rtx gen_ldr_got_tiny (rtx, rtx);
-extern rtx gen_aarch64_load_tp_hard (rtx);
-extern rtx gen_tlsie_small_si (rtx, rtx);
-extern rtx gen_tlsie_small_di (rtx, rtx);
-extern rtx gen_tlsie_small_sidi (rtx, rtx);
-extern rtx gen_tlsle_small_si (rtx, rtx, rtx);
-extern rtx gen_tlsle_small_di (rtx, rtx, rtx);
-extern rtx gen_tlsdesc_small_si (rtx);
-extern rtx gen_tlsdesc_small_di (rtx);
-extern rtx gen_stack_tie (rtx, rtx);
-extern rtx gen_stack_protect_set_si (rtx, rtx);
-extern rtx gen_stack_protect_set_di (rtx, rtx);
-extern rtx gen_stack_protect_test_si (rtx, rtx, rtx);
-extern rtx gen_stack_protect_test_di (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_dupv8qi (rtx, rtx);
-extern rtx gen_aarch64_simd_dupv16qi (rtx, rtx);
-extern rtx gen_aarch64_simd_dupv4hi (rtx, rtx);
-extern rtx gen_aarch64_simd_dupv8hi (rtx, rtx);
-extern rtx gen_aarch64_simd_dupv2si (rtx, rtx);
-extern rtx gen_aarch64_simd_dupv4si (rtx, rtx);
-extern rtx gen_aarch64_simd_dupv2di (rtx, rtx);
-extern rtx gen_aarch64_simd_dupv2sf (rtx, rtx);
-extern rtx gen_aarch64_simd_dupv4sf (rtx, rtx);
-extern rtx gen_aarch64_simd_dupv2df (rtx, rtx);
-extern rtx gen_aarch64_dup_lanev8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lanev16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lanev4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lanev8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lanev2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lanev4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lanev2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lanev2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lanev4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lanev2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lane_to_128v8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lane_to_64v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lane_to_128v4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lane_to_64v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lane_to_128v2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lane_to_64v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lane_to_64v2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lane_to_128v2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lane_to_64v4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_dup_lane_to_64v2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_mov_from_v16qilow (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_mov_from_v8hilow (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_mov_from_v4silow (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_mov_from_v2dilow (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_mov_from_v4sflow (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_mov_from_v2dflow (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_mov_from_v16qihigh (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_mov_from_v8hihigh (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_mov_from_v4sihigh (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_mov_from_v2dihigh (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_mov_from_v4sfhigh (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_mov_from_v2dfhigh (rtx, rtx, rtx);
-extern rtx gen_ornv8qi3 (rtx, rtx, rtx);
-extern rtx gen_ornv16qi3 (rtx, rtx, rtx);
-extern rtx gen_ornv4hi3 (rtx, rtx, rtx);
-extern rtx gen_ornv8hi3 (rtx, rtx, rtx);
-extern rtx gen_ornv2si3 (rtx, rtx, rtx);
-extern rtx gen_ornv4si3 (rtx, rtx, rtx);
-extern rtx gen_ornv2di3 (rtx, rtx, rtx);
-extern rtx gen_bicv8qi3 (rtx, rtx, rtx);
-extern rtx gen_bicv16qi3 (rtx, rtx, rtx);
-extern rtx gen_bicv4hi3 (rtx, rtx, rtx);
-extern rtx gen_bicv8hi3 (rtx, rtx, rtx);
-extern rtx gen_bicv2si3 (rtx, rtx, rtx);
-extern rtx gen_bicv4si3 (rtx, rtx, rtx);
-extern rtx gen_bicv2di3 (rtx, rtx, rtx);
-extern rtx gen_addv8qi3 (rtx, rtx, rtx);
-extern rtx gen_addv16qi3 (rtx, rtx, rtx);
-extern rtx gen_addv4hi3 (rtx, rtx, rtx);
-extern rtx gen_addv8hi3 (rtx, rtx, rtx);
-extern rtx gen_addv2si3 (rtx, rtx, rtx);
-extern rtx gen_addv4si3 (rtx, rtx, rtx);
-extern rtx gen_addv2di3 (rtx, rtx, rtx);
-extern rtx gen_subv8qi3 (rtx, rtx, rtx);
-extern rtx gen_subv16qi3 (rtx, rtx, rtx);
-extern rtx gen_subv4hi3 (rtx, rtx, rtx);
-extern rtx gen_subv8hi3 (rtx, rtx, rtx);
-extern rtx gen_subv2si3 (rtx, rtx, rtx);
-extern rtx gen_subv4si3 (rtx, rtx, rtx);
-extern rtx gen_subv2di3 (rtx, rtx, rtx);
-extern rtx gen_mulv8qi3 (rtx, rtx, rtx);
-extern rtx gen_mulv16qi3 (rtx, rtx, rtx);
-extern rtx gen_mulv4hi3 (rtx, rtx, rtx);
-extern rtx gen_mulv8hi3 (rtx, rtx, rtx);
-extern rtx gen_mulv2si3 (rtx, rtx, rtx);
-extern rtx gen_mulv4si3 (rtx, rtx, rtx);
-extern rtx gen_bswapv4hi (rtx, rtx);
-extern rtx gen_bswapv8hi (rtx, rtx);
-extern rtx gen_bswapv2si (rtx, rtx);
-extern rtx gen_bswapv4si (rtx, rtx);
-extern rtx gen_bswapv2di (rtx, rtx);
-extern rtx gen_negv8qi2 (rtx, rtx);
-extern rtx gen_negv16qi2 (rtx, rtx);
-extern rtx gen_negv4hi2 (rtx, rtx);
-extern rtx gen_negv8hi2 (rtx, rtx);
-extern rtx gen_negv2si2 (rtx, rtx);
-extern rtx gen_negv4si2 (rtx, rtx);
-extern rtx gen_negv2di2 (rtx, rtx);
-extern rtx gen_absv8qi2 (rtx, rtx);
-extern rtx gen_absv16qi2 (rtx, rtx);
-extern rtx gen_absv4hi2 (rtx, rtx);
-extern rtx gen_absv8hi2 (rtx, rtx);
-extern rtx gen_absv2si2 (rtx, rtx);
-extern rtx gen_absv4si2 (rtx, rtx);
-extern rtx gen_absv2di2 (rtx, rtx);
-extern rtx gen_abdv8qi_3 (rtx, rtx, rtx);
-extern rtx gen_abdv16qi_3 (rtx, rtx, rtx);
-extern rtx gen_abdv4hi_3 (rtx, rtx, rtx);
-extern rtx gen_abdv8hi_3 (rtx, rtx, rtx);
-extern rtx gen_abdv2si_3 (rtx, rtx, rtx);
-extern rtx gen_abdv4si_3 (rtx, rtx, rtx);
-extern rtx gen_abav8qi_3 (rtx, rtx, rtx, rtx);
-extern rtx gen_abav16qi_3 (rtx, rtx, rtx, rtx);
-extern rtx gen_abav4hi_3 (rtx, rtx, rtx, rtx);
-extern rtx gen_abav8hi_3 (rtx, rtx, rtx, rtx);
-extern rtx gen_abav2si_3 (rtx, rtx, rtx, rtx);
-extern rtx gen_abav4si_3 (rtx, rtx, rtx, rtx);
-extern rtx gen_fabdv2sf_3 (rtx, rtx, rtx);
-extern rtx gen_fabdv4sf_3 (rtx, rtx, rtx);
-extern rtx gen_fabdv2df_3 (rtx, rtx, rtx);
-extern rtx gen_andv8qi3 (rtx, rtx, rtx);
-extern rtx gen_andv16qi3 (rtx, rtx, rtx);
-extern rtx gen_andv4hi3 (rtx, rtx, rtx);
-extern rtx gen_andv8hi3 (rtx, rtx, rtx);
-extern rtx gen_andv2si3 (rtx, rtx, rtx);
-extern rtx gen_andv4si3 (rtx, rtx, rtx);
-extern rtx gen_andv2di3 (rtx, rtx, rtx);
-extern rtx gen_iorv8qi3 (rtx, rtx, rtx);
-extern rtx gen_iorv16qi3 (rtx, rtx, rtx);
-extern rtx gen_iorv4hi3 (rtx, rtx, rtx);
-extern rtx gen_iorv8hi3 (rtx, rtx, rtx);
-extern rtx gen_iorv2si3 (rtx, rtx, rtx);
-extern rtx gen_iorv4si3 (rtx, rtx, rtx);
-extern rtx gen_iorv2di3 (rtx, rtx, rtx);
-extern rtx gen_xorv8qi3 (rtx, rtx, rtx);
-extern rtx gen_xorv16qi3 (rtx, rtx, rtx);
-extern rtx gen_xorv4hi3 (rtx, rtx, rtx);
-extern rtx gen_xorv8hi3 (rtx, rtx, rtx);
-extern rtx gen_xorv2si3 (rtx, rtx, rtx);
-extern rtx gen_xorv4si3 (rtx, rtx, rtx);
-extern rtx gen_xorv2di3 (rtx, rtx, rtx);
-extern rtx gen_one_cmplv8qi2 (rtx, rtx);
-extern rtx gen_one_cmplv16qi2 (rtx, rtx);
-extern rtx gen_one_cmplv4hi2 (rtx, rtx);
-extern rtx gen_one_cmplv8hi2 (rtx, rtx);
-extern rtx gen_one_cmplv2si2 (rtx, rtx);
-extern rtx gen_one_cmplv4si2 (rtx, rtx);
-extern rtx gen_one_cmplv2di2 (rtx, rtx);
-extern rtx gen_aarch64_simd_vec_setv8qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_setv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_setv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_setv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_setv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_setv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_lshrv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_lshrv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_lshrv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_lshrv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_lshrv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_lshrv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_lshrv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_ashrv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_ashrv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_ashrv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_ashrv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_ashrv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_ashrv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_ashrv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_imm_shlv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_imm_shlv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_imm_shlv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_imm_shlv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_imm_shlv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_imm_shlv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_imm_shlv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_sshlv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_sshlv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_sshlv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_sshlv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_sshlv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_sshlv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_sshlv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_shlv8qi_unsigned (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_shlv16qi_unsigned (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_shlv4hi_unsigned (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_shlv8hi_unsigned (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_shlv2si_unsigned (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_shlv4si_unsigned (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_shlv2di_unsigned (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_shlv8qi_signed (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_shlv16qi_signed (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_shlv4hi_signed (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_shlv8hi_signed (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_shlv2si_signed (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_shlv4si_signed (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_reg_shlv2di_signed (rtx, rtx, rtx);
-extern rtx gen_aarch64_sshr_simddi (rtx, rtx);
-extern rtx gen_aarch64_ushr_simddi (rtx, rtx);
-extern rtx gen_aarch64_simd_vec_setv2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_setv2sf (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_setv4sf (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_setv2df (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_mlav8qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_mlav16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_mlav4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_mlav8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_mlav2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_mlav4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_mlsv8qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_mlsv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_mlsv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_mlsv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_mlsv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_mlsv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_smaxv8qi3 (rtx, rtx, rtx);
-extern rtx gen_sminv8qi3 (rtx, rtx, rtx);
-extern rtx gen_umaxv8qi3 (rtx, rtx, rtx);
-extern rtx gen_uminv8qi3 (rtx, rtx, rtx);
-extern rtx gen_smaxv16qi3 (rtx, rtx, rtx);
-extern rtx gen_sminv16qi3 (rtx, rtx, rtx);
-extern rtx gen_umaxv16qi3 (rtx, rtx, rtx);
-extern rtx gen_uminv16qi3 (rtx, rtx, rtx);
-extern rtx gen_smaxv4hi3 (rtx, rtx, rtx);
-extern rtx gen_sminv4hi3 (rtx, rtx, rtx);
-extern rtx gen_umaxv4hi3 (rtx, rtx, rtx);
-extern rtx gen_uminv4hi3 (rtx, rtx, rtx);
-extern rtx gen_smaxv8hi3 (rtx, rtx, rtx);
-extern rtx gen_sminv8hi3 (rtx, rtx, rtx);
-extern rtx gen_umaxv8hi3 (rtx, rtx, rtx);
-extern rtx gen_uminv8hi3 (rtx, rtx, rtx);
-extern rtx gen_smaxv2si3 (rtx, rtx, rtx);
-extern rtx gen_sminv2si3 (rtx, rtx, rtx);
-extern rtx gen_umaxv2si3 (rtx, rtx, rtx);
-extern rtx gen_uminv2si3 (rtx, rtx, rtx);
-extern rtx gen_smaxv4si3 (rtx, rtx, rtx);
-extern rtx gen_sminv4si3 (rtx, rtx, rtx);
-extern rtx gen_umaxv4si3 (rtx, rtx, rtx);
-extern rtx gen_uminv4si3 (rtx, rtx, rtx);
-extern rtx gen_move_lo_quad_v16qi (rtx, rtx);
-extern rtx gen_move_lo_quad_v8hi (rtx, rtx);
-extern rtx gen_move_lo_quad_v4si (rtx, rtx);
-extern rtx gen_move_lo_quad_v2di (rtx, rtx);
-extern rtx gen_move_lo_quad_v4sf (rtx, rtx);
-extern rtx gen_move_lo_quad_v2df (rtx, rtx);
-extern rtx gen_aarch64_simd_move_hi_quad_v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_move_hi_quad_v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_move_hi_quad_v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_move_hi_quad_v2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_move_hi_quad_v4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_move_hi_quad_v2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_pack_trunc_v8hi (rtx, rtx);
-extern rtx gen_aarch64_simd_vec_pack_trunc_v4si (rtx, rtx);
-extern rtx gen_aarch64_simd_vec_pack_trunc_v2di (rtx, rtx);
-extern rtx gen_vec_pack_trunc_v8hi (rtx, rtx, rtx);
-extern rtx gen_vec_pack_trunc_v4si (rtx, rtx, rtx);
-extern rtx gen_vec_pack_trunc_v2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_unpacks_lo_v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_unpacku_lo_v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_unpacks_lo_v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_unpacku_lo_v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_unpacks_lo_v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_unpacku_lo_v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_unpacks_hi_v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_unpacku_hi_v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_unpacks_hi_v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_unpacku_hi_v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_unpacks_hi_v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_unpacku_hi_v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_smult_lo_v16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_umult_lo_v16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_smult_lo_v8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_umult_lo_v8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_smult_lo_v4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_umult_lo_v4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_smult_hi_v16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_umult_hi_v16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_smult_hi_v8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_umult_hi_v8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_smult_hi_v4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_vec_umult_hi_v4si (rtx, rtx, rtx, rtx);
-extern rtx gen_addv2sf3 (rtx, rtx, rtx);
-extern rtx gen_addv4sf3 (rtx, rtx, rtx);
-extern rtx gen_addv2df3 (rtx, rtx, rtx);
-extern rtx gen_subv2sf3 (rtx, rtx, rtx);
-extern rtx gen_subv4sf3 (rtx, rtx, rtx);
-extern rtx gen_subv2df3 (rtx, rtx, rtx);
-extern rtx gen_mulv2sf3 (rtx, rtx, rtx);
-extern rtx gen_mulv4sf3 (rtx, rtx, rtx);
-extern rtx gen_mulv2df3 (rtx, rtx, rtx);
-extern rtx gen_divv2sf3 (rtx, rtx, rtx);
-extern rtx gen_divv4sf3 (rtx, rtx, rtx);
-extern rtx gen_divv2df3 (rtx, rtx, rtx);
-extern rtx gen_negv2sf2 (rtx, rtx);
-extern rtx gen_negv4sf2 (rtx, rtx);
-extern rtx gen_negv2df2 (rtx, rtx);
-extern rtx gen_absv2sf2 (rtx, rtx);
-extern rtx gen_absv4sf2 (rtx, rtx);
-extern rtx gen_absv2df2 (rtx, rtx);
-extern rtx gen_fmav2sf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_fmav4sf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_fmav2df4 (rtx, rtx, rtx, rtx);
-extern rtx gen_fnmav2sf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_fnmav4sf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_fnmav2df4 (rtx, rtx, rtx, rtx);
-extern rtx gen_btruncv2sf2 (rtx, rtx);
-extern rtx gen_ceilv2sf2 (rtx, rtx);
-extern rtx gen_floorv2sf2 (rtx, rtx);
-extern rtx gen_frintnv2sf2 (rtx, rtx);
-extern rtx gen_nearbyintv2sf2 (rtx, rtx);
-extern rtx gen_rintv2sf2 (rtx, rtx);
-extern rtx gen_roundv2sf2 (rtx, rtx);
-extern rtx gen_btruncv4sf2 (rtx, rtx);
-extern rtx gen_ceilv4sf2 (rtx, rtx);
-extern rtx gen_floorv4sf2 (rtx, rtx);
-extern rtx gen_frintnv4sf2 (rtx, rtx);
-extern rtx gen_nearbyintv4sf2 (rtx, rtx);
-extern rtx gen_rintv4sf2 (rtx, rtx);
-extern rtx gen_roundv4sf2 (rtx, rtx);
-extern rtx gen_btruncv2df2 (rtx, rtx);
-extern rtx gen_ceilv2df2 (rtx, rtx);
-extern rtx gen_floorv2df2 (rtx, rtx);
-extern rtx gen_frintnv2df2 (rtx, rtx);
-extern rtx gen_nearbyintv2df2 (rtx, rtx);
-extern rtx gen_rintv2df2 (rtx, rtx);
-extern rtx gen_roundv2df2 (rtx, rtx);
-extern rtx gen_lbtruncv2sfv2si2 (rtx, rtx);
-extern rtx gen_lceilv2sfv2si2 (rtx, rtx);
-extern rtx gen_lfloorv2sfv2si2 (rtx, rtx);
-extern rtx gen_lroundv2sfv2si2 (rtx, rtx);
-extern rtx gen_lfrintnv2sfv2si2 (rtx, rtx);
-extern rtx gen_lbtruncuv2sfv2si2 (rtx, rtx);
-extern rtx gen_lceiluv2sfv2si2 (rtx, rtx);
-extern rtx gen_lflooruv2sfv2si2 (rtx, rtx);
-extern rtx gen_lrounduv2sfv2si2 (rtx, rtx);
-extern rtx gen_lfrintnuv2sfv2si2 (rtx, rtx);
-extern rtx gen_lbtruncv4sfv4si2 (rtx, rtx);
-extern rtx gen_lceilv4sfv4si2 (rtx, rtx);
-extern rtx gen_lfloorv4sfv4si2 (rtx, rtx);
-extern rtx gen_lroundv4sfv4si2 (rtx, rtx);
-extern rtx gen_lfrintnv4sfv4si2 (rtx, rtx);
-extern rtx gen_lbtruncuv4sfv4si2 (rtx, rtx);
-extern rtx gen_lceiluv4sfv4si2 (rtx, rtx);
-extern rtx gen_lflooruv4sfv4si2 (rtx, rtx);
-extern rtx gen_lrounduv4sfv4si2 (rtx, rtx);
-extern rtx gen_lfrintnuv4sfv4si2 (rtx, rtx);
-extern rtx gen_lbtruncv2dfv2di2 (rtx, rtx);
-extern rtx gen_lceilv2dfv2di2 (rtx, rtx);
-extern rtx gen_lfloorv2dfv2di2 (rtx, rtx);
-extern rtx gen_lroundv2dfv2di2 (rtx, rtx);
-extern rtx gen_lfrintnv2dfv2di2 (rtx, rtx);
-extern rtx gen_lbtruncuv2dfv2di2 (rtx, rtx);
-extern rtx gen_lceiluv2dfv2di2 (rtx, rtx);
-extern rtx gen_lflooruv2dfv2di2 (rtx, rtx);
-extern rtx gen_lrounduv2dfv2di2 (rtx, rtx);
-extern rtx gen_lfrintnuv2dfv2di2 (rtx, rtx);
-extern rtx gen_floatv2siv2sf2 (rtx, rtx);
-extern rtx gen_floatunsv2siv2sf2 (rtx, rtx);
-extern rtx gen_floatv4siv4sf2 (rtx, rtx);
-extern rtx gen_floatunsv4siv4sf2 (rtx, rtx);
-extern rtx gen_floatv2div2df2 (rtx, rtx);
-extern rtx gen_floatunsv2div2df2 (rtx, rtx);
-extern rtx gen_vec_unpacks_lo_v4sf (rtx, rtx);
-extern rtx gen_aarch64_float_extend_lo_v2df (rtx, rtx);
-extern rtx gen_vec_unpacks_hi_v4sf (rtx, rtx);
-extern rtx gen_aarch64_float_truncate_lo_v2sf (rtx, rtx);
-extern rtx gen_aarch64_float_truncate_hi_v4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_vmlsv2sf (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vmlsv4sf (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vmlsv2df (rtx, rtx, rtx, rtx);
-extern rtx gen_smaxv2sf3 (rtx, rtx, rtx);
-extern rtx gen_sminv2sf3 (rtx, rtx, rtx);
-extern rtx gen_smaxv4sf3 (rtx, rtx, rtx);
-extern rtx gen_sminv4sf3 (rtx, rtx, rtx);
-extern rtx gen_smaxv2df3 (rtx, rtx, rtx);
-extern rtx gen_sminv2df3 (rtx, rtx, rtx);
-extern rtx gen_smax_nanv2sf3 (rtx, rtx, rtx);
-extern rtx gen_smin_nanv2sf3 (rtx, rtx, rtx);
-extern rtx gen_smax_nanv4sf3 (rtx, rtx, rtx);
-extern rtx gen_smin_nanv4sf3 (rtx, rtx, rtx);
-extern rtx gen_smax_nanv2df3 (rtx, rtx, rtx);
-extern rtx gen_smin_nanv2df3 (rtx, rtx, rtx);
-extern rtx gen_reduc_splus_v8qi (rtx, rtx);
-extern rtx gen_reduc_uplus_v8qi (rtx, rtx);
-extern rtx gen_reduc_splus_v16qi (rtx, rtx);
-extern rtx gen_reduc_uplus_v16qi (rtx, rtx);
-extern rtx gen_reduc_splus_v4hi (rtx, rtx);
-extern rtx gen_reduc_uplus_v4hi (rtx, rtx);
-extern rtx gen_reduc_splus_v8hi (rtx, rtx);
-extern rtx gen_reduc_uplus_v8hi (rtx, rtx);
-extern rtx gen_reduc_splus_v4si (rtx, rtx);
-extern rtx gen_reduc_uplus_v4si (rtx, rtx);
-extern rtx gen_reduc_splus_v2di (rtx, rtx);
-extern rtx gen_reduc_uplus_v2di (rtx, rtx);
-extern rtx gen_reduc_splus_v2si (rtx, rtx);
-extern rtx gen_reduc_uplus_v2si (rtx, rtx);
-extern rtx gen_reduc_splus_v2sf (rtx, rtx);
-extern rtx gen_reduc_splus_v2df (rtx, rtx);
-extern rtx gen_aarch64_addpv4sf (rtx, rtx);
-extern rtx gen_clzv8qi2 (rtx, rtx);
-extern rtx gen_clzv16qi2 (rtx, rtx);
-extern rtx gen_clzv4hi2 (rtx, rtx);
-extern rtx gen_clzv8hi2 (rtx, rtx);
-extern rtx gen_clzv2si2 (rtx, rtx);
-extern rtx gen_clzv4si2 (rtx, rtx);
-extern rtx gen_reduc_umax_v8qi (rtx, rtx);
-extern rtx gen_reduc_umin_v8qi (rtx, rtx);
-extern rtx gen_reduc_smax_v8qi (rtx, rtx);
-extern rtx gen_reduc_smin_v8qi (rtx, rtx);
-extern rtx gen_reduc_umax_v16qi (rtx, rtx);
-extern rtx gen_reduc_umin_v16qi (rtx, rtx);
-extern rtx gen_reduc_smax_v16qi (rtx, rtx);
-extern rtx gen_reduc_smin_v16qi (rtx, rtx);
-extern rtx gen_reduc_umax_v4hi (rtx, rtx);
-extern rtx gen_reduc_umin_v4hi (rtx, rtx);
-extern rtx gen_reduc_smax_v4hi (rtx, rtx);
-extern rtx gen_reduc_smin_v4hi (rtx, rtx);
-extern rtx gen_reduc_umax_v8hi (rtx, rtx);
-extern rtx gen_reduc_umin_v8hi (rtx, rtx);
-extern rtx gen_reduc_smax_v8hi (rtx, rtx);
-extern rtx gen_reduc_smin_v8hi (rtx, rtx);
-extern rtx gen_reduc_umax_v4si (rtx, rtx);
-extern rtx gen_reduc_umin_v4si (rtx, rtx);
-extern rtx gen_reduc_smax_v4si (rtx, rtx);
-extern rtx gen_reduc_smin_v4si (rtx, rtx);
-extern rtx gen_reduc_umax_v2si (rtx, rtx);
-extern rtx gen_reduc_umin_v2si (rtx, rtx);
-extern rtx gen_reduc_smax_v2si (rtx, rtx);
-extern rtx gen_reduc_smin_v2si (rtx, rtx);
-extern rtx gen_reduc_smax_nan_v2sf (rtx, rtx);
-extern rtx gen_reduc_smin_nan_v2sf (rtx, rtx);
-extern rtx gen_reduc_smax_v2sf (rtx, rtx);
-extern rtx gen_reduc_smin_v2sf (rtx, rtx);
-extern rtx gen_reduc_smax_nan_v2df (rtx, rtx);
-extern rtx gen_reduc_smin_nan_v2df (rtx, rtx);
-extern rtx gen_reduc_smax_v2df (rtx, rtx);
-extern rtx gen_reduc_smin_v2df (rtx, rtx);
-extern rtx gen_reduc_smax_nan_v4sf (rtx, rtx);
-extern rtx gen_reduc_smin_nan_v4sf (rtx, rtx);
-extern rtx gen_reduc_smax_v4sf (rtx, rtx);
-extern rtx gen_reduc_smin_v4sf (rtx, rtx);
-extern rtx gen_aarch64_simd_bslv8qi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv16qi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv4hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv8hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv2si_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv4si_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv2di_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv2sf_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv4sf_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv2df_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bsldi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bsldf_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_get_lanev8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_lanev16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_lanev4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_lanev8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_lanev2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_lanev4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_lanev2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_lanev2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_lanev4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_lanev2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_combinev8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_combinev4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_combinev2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_combinev2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_combinedi (rtx, rtx, rtx);
-extern rtx gen_aarch64_combinedf (rtx, rtx, rtx);
-extern rtx gen_aarch64_saddlv16qi_hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssublv16qi_hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddlv16qi_hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usublv16qi_hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_saddlv8hi_hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssublv8hi_hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddlv8hi_hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usublv8hi_hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_saddlv4si_hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssublv4si_hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddlv4si_hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usublv4si_hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_saddlv16qi_lo_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssublv16qi_lo_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddlv16qi_lo_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usublv16qi_lo_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_saddlv8hi_lo_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssublv8hi_lo_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddlv8hi_lo_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usublv8hi_lo_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_saddlv4si_lo_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssublv4si_lo_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddlv4si_lo_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usublv4si_lo_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_saddlv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ssublv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddlv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usublv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_saddlv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ssublv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddlv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usublv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_saddlv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_ssublv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddlv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_usublv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_saddwv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ssubwv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddwv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usubwv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_saddwv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ssubwv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddwv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usubwv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_saddwv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_ssubwv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddwv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_usubwv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_saddw2v16qi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssubw2v16qi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddw2v16qi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usubw2v16qi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_saddw2v8hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssubw2v8hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddw2v8hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usubw2v8hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_saddw2v4si_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssubw2v4si_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddw2v4si_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usubw2v4si_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_shaddv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uhaddv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srhaddv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urhaddv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_shsubv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uhsubv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srhsubv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urhsubv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_shaddv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uhaddv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srhaddv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urhaddv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_shsubv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uhsubv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srhsubv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urhsubv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_shaddv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uhaddv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srhaddv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urhaddv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_shsubv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uhsubv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srhsubv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urhsubv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_shaddv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uhaddv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srhaddv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urhaddv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_shsubv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uhsubv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srhsubv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urhsubv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_shaddv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uhaddv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_srhaddv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_urhaddv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_shsubv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uhsubv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_srhsubv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_urhsubv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_shaddv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uhaddv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_srhaddv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_urhaddv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_shsubv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uhsubv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_srhsubv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_urhsubv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_addhnv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_raddhnv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_subhnv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_rsubhnv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_addhnv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_raddhnv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_subhnv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_rsubhnv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_addhnv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_raddhnv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_subhnv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_rsubhnv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_addhn2v8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_raddhn2v8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_subhn2v8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_rsubhn2v8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_addhn2v4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_raddhn2v4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_subhn2v4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_rsubhn2v4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_addhn2v2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_raddhn2v2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_subhn2v2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_rsubhn2v2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_pmulv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_pmulv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqaddv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqaddv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqsubv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqsubv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqaddv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqaddv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqsubv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqsubv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqaddv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqaddv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqsubv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqsubv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqaddv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqaddv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqsubv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqsubv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqaddv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqaddv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqsubv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqsubv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqaddv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqaddv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqsubv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqsubv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqaddv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqaddv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqsubv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqsubv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqaddqi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqaddqi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqsubqi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqsubqi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqaddhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqaddhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqsubhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqsubhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqaddsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqaddsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqsubsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqsubsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqadddi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqadddi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqsubdi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqsubdi (rtx, rtx, rtx);
-extern rtx gen_aarch64_suqaddv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usqaddv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_suqaddv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usqaddv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_suqaddv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usqaddv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_suqaddv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usqaddv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_suqaddv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_usqaddv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_suqaddv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_usqaddv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_suqaddv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_usqaddv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_suqaddqi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usqaddqi (rtx, rtx, rtx);
-extern rtx gen_aarch64_suqaddhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usqaddhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_suqaddsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usqaddsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_suqadddi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usqadddi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqmovunv8hi (rtx, rtx);
-extern rtx gen_aarch64_sqmovunv4si (rtx, rtx);
-extern rtx gen_aarch64_sqmovunv2di (rtx, rtx);
-extern rtx gen_aarch64_sqmovunhi (rtx, rtx);
-extern rtx gen_aarch64_sqmovunsi (rtx, rtx);
-extern rtx gen_aarch64_sqmovundi (rtx, rtx);
-extern rtx gen_aarch64_sqmovnv8hi (rtx, rtx);
-extern rtx gen_aarch64_uqmovnv8hi (rtx, rtx);
-extern rtx gen_aarch64_sqmovnv4si (rtx, rtx);
-extern rtx gen_aarch64_uqmovnv4si (rtx, rtx);
-extern rtx gen_aarch64_sqmovnv2di (rtx, rtx);
-extern rtx gen_aarch64_uqmovnv2di (rtx, rtx);
-extern rtx gen_aarch64_sqmovnhi (rtx, rtx);
-extern rtx gen_aarch64_uqmovnhi (rtx, rtx);
-extern rtx gen_aarch64_sqmovnsi (rtx, rtx);
-extern rtx gen_aarch64_uqmovnsi (rtx, rtx);
-extern rtx gen_aarch64_sqmovndi (rtx, rtx);
-extern rtx gen_aarch64_uqmovndi (rtx, rtx);
-extern rtx gen_aarch64_sqnegv8qi (rtx, rtx);
-extern rtx gen_aarch64_sqabsv8qi (rtx, rtx);
-extern rtx gen_aarch64_sqnegv16qi (rtx, rtx);
-extern rtx gen_aarch64_sqabsv16qi (rtx, rtx);
-extern rtx gen_aarch64_sqnegv4hi (rtx, rtx);
-extern rtx gen_aarch64_sqabsv4hi (rtx, rtx);
-extern rtx gen_aarch64_sqnegv8hi (rtx, rtx);
-extern rtx gen_aarch64_sqabsv8hi (rtx, rtx);
-extern rtx gen_aarch64_sqnegv2si (rtx, rtx);
-extern rtx gen_aarch64_sqabsv2si (rtx, rtx);
-extern rtx gen_aarch64_sqnegv4si (rtx, rtx);
-extern rtx gen_aarch64_sqabsv4si (rtx, rtx);
-extern rtx gen_aarch64_sqnegv2di (rtx, rtx);
-extern rtx gen_aarch64_sqabsv2di (rtx, rtx);
-extern rtx gen_aarch64_sqnegqi (rtx, rtx);
-extern rtx gen_aarch64_sqabsqi (rtx, rtx);
-extern rtx gen_aarch64_sqneghi (rtx, rtx);
-extern rtx gen_aarch64_sqabshi (rtx, rtx);
-extern rtx gen_aarch64_sqnegsi (rtx, rtx);
-extern rtx gen_aarch64_sqabssi (rtx, rtx);
-extern rtx gen_aarch64_sqnegdi (rtx, rtx);
-extern rtx gen_aarch64_sqabsdi (rtx, rtx);
-extern rtx gen_aarch64_sqdmulhv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulhv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulhv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulhv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulhv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulhv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulhv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulhv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulhhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulhhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulhsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulhsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulh_lanev4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulh_lanev4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulh_lanev8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulh_lanev8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulh_lanev2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulh_lanev2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulh_lanev4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulh_lanev4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulh_laneqv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulh_laneqv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulh_laneqv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulh_laneqv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulh_laneqv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulh_laneqv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulh_laneqv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulh_laneqv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulh_lanehi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulh_lanehi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmulh_lanesi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrdmulh_lanesi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlalv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlslv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlalv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlslv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlalhi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlslhi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlalsi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlslsi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal_lanev4hi_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl_lanev4hi_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal_lanev2si_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl_lanev2si_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal_lanehi_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl_lanehi_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal_lanesi_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl_lanesi_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal_nv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl_nv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal_nv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl_nv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal2v8hi_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl2v8hi_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal2v4si_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl2v4si_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal2_lanev8hi_internal (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl2_lanev8hi_internal (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal2_lanev4si_internal (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl2_lanev4si_internal (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal2_nv8hi_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl2_nv8hi_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal2_nv4si_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl2_nv4si_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmullv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmullv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmullhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmullsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull_lanev4hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull_lanev2si_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull_lanehi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull_lanesi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull_nv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull_nv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull2v8hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull2v4si_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull2_lanev8hi_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull2_lanev4si_internal (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull2_nv8hi_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull2_nv4si_internal (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sshlv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ushlv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshlv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshlv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sshlv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ushlv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshlv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshlv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sshlv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ushlv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshlv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshlv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sshlv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ushlv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshlv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshlv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sshlv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_ushlv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshlv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshlv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sshlv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_ushlv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshlv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshlv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sshlv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_ushlv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshlv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshlv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_sshldi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ushldi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshldi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshldi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshlv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshlv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshlv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshlv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshlv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshlv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshlv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshlv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshlv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshlv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshlv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshlv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshlv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshlv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshlv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshlv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshlv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshlv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshlv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshlv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshlv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlqi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshlqi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshlqi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshlqi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshlhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshlhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshlhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshlsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshlsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshlsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshldi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshldi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshldi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshldi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sshll_nv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ushll_nv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sshll_nv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ushll_nv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sshll_nv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_ushll_nv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sshll2_nv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ushll2_nv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sshll2_nv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ushll2_nv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sshll2_nv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_ushll2_nv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshr_nv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshr_nv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshr_nv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshr_nv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshr_nv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshr_nv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshr_nv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshr_nv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshr_nv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshr_nv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshr_nv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshr_nv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshr_nv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshr_nv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_srshr_ndi (rtx, rtx, rtx);
-extern rtx gen_aarch64_urshr_ndi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ssra_nv8qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usra_nv8qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_srsra_nv8qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ursra_nv8qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssra_nv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usra_nv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_srsra_nv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ursra_nv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssra_nv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usra_nv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_srsra_nv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ursra_nv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssra_nv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usra_nv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_srsra_nv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ursra_nv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssra_nv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usra_nv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_srsra_nv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ursra_nv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssra_nv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usra_nv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_srsra_nv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ursra_nv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssra_nv2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usra_nv2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_srsra_nv2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ursra_nv2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssra_ndi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usra_ndi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_srsra_ndi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ursra_ndi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssli_nv8qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usli_nv8qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssri_nv8qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usri_nv8qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssli_nv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usli_nv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssri_nv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usri_nv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssli_nv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usli_nv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssri_nv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usri_nv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssli_nv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usli_nv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssri_nv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usri_nv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssli_nv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usli_nv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssri_nv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usri_nv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssli_nv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usli_nv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssri_nv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usri_nv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssli_nv2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usli_nv2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssri_nv2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usri_nv2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssli_ndi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usli_ndi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_ssri_ndi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_usri_ndi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlu_nv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshl_nv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshl_nv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlu_nv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshl_nv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshl_nv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlu_nv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshl_nv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshl_nv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlu_nv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshl_nv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshl_nv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlu_nv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshl_nv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshl_nv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlu_nv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshl_nv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshl_nv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlu_nv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshl_nv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshl_nv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlu_nqi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshl_nqi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshl_nqi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlu_nhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshl_nhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshl_nhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlu_nsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshl_nsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshl_nsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshlu_ndi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshl_ndi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshl_ndi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshrun_nv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshrun_nv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshrn_nv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshrn_nv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshrn_nv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshrn_nv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshrun_nv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshrun_nv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshrn_nv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshrn_nv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshrn_nv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshrn_nv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshrun_nv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshrun_nv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshrn_nv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshrn_nv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshrn_nv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshrn_nv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshrun_nhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshrun_nhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshrn_nhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshrn_nhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshrn_nhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshrn_nhi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshrun_nsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshrun_nsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshrn_nsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshrn_nsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshrn_nsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshrn_nsi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshrun_ndi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshrun_ndi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqshrn_ndi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqshrn_ndi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqrshrn_ndi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uqrshrn_ndi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmlev8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmeqv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgev8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmlev16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmeqv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgev16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmlev4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmeqv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgev4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmlev8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmeqv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgev8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmlev2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmeqv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgev2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmlev4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmeqv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgev4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmlev2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmeqv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgev2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltdi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmledi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmeqdi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgedi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtdi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltuv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmleuv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgeuv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtuv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltuv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmleuv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgeuv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtuv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltuv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmleuv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgeuv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtuv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltuv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmleuv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgeuv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtuv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltuv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmleuv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgeuv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtuv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltuv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmleuv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgeuv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtuv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltuv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmleuv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgeuv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtuv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltudi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmleudi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgeudi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtudi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmtstv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmtstv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmtstv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmtstv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmtstv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmtstv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmtstv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmtstdi (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltv2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmlev2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmeqv2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgev2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtv2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltv4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmlev4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmeqv4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgev4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtv4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltv2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmlev2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmeqv2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgev2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtv2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltsf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmlesf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmeqsf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgesf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtsf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmltdf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmledf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmeqdf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgedf (rtx, rtx, rtx);
-extern rtx gen_aarch64_cmgtdf (rtx, rtx, rtx);
-extern rtx gen_aarch64_addpv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_addpv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_addpv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_addpdi (rtx, rtx);
-extern rtx gen_sqrtv2sf2 (rtx, rtx);
-extern rtx gen_sqrtv4sf2 (rtx, rtx);
-extern rtx gen_sqrtv2df2 (rtx, rtx);
-extern rtx gen_vec_load_lanesoiv16qi (rtx, rtx);
-extern rtx gen_vec_load_lanesoiv8hi (rtx, rtx);
-extern rtx gen_vec_load_lanesoiv4si (rtx, rtx);
-extern rtx gen_vec_load_lanesoiv2di (rtx, rtx);
-extern rtx gen_vec_load_lanesoiv4sf (rtx, rtx);
-extern rtx gen_vec_load_lanesoiv2df (rtx, rtx);
-extern rtx gen_vec_store_lanesoiv16qi (rtx, rtx);
-extern rtx gen_vec_store_lanesoiv8hi (rtx, rtx);
-extern rtx gen_vec_store_lanesoiv4si (rtx, rtx);
-extern rtx gen_vec_store_lanesoiv2di (rtx, rtx);
-extern rtx gen_vec_store_lanesoiv4sf (rtx, rtx);
-extern rtx gen_vec_store_lanesoiv2df (rtx, rtx);
-extern rtx gen_vec_store_lanesoi_lanev16qi (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesoi_lanev8hi (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesoi_lanev4si (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesoi_lanev2di (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesoi_lanev4sf (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesoi_lanev2df (rtx, rtx, rtx);
-extern rtx gen_vec_load_lanesciv16qi (rtx, rtx);
-extern rtx gen_vec_load_lanesciv8hi (rtx, rtx);
-extern rtx gen_vec_load_lanesciv4si (rtx, rtx);
-extern rtx gen_vec_load_lanesciv2di (rtx, rtx);
-extern rtx gen_vec_load_lanesciv4sf (rtx, rtx);
-extern rtx gen_vec_load_lanesciv2df (rtx, rtx);
-extern rtx gen_vec_store_lanesciv16qi (rtx, rtx);
-extern rtx gen_vec_store_lanesciv8hi (rtx, rtx);
-extern rtx gen_vec_store_lanesciv4si (rtx, rtx);
-extern rtx gen_vec_store_lanesciv2di (rtx, rtx);
-extern rtx gen_vec_store_lanesciv4sf (rtx, rtx);
-extern rtx gen_vec_store_lanesciv2df (rtx, rtx);
-extern rtx gen_vec_store_lanesci_lanev16qi (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesci_lanev8hi (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesci_lanev4si (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesci_lanev2di (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesci_lanev4sf (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesci_lanev2df (rtx, rtx, rtx);
-extern rtx gen_vec_load_lanesxiv16qi (rtx, rtx);
-extern rtx gen_vec_load_lanesxiv8hi (rtx, rtx);
-extern rtx gen_vec_load_lanesxiv4si (rtx, rtx);
-extern rtx gen_vec_load_lanesxiv2di (rtx, rtx);
-extern rtx gen_vec_load_lanesxiv4sf (rtx, rtx);
-extern rtx gen_vec_load_lanesxiv2df (rtx, rtx);
-extern rtx gen_vec_store_lanesxiv16qi (rtx, rtx);
-extern rtx gen_vec_store_lanesxiv8hi (rtx, rtx);
-extern rtx gen_vec_store_lanesxiv4si (rtx, rtx);
-extern rtx gen_vec_store_lanesxiv2di (rtx, rtx);
-extern rtx gen_vec_store_lanesxiv4sf (rtx, rtx);
-extern rtx gen_vec_store_lanesxiv2df (rtx, rtx);
-extern rtx gen_vec_store_lanesxi_lanev16qi (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesxi_lanev8hi (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesxi_lanev4si (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesxi_lanev2di (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesxi_lanev4sf (rtx, rtx, rtx);
-extern rtx gen_vec_store_lanesxi_lanev2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_be_ld1v8qi (rtx, rtx);
-extern rtx gen_aarch64_be_ld1v16qi (rtx, rtx);
-extern rtx gen_aarch64_be_ld1v4hi (rtx, rtx);
-extern rtx gen_aarch64_be_ld1v8hi (rtx, rtx);
-extern rtx gen_aarch64_be_ld1v2si (rtx, rtx);
-extern rtx gen_aarch64_be_ld1v4si (rtx, rtx);
-extern rtx gen_aarch64_be_ld1v2di (rtx, rtx);
-extern rtx gen_aarch64_be_ld1v2sf (rtx, rtx);
-extern rtx gen_aarch64_be_ld1v4sf (rtx, rtx);
-extern rtx gen_aarch64_be_ld1v2df (rtx, rtx);
-extern rtx gen_aarch64_be_ld1di (rtx, rtx);
-extern rtx gen_aarch64_be_st1v8qi (rtx, rtx);
-extern rtx gen_aarch64_be_st1v16qi (rtx, rtx);
-extern rtx gen_aarch64_be_st1v4hi (rtx, rtx);
-extern rtx gen_aarch64_be_st1v8hi (rtx, rtx);
-extern rtx gen_aarch64_be_st1v2si (rtx, rtx);
-extern rtx gen_aarch64_be_st1v4si (rtx, rtx);
-extern rtx gen_aarch64_be_st1v2di (rtx, rtx);
-extern rtx gen_aarch64_be_st1v2sf (rtx, rtx);
-extern rtx gen_aarch64_be_st1v4sf (rtx, rtx);
-extern rtx gen_aarch64_be_st1v2df (rtx, rtx);
-extern rtx gen_aarch64_be_st1di (rtx, rtx);
-extern rtx gen_aarch64_ld2v8qi_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld2v4hi_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld2v2si_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld2v2sf_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld2di_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld2df_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld3v8qi_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld3v4hi_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld3v2si_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld3v2sf_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld3di_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld3df_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld4v8qi_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld4v4hi_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld4v2si_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld4v2sf_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld4di_dreg (rtx, rtx);
-extern rtx gen_aarch64_ld4df_dreg (rtx, rtx);
-extern rtx gen_aarch64_tbl1v8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_tbl1v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_tbl2v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_combinev16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip1v8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip2v8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn1v8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn2v8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp1v8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp2v8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip1v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip2v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn1v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn2v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp1v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp2v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip1v4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip2v4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn1v4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn2v4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp1v4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp2v4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip1v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip2v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn1v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn2v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp1v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp2v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip1v2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip2v2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn1v2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn2v2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp1v2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp2v2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip1v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip2v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn1v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn2v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp1v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp2v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip1v2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip2v2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn1v2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn2v2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp1v2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp2v2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip1v2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip2v2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn1v2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn2v2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp1v2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp2v2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip1v4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip2v4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn1v4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn2v4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp1v4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp2v4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip1v2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_zip2v2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn1v2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_trn2v2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp1v2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_uzp2v2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_st2v8qi_dreg (rtx, rtx);
-extern rtx gen_aarch64_st2v4hi_dreg (rtx, rtx);
-extern rtx gen_aarch64_st2v2si_dreg (rtx, rtx);
-extern rtx gen_aarch64_st2v2sf_dreg (rtx, rtx);
-extern rtx gen_aarch64_st2di_dreg (rtx, rtx);
-extern rtx gen_aarch64_st2df_dreg (rtx, rtx);
-extern rtx gen_aarch64_st3v8qi_dreg (rtx, rtx);
-extern rtx gen_aarch64_st3v4hi_dreg (rtx, rtx);
-extern rtx gen_aarch64_st3v2si_dreg (rtx, rtx);
-extern rtx gen_aarch64_st3v2sf_dreg (rtx, rtx);
-extern rtx gen_aarch64_st3di_dreg (rtx, rtx);
-extern rtx gen_aarch64_st3df_dreg (rtx, rtx);
-extern rtx gen_aarch64_st4v8qi_dreg (rtx, rtx);
-extern rtx gen_aarch64_st4v4hi_dreg (rtx, rtx);
-extern rtx gen_aarch64_st4v2si_dreg (rtx, rtx);
-extern rtx gen_aarch64_st4v2sf_dreg (rtx, rtx);
-extern rtx gen_aarch64_st4di_dreg (rtx, rtx);
-extern rtx gen_aarch64_st4df_dreg (rtx, rtx);
-extern rtx gen_aarch64_frecpev2sf (rtx, rtx);
-extern rtx gen_aarch64_frecpev4sf (rtx, rtx);
-extern rtx gen_aarch64_frecpev2df (rtx, rtx);
-extern rtx gen_aarch64_frecpesf (rtx, rtx);
-extern rtx gen_aarch64_frecpxsf (rtx, rtx);
-extern rtx gen_aarch64_frecpedf (rtx, rtx);
-extern rtx gen_aarch64_frecpxdf (rtx, rtx);
-extern rtx gen_aarch64_frecpsv2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_frecpsv4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_frecpsv2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_frecpssf (rtx, rtx, rtx);
-extern rtx gen_aarch64_frecpsdf (rtx, rtx, rtx);
-extern rtx gen_aarch64_crypto_aesev16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_crypto_aesdv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_crypto_aesmcv16qi (rtx, rtx);
-extern rtx gen_aarch64_crypto_aesimcv16qi (rtx, rtx);
-extern rtx gen_aarch64_crypto_sha1hsi (rtx, rtx);
-extern rtx gen_aarch64_crypto_sha1su1v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_crypto_sha1cv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_crypto_sha1mv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_crypto_sha1pv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_crypto_sha1su0v4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_crypto_sha256hv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_crypto_sha256h2v4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_crypto_sha256su0v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_crypto_sha256su1v4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_crypto_pmulldi (rtx, rtx, rtx);
-extern rtx gen_aarch64_crypto_pmullv2di (rtx, rtx, rtx);
-extern rtx gen_atomic_compare_and_swapqi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_compare_and_swaphi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_compare_and_swapsi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_compare_and_swapdi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_exchangeqi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_exchangehi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_exchangesi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_exchangedi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_addqi (rtx, rtx, rtx);
-extern rtx gen_atomic_subqi (rtx, rtx, rtx);
-extern rtx gen_atomic_orqi (rtx, rtx, rtx);
-extern rtx gen_atomic_xorqi (rtx, rtx, rtx);
-extern rtx gen_atomic_andqi (rtx, rtx, rtx);
-extern rtx gen_atomic_addhi (rtx, rtx, rtx);
-extern rtx gen_atomic_subhi (rtx, rtx, rtx);
-extern rtx gen_atomic_orhi (rtx, rtx, rtx);
-extern rtx gen_atomic_xorhi (rtx, rtx, rtx);
-extern rtx gen_atomic_andhi (rtx, rtx, rtx);
-extern rtx gen_atomic_addsi (rtx, rtx, rtx);
-extern rtx gen_atomic_subsi (rtx, rtx, rtx);
-extern rtx gen_atomic_orsi (rtx, rtx, rtx);
-extern rtx gen_atomic_xorsi (rtx, rtx, rtx);
-extern rtx gen_atomic_andsi (rtx, rtx, rtx);
-extern rtx gen_atomic_adddi (rtx, rtx, rtx);
-extern rtx gen_atomic_subdi (rtx, rtx, rtx);
-extern rtx gen_atomic_ordi (rtx, rtx, rtx);
-extern rtx gen_atomic_xordi (rtx, rtx, rtx);
-extern rtx gen_atomic_anddi (rtx, rtx, rtx);
-extern rtx gen_atomic_nandqi (rtx, rtx, rtx);
-extern rtx gen_atomic_nandhi (rtx, rtx, rtx);
-extern rtx gen_atomic_nandsi (rtx, rtx, rtx);
-extern rtx gen_atomic_nanddi (rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_addqi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_subqi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_orqi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_xorqi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_andqi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_addhi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_subhi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_orhi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_xorhi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_andhi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_addsi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_subsi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_orsi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_xorsi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_andsi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_adddi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_subdi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_ordi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_xordi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_anddi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_nandqi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_nandhi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_nandsi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_fetch_nanddi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_add_fetchqi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_sub_fetchqi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_or_fetchqi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_xor_fetchqi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_and_fetchqi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_add_fetchhi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_sub_fetchhi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_or_fetchhi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_xor_fetchhi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_and_fetchhi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_add_fetchsi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_sub_fetchsi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_or_fetchsi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_xor_fetchsi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_and_fetchsi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_add_fetchdi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_sub_fetchdi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_or_fetchdi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_xor_fetchdi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_and_fetchdi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_nand_fetchqi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_nand_fetchhi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_nand_fetchsi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_nand_fetchdi (rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_loadqi (rtx, rtx, rtx);
-extern rtx gen_atomic_loadhi (rtx, rtx, rtx);
-extern rtx gen_atomic_loadsi (rtx, rtx, rtx);
-extern rtx gen_atomic_loaddi (rtx, rtx, rtx);
-extern rtx gen_atomic_storeqi (rtx, rtx, rtx);
-extern rtx gen_atomic_storehi (rtx, rtx, rtx);
-extern rtx gen_atomic_storesi (rtx, rtx, rtx);
-extern rtx gen_atomic_storedi (rtx, rtx, rtx);
-extern rtx gen_aarch64_load_exclusiveqi (rtx, rtx, rtx);
-extern rtx gen_aarch64_load_exclusivehi (rtx, rtx, rtx);
-extern rtx gen_aarch64_load_exclusivesi (rtx, rtx, rtx);
-extern rtx gen_aarch64_load_exclusivedi (rtx, rtx, rtx);
-extern rtx gen_aarch64_store_exclusiveqi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_store_exclusivehi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_store_exclusivesi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_store_exclusivedi (rtx, rtx, rtx, rtx);
-extern rtx gen_cbranchsi4 (rtx, rtx, rtx, rtx);
-extern rtx gen_cbranchdi4 (rtx, rtx, rtx, rtx);
-extern rtx gen_cbranchsf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_cbranchdf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_casesi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_prologue (void);
-extern rtx gen_epilogue (void);
-extern rtx gen_sibcall_epilogue (void);
+extern rtx gen_indirect_jump (rtx);
+extern rtx gen_jump (rtx);
+extern rtx gen_casesi_dispatch (rtx, rtx, rtx);
+extern rtx gen_nop (void);
+extern rtx gen_trap (void);
+extern rtx gen_eh_return (rtx);
+extern rtx gen_insv_immsi (rtx, rtx, rtx);
+extern rtx gen_insv_immdi (rtx, rtx, rtx);
+extern rtx gen_load_pairsi (rtx, rtx, rtx, rtx);
+extern rtx gen_load_pairdi (rtx, rtx, rtx, rtx);
+extern rtx gen_store_pairsi (rtx, rtx, rtx, rtx);
+extern rtx gen_store_pairdi (rtx, rtx, rtx, rtx);
+extern rtx gen_load_pairsf (rtx, rtx, rtx, rtx);
+extern rtx gen_load_pairdf (rtx, rtx, rtx, rtx);
+extern rtx gen_store_pairsf (rtx, rtx, rtx, rtx);
+extern rtx gen_store_pairdf (rtx, rtx, rtx, rtx);
+extern rtx gen_loadwb_pairsi_si (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_loadwb_pairsi_di (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_loadwb_pairdi_si (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_loadwb_pairdi_di (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_storewb_pairsi_si (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_storewb_pairsi_di (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_storewb_pairdi_si (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_storewb_pairdi_di (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_addsi3_compare0 (rtx, rtx, rtx);
+extern rtx gen_adddi3_compare0 (rtx, rtx, rtx);
+extern rtx gen_addsi3_carryin (rtx, rtx, rtx);
+extern rtx gen_adddi3_carryin (rtx, rtx, rtx);
+extern rtx gen_subsi3 (rtx, rtx, rtx);
+extern rtx gen_subdi3 (rtx, rtx, rtx);
+extern rtx gen_subsi3_compare0 (rtx, rtx, rtx);
+extern rtx gen_subdi3_compare0 (rtx, rtx, rtx);
+extern rtx gen_subsi3_carryin (rtx, rtx, rtx);
+extern rtx gen_subdi3_carryin (rtx, rtx, rtx);
+extern rtx gen_absdi2 (rtx, rtx);
+extern rtx gen_negsi2 (rtx, rtx);
+extern rtx gen_negdi2 (rtx, rtx);
+extern rtx gen_mulsi3 (rtx, rtx, rtx);
+extern rtx gen_muldi3 (rtx, rtx, rtx);
+extern rtx gen_maddsi (rtx, rtx, rtx, rtx);
+extern rtx gen_madddi (rtx, rtx, rtx, rtx);
+extern rtx gen_mulsidi3 (rtx, rtx, rtx);
+extern rtx gen_umulsidi3 (rtx, rtx, rtx);
+extern rtx gen_maddsidi4 (rtx, rtx, rtx, rtx);
+extern rtx gen_umaddsidi4 (rtx, rtx, rtx, rtx);
+extern rtx gen_msubsidi4 (rtx, rtx, rtx, rtx);
+extern rtx gen_umsubsidi4 (rtx, rtx, rtx, rtx);
+extern rtx gen_smuldi3_highpart (rtx, rtx, rtx);
+extern rtx gen_umuldi3_highpart (rtx, rtx, rtx);
+extern rtx gen_divsi3 (rtx, rtx, rtx);
+extern rtx gen_udivsi3 (rtx, rtx, rtx);
+extern rtx gen_divdi3 (rtx, rtx, rtx);
+extern rtx gen_udivdi3 (rtx, rtx, rtx);
+extern rtx gen_cstoreqi_neg (rtx, rtx, rtx);
+extern rtx gen_cstorehi_neg (rtx, rtx, rtx);
+extern rtx gen_cstoresi_neg (rtx, rtx, rtx);
+extern rtx gen_cstoredi_neg (rtx, rtx, rtx);
+extern rtx gen_aarch64_crc32b (rtx, rtx, rtx);
+extern rtx gen_aarch64_crc32h (rtx, rtx, rtx);
+extern rtx gen_aarch64_crc32w (rtx, rtx, rtx);
+extern rtx gen_aarch64_crc32x (rtx, rtx, rtx);
+extern rtx gen_aarch64_crc32cb (rtx, rtx, rtx);
+extern rtx gen_aarch64_crc32ch (rtx, rtx, rtx);
+extern rtx gen_aarch64_crc32cw (rtx, rtx, rtx);
+extern rtx gen_aarch64_crc32cx (rtx, rtx, rtx);
+extern rtx gen_csinc3si_insn (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_csinc3di_insn (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_andsi3 (rtx, rtx, rtx);
+extern rtx gen_iorsi3 (rtx, rtx, rtx);
+extern rtx gen_xorsi3 (rtx, rtx, rtx);
+extern rtx gen_anddi3 (rtx, rtx, rtx);
+extern rtx gen_iordi3 (rtx, rtx, rtx);
+extern rtx gen_xordi3 (rtx, rtx, rtx);
+extern rtx gen_one_cmplsi2 (rtx, rtx);
+extern rtx gen_one_cmpldi2 (rtx, rtx);
+extern rtx gen_clzsi2 (rtx, rtx);
+extern rtx gen_clzdi2 (rtx, rtx);
+extern rtx gen_clrsbsi2 (rtx, rtx);
+extern rtx gen_clrsbdi2 (rtx, rtx);
+extern rtx gen_rbitsi2 (rtx, rtx);
+extern rtx gen_rbitdi2 (rtx, rtx);
+extern rtx gen_bswapsi2 (rtx, rtx);
+extern rtx gen_bswapdi2 (rtx, rtx);
+extern rtx gen_bswaphi2 (rtx, rtx);
+extern rtx gen_rev16si2 (rtx, rtx, rtx, rtx);
+extern rtx gen_rev16di2 (rtx, rtx, rtx, rtx);
+extern rtx gen_rev16si2_alt (rtx, rtx, rtx, rtx);
+extern rtx gen_rev16di2_alt (rtx, rtx, rtx, rtx);
+extern rtx gen_btruncsf2 (rtx, rtx);
+extern rtx gen_ceilsf2 (rtx, rtx);
+extern rtx gen_floorsf2 (rtx, rtx);
+extern rtx gen_frintnsf2 (rtx, rtx);
+extern rtx gen_nearbyintsf2 (rtx, rtx);
+extern rtx gen_rintsf2 (rtx, rtx);
+extern rtx gen_roundsf2 (rtx, rtx);
+extern rtx gen_btruncdf2 (rtx, rtx);
+extern rtx gen_ceildf2 (rtx, rtx);
+extern rtx gen_floordf2 (rtx, rtx);
+extern rtx gen_frintndf2 (rtx, rtx);
+extern rtx gen_nearbyintdf2 (rtx, rtx);
+extern rtx gen_rintdf2 (rtx, rtx);
+extern rtx gen_rounddf2 (rtx, rtx);
+extern rtx gen_lbtruncsfsi2 (rtx, rtx);
+extern rtx gen_lceilsfsi2 (rtx, rtx);
+extern rtx gen_lfloorsfsi2 (rtx, rtx);
+extern rtx gen_lroundsfsi2 (rtx, rtx);
+extern rtx gen_lfrintnsfsi2 (rtx, rtx);
+extern rtx gen_lbtruncusfsi2 (rtx, rtx);
+extern rtx gen_lceilusfsi2 (rtx, rtx);
+extern rtx gen_lfloorusfsi2 (rtx, rtx);
+extern rtx gen_lroundusfsi2 (rtx, rtx);
+extern rtx gen_lfrintnusfsi2 (rtx, rtx);
+extern rtx gen_lbtruncdfsi2 (rtx, rtx);
+extern rtx gen_lceildfsi2 (rtx, rtx);
+extern rtx gen_lfloordfsi2 (rtx, rtx);
+extern rtx gen_lrounddfsi2 (rtx, rtx);
+extern rtx gen_lfrintndfsi2 (rtx, rtx);
+extern rtx gen_lbtruncudfsi2 (rtx, rtx);
+extern rtx gen_lceiludfsi2 (rtx, rtx);
+extern rtx gen_lfloorudfsi2 (rtx, rtx);
+extern rtx gen_lroundudfsi2 (rtx, rtx);
+extern rtx gen_lfrintnudfsi2 (rtx, rtx);
+extern rtx gen_lbtruncsfdi2 (rtx, rtx);
+extern rtx gen_lceilsfdi2 (rtx, rtx);
+extern rtx gen_lfloorsfdi2 (rtx, rtx);
+extern rtx gen_lroundsfdi2 (rtx, rtx);
+extern rtx gen_lfrintnsfdi2 (rtx, rtx);
+extern rtx gen_lbtruncusfdi2 (rtx, rtx);
+extern rtx gen_lceilusfdi2 (rtx, rtx);
+extern rtx gen_lfloorusfdi2 (rtx, rtx);
+extern rtx gen_lroundusfdi2 (rtx, rtx);
+extern rtx gen_lfrintnusfdi2 (rtx, rtx);
+extern rtx gen_lbtruncdfdi2 (rtx, rtx);
+extern rtx gen_lceildfdi2 (rtx, rtx);
+extern rtx gen_lfloordfdi2 (rtx, rtx);
+extern rtx gen_lrounddfdi2 (rtx, rtx);
+extern rtx gen_lfrintndfdi2 (rtx, rtx);
+extern rtx gen_lbtruncudfdi2 (rtx, rtx);
+extern rtx gen_lceiludfdi2 (rtx, rtx);
+extern rtx gen_lfloorudfdi2 (rtx, rtx);
+extern rtx gen_lroundudfdi2 (rtx, rtx);
+extern rtx gen_lfrintnudfdi2 (rtx, rtx);
+extern rtx gen_fmasf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_fmadf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_fnmasf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_fnmadf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_fmssf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_fmsdf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_fnmssf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_fnmsdf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_extendsfdf2 (rtx, rtx);
+extern rtx gen_truncdfsf2 (rtx, rtx);
+extern rtx gen_fix_truncsfsi2 (rtx, rtx);
+extern rtx gen_fix_truncdfsi2 (rtx, rtx);
+extern rtx gen_fix_truncsfdi2 (rtx, rtx);
+extern rtx gen_fix_truncdfdi2 (rtx, rtx);
+extern rtx gen_fixuns_truncsfsi2 (rtx, rtx);
+extern rtx gen_fixuns_truncdfsi2 (rtx, rtx);
+extern rtx gen_fixuns_truncsfdi2 (rtx, rtx);
+extern rtx gen_fixuns_truncdfdi2 (rtx, rtx);
+extern rtx gen_floatsisf2 (rtx, rtx);
+extern rtx gen_floatsidf2 (rtx, rtx);
+extern rtx gen_floatdisf2 (rtx, rtx);
+extern rtx gen_floatdidf2 (rtx, rtx);
+extern rtx gen_floatunssisf2 (rtx, rtx);
+extern rtx gen_floatunssidf2 (rtx, rtx);
+extern rtx gen_floatunsdisf2 (rtx, rtx);
+extern rtx gen_floatunsdidf2 (rtx, rtx);
+extern rtx gen_addsf3 (rtx, rtx, rtx);
+extern rtx gen_adddf3 (rtx, rtx, rtx);
+extern rtx gen_subsf3 (rtx, rtx, rtx);
+extern rtx gen_subdf3 (rtx, rtx, rtx);
+extern rtx gen_mulsf3 (rtx, rtx, rtx);
+extern rtx gen_muldf3 (rtx, rtx, rtx);
+extern rtx gen_divsf3 (rtx, rtx, rtx);
+extern rtx gen_divdf3 (rtx, rtx, rtx);
+extern rtx gen_negsf2 (rtx, rtx);
+extern rtx gen_negdf2 (rtx, rtx);
+extern rtx gen_sqrtsf2 (rtx, rtx);
+extern rtx gen_sqrtdf2 (rtx, rtx);
+extern rtx gen_abssf2 (rtx, rtx);
+extern rtx gen_absdf2 (rtx, rtx);
+extern rtx gen_smaxsf3 (rtx, rtx, rtx);
+extern rtx gen_smaxdf3 (rtx, rtx, rtx);
+extern rtx gen_sminsf3 (rtx, rtx, rtx);
+extern rtx gen_smindf3 (rtx, rtx, rtx);
+extern rtx gen_aarch64_movdi_tilow (rtx, rtx);
+extern rtx gen_aarch64_movdi_tflow (rtx, rtx);
+extern rtx gen_aarch64_movdi_tihigh (rtx, rtx);
+extern rtx gen_aarch64_movdi_tfhigh (rtx, rtx);
+extern rtx gen_aarch64_movtihigh_di (rtx, rtx);
+extern rtx gen_aarch64_movtfhigh_di (rtx, rtx);
+extern rtx gen_aarch64_movtilow_di (rtx, rtx);
+extern rtx gen_aarch64_movtflow_di (rtx, rtx);
+extern rtx gen_aarch64_movtilow_tilow (rtx, rtx);
+extern rtx gen_add_losym_si (rtx, rtx, rtx);
+extern rtx gen_add_losym_di (rtx, rtx, rtx);
+extern rtx gen_ldr_got_small_si (rtx, rtx, rtx);
+extern rtx gen_ldr_got_small_di (rtx, rtx, rtx);
+extern rtx gen_ldr_got_small_sidi (rtx, rtx, rtx);
+extern rtx gen_ldr_got_tiny (rtx, rtx);
+extern rtx gen_aarch64_load_tp_hard (rtx);
+extern rtx gen_tlsie_small_si (rtx, rtx);
+extern rtx gen_tlsie_small_di (rtx, rtx);
+extern rtx gen_tlsie_small_sidi (rtx, rtx);
+extern rtx gen_tlsle_small_si (rtx, rtx, rtx);
+extern rtx gen_tlsle_small_di (rtx, rtx, rtx);
+extern rtx gen_tlsdesc_small_si (rtx);
+extern rtx gen_tlsdesc_small_di (rtx);
+extern rtx gen_stack_tie (rtx, rtx);
+extern rtx gen_stack_protect_set_si (rtx, rtx);
+extern rtx gen_stack_protect_set_di (rtx, rtx);
+extern rtx gen_stack_protect_test_si (rtx, rtx, rtx);
+extern rtx gen_stack_protect_test_di (rtx, rtx, rtx);
+extern rtx gen_set_fpcr (rtx);
+extern rtx gen_get_fpcr (rtx);
+extern rtx gen_set_fpsr (rtx);
+extern rtx gen_get_fpsr (rtx);
+extern rtx gen_aarch64_simd_dupv8qi (rtx, rtx);
+extern rtx gen_aarch64_simd_dupv16qi (rtx, rtx);
+extern rtx gen_aarch64_simd_dupv4hi (rtx, rtx);
+extern rtx gen_aarch64_simd_dupv8hi (rtx, rtx);
+extern rtx gen_aarch64_simd_dupv2si (rtx, rtx);
+extern rtx gen_aarch64_simd_dupv4si (rtx, rtx);
+extern rtx gen_aarch64_simd_dupv2di (rtx, rtx);
+extern rtx gen_aarch64_simd_dupv2sf (rtx, rtx);
+extern rtx gen_aarch64_simd_dupv4sf (rtx, rtx);
+extern rtx gen_aarch64_simd_dupv2df (rtx, rtx);
+extern rtx gen_aarch64_dup_lanev8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lanev16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lanev4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lanev8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lanev2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lanev4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lanev2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lanev2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lanev4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lanev2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lane_to_128v8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lane_to_64v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lane_to_128v4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lane_to_64v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lane_to_128v2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lane_to_64v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lane_to_64v2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lane_to_128v2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lane_to_64v4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_dup_lane_to_64v2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_mov_from_v16qilow (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_mov_from_v8hilow (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_mov_from_v4silow (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_mov_from_v2dilow (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_mov_from_v4sflow (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_mov_from_v2dflow (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_mov_from_v16qihigh (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_mov_from_v8hihigh (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_mov_from_v4sihigh (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_mov_from_v2dihigh (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_mov_from_v4sfhigh (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_mov_from_v2dfhigh (rtx, rtx, rtx);
+extern rtx gen_ornv8qi3 (rtx, rtx, rtx);
+extern rtx gen_ornv16qi3 (rtx, rtx, rtx);
+extern rtx gen_ornv4hi3 (rtx, rtx, rtx);
+extern rtx gen_ornv8hi3 (rtx, rtx, rtx);
+extern rtx gen_ornv2si3 (rtx, rtx, rtx);
+extern rtx gen_ornv4si3 (rtx, rtx, rtx);
+extern rtx gen_ornv2di3 (rtx, rtx, rtx);
+extern rtx gen_bicv8qi3 (rtx, rtx, rtx);
+extern rtx gen_bicv16qi3 (rtx, rtx, rtx);
+extern rtx gen_bicv4hi3 (rtx, rtx, rtx);
+extern rtx gen_bicv8hi3 (rtx, rtx, rtx);
+extern rtx gen_bicv2si3 (rtx, rtx, rtx);
+extern rtx gen_bicv4si3 (rtx, rtx, rtx);
+extern rtx gen_bicv2di3 (rtx, rtx, rtx);
+extern rtx gen_addv8qi3 (rtx, rtx, rtx);
+extern rtx gen_addv16qi3 (rtx, rtx, rtx);
+extern rtx gen_addv4hi3 (rtx, rtx, rtx);
+extern rtx gen_addv8hi3 (rtx, rtx, rtx);
+extern rtx gen_addv2si3 (rtx, rtx, rtx);
+extern rtx gen_addv4si3 (rtx, rtx, rtx);
+extern rtx gen_addv2di3 (rtx, rtx, rtx);
+extern rtx gen_subv8qi3 (rtx, rtx, rtx);
+extern rtx gen_subv16qi3 (rtx, rtx, rtx);
+extern rtx gen_subv4hi3 (rtx, rtx, rtx);
+extern rtx gen_subv8hi3 (rtx, rtx, rtx);
+extern rtx gen_subv2si3 (rtx, rtx, rtx);
+extern rtx gen_subv4si3 (rtx, rtx, rtx);
+extern rtx gen_subv2di3 (rtx, rtx, rtx);
+extern rtx gen_mulv8qi3 (rtx, rtx, rtx);
+extern rtx gen_mulv16qi3 (rtx, rtx, rtx);
+extern rtx gen_mulv4hi3 (rtx, rtx, rtx);
+extern rtx gen_mulv8hi3 (rtx, rtx, rtx);
+extern rtx gen_mulv2si3 (rtx, rtx, rtx);
+extern rtx gen_mulv4si3 (rtx, rtx, rtx);
+extern rtx gen_bswapv4hi (rtx, rtx);
+extern rtx gen_bswapv8hi (rtx, rtx);
+extern rtx gen_bswapv2si (rtx, rtx);
+extern rtx gen_bswapv4si (rtx, rtx);
+extern rtx gen_bswapv2di (rtx, rtx);
+extern rtx gen_negv8qi2 (rtx, rtx);
+extern rtx gen_negv16qi2 (rtx, rtx);
+extern rtx gen_negv4hi2 (rtx, rtx);
+extern rtx gen_negv8hi2 (rtx, rtx);
+extern rtx gen_negv2si2 (rtx, rtx);
+extern rtx gen_negv4si2 (rtx, rtx);
+extern rtx gen_negv2di2 (rtx, rtx);
+extern rtx gen_absv8qi2 (rtx, rtx);
+extern rtx gen_absv16qi2 (rtx, rtx);
+extern rtx gen_absv4hi2 (rtx, rtx);
+extern rtx gen_absv8hi2 (rtx, rtx);
+extern rtx gen_absv2si2 (rtx, rtx);
+extern rtx gen_absv4si2 (rtx, rtx);
+extern rtx gen_absv2di2 (rtx, rtx);
+extern rtx gen_abdv8qi_3 (rtx, rtx, rtx);
+extern rtx gen_abdv16qi_3 (rtx, rtx, rtx);
+extern rtx gen_abdv4hi_3 (rtx, rtx, rtx);
+extern rtx gen_abdv8hi_3 (rtx, rtx, rtx);
+extern rtx gen_abdv2si_3 (rtx, rtx, rtx);
+extern rtx gen_abdv4si_3 (rtx, rtx, rtx);
+extern rtx gen_abav8qi_3 (rtx, rtx, rtx, rtx);
+extern rtx gen_abav16qi_3 (rtx, rtx, rtx, rtx);
+extern rtx gen_abav4hi_3 (rtx, rtx, rtx, rtx);
+extern rtx gen_abav8hi_3 (rtx, rtx, rtx, rtx);
+extern rtx gen_abav2si_3 (rtx, rtx, rtx, rtx);
+extern rtx gen_abav4si_3 (rtx, rtx, rtx, rtx);
+extern rtx gen_fabdv2sf_3 (rtx, rtx, rtx);
+extern rtx gen_fabdv4sf_3 (rtx, rtx, rtx);
+extern rtx gen_fabdv2df_3 (rtx, rtx, rtx);
+extern rtx gen_andv8qi3 (rtx, rtx, rtx);
+extern rtx gen_andv16qi3 (rtx, rtx, rtx);
+extern rtx gen_andv4hi3 (rtx, rtx, rtx);
+extern rtx gen_andv8hi3 (rtx, rtx, rtx);
+extern rtx gen_andv2si3 (rtx, rtx, rtx);
+extern rtx gen_andv4si3 (rtx, rtx, rtx);
+extern rtx gen_andv2di3 (rtx, rtx, rtx);
+extern rtx gen_iorv8qi3 (rtx, rtx, rtx);
+extern rtx gen_iorv16qi3 (rtx, rtx, rtx);
+extern rtx gen_iorv4hi3 (rtx, rtx, rtx);
+extern rtx gen_iorv8hi3 (rtx, rtx, rtx);
+extern rtx gen_iorv2si3 (rtx, rtx, rtx);
+extern rtx gen_iorv4si3 (rtx, rtx, rtx);
+extern rtx gen_iorv2di3 (rtx, rtx, rtx);
+extern rtx gen_xorv8qi3 (rtx, rtx, rtx);
+extern rtx gen_xorv16qi3 (rtx, rtx, rtx);
+extern rtx gen_xorv4hi3 (rtx, rtx, rtx);
+extern rtx gen_xorv8hi3 (rtx, rtx, rtx);
+extern rtx gen_xorv2si3 (rtx, rtx, rtx);
+extern rtx gen_xorv4si3 (rtx, rtx, rtx);
+extern rtx gen_xorv2di3 (rtx, rtx, rtx);
+extern rtx gen_one_cmplv8qi2 (rtx, rtx);
+extern rtx gen_one_cmplv16qi2 (rtx, rtx);
+extern rtx gen_one_cmplv4hi2 (rtx, rtx);
+extern rtx gen_one_cmplv8hi2 (rtx, rtx);
+extern rtx gen_one_cmplv2si2 (rtx, rtx);
+extern rtx gen_one_cmplv4si2 (rtx, rtx);
+extern rtx gen_one_cmplv2di2 (rtx, rtx);
+extern rtx gen_aarch64_simd_vec_setv8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_setv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_setv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_setv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_setv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_setv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_lshrv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_lshrv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_lshrv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_lshrv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_lshrv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_lshrv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_lshrv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_ashrv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_ashrv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_ashrv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_ashrv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_ashrv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_ashrv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_ashrv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_imm_shlv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_imm_shlv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_imm_shlv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_imm_shlv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_imm_shlv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_imm_shlv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_imm_shlv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_sshlv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_sshlv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_sshlv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_sshlv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_sshlv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_sshlv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_sshlv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_shlv8qi_unsigned (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_shlv16qi_unsigned (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_shlv4hi_unsigned (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_shlv8hi_unsigned (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_shlv2si_unsigned (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_shlv4si_unsigned (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_shlv2di_unsigned (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_shlv8qi_signed (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_shlv16qi_signed (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_shlv4hi_signed (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_shlv8hi_signed (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_shlv2si_signed (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_shlv4si_signed (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_reg_shlv2di_signed (rtx, rtx, rtx);
+extern rtx gen_aarch64_sshr_simddi (rtx, rtx);
+extern rtx gen_aarch64_ushr_simddi (rtx, rtx);
+extern rtx gen_aarch64_simd_vec_setv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_setv2sf (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_setv4sf (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_setv2df (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_mlav8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_mlav16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_mlav4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_mlav8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_mlav2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_mlav4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_mlsv8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_mlsv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_mlsv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_mlsv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_mlsv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_mlsv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_smaxv8qi3 (rtx, rtx, rtx);
+extern rtx gen_sminv8qi3 (rtx, rtx, rtx);
+extern rtx gen_umaxv8qi3 (rtx, rtx, rtx);
+extern rtx gen_uminv8qi3 (rtx, rtx, rtx);
+extern rtx gen_smaxv16qi3 (rtx, rtx, rtx);
+extern rtx gen_sminv16qi3 (rtx, rtx, rtx);
+extern rtx gen_umaxv16qi3 (rtx, rtx, rtx);
+extern rtx gen_uminv16qi3 (rtx, rtx, rtx);
+extern rtx gen_smaxv4hi3 (rtx, rtx, rtx);
+extern rtx gen_sminv4hi3 (rtx, rtx, rtx);
+extern rtx gen_umaxv4hi3 (rtx, rtx, rtx);
+extern rtx gen_uminv4hi3 (rtx, rtx, rtx);
+extern rtx gen_smaxv8hi3 (rtx, rtx, rtx);
+extern rtx gen_sminv8hi3 (rtx, rtx, rtx);
+extern rtx gen_umaxv8hi3 (rtx, rtx, rtx);
+extern rtx gen_uminv8hi3 (rtx, rtx, rtx);
+extern rtx gen_smaxv2si3 (rtx, rtx, rtx);
+extern rtx gen_sminv2si3 (rtx, rtx, rtx);
+extern rtx gen_umaxv2si3 (rtx, rtx, rtx);
+extern rtx gen_uminv2si3 (rtx, rtx, rtx);
+extern rtx gen_smaxv4si3 (rtx, rtx, rtx);
+extern rtx gen_sminv4si3 (rtx, rtx, rtx);
+extern rtx gen_umaxv4si3 (rtx, rtx, rtx);
+extern rtx gen_uminv4si3 (rtx, rtx, rtx);
+extern rtx gen_move_lo_quad_internal_v16qi (rtx, rtx);
+extern rtx gen_move_lo_quad_internal_v8hi (rtx, rtx);
+extern rtx gen_move_lo_quad_internal_v4si (rtx, rtx);
+extern rtx gen_move_lo_quad_internal_v2di (rtx, rtx);
+extern rtx gen_move_lo_quad_internal_v4sf (rtx, rtx);
+extern rtx gen_move_lo_quad_internal_v2df (rtx, rtx);
+extern rtx gen_move_lo_quad_internal_be_v16qi (rtx, rtx);
+extern rtx gen_move_lo_quad_internal_be_v8hi (rtx, rtx);
+extern rtx gen_move_lo_quad_internal_be_v4si (rtx, rtx);
+extern rtx gen_move_lo_quad_internal_be_v2di (rtx, rtx);
+extern rtx gen_move_lo_quad_internal_be_v4sf (rtx, rtx);
+extern rtx gen_move_lo_quad_internal_be_v2df (rtx, rtx);
+extern rtx gen_aarch64_simd_move_hi_quad_v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_move_hi_quad_v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_move_hi_quad_v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_move_hi_quad_v2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_move_hi_quad_v4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_move_hi_quad_v2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_move_hi_quad_be_v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_move_hi_quad_be_v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_move_hi_quad_be_v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_move_hi_quad_be_v2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_move_hi_quad_be_v4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_move_hi_quad_be_v2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_pack_trunc_v8hi (rtx, rtx);
+extern rtx gen_aarch64_simd_vec_pack_trunc_v4si (rtx, rtx);
+extern rtx gen_aarch64_simd_vec_pack_trunc_v2di (rtx, rtx);
+extern rtx gen_vec_pack_trunc_v8hi (rtx, rtx, rtx);
+extern rtx gen_vec_pack_trunc_v4si (rtx, rtx, rtx);
+extern rtx gen_vec_pack_trunc_v2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_unpacks_lo_v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_unpacku_lo_v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_unpacks_lo_v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_unpacku_lo_v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_unpacks_lo_v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_unpacku_lo_v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_unpacks_hi_v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_unpacku_hi_v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_unpacks_hi_v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_unpacku_hi_v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_unpacks_hi_v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_unpacku_hi_v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_smult_lo_v16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_umult_lo_v16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_smult_lo_v8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_umult_lo_v8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_smult_lo_v4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_umult_lo_v4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_smult_hi_v16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_umult_hi_v16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_smult_hi_v8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_umult_hi_v8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_smult_hi_v4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_vec_umult_hi_v4si (rtx, rtx, rtx, rtx);
+extern rtx gen_addv2sf3 (rtx, rtx, rtx);
+extern rtx gen_addv4sf3 (rtx, rtx, rtx);
+extern rtx gen_addv2df3 (rtx, rtx, rtx);
+extern rtx gen_subv2sf3 (rtx, rtx, rtx);
+extern rtx gen_subv4sf3 (rtx, rtx, rtx);
+extern rtx gen_subv2df3 (rtx, rtx, rtx);
+extern rtx gen_mulv2sf3 (rtx, rtx, rtx);
+extern rtx gen_mulv4sf3 (rtx, rtx, rtx);
+extern rtx gen_mulv2df3 (rtx, rtx, rtx);
+extern rtx gen_divv2sf3 (rtx, rtx, rtx);
+extern rtx gen_divv4sf3 (rtx, rtx, rtx);
+extern rtx gen_divv2df3 (rtx, rtx, rtx);
+extern rtx gen_negv2sf2 (rtx, rtx);
+extern rtx gen_negv4sf2 (rtx, rtx);
+extern rtx gen_negv2df2 (rtx, rtx);
+extern rtx gen_absv2sf2 (rtx, rtx);
+extern rtx gen_absv4sf2 (rtx, rtx);
+extern rtx gen_absv2df2 (rtx, rtx);
+extern rtx gen_fmav2sf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_fmav4sf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_fmav2df4 (rtx, rtx, rtx, rtx);
+extern rtx gen_fnmav2sf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_fnmav4sf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_fnmav2df4 (rtx, rtx, rtx, rtx);
+extern rtx gen_btruncv2sf2 (rtx, rtx);
+extern rtx gen_ceilv2sf2 (rtx, rtx);
+extern rtx gen_floorv2sf2 (rtx, rtx);
+extern rtx gen_frintnv2sf2 (rtx, rtx);
+extern rtx gen_nearbyintv2sf2 (rtx, rtx);
+extern rtx gen_rintv2sf2 (rtx, rtx);
+extern rtx gen_roundv2sf2 (rtx, rtx);
+extern rtx gen_btruncv4sf2 (rtx, rtx);
+extern rtx gen_ceilv4sf2 (rtx, rtx);
+extern rtx gen_floorv4sf2 (rtx, rtx);
+extern rtx gen_frintnv4sf2 (rtx, rtx);
+extern rtx gen_nearbyintv4sf2 (rtx, rtx);
+extern rtx gen_rintv4sf2 (rtx, rtx);
+extern rtx gen_roundv4sf2 (rtx, rtx);
+extern rtx gen_btruncv2df2 (rtx, rtx);
+extern rtx gen_ceilv2df2 (rtx, rtx);
+extern rtx gen_floorv2df2 (rtx, rtx);
+extern rtx gen_frintnv2df2 (rtx, rtx);
+extern rtx gen_nearbyintv2df2 (rtx, rtx);
+extern rtx gen_rintv2df2 (rtx, rtx);
+extern rtx gen_roundv2df2 (rtx, rtx);
+extern rtx gen_lbtruncv2sfv2si2 (rtx, rtx);
+extern rtx gen_lceilv2sfv2si2 (rtx, rtx);
+extern rtx gen_lfloorv2sfv2si2 (rtx, rtx);
+extern rtx gen_lroundv2sfv2si2 (rtx, rtx);
+extern rtx gen_lfrintnv2sfv2si2 (rtx, rtx);
+extern rtx gen_lbtruncuv2sfv2si2 (rtx, rtx);
+extern rtx gen_lceiluv2sfv2si2 (rtx, rtx);
+extern rtx gen_lflooruv2sfv2si2 (rtx, rtx);
+extern rtx gen_lrounduv2sfv2si2 (rtx, rtx);
+extern rtx gen_lfrintnuv2sfv2si2 (rtx, rtx);
+extern rtx gen_lbtruncv4sfv4si2 (rtx, rtx);
+extern rtx gen_lceilv4sfv4si2 (rtx, rtx);
+extern rtx gen_lfloorv4sfv4si2 (rtx, rtx);
+extern rtx gen_lroundv4sfv4si2 (rtx, rtx);
+extern rtx gen_lfrintnv4sfv4si2 (rtx, rtx);
+extern rtx gen_lbtruncuv4sfv4si2 (rtx, rtx);
+extern rtx gen_lceiluv4sfv4si2 (rtx, rtx);
+extern rtx gen_lflooruv4sfv4si2 (rtx, rtx);
+extern rtx gen_lrounduv4sfv4si2 (rtx, rtx);
+extern rtx gen_lfrintnuv4sfv4si2 (rtx, rtx);
+extern rtx gen_lbtruncv2dfv2di2 (rtx, rtx);
+extern rtx gen_lceilv2dfv2di2 (rtx, rtx);
+extern rtx gen_lfloorv2dfv2di2 (rtx, rtx);
+extern rtx gen_lroundv2dfv2di2 (rtx, rtx);
+extern rtx gen_lfrintnv2dfv2di2 (rtx, rtx);
+extern rtx gen_lbtruncuv2dfv2di2 (rtx, rtx);
+extern rtx gen_lceiluv2dfv2di2 (rtx, rtx);
+extern rtx gen_lflooruv2dfv2di2 (rtx, rtx);
+extern rtx gen_lrounduv2dfv2di2 (rtx, rtx);
+extern rtx gen_lfrintnuv2dfv2di2 (rtx, rtx);
+extern rtx gen_floatv2siv2sf2 (rtx, rtx);
+extern rtx gen_floatunsv2siv2sf2 (rtx, rtx);
+extern rtx gen_floatv4siv4sf2 (rtx, rtx);
+extern rtx gen_floatunsv4siv4sf2 (rtx, rtx);
+extern rtx gen_floatv2div2df2 (rtx, rtx);
+extern rtx gen_floatunsv2div2df2 (rtx, rtx);
+extern rtx gen_vec_unpacks_lo_v4sf (rtx, rtx);
+extern rtx gen_aarch64_float_extend_lo_v2df (rtx, rtx);
+extern rtx gen_vec_unpacks_hi_v4sf (rtx, rtx);
+extern rtx gen_aarch64_float_truncate_lo_v2sf (rtx, rtx);
+extern rtx gen_aarch64_float_truncate_hi_v4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_vmlsv2sf (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vmlsv4sf (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vmlsv2df (rtx, rtx, rtx, rtx);
+extern rtx gen_smaxv2sf3 (rtx, rtx, rtx);
+extern rtx gen_sminv2sf3 (rtx, rtx, rtx);
+extern rtx gen_smaxv4sf3 (rtx, rtx, rtx);
+extern rtx gen_sminv4sf3 (rtx, rtx, rtx);
+extern rtx gen_smaxv2df3 (rtx, rtx, rtx);
+extern rtx gen_sminv2df3 (rtx, rtx, rtx);
+extern rtx gen_smax_nanv2sf3 (rtx, rtx, rtx);
+extern rtx gen_smin_nanv2sf3 (rtx, rtx, rtx);
+extern rtx gen_smax_nanv4sf3 (rtx, rtx, rtx);
+extern rtx gen_smin_nanv4sf3 (rtx, rtx, rtx);
+extern rtx gen_smax_nanv2df3 (rtx, rtx, rtx);
+extern rtx gen_smin_nanv2df3 (rtx, rtx, rtx);
+extern rtx gen_reduc_splus_v8qi (rtx, rtx);
+extern rtx gen_reduc_uplus_v8qi (rtx, rtx);
+extern rtx gen_reduc_splus_v16qi (rtx, rtx);
+extern rtx gen_reduc_uplus_v16qi (rtx, rtx);
+extern rtx gen_reduc_splus_v4hi (rtx, rtx);
+extern rtx gen_reduc_uplus_v4hi (rtx, rtx);
+extern rtx gen_reduc_splus_v8hi (rtx, rtx);
+extern rtx gen_reduc_uplus_v8hi (rtx, rtx);
+extern rtx gen_reduc_splus_v4si (rtx, rtx);
+extern rtx gen_reduc_uplus_v4si (rtx, rtx);
+extern rtx gen_reduc_splus_v2di (rtx, rtx);
+extern rtx gen_reduc_uplus_v2di (rtx, rtx);
+extern rtx gen_reduc_splus_v2si (rtx, rtx);
+extern rtx gen_reduc_uplus_v2si (rtx, rtx);
+extern rtx gen_reduc_splus_v2sf (rtx, rtx);
+extern rtx gen_reduc_splus_v2df (rtx, rtx);
+extern rtx gen_aarch64_addpv4sf (rtx, rtx);
+extern rtx gen_clzv8qi2 (rtx, rtx);
+extern rtx gen_clzv16qi2 (rtx, rtx);
+extern rtx gen_clzv4hi2 (rtx, rtx);
+extern rtx gen_clzv8hi2 (rtx, rtx);
+extern rtx gen_clzv2si2 (rtx, rtx);
+extern rtx gen_clzv4si2 (rtx, rtx);
+extern rtx gen_reduc_umax_v8qi (rtx, rtx);
+extern rtx gen_reduc_umin_v8qi (rtx, rtx);
+extern rtx gen_reduc_smax_v8qi (rtx, rtx);
+extern rtx gen_reduc_smin_v8qi (rtx, rtx);
+extern rtx gen_reduc_umax_v16qi (rtx, rtx);
+extern rtx gen_reduc_umin_v16qi (rtx, rtx);
+extern rtx gen_reduc_smax_v16qi (rtx, rtx);
+extern rtx gen_reduc_smin_v16qi (rtx, rtx);
+extern rtx gen_reduc_umax_v4hi (rtx, rtx);
+extern rtx gen_reduc_umin_v4hi (rtx, rtx);
+extern rtx gen_reduc_smax_v4hi (rtx, rtx);
+extern rtx gen_reduc_smin_v4hi (rtx, rtx);
+extern rtx gen_reduc_umax_v8hi (rtx, rtx);
+extern rtx gen_reduc_umin_v8hi (rtx, rtx);
+extern rtx gen_reduc_smax_v8hi (rtx, rtx);
+extern rtx gen_reduc_smin_v8hi (rtx, rtx);
+extern rtx gen_reduc_umax_v4si (rtx, rtx);
+extern rtx gen_reduc_umin_v4si (rtx, rtx);
+extern rtx gen_reduc_smax_v4si (rtx, rtx);
+extern rtx gen_reduc_smin_v4si (rtx, rtx);
+extern rtx gen_reduc_umax_v2si (rtx, rtx);
+extern rtx gen_reduc_umin_v2si (rtx, rtx);
+extern rtx gen_reduc_smax_v2si (rtx, rtx);
+extern rtx gen_reduc_smin_v2si (rtx, rtx);
+extern rtx gen_reduc_smax_nan_v2sf (rtx, rtx);
+extern rtx gen_reduc_smin_nan_v2sf (rtx, rtx);
+extern rtx gen_reduc_smax_v2sf (rtx, rtx);
+extern rtx gen_reduc_smin_v2sf (rtx, rtx);
+extern rtx gen_reduc_smax_nan_v2df (rtx, rtx);
+extern rtx gen_reduc_smin_nan_v2df (rtx, rtx);
+extern rtx gen_reduc_smax_v2df (rtx, rtx);
+extern rtx gen_reduc_smin_v2df (rtx, rtx);
+extern rtx gen_reduc_smax_nan_v4sf (rtx, rtx);
+extern rtx gen_reduc_smin_nan_v4sf (rtx, rtx);
+extern rtx gen_reduc_smax_v4sf (rtx, rtx);
+extern rtx gen_reduc_smin_v4sf (rtx, rtx);
+extern rtx gen_aarch64_simd_bslv8qi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv16qi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv4hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv8hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv2si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv4si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv2di_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv2sf_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv4sf_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv2df_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bsldi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bsldf_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_get_lanev8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_lanev16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_lanev4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_lanev8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_lanev2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_lanev4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_lanev2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_lanev2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_lanev4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_lanev2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_combine_internalv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_combine_internalv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_combine_internalv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_combine_internalv2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_combine_internaldi (rtx, rtx, rtx);
+extern rtx gen_aarch64_combine_internaldf (rtx, rtx, rtx);
+extern rtx gen_aarch64_saddlv16qi_hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssublv16qi_hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddlv16qi_hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usublv16qi_hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_saddlv8hi_hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssublv8hi_hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddlv8hi_hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usublv8hi_hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_saddlv4si_hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssublv4si_hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddlv4si_hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usublv4si_hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_saddlv16qi_lo_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssublv16qi_lo_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddlv16qi_lo_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usublv16qi_lo_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_saddlv8hi_lo_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssublv8hi_lo_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddlv8hi_lo_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usublv8hi_lo_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_saddlv4si_lo_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssublv4si_lo_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddlv4si_lo_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usublv4si_lo_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_saddlv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ssublv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddlv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usublv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_saddlv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ssublv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddlv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usublv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_saddlv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_ssublv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddlv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_usublv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_saddwv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ssubwv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddwv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usubwv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_saddwv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ssubwv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddwv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usubwv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_saddwv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_ssubwv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddwv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_usubwv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_saddw2v16qi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssubw2v16qi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddw2v16qi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usubw2v16qi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_saddw2v8hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssubw2v8hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddw2v8hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usubw2v8hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_saddw2v4si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssubw2v4si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddw2v4si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usubw2v4si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_shaddv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uhaddv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srhaddv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urhaddv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_shsubv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uhsubv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srhsubv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urhsubv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_shaddv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uhaddv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srhaddv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urhaddv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_shsubv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uhsubv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srhsubv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urhsubv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_shaddv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uhaddv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srhaddv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urhaddv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_shsubv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uhsubv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srhsubv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urhsubv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_shaddv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uhaddv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srhaddv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urhaddv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_shsubv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uhsubv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srhsubv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urhsubv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_shaddv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uhaddv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_srhaddv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_urhaddv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_shsubv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uhsubv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_srhsubv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_urhsubv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_shaddv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uhaddv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_srhaddv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_urhaddv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_shsubv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uhsubv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_srhsubv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_urhsubv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_addhnv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_raddhnv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_subhnv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_rsubhnv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_addhnv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_raddhnv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_subhnv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_rsubhnv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_addhnv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_raddhnv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_subhnv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_rsubhnv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_addhn2v8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_raddhn2v8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_subhn2v8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_rsubhn2v8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_addhn2v4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_raddhn2v4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_subhn2v4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_rsubhn2v4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_addhn2v2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_raddhn2v2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_subhn2v2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_rsubhn2v2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_pmulv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_pmulv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqaddv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqaddv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqsubv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqsubv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqaddv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqaddv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqsubv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqsubv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqaddv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqaddv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqsubv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqsubv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqaddv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqaddv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqsubv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqsubv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqaddv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqaddv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqsubv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqsubv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqaddv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqaddv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqsubv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqsubv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqaddv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqaddv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqsubv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqsubv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqaddqi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqaddqi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqsubqi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqsubqi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqaddhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqaddhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqsubhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqsubhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqaddsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqaddsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqsubsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqsubsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqadddi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqadddi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqsubdi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqsubdi (rtx, rtx, rtx);
+extern rtx gen_aarch64_suqaddv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usqaddv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_suqaddv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usqaddv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_suqaddv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usqaddv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_suqaddv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usqaddv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_suqaddv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_usqaddv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_suqaddv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_usqaddv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_suqaddv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_usqaddv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_suqaddqi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usqaddqi (rtx, rtx, rtx);
+extern rtx gen_aarch64_suqaddhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usqaddhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_suqaddsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usqaddsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_suqadddi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usqadddi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqmovunv8hi (rtx, rtx);
+extern rtx gen_aarch64_sqmovunv4si (rtx, rtx);
+extern rtx gen_aarch64_sqmovunv2di (rtx, rtx);
+extern rtx gen_aarch64_sqmovunhi (rtx, rtx);
+extern rtx gen_aarch64_sqmovunsi (rtx, rtx);
+extern rtx gen_aarch64_sqmovundi (rtx, rtx);
+extern rtx gen_aarch64_sqmovnv8hi (rtx, rtx);
+extern rtx gen_aarch64_uqmovnv8hi (rtx, rtx);
+extern rtx gen_aarch64_sqmovnv4si (rtx, rtx);
+extern rtx gen_aarch64_uqmovnv4si (rtx, rtx);
+extern rtx gen_aarch64_sqmovnv2di (rtx, rtx);
+extern rtx gen_aarch64_uqmovnv2di (rtx, rtx);
+extern rtx gen_aarch64_sqmovnhi (rtx, rtx);
+extern rtx gen_aarch64_uqmovnhi (rtx, rtx);
+extern rtx gen_aarch64_sqmovnsi (rtx, rtx);
+extern rtx gen_aarch64_uqmovnsi (rtx, rtx);
+extern rtx gen_aarch64_sqmovndi (rtx, rtx);
+extern rtx gen_aarch64_uqmovndi (rtx, rtx);
+extern rtx gen_aarch64_sqnegv8qi (rtx, rtx);
+extern rtx gen_aarch64_sqabsv8qi (rtx, rtx);
+extern rtx gen_aarch64_sqnegv16qi (rtx, rtx);
+extern rtx gen_aarch64_sqabsv16qi (rtx, rtx);
+extern rtx gen_aarch64_sqnegv4hi (rtx, rtx);
+extern rtx gen_aarch64_sqabsv4hi (rtx, rtx);
+extern rtx gen_aarch64_sqnegv8hi (rtx, rtx);
+extern rtx gen_aarch64_sqabsv8hi (rtx, rtx);
+extern rtx gen_aarch64_sqnegv2si (rtx, rtx);
+extern rtx gen_aarch64_sqabsv2si (rtx, rtx);
+extern rtx gen_aarch64_sqnegv4si (rtx, rtx);
+extern rtx gen_aarch64_sqabsv4si (rtx, rtx);
+extern rtx gen_aarch64_sqnegv2di (rtx, rtx);
+extern rtx gen_aarch64_sqabsv2di (rtx, rtx);
+extern rtx gen_aarch64_sqnegqi (rtx, rtx);
+extern rtx gen_aarch64_sqabsqi (rtx, rtx);
+extern rtx gen_aarch64_sqneghi (rtx, rtx);
+extern rtx gen_aarch64_sqabshi (rtx, rtx);
+extern rtx gen_aarch64_sqnegsi (rtx, rtx);
+extern rtx gen_aarch64_sqabssi (rtx, rtx);
+extern rtx gen_aarch64_sqnegdi (rtx, rtx);
+extern rtx gen_aarch64_sqabsdi (rtx, rtx);
+extern rtx gen_aarch64_sqdmulhv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulhv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulhv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulhv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulhv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulhv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulhv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulhv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulhhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulhhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulhsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulhsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_lanev4hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_lanev4hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_lanev8hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_lanev8hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_lanev2si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_lanev2si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_lanev4si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_lanev4si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_laneqv4hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_laneqv4hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_laneqv8hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_laneqv8hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_laneqv2si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_laneqv2si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_laneqv4si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_laneqv4si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_lanehi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_lanehi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_lanesi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_lanesi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlalv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlslv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlalv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlslv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlalhi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlslhi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlalsi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlslsi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_lanev4hi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_lanev4hi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_lanev2si_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_lanev2si_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_laneqv4hi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_laneqv4hi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_laneqv2si_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_laneqv2si_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_lanehi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_lanehi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_lanesi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_lanesi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_laneqhi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_laneqhi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_laneqsi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_laneqsi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_nv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_nv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_nv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_nv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2v8hi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2v8hi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2v4si_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2v4si_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2_lanev8hi_internal (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2_lanev8hi_internal (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2_lanev4si_internal (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2_lanev4si_internal (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2_laneqv8hi_internal (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2_laneqv8hi_internal (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2_laneqv4si_internal (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2_laneqv4si_internal (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2_nv8hi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2_nv8hi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2_nv4si_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2_nv4si_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmullv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmullv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmullhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmullsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_lanev4hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_lanev2si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_laneqv4hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_laneqv2si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_lanehi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_lanesi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_laneqhi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_laneqsi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_nv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_nv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2v8hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2v4si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2_lanev8hi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2_lanev4si_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2_laneqv8hi_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2_laneqv4si_internal (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2_nv8hi_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2_nv4si_internal (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sshlv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ushlv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshlv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshlv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sshlv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ushlv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshlv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshlv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sshlv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ushlv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshlv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshlv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sshlv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ushlv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshlv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshlv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sshlv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_ushlv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshlv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshlv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sshlv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_ushlv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshlv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshlv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sshlv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_ushlv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshlv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshlv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_sshldi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ushldi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshldi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshldi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshlv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshlv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshlv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshlv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshlv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshlv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshlv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshlv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshlv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshlv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshlv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshlv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshlv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshlv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshlv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshlv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshlv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshlv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshlv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshlv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshlv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlqi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshlqi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshlqi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshlqi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshlhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshlhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshlhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshlsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshlsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshlsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshldi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshldi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshldi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshldi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sshll_nv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ushll_nv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sshll_nv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ushll_nv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sshll_nv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_ushll_nv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sshll2_nv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ushll2_nv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sshll2_nv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ushll2_nv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sshll2_nv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_ushll2_nv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshr_nv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshr_nv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshr_nv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshr_nv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshr_nv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshr_nv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshr_nv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshr_nv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshr_nv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshr_nv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshr_nv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshr_nv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshr_nv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshr_nv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_srshr_ndi (rtx, rtx, rtx);
+extern rtx gen_aarch64_urshr_ndi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ssra_nv8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usra_nv8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_srsra_nv8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ursra_nv8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssra_nv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usra_nv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_srsra_nv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ursra_nv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssra_nv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usra_nv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_srsra_nv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ursra_nv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssra_nv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usra_nv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_srsra_nv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ursra_nv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssra_nv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usra_nv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_srsra_nv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ursra_nv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssra_nv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usra_nv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_srsra_nv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ursra_nv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssra_nv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usra_nv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_srsra_nv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ursra_nv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssra_ndi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usra_ndi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_srsra_ndi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ursra_ndi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssli_nv8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usli_nv8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssri_nv8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usri_nv8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssli_nv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usli_nv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssri_nv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usri_nv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssli_nv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usli_nv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssri_nv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usri_nv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssli_nv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usli_nv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssri_nv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usri_nv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssli_nv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usli_nv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssri_nv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usri_nv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssli_nv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usli_nv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssri_nv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usri_nv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssli_nv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usli_nv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssri_nv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usri_nv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssli_ndi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usli_ndi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_ssri_ndi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_usri_ndi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlu_nv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshl_nv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshl_nv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlu_nv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshl_nv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshl_nv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlu_nv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshl_nv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshl_nv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlu_nv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshl_nv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshl_nv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlu_nv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshl_nv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshl_nv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlu_nv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshl_nv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshl_nv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlu_nv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshl_nv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshl_nv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlu_nqi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshl_nqi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshl_nqi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlu_nhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshl_nhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshl_nhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlu_nsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshl_nsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshl_nsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshlu_ndi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshl_ndi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshl_ndi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshrun_nv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshrun_nv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshrn_nv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshrn_nv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshrn_nv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshrn_nv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshrun_nv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshrun_nv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshrn_nv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshrn_nv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshrn_nv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshrn_nv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshrun_nv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshrun_nv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshrn_nv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshrn_nv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshrn_nv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshrn_nv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshrun_nhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshrun_nhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshrn_nhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshrn_nhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshrn_nhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshrn_nhi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshrun_nsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshrun_nsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshrn_nsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshrn_nsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshrn_nsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshrn_nsi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshrun_ndi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshrun_ndi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqshrn_ndi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqshrn_ndi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrshrn_ndi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uqrshrn_ndi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmlev8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmeqv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgev8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmlev16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmeqv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgev16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmlev4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmeqv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgev4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmlev8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmeqv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgev8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmlev2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmeqv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgev2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmlev4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmeqv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgev4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmlev2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmeqv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgev2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltdi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmledi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmeqdi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgedi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtdi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltuv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmleuv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgeuv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtuv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltuv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmleuv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgeuv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtuv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltuv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmleuv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgeuv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtuv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltuv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmleuv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgeuv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtuv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltuv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmleuv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgeuv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtuv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltuv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmleuv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgeuv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtuv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltuv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmleuv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgeuv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtuv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltudi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmleudi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgeudi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtudi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmtstv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmtstv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmtstv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmtstv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmtstv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmtstv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmtstv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmtstdi (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltv2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmlev2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmeqv2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgev2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtv2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltv4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmlev4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmeqv4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgev4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtv4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltv2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmlev2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmeqv2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgev2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtv2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltsf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmlesf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmeqsf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgesf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtsf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmltdf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmledf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmeqdf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgedf (rtx, rtx, rtx);
+extern rtx gen_aarch64_cmgtdf (rtx, rtx, rtx);
+extern rtx gen_aarch64_addpv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_addpv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_addpv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_addpdi (rtx, rtx);
+extern rtx gen_sqrtv2sf2 (rtx, rtx);
+extern rtx gen_sqrtv4sf2 (rtx, rtx);
+extern rtx gen_sqrtv2df2 (rtx, rtx);
+extern rtx gen_vec_load_lanesoiv16qi (rtx, rtx);
+extern rtx gen_vec_load_lanesoiv8hi (rtx, rtx);
+extern rtx gen_vec_load_lanesoiv4si (rtx, rtx);
+extern rtx gen_vec_load_lanesoiv2di (rtx, rtx);
+extern rtx gen_vec_load_lanesoiv4sf (rtx, rtx);
+extern rtx gen_vec_load_lanesoiv2df (rtx, rtx);
+extern rtx gen_vec_store_lanesoiv16qi (rtx, rtx);
+extern rtx gen_vec_store_lanesoiv8hi (rtx, rtx);
+extern rtx gen_vec_store_lanesoiv4si (rtx, rtx);
+extern rtx gen_vec_store_lanesoiv2di (rtx, rtx);
+extern rtx gen_vec_store_lanesoiv4sf (rtx, rtx);
+extern rtx gen_vec_store_lanesoiv2df (rtx, rtx);
+extern rtx gen_vec_store_lanesoi_lanev16qi (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesoi_lanev8hi (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesoi_lanev4si (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesoi_lanev2di (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesoi_lanev4sf (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesoi_lanev2df (rtx, rtx, rtx);
+extern rtx gen_vec_load_lanesciv16qi (rtx, rtx);
+extern rtx gen_vec_load_lanesciv8hi (rtx, rtx);
+extern rtx gen_vec_load_lanesciv4si (rtx, rtx);
+extern rtx gen_vec_load_lanesciv2di (rtx, rtx);
+extern rtx gen_vec_load_lanesciv4sf (rtx, rtx);
+extern rtx gen_vec_load_lanesciv2df (rtx, rtx);
+extern rtx gen_vec_store_lanesciv16qi (rtx, rtx);
+extern rtx gen_vec_store_lanesciv8hi (rtx, rtx);
+extern rtx gen_vec_store_lanesciv4si (rtx, rtx);
+extern rtx gen_vec_store_lanesciv2di (rtx, rtx);
+extern rtx gen_vec_store_lanesciv4sf (rtx, rtx);
+extern rtx gen_vec_store_lanesciv2df (rtx, rtx);
+extern rtx gen_vec_store_lanesci_lanev16qi (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesci_lanev8hi (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesci_lanev4si (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesci_lanev2di (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesci_lanev4sf (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesci_lanev2df (rtx, rtx, rtx);
+extern rtx gen_vec_load_lanesxiv16qi (rtx, rtx);
+extern rtx gen_vec_load_lanesxiv8hi (rtx, rtx);
+extern rtx gen_vec_load_lanesxiv4si (rtx, rtx);
+extern rtx gen_vec_load_lanesxiv2di (rtx, rtx);
+extern rtx gen_vec_load_lanesxiv4sf (rtx, rtx);
+extern rtx gen_vec_load_lanesxiv2df (rtx, rtx);
+extern rtx gen_vec_store_lanesxiv16qi (rtx, rtx);
+extern rtx gen_vec_store_lanesxiv8hi (rtx, rtx);
+extern rtx gen_vec_store_lanesxiv4si (rtx, rtx);
+extern rtx gen_vec_store_lanesxiv2di (rtx, rtx);
+extern rtx gen_vec_store_lanesxiv4sf (rtx, rtx);
+extern rtx gen_vec_store_lanesxiv2df (rtx, rtx);
+extern rtx gen_vec_store_lanesxi_lanev16qi (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesxi_lanev8hi (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesxi_lanev4si (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesxi_lanev2di (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesxi_lanev4sf (rtx, rtx, rtx);
+extern rtx gen_vec_store_lanesxi_lanev2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_be_ld1v8qi (rtx, rtx);
+extern rtx gen_aarch64_be_ld1v16qi (rtx, rtx);
+extern rtx gen_aarch64_be_ld1v4hi (rtx, rtx);
+extern rtx gen_aarch64_be_ld1v8hi (rtx, rtx);
+extern rtx gen_aarch64_be_ld1v2si (rtx, rtx);
+extern rtx gen_aarch64_be_ld1v4si (rtx, rtx);
+extern rtx gen_aarch64_be_ld1v2di (rtx, rtx);
+extern rtx gen_aarch64_be_ld1v2sf (rtx, rtx);
+extern rtx gen_aarch64_be_ld1v4sf (rtx, rtx);
+extern rtx gen_aarch64_be_ld1v2df (rtx, rtx);
+extern rtx gen_aarch64_be_ld1di (rtx, rtx);
+extern rtx gen_aarch64_be_st1v8qi (rtx, rtx);
+extern rtx gen_aarch64_be_st1v16qi (rtx, rtx);
+extern rtx gen_aarch64_be_st1v4hi (rtx, rtx);
+extern rtx gen_aarch64_be_st1v8hi (rtx, rtx);
+extern rtx gen_aarch64_be_st1v2si (rtx, rtx);
+extern rtx gen_aarch64_be_st1v4si (rtx, rtx);
+extern rtx gen_aarch64_be_st1v2di (rtx, rtx);
+extern rtx gen_aarch64_be_st1v2sf (rtx, rtx);
+extern rtx gen_aarch64_be_st1v4sf (rtx, rtx);
+extern rtx gen_aarch64_be_st1v2df (rtx, rtx);
+extern rtx gen_aarch64_be_st1di (rtx, rtx);
+extern rtx gen_aarch64_ld2v8qi_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld2v4hi_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld2v2si_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld2v2sf_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld2di_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld2df_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld3v8qi_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld3v4hi_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld3v2si_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld3v2sf_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld3di_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld3df_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld4v8qi_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld4v4hi_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld4v2si_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld4v2sf_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld4di_dreg (rtx, rtx);
+extern rtx gen_aarch64_ld4df_dreg (rtx, rtx);
+extern rtx gen_aarch64_tbl1v8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_tbl1v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_tbl2v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_combinev16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip1v8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip2v8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn1v8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn2v8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp1v8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp2v8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip1v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip2v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn1v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn2v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp1v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp2v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip1v4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip2v4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn1v4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn2v4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp1v4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp2v4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip1v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip2v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn1v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn2v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp1v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp2v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip1v2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip2v2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn1v2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn2v2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp1v2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp2v2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip1v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip2v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn1v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn2v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp1v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp2v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip1v2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip2v2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn1v2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn2v2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp1v2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp2v2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip1v2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip2v2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn1v2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn2v2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp1v2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp2v2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip1v4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip2v4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn1v4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn2v4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp1v4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp2v4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip1v2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_zip2v2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn1v2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_trn2v2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp1v2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_uzp2v2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_extv8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_extv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_extv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_extv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_extv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_extv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_extv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_extv2sf (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_extv4sf (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_extv2df (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_rev64v8qi (rtx, rtx);
+extern rtx gen_aarch64_rev32v8qi (rtx, rtx);
+extern rtx gen_aarch64_rev16v8qi (rtx, rtx);
+extern rtx gen_aarch64_rev64v16qi (rtx, rtx);
+extern rtx gen_aarch64_rev32v16qi (rtx, rtx);
+extern rtx gen_aarch64_rev16v16qi (rtx, rtx);
+extern rtx gen_aarch64_rev64v4hi (rtx, rtx);
+extern rtx gen_aarch64_rev32v4hi (rtx, rtx);
+extern rtx gen_aarch64_rev16v4hi (rtx, rtx);
+extern rtx gen_aarch64_rev64v8hi (rtx, rtx);
+extern rtx gen_aarch64_rev32v8hi (rtx, rtx);
+extern rtx gen_aarch64_rev16v8hi (rtx, rtx);
+extern rtx gen_aarch64_rev64v2si (rtx, rtx);
+extern rtx gen_aarch64_rev32v2si (rtx, rtx);
+extern rtx gen_aarch64_rev16v2si (rtx, rtx);
+extern rtx gen_aarch64_rev64v4si (rtx, rtx);
+extern rtx gen_aarch64_rev32v4si (rtx, rtx);
+extern rtx gen_aarch64_rev16v4si (rtx, rtx);
+extern rtx gen_aarch64_rev64v2di (rtx, rtx);
+extern rtx gen_aarch64_rev32v2di (rtx, rtx);
+extern rtx gen_aarch64_rev16v2di (rtx, rtx);
+extern rtx gen_aarch64_rev64v2sf (rtx, rtx);
+extern rtx gen_aarch64_rev32v2sf (rtx, rtx);
+extern rtx gen_aarch64_rev16v2sf (rtx, rtx);
+extern rtx gen_aarch64_rev64v4sf (rtx, rtx);
+extern rtx gen_aarch64_rev32v4sf (rtx, rtx);
+extern rtx gen_aarch64_rev16v4sf (rtx, rtx);
+extern rtx gen_aarch64_rev64v2df (rtx, rtx);
+extern rtx gen_aarch64_rev32v2df (rtx, rtx);
+extern rtx gen_aarch64_rev16v2df (rtx, rtx);
+extern rtx gen_aarch64_st2v8qi_dreg (rtx, rtx);
+extern rtx gen_aarch64_st2v4hi_dreg (rtx, rtx);
+extern rtx gen_aarch64_st2v2si_dreg (rtx, rtx);
+extern rtx gen_aarch64_st2v2sf_dreg (rtx, rtx);
+extern rtx gen_aarch64_st2di_dreg (rtx, rtx);
+extern rtx gen_aarch64_st2df_dreg (rtx, rtx);
+extern rtx gen_aarch64_st3v8qi_dreg (rtx, rtx);
+extern rtx gen_aarch64_st3v4hi_dreg (rtx, rtx);
+extern rtx gen_aarch64_st3v2si_dreg (rtx, rtx);
+extern rtx gen_aarch64_st3v2sf_dreg (rtx, rtx);
+extern rtx gen_aarch64_st3di_dreg (rtx, rtx);
+extern rtx gen_aarch64_st3df_dreg (rtx, rtx);
+extern rtx gen_aarch64_st4v8qi_dreg (rtx, rtx);
+extern rtx gen_aarch64_st4v4hi_dreg (rtx, rtx);
+extern rtx gen_aarch64_st4v2si_dreg (rtx, rtx);
+extern rtx gen_aarch64_st4v2sf_dreg (rtx, rtx);
+extern rtx gen_aarch64_st4di_dreg (rtx, rtx);
+extern rtx gen_aarch64_st4df_dreg (rtx, rtx);
+extern rtx gen_aarch64_frecpev2sf (rtx, rtx);
+extern rtx gen_aarch64_frecpev4sf (rtx, rtx);
+extern rtx gen_aarch64_frecpev2df (rtx, rtx);
+extern rtx gen_aarch64_frecpesf (rtx, rtx);
+extern rtx gen_aarch64_frecpxsf (rtx, rtx);
+extern rtx gen_aarch64_frecpedf (rtx, rtx);
+extern rtx gen_aarch64_frecpxdf (rtx, rtx);
+extern rtx gen_aarch64_frecpsv2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_frecpsv4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_frecpsv2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_frecpssf (rtx, rtx, rtx);
+extern rtx gen_aarch64_frecpsdf (rtx, rtx, rtx);
+extern rtx gen_aarch64_crypto_aesev16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_crypto_aesdv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_crypto_aesmcv16qi (rtx, rtx);
+extern rtx gen_aarch64_crypto_aesimcv16qi (rtx, rtx);
+extern rtx gen_aarch64_crypto_sha1hsi (rtx, rtx);
+extern rtx gen_aarch64_crypto_sha1su1v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_crypto_sha1cv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_crypto_sha1mv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_crypto_sha1pv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_crypto_sha1su0v4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_crypto_sha256hv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_crypto_sha256h2v4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_crypto_sha256su0v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_crypto_sha256su1v4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_crypto_pmulldi (rtx, rtx, rtx);
+extern rtx gen_aarch64_crypto_pmullv2di (rtx, rtx, rtx);
+extern rtx gen_atomic_compare_and_swapqi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_compare_and_swaphi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_compare_and_swapsi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_compare_and_swapdi_1 (rtx, rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_exchangeqi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_exchangehi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_exchangesi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_exchangedi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_addqi (rtx, rtx, rtx);
+extern rtx gen_atomic_subqi (rtx, rtx, rtx);
+extern rtx gen_atomic_orqi (rtx, rtx, rtx);
+extern rtx gen_atomic_xorqi (rtx, rtx, rtx);
+extern rtx gen_atomic_andqi (rtx, rtx, rtx);
+extern rtx gen_atomic_addhi (rtx, rtx, rtx);
+extern rtx gen_atomic_subhi (rtx, rtx, rtx);
+extern rtx gen_atomic_orhi (rtx, rtx, rtx);
+extern rtx gen_atomic_xorhi (rtx, rtx, rtx);
+extern rtx gen_atomic_andhi (rtx, rtx, rtx);
+extern rtx gen_atomic_addsi (rtx, rtx, rtx);
+extern rtx gen_atomic_subsi (rtx, rtx, rtx);
+extern rtx gen_atomic_orsi (rtx, rtx, rtx);
+extern rtx gen_atomic_xorsi (rtx, rtx, rtx);
+extern rtx gen_atomic_andsi (rtx, rtx, rtx);
+extern rtx gen_atomic_adddi (rtx, rtx, rtx);
+extern rtx gen_atomic_subdi (rtx, rtx, rtx);
+extern rtx gen_atomic_ordi (rtx, rtx, rtx);
+extern rtx gen_atomic_xordi (rtx, rtx, rtx);
+extern rtx gen_atomic_anddi (rtx, rtx, rtx);
+extern rtx gen_atomic_nandqi (rtx, rtx, rtx);
+extern rtx gen_atomic_nandhi (rtx, rtx, rtx);
+extern rtx gen_atomic_nandsi (rtx, rtx, rtx);
+extern rtx gen_atomic_nanddi (rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_addqi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_subqi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_orqi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_xorqi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_andqi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_addhi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_subhi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_orhi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_xorhi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_andhi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_addsi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_subsi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_orsi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_xorsi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_andsi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_adddi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_subdi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_ordi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_xordi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_anddi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_nandqi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_nandhi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_nandsi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_fetch_nanddi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_add_fetchqi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_sub_fetchqi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_or_fetchqi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_xor_fetchqi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_and_fetchqi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_add_fetchhi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_sub_fetchhi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_or_fetchhi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_xor_fetchhi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_and_fetchhi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_add_fetchsi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_sub_fetchsi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_or_fetchsi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_xor_fetchsi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_and_fetchsi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_add_fetchdi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_sub_fetchdi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_or_fetchdi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_xor_fetchdi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_and_fetchdi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_nand_fetchqi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_nand_fetchhi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_nand_fetchsi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_nand_fetchdi (rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_loadqi (rtx, rtx, rtx);
+extern rtx gen_atomic_loadhi (rtx, rtx, rtx);
+extern rtx gen_atomic_loadsi (rtx, rtx, rtx);
+extern rtx gen_atomic_loaddi (rtx, rtx, rtx);
+extern rtx gen_atomic_storeqi (rtx, rtx, rtx);
+extern rtx gen_atomic_storehi (rtx, rtx, rtx);
+extern rtx gen_atomic_storesi (rtx, rtx, rtx);
+extern rtx gen_atomic_storedi (rtx, rtx, rtx);
+extern rtx gen_aarch64_load_exclusiveqi (rtx, rtx, rtx);
+extern rtx gen_aarch64_load_exclusivehi (rtx, rtx, rtx);
+extern rtx gen_aarch64_load_exclusivesi (rtx, rtx, rtx);
+extern rtx gen_aarch64_load_exclusivedi (rtx, rtx, rtx);
+extern rtx gen_aarch64_store_exclusiveqi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_store_exclusivehi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_store_exclusivesi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_store_exclusivedi (rtx, rtx, rtx, rtx);
+extern rtx gen_cbranchsi4 (rtx, rtx, rtx, rtx);
+extern rtx gen_cbranchdi4 (rtx, rtx, rtx, rtx);
+extern rtx gen_cbranchsf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_cbranchdf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_casesi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_prologue (void);
+extern rtx gen_epilogue (void);
+extern rtx gen_sibcall_epilogue (void);
#define GEN_CALL(A, B, C, D) gen_call ((A), (B), (C))
-extern rtx gen_call (rtx, rtx, rtx);
+extern rtx gen_call (rtx, rtx, rtx);
#define GEN_CALL_VALUE(A, B, C, D, E) gen_call_value ((A), (B), (C), (D))
-extern rtx gen_call_value (rtx, rtx, rtx, rtx);
+extern rtx gen_call_value (rtx, rtx, rtx, rtx);
#define GEN_SIBCALL(A, B, C, D) gen_sibcall ((A), (B), (C))
-extern rtx gen_sibcall (rtx, rtx, rtx);
+extern rtx gen_sibcall (rtx, rtx, rtx);
#define GEN_SIBCALL_VALUE(A, B, C, D, E) gen_sibcall_value ((A), (B), (C), (D))
-extern rtx gen_sibcall_value (rtx, rtx, rtx, rtx);
-extern rtx gen_untyped_call (rtx, rtx, rtx);
-extern rtx gen_movqi (rtx, rtx);
-extern rtx gen_movhi (rtx, rtx);
-extern rtx gen_movsi (rtx, rtx);
-extern rtx gen_movdi (rtx, rtx);
-extern rtx gen_movti (rtx, rtx);
-extern rtx gen_movsf (rtx, rtx);
-extern rtx gen_movdf (rtx, rtx);
-extern rtx gen_movtf (rtx, rtx);
-extern rtx gen_extendsidi2 (rtx, rtx);
-extern rtx gen_zero_extendsidi2 (rtx, rtx);
-extern rtx gen_extendqisi2 (rtx, rtx);
-extern rtx gen_zero_extendqisi2 (rtx, rtx);
-extern rtx gen_extendqidi2 (rtx, rtx);
-extern rtx gen_zero_extendqidi2 (rtx, rtx);
-extern rtx gen_extendhisi2 (rtx, rtx);
-extern rtx gen_zero_extendhisi2 (rtx, rtx);
-extern rtx gen_extendhidi2 (rtx, rtx);
-extern rtx gen_zero_extendhidi2 (rtx, rtx);
-extern rtx gen_extendqihi2 (rtx, rtx);
-extern rtx gen_zero_extendqihi2 (rtx, rtx);
-extern rtx gen_addsi3 (rtx, rtx, rtx);
-extern rtx gen_adddi3 (rtx, rtx, rtx);
-extern rtx gen_addti3 (rtx, rtx, rtx);
-extern rtx gen_subti3 (rtx, rtx, rtx);
-extern rtx gen_mulditi3 (rtx, rtx, rtx);
-extern rtx gen_umulditi3 (rtx, rtx, rtx);
-extern rtx gen_multi3 (rtx, rtx, rtx);
-extern rtx gen_cstoresi4 (rtx, rtx, rtx, rtx);
-extern rtx gen_cstoredi4 (rtx, rtx, rtx, rtx);
-extern rtx gen_cstoresf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_cstoredf4 (rtx, rtx, rtx, rtx);
-extern rtx gen_cmovsi6 (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_cmovdi6 (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_cmovsf6 (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_cmovdf6 (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_movqicc (rtx, rtx, rtx, rtx);
-extern rtx gen_movhicc (rtx, rtx, rtx, rtx);
-extern rtx gen_movsicc (rtx, rtx, rtx, rtx);
-extern rtx gen_movdicc (rtx, rtx, rtx, rtx);
-extern rtx gen_movsfsicc (rtx, rtx, rtx, rtx);
-extern rtx gen_movdfsicc (rtx, rtx, rtx, rtx);
-extern rtx gen_movsfdicc (rtx, rtx, rtx, rtx);
-extern rtx gen_movdfdicc (rtx, rtx, rtx, rtx);
-extern rtx gen_movsfcc (rtx, rtx, rtx, rtx);
-extern rtx gen_movdfcc (rtx, rtx, rtx, rtx);
-extern rtx gen_ffssi2 (rtx, rtx);
-extern rtx gen_ffsdi2 (rtx, rtx);
-extern rtx gen_ctzsi2 (rtx, rtx);
-extern rtx gen_ctzdi2 (rtx, rtx);
-extern rtx gen_ashlsi3 (rtx, rtx, rtx);
-extern rtx gen_ashrsi3 (rtx, rtx, rtx);
-extern rtx gen_lshrsi3 (rtx, rtx, rtx);
-extern rtx gen_ashldi3 (rtx, rtx, rtx);
-extern rtx gen_ashrdi3 (rtx, rtx, rtx);
-extern rtx gen_lshrdi3 (rtx, rtx, rtx);
-extern rtx gen_ashlqi3 (rtx, rtx, rtx);
-extern rtx gen_ashlhi3 (rtx, rtx, rtx);
-extern rtx gen_rotrsi3 (rtx, rtx, rtx);
-extern rtx gen_rotrdi3 (rtx, rtx, rtx);
-extern rtx gen_rotlsi3 (rtx, rtx, rtx);
-extern rtx gen_rotldi3 (rtx, rtx, rtx);
-extern rtx gen_extv (rtx, rtx, rtx, rtx);
-extern rtx gen_extzv (rtx, rtx, rtx, rtx);
-extern rtx gen_insvsi (rtx, rtx, rtx, rtx);
-extern rtx gen_insvdi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_reload_movti (rtx, rtx, rtx);
-extern rtx gen_aarch64_reload_movtf (rtx, rtx, rtx);
-extern rtx gen_add_losym (rtx, rtx, rtx);
-extern rtx gen_tlsgd_small (rtx, rtx);
-extern rtx gen_tlsle_small (rtx, rtx, rtx);
-extern rtx gen_get_thread_pointerdi (rtx);
-extern rtx gen_stack_protect_set (rtx, rtx);
-extern rtx gen_stack_protect_test (rtx, rtx, rtx);
-extern rtx gen_movv8qi (rtx, rtx);
-extern rtx gen_movv16qi (rtx, rtx);
-extern rtx gen_movv4hi (rtx, rtx);
-extern rtx gen_movv8hi (rtx, rtx);
-extern rtx gen_movv2si (rtx, rtx);
-extern rtx gen_movv4si (rtx, rtx);
-extern rtx gen_movv2di (rtx, rtx);
-extern rtx gen_movv2sf (rtx, rtx);
-extern rtx gen_movv4sf (rtx, rtx);
-extern rtx gen_movv2df (rtx, rtx);
-extern rtx gen_movmisalignv8qi (rtx, rtx);
-extern rtx gen_movmisalignv16qi (rtx, rtx);
-extern rtx gen_movmisalignv4hi (rtx, rtx);
-extern rtx gen_movmisalignv8hi (rtx, rtx);
-extern rtx gen_movmisalignv2si (rtx, rtx);
-extern rtx gen_movmisalignv4si (rtx, rtx);
-extern rtx gen_movmisalignv2di (rtx, rtx);
-extern rtx gen_movmisalignv2sf (rtx, rtx);
-extern rtx gen_movmisalignv4sf (rtx, rtx);
-extern rtx gen_movmisalignv2df (rtx, rtx);
-extern rtx gen_aarch64_split_simd_movv16qi (rtx, rtx);
-extern rtx gen_aarch64_split_simd_movv8hi (rtx, rtx);
-extern rtx gen_aarch64_split_simd_movv4si (rtx, rtx);
-extern rtx gen_aarch64_split_simd_movv2di (rtx, rtx);
-extern rtx gen_aarch64_split_simd_movv4sf (rtx, rtx);
-extern rtx gen_aarch64_split_simd_movv2df (rtx, rtx);
-extern rtx gen_ashlv8qi3 (rtx, rtx, rtx);
-extern rtx gen_ashlv16qi3 (rtx, rtx, rtx);
-extern rtx gen_ashlv4hi3 (rtx, rtx, rtx);
-extern rtx gen_ashlv8hi3 (rtx, rtx, rtx);
-extern rtx gen_ashlv2si3 (rtx, rtx, rtx);
-extern rtx gen_ashlv4si3 (rtx, rtx, rtx);
-extern rtx gen_ashlv2di3 (rtx, rtx, rtx);
-extern rtx gen_lshrv8qi3 (rtx, rtx, rtx);
-extern rtx gen_lshrv16qi3 (rtx, rtx, rtx);
-extern rtx gen_lshrv4hi3 (rtx, rtx, rtx);
-extern rtx gen_lshrv8hi3 (rtx, rtx, rtx);
-extern rtx gen_lshrv2si3 (rtx, rtx, rtx);
-extern rtx gen_lshrv4si3 (rtx, rtx, rtx);
-extern rtx gen_lshrv2di3 (rtx, rtx, rtx);
-extern rtx gen_ashrv8qi3 (rtx, rtx, rtx);
-extern rtx gen_ashrv16qi3 (rtx, rtx, rtx);
-extern rtx gen_ashrv4hi3 (rtx, rtx, rtx);
-extern rtx gen_ashrv8hi3 (rtx, rtx, rtx);
-extern rtx gen_ashrv2si3 (rtx, rtx, rtx);
-extern rtx gen_ashrv4si3 (rtx, rtx, rtx);
-extern rtx gen_ashrv2di3 (rtx, rtx, rtx);
-extern rtx gen_vashlv8qi3 (rtx, rtx, rtx);
-extern rtx gen_vashlv16qi3 (rtx, rtx, rtx);
-extern rtx gen_vashlv4hi3 (rtx, rtx, rtx);
-extern rtx gen_vashlv8hi3 (rtx, rtx, rtx);
-extern rtx gen_vashlv2si3 (rtx, rtx, rtx);
-extern rtx gen_vashlv4si3 (rtx, rtx, rtx);
-extern rtx gen_vashlv2di3 (rtx, rtx, rtx);
-extern rtx gen_vashrv8qi3 (rtx, rtx, rtx);
-extern rtx gen_vashrv16qi3 (rtx, rtx, rtx);
-extern rtx gen_vashrv4hi3 (rtx, rtx, rtx);
-extern rtx gen_vashrv8hi3 (rtx, rtx, rtx);
-extern rtx gen_vashrv2si3 (rtx, rtx, rtx);
-extern rtx gen_vashrv4si3 (rtx, rtx, rtx);
-extern rtx gen_aarch64_ashr_simddi (rtx, rtx, rtx);
-extern rtx gen_vlshrv8qi3 (rtx, rtx, rtx);
-extern rtx gen_vlshrv16qi3 (rtx, rtx, rtx);
-extern rtx gen_vlshrv4hi3 (rtx, rtx, rtx);
-extern rtx gen_vlshrv8hi3 (rtx, rtx, rtx);
-extern rtx gen_vlshrv2si3 (rtx, rtx, rtx);
-extern rtx gen_vlshrv4si3 (rtx, rtx, rtx);
-extern rtx gen_aarch64_lshr_simddi (rtx, rtx, rtx);
-extern rtx gen_vec_setv8qi (rtx, rtx, rtx);
-extern rtx gen_vec_setv16qi (rtx, rtx, rtx);
-extern rtx gen_vec_setv4hi (rtx, rtx, rtx);
-extern rtx gen_vec_setv8hi (rtx, rtx, rtx);
-extern rtx gen_vec_setv2si (rtx, rtx, rtx);
-extern rtx gen_vec_setv4si (rtx, rtx, rtx);
-extern rtx gen_vec_setv2di (rtx, rtx, rtx);
-extern rtx gen_vec_setv2sf (rtx, rtx, rtx);
-extern rtx gen_vec_setv4sf (rtx, rtx, rtx);
-extern rtx gen_vec_setv2df (rtx, rtx, rtx);
-extern rtx gen_move_hi_quad_v16qi (rtx, rtx);
-extern rtx gen_move_hi_quad_v8hi (rtx, rtx);
-extern rtx gen_move_hi_quad_v4si (rtx, rtx);
-extern rtx gen_move_hi_quad_v2di (rtx, rtx);
-extern rtx gen_move_hi_quad_v4sf (rtx, rtx);
-extern rtx gen_move_hi_quad_v2df (rtx, rtx);
-extern rtx gen_vec_pack_trunc_v4hi (rtx, rtx, rtx);
-extern rtx gen_vec_pack_trunc_v2si (rtx, rtx, rtx);
-extern rtx gen_vec_pack_trunc_di (rtx, rtx, rtx);
-extern rtx gen_vec_unpacks_hi_v16qi (rtx, rtx);
-extern rtx gen_vec_unpacku_hi_v16qi (rtx, rtx);
-extern rtx gen_vec_unpacks_hi_v8hi (rtx, rtx);
-extern rtx gen_vec_unpacku_hi_v8hi (rtx, rtx);
-extern rtx gen_vec_unpacks_hi_v4si (rtx, rtx);
-extern rtx gen_vec_unpacku_hi_v4si (rtx, rtx);
-extern rtx gen_vec_unpacks_lo_v16qi (rtx, rtx);
-extern rtx gen_vec_unpacku_lo_v16qi (rtx, rtx);
-extern rtx gen_vec_unpacks_lo_v8hi (rtx, rtx);
-extern rtx gen_vec_unpacku_lo_v8hi (rtx, rtx);
-extern rtx gen_vec_unpacks_lo_v4si (rtx, rtx);
-extern rtx gen_vec_unpacku_lo_v4si (rtx, rtx);
-extern rtx gen_vec_widen_smult_lo_v16qi (rtx, rtx, rtx);
-extern rtx gen_vec_widen_umult_lo_v16qi (rtx, rtx, rtx);
-extern rtx gen_vec_widen_smult_lo_v8hi (rtx, rtx, rtx);
-extern rtx gen_vec_widen_umult_lo_v8hi (rtx, rtx, rtx);
-extern rtx gen_vec_widen_smult_lo_v4si (rtx, rtx, rtx);
-extern rtx gen_vec_widen_umult_lo_v4si (rtx, rtx, rtx);
-extern rtx gen_vec_widen_smult_hi_v16qi (rtx, rtx, rtx);
-extern rtx gen_vec_widen_umult_hi_v16qi (rtx, rtx, rtx);
-extern rtx gen_vec_widen_smult_hi_v8hi (rtx, rtx, rtx);
-extern rtx gen_vec_widen_umult_hi_v8hi (rtx, rtx, rtx);
-extern rtx gen_vec_widen_smult_hi_v4si (rtx, rtx, rtx);
-extern rtx gen_vec_widen_umult_hi_v4si (rtx, rtx, rtx);
-extern rtx gen_fixv2sfv2si2 (rtx, rtx);
-extern rtx gen_fixunsv2sfv2si2 (rtx, rtx);
-extern rtx gen_fixv4sfv4si2 (rtx, rtx);
-extern rtx gen_fixunsv4sfv4si2 (rtx, rtx);
-extern rtx gen_fixv2dfv2di2 (rtx, rtx);
-extern rtx gen_fixunsv2dfv2di2 (rtx, rtx);
-extern rtx gen_fix_truncv2sfv2si2 (rtx, rtx);
-extern rtx gen_fixuns_truncv2sfv2si2 (rtx, rtx);
-extern rtx gen_fix_truncv4sfv4si2 (rtx, rtx);
-extern rtx gen_fixuns_truncv4sfv4si2 (rtx, rtx);
-extern rtx gen_fix_truncv2dfv2di2 (rtx, rtx);
-extern rtx gen_fixuns_truncv2dfv2di2 (rtx, rtx);
-extern rtx gen_ftruncv2sf2 (rtx, rtx);
-extern rtx gen_ftruncv4sf2 (rtx, rtx);
-extern rtx gen_ftruncv2df2 (rtx, rtx);
-extern rtx gen_vec_pack_trunc_v2df (rtx, rtx, rtx);
-extern rtx gen_vec_pack_trunc_df (rtx, rtx, rtx);
-extern rtx gen_reduc_splus_v4sf (rtx, rtx);
-extern rtx gen_aarch64_simd_bslv8qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv2sf (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv4sf (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bslv2df (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bsldi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_bsldf (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv8qiv8qi (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv16qiv16qi (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv4hiv4hi (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv8hiv8hi (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv2siv2si (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv4siv4si (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv2div2di (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv2sfv2sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv2siv2sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv4sfv2sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv4siv2sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv2dfv2sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv2div2sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv2sfv4sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv2siv4sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv4sfv4sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv4siv4sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv2dfv4sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv2div4sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv2sfv2df (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv2siv2df (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv4sfv2df (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv4siv2df (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv2dfv2df (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_vcond_internalv2div2df (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vcondv8qiv8qi (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vcondv16qiv16qi (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vcondv4hiv4hi (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vcondv8hiv8hi (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vcondv2siv2si (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vcondv4siv4si (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vcondv2div2di (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vcondv2sfv2sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vcondv4sfv4sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vcondv2dfv2df (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vcondv2siv2sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vcondv4siv4sf (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vcondv2div2df (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vconduv8qiv8qi (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vconduv16qiv16qi (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vconduv4hiv4hi (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vconduv8hiv8hi (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vconduv2siv2si (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vconduv4siv4si (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_vconduv2div2di (rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_createv8qi (rtx, rtx);
-extern rtx gen_aarch64_createv4hi (rtx, rtx);
-extern rtx gen_aarch64_createv2si (rtx, rtx);
-extern rtx gen_aarch64_createdi (rtx, rtx);
-extern rtx gen_aarch64_createdf (rtx, rtx);
-extern rtx gen_aarch64_createv2sf (rtx, rtx);
-extern rtx gen_aarch64_be_checked_get_lanev8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_be_checked_get_lanev16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_be_checked_get_lanev4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_be_checked_get_lanev8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_be_checked_get_lanev2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_be_checked_get_lanev4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_be_checked_get_lanev2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_be_checked_get_lanev2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_be_checked_get_lanev4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_be_checked_get_lanev2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_lanedi (rtx, rtx, rtx);
-extern rtx gen_aarch64_reinterpretv8qiv8qi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv8qiv4hi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv8qiv2si (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv8qiv2sf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv8qidi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv8qidf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4hiv8qi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4hiv4hi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4hiv2si (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4hiv2sf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4hidi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4hidf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2siv8qi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2siv4hi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2siv2si (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2siv2sf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2sidi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2sidf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2sfv8qi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2sfv4hi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2sfv2si (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2sfv2sf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2sfdi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2sfdf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretdiv8qi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretdiv4hi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretdiv2si (rtx, rtx);
-extern rtx gen_aarch64_reinterpretdidi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretdidf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretdiv2sf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretdfv8qi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretdfv4hi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretdfv2si (rtx, rtx);
-extern rtx gen_aarch64_reinterpretdfdi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretdfdf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretdfv2sf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv16qiv16qi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv16qiv8hi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv16qiv4si (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv16qiv2di (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv16qiv4sf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv16qiv2df (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv8hiv16qi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv8hiv8hi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv8hiv4si (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv8hiv2di (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv8hiv4sf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv8hiv2df (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4siv16qi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4siv8hi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4siv4si (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4siv2di (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4siv4sf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4siv2df (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4sfv16qi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4sfv8hi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4sfv4si (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4sfv2di (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4sfv4sf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv4sfv2df (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2div16qi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2div8hi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2div4si (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2div2di (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2div4sf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2div2df (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2dfv16qi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2dfv8hi (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2dfv4si (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2dfv2di (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2dfv4sf (rtx, rtx);
-extern rtx gen_aarch64_reinterpretv2dfv2df (rtx, rtx);
-extern rtx gen_aarch64_simd_combinev8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_combinev4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_combinev2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_combinev2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_combinedi (rtx, rtx, rtx);
-extern rtx gen_aarch64_simd_combinedf (rtx, rtx, rtx);
-extern rtx gen_aarch64_saddl2v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_saddl2v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_saddl2v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddl2v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddl2v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddl2v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_ssubl2v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ssubl2v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ssubl2v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_usubl2v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usubl2v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usubl2v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_saddw2v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_saddw2v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_saddw2v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddw2v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddw2v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_uaddw2v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_ssubw2v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ssubw2v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_ssubw2v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_usubw2v16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usubw2v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_usubw2v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal_lanev4hi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal_lanev2si (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal_lanehi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal_lanesi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal_laneqv4hi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal_laneqv2si (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal_laneqhi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal_laneqsi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl_lanev4hi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl_lanev2si (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl_lanehi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl_lanesi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl_laneqv4hi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl_laneqv2si (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl_laneqhi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl_laneqsi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal2v8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal2v4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl2v8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl2v4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal2_lanev8hi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal2_lanev4si (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal2_laneqv8hi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal2_laneqv4si (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl2_lanev8hi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl2_lanev4si (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl2_laneqv8hi (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl2_laneqv4si (rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal2_nv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlal2_nv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl2_nv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmlsl2_nv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull_lanev4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull_lanev2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull_lanehi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull_lanesi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull_laneqv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull_laneqv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull2v8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull2v4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull2_lanev8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull2_lanev4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull2_laneqv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull2_laneqv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull2_nv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_sqdmull2_nv4si (rtx, rtx, rtx);
-extern rtx gen_movoi (rtx, rtx);
-extern rtx gen_movci (rtx, rtx);
-extern rtx gen_movxi (rtx, rtx);
-extern rtx gen_aarch64_ld2v8qi (rtx, rtx);
-extern rtx gen_aarch64_ld2v4hi (rtx, rtx);
-extern rtx gen_aarch64_ld2v2si (rtx, rtx);
-extern rtx gen_aarch64_ld2v2sf (rtx, rtx);
-extern rtx gen_aarch64_ld2di (rtx, rtx);
-extern rtx gen_aarch64_ld2df (rtx, rtx);
-extern rtx gen_aarch64_ld3v8qi (rtx, rtx);
-extern rtx gen_aarch64_ld3v4hi (rtx, rtx);
-extern rtx gen_aarch64_ld3v2si (rtx, rtx);
-extern rtx gen_aarch64_ld3v2sf (rtx, rtx);
-extern rtx gen_aarch64_ld3di (rtx, rtx);
-extern rtx gen_aarch64_ld3df (rtx, rtx);
-extern rtx gen_aarch64_ld4v8qi (rtx, rtx);
-extern rtx gen_aarch64_ld4v4hi (rtx, rtx);
-extern rtx gen_aarch64_ld4v2si (rtx, rtx);
-extern rtx gen_aarch64_ld4v2sf (rtx, rtx);
-extern rtx gen_aarch64_ld4di (rtx, rtx);
-extern rtx gen_aarch64_ld4df (rtx, rtx);
-extern rtx gen_aarch64_ld1v8qi (rtx, rtx);
-extern rtx gen_aarch64_ld1v16qi (rtx, rtx);
-extern rtx gen_aarch64_ld1v4hi (rtx, rtx);
-extern rtx gen_aarch64_ld1v8hi (rtx, rtx);
-extern rtx gen_aarch64_ld1v2si (rtx, rtx);
-extern rtx gen_aarch64_ld1v4si (rtx, rtx);
-extern rtx gen_aarch64_ld1v2di (rtx, rtx);
-extern rtx gen_aarch64_ld1v2sf (rtx, rtx);
-extern rtx gen_aarch64_ld1v4sf (rtx, rtx);
-extern rtx gen_aarch64_ld1v2df (rtx, rtx);
-extern rtx gen_aarch64_ld2v16qi (rtx, rtx);
-extern rtx gen_aarch64_ld3v16qi (rtx, rtx);
-extern rtx gen_aarch64_ld4v16qi (rtx, rtx);
-extern rtx gen_aarch64_ld2v8hi (rtx, rtx);
-extern rtx gen_aarch64_ld3v8hi (rtx, rtx);
-extern rtx gen_aarch64_ld4v8hi (rtx, rtx);
-extern rtx gen_aarch64_ld2v4si (rtx, rtx);
-extern rtx gen_aarch64_ld3v4si (rtx, rtx);
-extern rtx gen_aarch64_ld4v4si (rtx, rtx);
-extern rtx gen_aarch64_ld2v2di (rtx, rtx);
-extern rtx gen_aarch64_ld3v2di (rtx, rtx);
-extern rtx gen_aarch64_ld4v2di (rtx, rtx);
-extern rtx gen_aarch64_ld2v4sf (rtx, rtx);
-extern rtx gen_aarch64_ld3v4sf (rtx, rtx);
-extern rtx gen_aarch64_ld4v4sf (rtx, rtx);
-extern rtx gen_aarch64_ld2v2df (rtx, rtx);
-extern rtx gen_aarch64_ld3v2df (rtx, rtx);
-extern rtx gen_aarch64_ld4v2df (rtx, rtx);
-extern rtx gen_aarch64_get_dregoiv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregoiv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregoiv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregoiv2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregoidi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregoidf (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregciv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregciv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregciv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregciv2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregcidi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregcidf (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregxiv8qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregxiv4hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregxiv2si (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregxiv2sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregxidi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_dregxidf (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregoiv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregciv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregxiv16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregoiv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregciv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregxiv8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregoiv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregciv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregxiv4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregoiv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregciv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregxiv2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregoiv4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregciv4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregxiv4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregoiv2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregciv2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_get_qregxiv2df (rtx, rtx, rtx);
-extern rtx gen_vec_perm_constv8qi (rtx, rtx, rtx, rtx);
-extern rtx gen_vec_perm_constv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_vec_perm_constv4hi (rtx, rtx, rtx, rtx);
-extern rtx gen_vec_perm_constv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_vec_perm_constv2si (rtx, rtx, rtx, rtx);
-extern rtx gen_vec_perm_constv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_vec_perm_constv2di (rtx, rtx, rtx, rtx);
-extern rtx gen_vec_perm_constv2sf (rtx, rtx, rtx, rtx);
-extern rtx gen_vec_perm_constv4sf (rtx, rtx, rtx, rtx);
-extern rtx gen_vec_perm_constv2df (rtx, rtx, rtx, rtx);
-extern rtx gen_vec_permv8qi (rtx, rtx, rtx, rtx);
-extern rtx gen_vec_permv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_st2v8qi (rtx, rtx);
-extern rtx gen_aarch64_st2v4hi (rtx, rtx);
-extern rtx gen_aarch64_st2v2si (rtx, rtx);
-extern rtx gen_aarch64_st2v2sf (rtx, rtx);
-extern rtx gen_aarch64_st2di (rtx, rtx);
-extern rtx gen_aarch64_st2df (rtx, rtx);
-extern rtx gen_aarch64_st3v8qi (rtx, rtx);
-extern rtx gen_aarch64_st3v4hi (rtx, rtx);
-extern rtx gen_aarch64_st3v2si (rtx, rtx);
-extern rtx gen_aarch64_st3v2sf (rtx, rtx);
-extern rtx gen_aarch64_st3di (rtx, rtx);
-extern rtx gen_aarch64_st3df (rtx, rtx);
-extern rtx gen_aarch64_st4v8qi (rtx, rtx);
-extern rtx gen_aarch64_st4v4hi (rtx, rtx);
-extern rtx gen_aarch64_st4v2si (rtx, rtx);
-extern rtx gen_aarch64_st4v2sf (rtx, rtx);
-extern rtx gen_aarch64_st4di (rtx, rtx);
-extern rtx gen_aarch64_st4df (rtx, rtx);
-extern rtx gen_aarch64_st2v16qi (rtx, rtx);
-extern rtx gen_aarch64_st3v16qi (rtx, rtx);
-extern rtx gen_aarch64_st4v16qi (rtx, rtx);
-extern rtx gen_aarch64_st2v8hi (rtx, rtx);
-extern rtx gen_aarch64_st3v8hi (rtx, rtx);
-extern rtx gen_aarch64_st4v8hi (rtx, rtx);
-extern rtx gen_aarch64_st2v4si (rtx, rtx);
-extern rtx gen_aarch64_st3v4si (rtx, rtx);
-extern rtx gen_aarch64_st4v4si (rtx, rtx);
-extern rtx gen_aarch64_st2v2di (rtx, rtx);
-extern rtx gen_aarch64_st3v2di (rtx, rtx);
-extern rtx gen_aarch64_st4v2di (rtx, rtx);
-extern rtx gen_aarch64_st2v4sf (rtx, rtx);
-extern rtx gen_aarch64_st3v4sf (rtx, rtx);
-extern rtx gen_aarch64_st4v4sf (rtx, rtx);
-extern rtx gen_aarch64_st2v2df (rtx, rtx);
-extern rtx gen_aarch64_st3v2df (rtx, rtx);
-extern rtx gen_aarch64_st4v2df (rtx, rtx);
-extern rtx gen_aarch64_st2_lanev16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_st2_lanev8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_st2_lanev4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_st2_lanev2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_st2_lanev4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_st2_lanev2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_st3_lanev16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_st3_lanev8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_st3_lanev4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_st3_lanev2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_st3_lanev4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_st3_lanev2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_st4_lanev16qi (rtx, rtx, rtx);
-extern rtx gen_aarch64_st4_lanev8hi (rtx, rtx, rtx);
-extern rtx gen_aarch64_st4_lanev4si (rtx, rtx, rtx);
-extern rtx gen_aarch64_st4_lanev2di (rtx, rtx, rtx);
-extern rtx gen_aarch64_st4_lanev4sf (rtx, rtx, rtx);
-extern rtx gen_aarch64_st4_lanev2df (rtx, rtx, rtx);
-extern rtx gen_aarch64_st1v8qi (rtx, rtx);
-extern rtx gen_aarch64_st1v16qi (rtx, rtx);
-extern rtx gen_aarch64_st1v4hi (rtx, rtx);
-extern rtx gen_aarch64_st1v8hi (rtx, rtx);
-extern rtx gen_aarch64_st1v2si (rtx, rtx);
-extern rtx gen_aarch64_st1v4si (rtx, rtx);
-extern rtx gen_aarch64_st1v2di (rtx, rtx);
-extern rtx gen_aarch64_st1v2sf (rtx, rtx);
-extern rtx gen_aarch64_st1v4sf (rtx, rtx);
-extern rtx gen_aarch64_st1v2df (rtx, rtx);
-extern rtx gen_aarch64_set_qregoiv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregciv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregxiv16qi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregoiv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregciv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregxiv8hi (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregoiv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregciv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregxiv4si (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregoiv2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregciv2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregxiv2di (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregoiv4sf (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregciv4sf (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregxiv4sf (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregoiv2df (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregciv2df (rtx, rtx, rtx, rtx);
-extern rtx gen_aarch64_set_qregxiv2df (rtx, rtx, rtx, rtx);
-extern rtx gen_vec_initv8qi (rtx, rtx);
-extern rtx gen_vec_initv16qi (rtx, rtx);
-extern rtx gen_vec_initv4hi (rtx, rtx);
-extern rtx gen_vec_initv8hi (rtx, rtx);
-extern rtx gen_vec_initv2si (rtx, rtx);
-extern rtx gen_vec_initv4si (rtx, rtx);
-extern rtx gen_vec_initv2di (rtx, rtx);
-extern rtx gen_vec_initv2sf (rtx, rtx);
-extern rtx gen_vec_initv4sf (rtx, rtx);
-extern rtx gen_vec_initv2df (rtx, rtx);
-extern rtx gen_vec_extractv8qi (rtx, rtx, rtx);
-extern rtx gen_vec_extractv16qi (rtx, rtx, rtx);
-extern rtx gen_vec_extractv4hi (rtx, rtx, rtx);
-extern rtx gen_vec_extractv8hi (rtx, rtx, rtx);
-extern rtx gen_vec_extractv2si (rtx, rtx, rtx);
-extern rtx gen_vec_extractv4si (rtx, rtx, rtx);
-extern rtx gen_vec_extractv2di (rtx, rtx, rtx);
-extern rtx gen_vec_extractv2sf (rtx, rtx, rtx);
-extern rtx gen_vec_extractv4sf (rtx, rtx, rtx);
-extern rtx gen_vec_extractv2df (rtx, rtx, rtx);
-extern rtx gen_atomic_compare_and_swapqi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_compare_and_swaphi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_compare_and_swapsi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_atomic_compare_and_swapdi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx);
-extern rtx gen_mem_thread_fence (rtx);
-extern rtx gen_dmb (rtx);
+extern rtx gen_sibcall_value (rtx, rtx, rtx, rtx);
+extern rtx gen_untyped_call (rtx, rtx, rtx);
+extern rtx gen_movqi (rtx, rtx);
+extern rtx gen_movhi (rtx, rtx);
+extern rtx gen_movsi (rtx, rtx);
+extern rtx gen_movdi (rtx, rtx);
+extern rtx gen_movti (rtx, rtx);
+extern rtx gen_movsf (rtx, rtx);
+extern rtx gen_movdf (rtx, rtx);
+extern rtx gen_movtf (rtx, rtx);
+extern rtx gen_movmemdi (rtx, rtx, rtx, rtx);
+extern rtx gen_extendsidi2 (rtx, rtx);
+extern rtx gen_zero_extendsidi2 (rtx, rtx);
+extern rtx gen_extendqisi2 (rtx, rtx);
+extern rtx gen_zero_extendqisi2 (rtx, rtx);
+extern rtx gen_extendqidi2 (rtx, rtx);
+extern rtx gen_zero_extendqidi2 (rtx, rtx);
+extern rtx gen_extendhisi2 (rtx, rtx);
+extern rtx gen_zero_extendhisi2 (rtx, rtx);
+extern rtx gen_extendhidi2 (rtx, rtx);
+extern rtx gen_zero_extendhidi2 (rtx, rtx);
+extern rtx gen_extendqihi2 (rtx, rtx);
+extern rtx gen_zero_extendqihi2 (rtx, rtx);
+extern rtx gen_addsi3 (rtx, rtx, rtx);
+extern rtx gen_adddi3 (rtx, rtx, rtx);
+extern rtx gen_addti3 (rtx, rtx, rtx);
+extern rtx gen_subti3 (rtx, rtx, rtx);
+extern rtx gen_mulditi3 (rtx, rtx, rtx);
+extern rtx gen_umulditi3 (rtx, rtx, rtx);
+extern rtx gen_multi3 (rtx, rtx, rtx);
+extern rtx gen_cstoresi4 (rtx, rtx, rtx, rtx);
+extern rtx gen_cstoredi4 (rtx, rtx, rtx, rtx);
+extern rtx gen_cstoresf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_cstoredf4 (rtx, rtx, rtx, rtx);
+extern rtx gen_cmovsi6 (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_cmovdi6 (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_cmovsf6 (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_cmovdf6 (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_movqicc (rtx, rtx, rtx, rtx);
+extern rtx gen_movhicc (rtx, rtx, rtx, rtx);
+extern rtx gen_movsicc (rtx, rtx, rtx, rtx);
+extern rtx gen_movdicc (rtx, rtx, rtx, rtx);
+extern rtx gen_movsfsicc (rtx, rtx, rtx, rtx);
+extern rtx gen_movdfsicc (rtx, rtx, rtx, rtx);
+extern rtx gen_movsfdicc (rtx, rtx, rtx, rtx);
+extern rtx gen_movdfdicc (rtx, rtx, rtx, rtx);
+extern rtx gen_movsfcc (rtx, rtx, rtx, rtx);
+extern rtx gen_movdfcc (rtx, rtx, rtx, rtx);
+extern rtx gen_ffssi2 (rtx, rtx);
+extern rtx gen_ffsdi2 (rtx, rtx);
+extern rtx gen_ctzsi2 (rtx, rtx);
+extern rtx gen_ctzdi2 (rtx, rtx);
+extern rtx gen_ashlsi3 (rtx, rtx, rtx);
+extern rtx gen_ashrsi3 (rtx, rtx, rtx);
+extern rtx gen_lshrsi3 (rtx, rtx, rtx);
+extern rtx gen_ashldi3 (rtx, rtx, rtx);
+extern rtx gen_ashrdi3 (rtx, rtx, rtx);
+extern rtx gen_lshrdi3 (rtx, rtx, rtx);
+extern rtx gen_ashlqi3 (rtx, rtx, rtx);
+extern rtx gen_ashlhi3 (rtx, rtx, rtx);
+extern rtx gen_rotrsi3 (rtx, rtx, rtx);
+extern rtx gen_rotrdi3 (rtx, rtx, rtx);
+extern rtx gen_rotlsi3 (rtx, rtx, rtx);
+extern rtx gen_rotldi3 (rtx, rtx, rtx);
+extern rtx gen_extv (rtx, rtx, rtx, rtx);
+extern rtx gen_extzv (rtx, rtx, rtx, rtx);
+extern rtx gen_insvsi (rtx, rtx, rtx, rtx);
+extern rtx gen_insvdi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_reload_movti (rtx, rtx, rtx);
+extern rtx gen_aarch64_reload_movtf (rtx, rtx, rtx);
+extern rtx gen_add_losym (rtx, rtx, rtx);
+extern rtx gen_tlsgd_small (rtx, rtx);
+extern rtx gen_tlsle_small (rtx, rtx, rtx);
+extern rtx gen_get_thread_pointerdi (rtx);
+extern rtx gen_stack_protect_set (rtx, rtx);
+extern rtx gen_stack_protect_test (rtx, rtx, rtx);
+extern rtx gen_movv8qi (rtx, rtx);
+extern rtx gen_movv16qi (rtx, rtx);
+extern rtx gen_movv4hi (rtx, rtx);
+extern rtx gen_movv8hi (rtx, rtx);
+extern rtx gen_movv2si (rtx, rtx);
+extern rtx gen_movv4si (rtx, rtx);
+extern rtx gen_movv2di (rtx, rtx);
+extern rtx gen_movv2sf (rtx, rtx);
+extern rtx gen_movv4sf (rtx, rtx);
+extern rtx gen_movv2df (rtx, rtx);
+extern rtx gen_movmisalignv8qi (rtx, rtx);
+extern rtx gen_movmisalignv16qi (rtx, rtx);
+extern rtx gen_movmisalignv4hi (rtx, rtx);
+extern rtx gen_movmisalignv8hi (rtx, rtx);
+extern rtx gen_movmisalignv2si (rtx, rtx);
+extern rtx gen_movmisalignv4si (rtx, rtx);
+extern rtx gen_movmisalignv2di (rtx, rtx);
+extern rtx gen_movmisalignv2sf (rtx, rtx);
+extern rtx gen_movmisalignv4sf (rtx, rtx);
+extern rtx gen_movmisalignv2df (rtx, rtx);
+extern rtx gen_aarch64_split_simd_movv16qi (rtx, rtx);
+extern rtx gen_aarch64_split_simd_movv8hi (rtx, rtx);
+extern rtx gen_aarch64_split_simd_movv4si (rtx, rtx);
+extern rtx gen_aarch64_split_simd_movv2di (rtx, rtx);
+extern rtx gen_aarch64_split_simd_movv4sf (rtx, rtx);
+extern rtx gen_aarch64_split_simd_movv2df (rtx, rtx);
+extern rtx gen_ashlv8qi3 (rtx, rtx, rtx);
+extern rtx gen_ashlv16qi3 (rtx, rtx, rtx);
+extern rtx gen_ashlv4hi3 (rtx, rtx, rtx);
+extern rtx gen_ashlv8hi3 (rtx, rtx, rtx);
+extern rtx gen_ashlv2si3 (rtx, rtx, rtx);
+extern rtx gen_ashlv4si3 (rtx, rtx, rtx);
+extern rtx gen_ashlv2di3 (rtx, rtx, rtx);
+extern rtx gen_lshrv8qi3 (rtx, rtx, rtx);
+extern rtx gen_lshrv16qi3 (rtx, rtx, rtx);
+extern rtx gen_lshrv4hi3 (rtx, rtx, rtx);
+extern rtx gen_lshrv8hi3 (rtx, rtx, rtx);
+extern rtx gen_lshrv2si3 (rtx, rtx, rtx);
+extern rtx gen_lshrv4si3 (rtx, rtx, rtx);
+extern rtx gen_lshrv2di3 (rtx, rtx, rtx);
+extern rtx gen_ashrv8qi3 (rtx, rtx, rtx);
+extern rtx gen_ashrv16qi3 (rtx, rtx, rtx);
+extern rtx gen_ashrv4hi3 (rtx, rtx, rtx);
+extern rtx gen_ashrv8hi3 (rtx, rtx, rtx);
+extern rtx gen_ashrv2si3 (rtx, rtx, rtx);
+extern rtx gen_ashrv4si3 (rtx, rtx, rtx);
+extern rtx gen_ashrv2di3 (rtx, rtx, rtx);
+extern rtx gen_vashlv8qi3 (rtx, rtx, rtx);
+extern rtx gen_vashlv16qi3 (rtx, rtx, rtx);
+extern rtx gen_vashlv4hi3 (rtx, rtx, rtx);
+extern rtx gen_vashlv8hi3 (rtx, rtx, rtx);
+extern rtx gen_vashlv2si3 (rtx, rtx, rtx);
+extern rtx gen_vashlv4si3 (rtx, rtx, rtx);
+extern rtx gen_vashlv2di3 (rtx, rtx, rtx);
+extern rtx gen_vashrv8qi3 (rtx, rtx, rtx);
+extern rtx gen_vashrv16qi3 (rtx, rtx, rtx);
+extern rtx gen_vashrv4hi3 (rtx, rtx, rtx);
+extern rtx gen_vashrv8hi3 (rtx, rtx, rtx);
+extern rtx gen_vashrv2si3 (rtx, rtx, rtx);
+extern rtx gen_vashrv4si3 (rtx, rtx, rtx);
+extern rtx gen_aarch64_ashr_simddi (rtx, rtx, rtx);
+extern rtx gen_vlshrv8qi3 (rtx, rtx, rtx);
+extern rtx gen_vlshrv16qi3 (rtx, rtx, rtx);
+extern rtx gen_vlshrv4hi3 (rtx, rtx, rtx);
+extern rtx gen_vlshrv8hi3 (rtx, rtx, rtx);
+extern rtx gen_vlshrv2si3 (rtx, rtx, rtx);
+extern rtx gen_vlshrv4si3 (rtx, rtx, rtx);
+extern rtx gen_aarch64_lshr_simddi (rtx, rtx, rtx);
+extern rtx gen_vec_setv8qi (rtx, rtx, rtx);
+extern rtx gen_vec_setv16qi (rtx, rtx, rtx);
+extern rtx gen_vec_setv4hi (rtx, rtx, rtx);
+extern rtx gen_vec_setv8hi (rtx, rtx, rtx);
+extern rtx gen_vec_setv2si (rtx, rtx, rtx);
+extern rtx gen_vec_setv4si (rtx, rtx, rtx);
+extern rtx gen_vec_setv2di (rtx, rtx, rtx);
+extern rtx gen_vec_setv2sf (rtx, rtx, rtx);
+extern rtx gen_vec_setv4sf (rtx, rtx, rtx);
+extern rtx gen_vec_setv2df (rtx, rtx, rtx);
+extern rtx gen_move_lo_quad_v16qi (rtx, rtx);
+extern rtx gen_move_lo_quad_v8hi (rtx, rtx);
+extern rtx gen_move_lo_quad_v4si (rtx, rtx);
+extern rtx gen_move_lo_quad_v2di (rtx, rtx);
+extern rtx gen_move_lo_quad_v4sf (rtx, rtx);
+extern rtx gen_move_lo_quad_v2df (rtx, rtx);
+extern rtx gen_move_hi_quad_v16qi (rtx, rtx);
+extern rtx gen_move_hi_quad_v8hi (rtx, rtx);
+extern rtx gen_move_hi_quad_v4si (rtx, rtx);
+extern rtx gen_move_hi_quad_v2di (rtx, rtx);
+extern rtx gen_move_hi_quad_v4sf (rtx, rtx);
+extern rtx gen_move_hi_quad_v2df (rtx, rtx);
+extern rtx gen_vec_pack_trunc_v4hi (rtx, rtx, rtx);
+extern rtx gen_vec_pack_trunc_v2si (rtx, rtx, rtx);
+extern rtx gen_vec_pack_trunc_di (rtx, rtx, rtx);
+extern rtx gen_vec_unpacks_hi_v16qi (rtx, rtx);
+extern rtx gen_vec_unpacku_hi_v16qi (rtx, rtx);
+extern rtx gen_vec_unpacks_hi_v8hi (rtx, rtx);
+extern rtx gen_vec_unpacku_hi_v8hi (rtx, rtx);
+extern rtx gen_vec_unpacks_hi_v4si (rtx, rtx);
+extern rtx gen_vec_unpacku_hi_v4si (rtx, rtx);
+extern rtx gen_vec_unpacks_lo_v16qi (rtx, rtx);
+extern rtx gen_vec_unpacku_lo_v16qi (rtx, rtx);
+extern rtx gen_vec_unpacks_lo_v8hi (rtx, rtx);
+extern rtx gen_vec_unpacku_lo_v8hi (rtx, rtx);
+extern rtx gen_vec_unpacks_lo_v4si (rtx, rtx);
+extern rtx gen_vec_unpacku_lo_v4si (rtx, rtx);
+extern rtx gen_vec_widen_smult_lo_v16qi (rtx, rtx, rtx);
+extern rtx gen_vec_widen_umult_lo_v16qi (rtx, rtx, rtx);
+extern rtx gen_vec_widen_smult_lo_v8hi (rtx, rtx, rtx);
+extern rtx gen_vec_widen_umult_lo_v8hi (rtx, rtx, rtx);
+extern rtx gen_vec_widen_smult_lo_v4si (rtx, rtx, rtx);
+extern rtx gen_vec_widen_umult_lo_v4si (rtx, rtx, rtx);
+extern rtx gen_vec_widen_smult_hi_v16qi (rtx, rtx, rtx);
+extern rtx gen_vec_widen_umult_hi_v16qi (rtx, rtx, rtx);
+extern rtx gen_vec_widen_smult_hi_v8hi (rtx, rtx, rtx);
+extern rtx gen_vec_widen_umult_hi_v8hi (rtx, rtx, rtx);
+extern rtx gen_vec_widen_smult_hi_v4si (rtx, rtx, rtx);
+extern rtx gen_vec_widen_umult_hi_v4si (rtx, rtx, rtx);
+extern rtx gen_fixv2sfv2si2 (rtx, rtx);
+extern rtx gen_fixunsv2sfv2si2 (rtx, rtx);
+extern rtx gen_fixv4sfv4si2 (rtx, rtx);
+extern rtx gen_fixunsv4sfv4si2 (rtx, rtx);
+extern rtx gen_fixv2dfv2di2 (rtx, rtx);
+extern rtx gen_fixunsv2dfv2di2 (rtx, rtx);
+extern rtx gen_fix_truncv2sfv2si2 (rtx, rtx);
+extern rtx gen_fixuns_truncv2sfv2si2 (rtx, rtx);
+extern rtx gen_fix_truncv4sfv4si2 (rtx, rtx);
+extern rtx gen_fixuns_truncv4sfv4si2 (rtx, rtx);
+extern rtx gen_fix_truncv2dfv2di2 (rtx, rtx);
+extern rtx gen_fixuns_truncv2dfv2di2 (rtx, rtx);
+extern rtx gen_ftruncv2sf2 (rtx, rtx);
+extern rtx gen_ftruncv4sf2 (rtx, rtx);
+extern rtx gen_ftruncv2df2 (rtx, rtx);
+extern rtx gen_vec_pack_trunc_v2df (rtx, rtx, rtx);
+extern rtx gen_vec_pack_trunc_df (rtx, rtx, rtx);
+extern rtx gen_reduc_splus_v4sf (rtx, rtx);
+extern rtx gen_aarch64_simd_bslv8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv2sf (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv4sf (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bslv2df (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bsldi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_bsldf (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv8qiv8qi (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv16qiv16qi (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv4hiv4hi (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv8hiv8hi (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv2siv2si (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv4siv4si (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv2div2di (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv2sfv2sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv2siv2sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv4sfv2sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv4siv2sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv2dfv2sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv2div2sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv2sfv4sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv2siv4sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv4sfv4sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv4siv4sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv2dfv4sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv2div4sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv2sfv2df (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv2siv2df (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv4sfv2df (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv4siv2df (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv2dfv2df (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_vcond_internalv2div2df (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vcondv8qiv8qi (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vcondv16qiv16qi (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vcondv4hiv4hi (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vcondv8hiv8hi (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vcondv2siv2si (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vcondv4siv4si (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vcondv2div2di (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vcondv2sfv2sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vcondv4sfv4sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vcondv2dfv2df (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vcondv2siv2sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vcondv4siv4sf (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vcondv2div2df (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vconduv8qiv8qi (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vconduv16qiv16qi (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vconduv4hiv4hi (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vconduv8hiv8hi (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vconduv2siv2si (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vconduv4siv4si (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_vconduv2div2di (rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_createv8qi (rtx, rtx);
+extern rtx gen_aarch64_createv4hi (rtx, rtx);
+extern rtx gen_aarch64_createv2si (rtx, rtx);
+extern rtx gen_aarch64_createdi (rtx, rtx);
+extern rtx gen_aarch64_createdf (rtx, rtx);
+extern rtx gen_aarch64_createv2sf (rtx, rtx);
+extern rtx gen_aarch64_be_checked_get_lanev8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_be_checked_get_lanev16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_be_checked_get_lanev4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_be_checked_get_lanev8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_be_checked_get_lanev2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_be_checked_get_lanev4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_be_checked_get_lanev2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_be_checked_get_lanev2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_be_checked_get_lanev4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_be_checked_get_lanev2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_lanedi (rtx, rtx, rtx);
+extern rtx gen_aarch64_reinterpretv8qiv8qi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv8qiv4hi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv8qiv2si (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv8qiv2sf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv8qidi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv8qidf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4hiv8qi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4hiv4hi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4hiv2si (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4hiv2sf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4hidi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4hidf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2siv8qi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2siv4hi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2siv2si (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2siv2sf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2sidi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2sidf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2sfv8qi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2sfv4hi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2sfv2si (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2sfv2sf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2sfdi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2sfdf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretdiv8qi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretdiv4hi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretdiv2si (rtx, rtx);
+extern rtx gen_aarch64_reinterpretdidi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretdidf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretdiv2sf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretdfv8qi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretdfv4hi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretdfv2si (rtx, rtx);
+extern rtx gen_aarch64_reinterpretdfdi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretdfdf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretdfv2sf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv16qiv16qi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv16qiv8hi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv16qiv4si (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv16qiv2di (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv16qiv4sf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv16qiv2df (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv8hiv16qi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv8hiv8hi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv8hiv4si (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv8hiv2di (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv8hiv4sf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv8hiv2df (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4siv16qi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4siv8hi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4siv4si (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4siv2di (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4siv4sf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4siv2df (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4sfv16qi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4sfv8hi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4sfv4si (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4sfv2di (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4sfv4sf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv4sfv2df (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2div16qi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2div8hi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2div4si (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2div2di (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2div4sf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2div2df (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2dfv16qi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2dfv8hi (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2dfv4si (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2dfv2di (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2dfv4sf (rtx, rtx);
+extern rtx gen_aarch64_reinterpretv2dfv2df (rtx, rtx);
+extern rtx gen_aarch64_combinev8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_combinev4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_combinev2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_combinev2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_combinedi (rtx, rtx, rtx);
+extern rtx gen_aarch64_combinedf (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_combinev8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_combinev4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_combinev2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_combinev2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_combinedi (rtx, rtx, rtx);
+extern rtx gen_aarch64_simd_combinedf (rtx, rtx, rtx);
+extern rtx gen_aarch64_saddl2v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_saddl2v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_saddl2v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddl2v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddl2v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddl2v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_ssubl2v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ssubl2v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ssubl2v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_usubl2v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usubl2v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usubl2v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_saddw2v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_saddw2v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_saddw2v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddw2v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddw2v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_uaddw2v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_ssubw2v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ssubw2v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_ssubw2v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_usubw2v16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usubw2v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_usubw2v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_lanev4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_lanev8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_lanev2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_lanev4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_lanev4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_lanev8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_lanev2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_lanev4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_laneqv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_laneqv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_laneqv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_laneqv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_laneqv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_laneqv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_laneqv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_laneqv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_lanehi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmulh_lanesi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_lanehi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqrdmulh_lanesi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_lanev4hi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_lanev2si (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_lanehi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_lanesi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_laneqv4hi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_laneqv2si (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_laneqhi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal_laneqsi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_lanev4hi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_lanev2si (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_lanehi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_lanesi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_laneqv4hi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_laneqv2si (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_laneqhi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl_laneqsi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2v8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2v4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2v8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2v4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2_lanev8hi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2_lanev4si (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2_laneqv8hi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2_laneqv4si (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2_lanev8hi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2_lanev4si (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2_laneqv8hi (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2_laneqv4si (rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2_nv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlal2_nv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2_nv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmlsl2_nv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_lanev4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_lanev2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_lanehi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_lanesi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_laneqv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull_laneqv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2v8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2v4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2_lanev8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2_lanev4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2_laneqv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2_laneqv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2_nv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_sqdmull2_nv4si (rtx, rtx, rtx);
+extern rtx gen_movoi (rtx, rtx);
+extern rtx gen_movci (rtx, rtx);
+extern rtx gen_movxi (rtx, rtx);
+extern rtx gen_aarch64_ld2v8qi (rtx, rtx);
+extern rtx gen_aarch64_ld2v4hi (rtx, rtx);
+extern rtx gen_aarch64_ld2v2si (rtx, rtx);
+extern rtx gen_aarch64_ld2v2sf (rtx, rtx);
+extern rtx gen_aarch64_ld2di (rtx, rtx);
+extern rtx gen_aarch64_ld2df (rtx, rtx);
+extern rtx gen_aarch64_ld3v8qi (rtx, rtx);
+extern rtx gen_aarch64_ld3v4hi (rtx, rtx);
+extern rtx gen_aarch64_ld3v2si (rtx, rtx);
+extern rtx gen_aarch64_ld3v2sf (rtx, rtx);
+extern rtx gen_aarch64_ld3di (rtx, rtx);
+extern rtx gen_aarch64_ld3df (rtx, rtx);
+extern rtx gen_aarch64_ld4v8qi (rtx, rtx);
+extern rtx gen_aarch64_ld4v4hi (rtx, rtx);
+extern rtx gen_aarch64_ld4v2si (rtx, rtx);
+extern rtx gen_aarch64_ld4v2sf (rtx, rtx);
+extern rtx gen_aarch64_ld4di (rtx, rtx);
+extern rtx gen_aarch64_ld4df (rtx, rtx);
+extern rtx gen_aarch64_ld1v8qi (rtx, rtx);
+extern rtx gen_aarch64_ld1v16qi (rtx, rtx);
+extern rtx gen_aarch64_ld1v4hi (rtx, rtx);
+extern rtx gen_aarch64_ld1v8hi (rtx, rtx);
+extern rtx gen_aarch64_ld1v2si (rtx, rtx);
+extern rtx gen_aarch64_ld1v4si (rtx, rtx);
+extern rtx gen_aarch64_ld1v2di (rtx, rtx);
+extern rtx gen_aarch64_ld1v2sf (rtx, rtx);
+extern rtx gen_aarch64_ld1v4sf (rtx, rtx);
+extern rtx gen_aarch64_ld1v2df (rtx, rtx);
+extern rtx gen_aarch64_ld2v16qi (rtx, rtx);
+extern rtx gen_aarch64_ld3v16qi (rtx, rtx);
+extern rtx gen_aarch64_ld4v16qi (rtx, rtx);
+extern rtx gen_aarch64_ld2v8hi (rtx, rtx);
+extern rtx gen_aarch64_ld3v8hi (rtx, rtx);
+extern rtx gen_aarch64_ld4v8hi (rtx, rtx);
+extern rtx gen_aarch64_ld2v4si (rtx, rtx);
+extern rtx gen_aarch64_ld3v4si (rtx, rtx);
+extern rtx gen_aarch64_ld4v4si (rtx, rtx);
+extern rtx gen_aarch64_ld2v2di (rtx, rtx);
+extern rtx gen_aarch64_ld3v2di (rtx, rtx);
+extern rtx gen_aarch64_ld4v2di (rtx, rtx);
+extern rtx gen_aarch64_ld2v4sf (rtx, rtx);
+extern rtx gen_aarch64_ld3v4sf (rtx, rtx);
+extern rtx gen_aarch64_ld4v4sf (rtx, rtx);
+extern rtx gen_aarch64_ld2v2df (rtx, rtx);
+extern rtx gen_aarch64_ld3v2df (rtx, rtx);
+extern rtx gen_aarch64_ld4v2df (rtx, rtx);
+extern rtx gen_aarch64_get_dregoiv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregoiv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregoiv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregoiv2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregoidi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregoidf (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregciv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregciv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregciv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregciv2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregcidi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregcidf (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregxiv8qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregxiv4hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregxiv2si (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregxiv2sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregxidi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_dregxidf (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregoiv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregciv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregxiv16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregoiv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregciv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregxiv8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregoiv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregciv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregxiv4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregoiv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregciv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregxiv2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregoiv4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregciv4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregxiv4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregoiv2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregciv2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_get_qregxiv2df (rtx, rtx, rtx);
+extern rtx gen_vec_perm_constv8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_vec_perm_constv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_vec_perm_constv4hi (rtx, rtx, rtx, rtx);
+extern rtx gen_vec_perm_constv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_vec_perm_constv2si (rtx, rtx, rtx, rtx);
+extern rtx gen_vec_perm_constv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_vec_perm_constv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_vec_perm_constv2sf (rtx, rtx, rtx, rtx);
+extern rtx gen_vec_perm_constv4sf (rtx, rtx, rtx, rtx);
+extern rtx gen_vec_perm_constv2df (rtx, rtx, rtx, rtx);
+extern rtx gen_vec_permv8qi (rtx, rtx, rtx, rtx);
+extern rtx gen_vec_permv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_im_lane_boundsi (rtx, rtx);
+extern rtx gen_aarch64_st2v8qi (rtx, rtx);
+extern rtx gen_aarch64_st2v4hi (rtx, rtx);
+extern rtx gen_aarch64_st2v2si (rtx, rtx);
+extern rtx gen_aarch64_st2v2sf (rtx, rtx);
+extern rtx gen_aarch64_st2di (rtx, rtx);
+extern rtx gen_aarch64_st2df (rtx, rtx);
+extern rtx gen_aarch64_st3v8qi (rtx, rtx);
+extern rtx gen_aarch64_st3v4hi (rtx, rtx);
+extern rtx gen_aarch64_st3v2si (rtx, rtx);
+extern rtx gen_aarch64_st3v2sf (rtx, rtx);
+extern rtx gen_aarch64_st3di (rtx, rtx);
+extern rtx gen_aarch64_st3df (rtx, rtx);
+extern rtx gen_aarch64_st4v8qi (rtx, rtx);
+extern rtx gen_aarch64_st4v4hi (rtx, rtx);
+extern rtx gen_aarch64_st4v2si (rtx, rtx);
+extern rtx gen_aarch64_st4v2sf (rtx, rtx);
+extern rtx gen_aarch64_st4di (rtx, rtx);
+extern rtx gen_aarch64_st4df (rtx, rtx);
+extern rtx gen_aarch64_st2v16qi (rtx, rtx);
+extern rtx gen_aarch64_st3v16qi (rtx, rtx);
+extern rtx gen_aarch64_st4v16qi (rtx, rtx);
+extern rtx gen_aarch64_st2v8hi (rtx, rtx);
+extern rtx gen_aarch64_st3v8hi (rtx, rtx);
+extern rtx gen_aarch64_st4v8hi (rtx, rtx);
+extern rtx gen_aarch64_st2v4si (rtx, rtx);
+extern rtx gen_aarch64_st3v4si (rtx, rtx);
+extern rtx gen_aarch64_st4v4si (rtx, rtx);
+extern rtx gen_aarch64_st2v2di (rtx, rtx);
+extern rtx gen_aarch64_st3v2di (rtx, rtx);
+extern rtx gen_aarch64_st4v2di (rtx, rtx);
+extern rtx gen_aarch64_st2v4sf (rtx, rtx);
+extern rtx gen_aarch64_st3v4sf (rtx, rtx);
+extern rtx gen_aarch64_st4v4sf (rtx, rtx);
+extern rtx gen_aarch64_st2v2df (rtx, rtx);
+extern rtx gen_aarch64_st3v2df (rtx, rtx);
+extern rtx gen_aarch64_st4v2df (rtx, rtx);
+extern rtx gen_aarch64_st2_lanev16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_st2_lanev8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_st2_lanev4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_st2_lanev2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_st2_lanev4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_st2_lanev2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_st3_lanev16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_st3_lanev8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_st3_lanev4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_st3_lanev2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_st3_lanev4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_st3_lanev2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_st4_lanev16qi (rtx, rtx, rtx);
+extern rtx gen_aarch64_st4_lanev8hi (rtx, rtx, rtx);
+extern rtx gen_aarch64_st4_lanev4si (rtx, rtx, rtx);
+extern rtx gen_aarch64_st4_lanev2di (rtx, rtx, rtx);
+extern rtx gen_aarch64_st4_lanev4sf (rtx, rtx, rtx);
+extern rtx gen_aarch64_st4_lanev2df (rtx, rtx, rtx);
+extern rtx gen_aarch64_st1v8qi (rtx, rtx);
+extern rtx gen_aarch64_st1v16qi (rtx, rtx);
+extern rtx gen_aarch64_st1v4hi (rtx, rtx);
+extern rtx gen_aarch64_st1v8hi (rtx, rtx);
+extern rtx gen_aarch64_st1v2si (rtx, rtx);
+extern rtx gen_aarch64_st1v4si (rtx, rtx);
+extern rtx gen_aarch64_st1v2di (rtx, rtx);
+extern rtx gen_aarch64_st1v2sf (rtx, rtx);
+extern rtx gen_aarch64_st1v4sf (rtx, rtx);
+extern rtx gen_aarch64_st1v2df (rtx, rtx);
+extern rtx gen_aarch64_set_qregoiv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregciv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregxiv16qi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregoiv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregciv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregxiv8hi (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregoiv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregciv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregxiv4si (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregoiv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregciv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregxiv2di (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregoiv4sf (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregciv4sf (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregxiv4sf (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregoiv2df (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregciv2df (rtx, rtx, rtx, rtx);
+extern rtx gen_aarch64_set_qregxiv2df (rtx, rtx, rtx, rtx);
+extern rtx gen_vec_initv8qi (rtx, rtx);
+extern rtx gen_vec_initv16qi (rtx, rtx);
+extern rtx gen_vec_initv4hi (rtx, rtx);
+extern rtx gen_vec_initv8hi (rtx, rtx);
+extern rtx gen_vec_initv2si (rtx, rtx);
+extern rtx gen_vec_initv4si (rtx, rtx);
+extern rtx gen_vec_initv2di (rtx, rtx);
+extern rtx gen_vec_initv2sf (rtx, rtx);
+extern rtx gen_vec_initv4sf (rtx, rtx);
+extern rtx gen_vec_initv2df (rtx, rtx);
+extern rtx gen_vec_extractv8qi (rtx, rtx, rtx);
+extern rtx gen_vec_extractv16qi (rtx, rtx, rtx);
+extern rtx gen_vec_extractv4hi (rtx, rtx, rtx);
+extern rtx gen_vec_extractv8hi (rtx, rtx, rtx);
+extern rtx gen_vec_extractv2si (rtx, rtx, rtx);
+extern rtx gen_vec_extractv4si (rtx, rtx, rtx);
+extern rtx gen_vec_extractv2di (rtx, rtx, rtx);
+extern rtx gen_vec_extractv2sf (rtx, rtx, rtx);
+extern rtx gen_vec_extractv4sf (rtx, rtx, rtx);
+extern rtx gen_vec_extractv2df (rtx, rtx, rtx);
+extern rtx gen_atomic_compare_and_swapqi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_compare_and_swaphi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_compare_and_swapsi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_atomic_compare_and_swapdi (rtx, rtx, rtx, rtx, rtx, rtx, rtx, rtx);
+extern rtx gen_mem_thread_fence (rtx);
+extern rtx gen_dmb (rtx);
#endif /* GCC_INSN_FLAGS_H */
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-modes.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-modes.h
index e0d2258..e6b5b0a 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-modes.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/insn-modes.h
@@ -22,10 +22,10 @@ enum machine_mode
SImode, /* machmode.def:189 */
DImode, /* machmode.def:190 */
TImode, /* machmode.def:191 */
- EImode, /* config/aarch64/aarch64-modes.def:40 */
- OImode, /* config/aarch64/aarch64-modes.def:36 */
- CImode, /* config/aarch64/aarch64-modes.def:41 */
- XImode, /* config/aarch64/aarch64-modes.def:42 */
+ EImode, /* config/aarch64/aarch64-modes.def:41 */
+ OImode, /* config/aarch64/aarch64-modes.def:37 */
+ CImode, /* config/aarch64/aarch64-modes.def:42 */
+ XImode, /* config/aarch64/aarch64-modes.def:43 */
QQmode, /* machmode.def:210 */
HQmode, /* machmode.def:211 */
SQmode, /* machmode.def:212 */
@@ -46,7 +46,7 @@ enum machine_mode
UTAmode, /* machmode.def:230 */
SFmode, /* machmode.def:202 */
DFmode, /* machmode.def:203 */
- TFmode, /* config/aarch64/aarch64-modes.def:55 */
+ TFmode, /* config/aarch64/aarch64-modes.def:56 */
SDmode, /* machmode.def:242 */
DDmode, /* machmode.def:243 */
TDmode, /* machmode.def:244 */
@@ -69,32 +69,33 @@ enum machine_mode
V8HImode, /* config/aarch64/aarch64-modes.def:31 */
V4SImode, /* config/aarch64/aarch64-modes.def:31 */
V2DImode, /* config/aarch64/aarch64-modes.def:31 */
- V32QImode, /* config/aarch64/aarch64-modes.def:45 */
- V16HImode, /* config/aarch64/aarch64-modes.def:45 */
- V8SImode, /* config/aarch64/aarch64-modes.def:45 */
- V4DImode, /* config/aarch64/aarch64-modes.def:45 */
- V2TImode, /* config/aarch64/aarch64-modes.def:45 */
- V48QImode, /* config/aarch64/aarch64-modes.def:48 */
- V24HImode, /* config/aarch64/aarch64-modes.def:48 */
- V12SImode, /* config/aarch64/aarch64-modes.def:48 */
- V6DImode, /* config/aarch64/aarch64-modes.def:48 */
- V3TImode, /* config/aarch64/aarch64-modes.def:48 */
- V2EImode, /* config/aarch64/aarch64-modes.def:48 */
- V64QImode, /* config/aarch64/aarch64-modes.def:51 */
- V32HImode, /* config/aarch64/aarch64-modes.def:51 */
- V16SImode, /* config/aarch64/aarch64-modes.def:51 */
- V8DImode, /* config/aarch64/aarch64-modes.def:51 */
- V4TImode, /* config/aarch64/aarch64-modes.def:51 */
- V2OImode, /* config/aarch64/aarch64-modes.def:51 */
+ V32QImode, /* config/aarch64/aarch64-modes.def:46 */
+ V16HImode, /* config/aarch64/aarch64-modes.def:46 */
+ V8SImode, /* config/aarch64/aarch64-modes.def:46 */
+ V4DImode, /* config/aarch64/aarch64-modes.def:46 */
+ V2TImode, /* config/aarch64/aarch64-modes.def:46 */
+ V48QImode, /* config/aarch64/aarch64-modes.def:49 */
+ V24HImode, /* config/aarch64/aarch64-modes.def:49 */
+ V12SImode, /* config/aarch64/aarch64-modes.def:49 */
+ V6DImode, /* config/aarch64/aarch64-modes.def:49 */
+ V3TImode, /* config/aarch64/aarch64-modes.def:49 */
+ V2EImode, /* config/aarch64/aarch64-modes.def:49 */
+ V64QImode, /* config/aarch64/aarch64-modes.def:52 */
+ V32HImode, /* config/aarch64/aarch64-modes.def:52 */
+ V16SImode, /* config/aarch64/aarch64-modes.def:52 */
+ V8DImode, /* config/aarch64/aarch64-modes.def:52 */
+ V4TImode, /* config/aarch64/aarch64-modes.def:52 */
+ V2OImode, /* config/aarch64/aarch64-modes.def:52 */
V2SFmode, /* config/aarch64/aarch64-modes.def:32 */
+ V1DFmode, /* config/aarch64/aarch64-modes.def:34 */
V4SFmode, /* config/aarch64/aarch64-modes.def:33 */
V2DFmode, /* config/aarch64/aarch64-modes.def:33 */
- V8SFmode, /* config/aarch64/aarch64-modes.def:46 */
- V4DFmode, /* config/aarch64/aarch64-modes.def:46 */
- V12SFmode, /* config/aarch64/aarch64-modes.def:49 */
- V6DFmode, /* config/aarch64/aarch64-modes.def:49 */
- V16SFmode, /* config/aarch64/aarch64-modes.def:52 */
- V8DFmode, /* config/aarch64/aarch64-modes.def:52 */
+ V8SFmode, /* config/aarch64/aarch64-modes.def:47 */
+ V4DFmode, /* config/aarch64/aarch64-modes.def:47 */
+ V12SFmode, /* config/aarch64/aarch64-modes.def:50 */
+ V6DFmode, /* config/aarch64/aarch64-modes.def:50 */
+ V16SFmode, /* config/aarch64/aarch64-modes.def:53 */
+ V8DFmode, /* config/aarch64/aarch64-modes.def:53 */
MAX_MACHINE_MODE,
MIN_MODE_RANDOM = VOIDmode,
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/ipa-prop.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/ipa-prop.h
index 8fdd92c..70185b2 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/ipa-prop.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/ipa-prop.h
@@ -585,6 +585,7 @@ tree ipa_get_indirect_edge_target (struct cgraph_edge *ie,
struct cgraph_edge *ipa_make_edge_direct_to_target (struct cgraph_edge *, tree);
tree ipa_binfo_from_known_type_jfunc (struct ipa_jump_func *);
tree ipa_intraprocedural_devirtualization (gimple);
+tree ipa_impossible_devirt_target (struct cgraph_edge *, tree);
/* Functions related to both. */
void ipa_analyze_node (struct cgraph_node *);
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/langhooks.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/langhooks.h
index c848b0c..b5997ee 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/langhooks.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/langhooks.h
@@ -219,12 +219,16 @@ struct lang_hooks_for_decls
/* Similarly, except use an assignment operator instead. */
tree (*omp_clause_assign_op) (tree clause, tree dst, tree src);
+ /* Build and return code for a constructor of DST that sets it to
+ SRC + ADD. */
+ tree (*omp_clause_linear_ctor) (tree clause, tree dst, tree src, tree add);
+
/* Build and return code destructing DECL. Return NULL if nothing
to be done. */
tree (*omp_clause_dtor) (tree clause, tree decl);
/* Do language specific checking on an implicitly determined clause. */
- void (*omp_finish_clause) (tree clause);
+ void (*omp_finish_clause) (tree clause, gimple_seq *pre_p);
};
/* Language hooks related to LTO serialization. */
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/plugin-version.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/plugin-version.h
index 653de0e..591112e 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/plugin-version.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/plugin-version.h
@@ -6,7 +6,7 @@
#define GCCPLUGIN_VERSION (GCCPLUGIN_VERSION_MAJOR*1000 + GCCPLUGIN_VERSION_MINOR)
static char basever[] = "4.9.1";
-static char datestamp[] = "20140529";
+static char datestamp[] = "20140710";
static char devphase[] = "prerelease";
static char revision[] = "";
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tm-preds.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tm-preds.h
index e31cbef..127ca26 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tm-preds.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tm-preds.h
@@ -1,5 +1,5 @@
/* Generated automatically by the program 'build/genpreds'
- from the machine description file '/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.06/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.06/gcc/config/aarch64/aarch64.md'. */
+ from the machine description file '/mnt/jenkins/workspace/bernhard.rosenkranzer_toolchain64-4.9-2014.07/build/objdir/../build/../gcc/gcc-linaro-4.9-2014.07/gcc/config/aarch64/aarch64.md'. */
#ifndef GCC_TM_PREDS_H
#define GCC_TM_PREDS_H
@@ -22,6 +22,7 @@ extern int indirect_operand (rtx, enum machine_mode);
extern int ordered_comparison_operator (rtx, enum machine_mode);
extern int comparison_operator (rtx, enum machine_mode);
extern int cc_register (rtx, enum machine_mode);
+extern int aarch64_call_insn_operand (rtx, enum machine_mode);
extern int aarch64_simd_register (rtx, enum machine_mode);
extern int aarch64_reg_or_zero (rtx, enum machine_mode);
extern int aarch64_reg_or_fp_zero (rtx, enum machine_mode);
@@ -67,6 +68,7 @@ enum constraint_num
{
CONSTRAINT__UNKNOWN = 0,
CONSTRAINT_k,
+ CONSTRAINT_Ucs,
CONSTRAINT_w,
CONSTRAINT_x,
CONSTRAINT_I,
@@ -81,6 +83,7 @@ enum constraint_num
CONSTRAINT_Ush,
CONSTRAINT_Uss,
CONSTRAINT_Usd,
+ CONSTRAINT_Usf,
CONSTRAINT_UsM,
CONSTRAINT_Ui1,
CONSTRAINT_Ui3,
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree-core.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree-core.h
index 1719c7e..40443d4 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree-core.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree-core.h
@@ -987,7 +987,7 @@ struct GTY(()) tree_base {
SSA_NAME_IN_FREELIST in
SSA_NAME
- VAR_DECL_NONALIASED in
+ DECL_NONALIASED in
VAR_DECL
deprecated_flag:
@@ -1127,6 +1127,11 @@ enum omp_clause_map_kind
array sections. OMP_CLAUSE_SIZE for these is not the pointer size,
which is implicitly POINTER_SIZE / BITS_PER_UNIT, but the bias. */
OMP_CLAUSE_MAP_POINTER,
+ /* Also internal, behaves like OMP_CLAUS_MAP_TO, but additionally any
+ OMP_CLAUSE_MAP_POINTER records consecutive after it which have addresses
+ falling into that range will not be ignored if OMP_CLAUSE_MAP_TO_PSET
+ wasn't mapped already. */
+ OMP_CLAUSE_MAP_TO_PSET,
OMP_CLAUSE_MAP_LAST
};
diff --git a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree.h b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree.h
index 90e5e27..d73bc52 100644
--- a/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree.h
+++ b/lib/gcc/aarch64-linux-android/4.9.1/plugin/include/tree.h
@@ -1327,9 +1327,17 @@ extern void protected_set_expr_location (tree, location_t);
#define OMP_CLAUSE_LINEAR_VARIABLE_STRIDE(NODE) \
TREE_PROTECTED (OMP_CLAUSE_SUBCODE_CHECK (NODE, OMP_CLAUSE_LINEAR))
+/* True if a LINEAR clause is for an array or allocatable variable that
+ needs special handling by the frontend. */
+#define OMP_CLAUSE_LINEAR_ARRAY(NODE) \
+ (OMP_CLAUSE_SUBCODE_CHECK (NODE, OMP_CLAUSE_LINEAR)->base.deprecated_flag)
+
#define OMP_CLAUSE_LINEAR_STEP(NODE) \
OMP_CLAUSE_OPERAND (OMP_CLAUSE_SUBCODE_CHECK (NODE, OMP_CLAUSE_LINEAR), 1)
+#define OMP_CLAUSE_LINEAR_STMT(NODE) \
+ OMP_CLAUSE_OPERAND (OMP_CLAUSE_SUBCODE_CHECK (NODE, OMP_CLAUSE_LINEAR), 2)
+
#define OMP_CLAUSE_LINEAR_GIMPLE_SEQ(NODE) \
(OMP_CLAUSE_CHECK (NODE))->omp_clause.gimple_reduction_init