1/*===---- arm_neon.h - ARM Neon intrinsics ---------------------------------===
2 *
3 * Permission is hereby granted, free of charge, to any person obtaining a copy
4 * of this software and associated documentation files (the "Software"), to deal
5 * in the Software without restriction, including without limitation the rights
6 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
7 * copies of the Software, and to permit persons to whom the Software is
8 * furnished to do so, subject to the following conditions:
9 *
10 * The above copyright notice and this permission notice shall be included in
11 * all copies or substantial portions of the Software.
12 *
13 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
14 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
15 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
16 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
17 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
18 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
19 * THE SOFTWARE.
20 *
21 *===-----------------------------------------------------------------------===
22 */
23
24#ifndef __ARM_NEON_H
25#define __ARM_NEON_H
26
27#if !defined(__ARM_NEON)
28#error "NEON support not enabled"
29#endif
30
31#include <stdint.h>
32
33typedef float float32_t;
34typedef __fp16 float16_t;
35#ifdef __aarch64__
36typedef double float64_t;
37#endif
38
39#ifdef __aarch64__
40typedef uint8_t poly8_t;
41typedef uint16_t poly16_t;
42typedef uint64_t poly64_t;
43typedef __uint128_t poly128_t;
44#else
45typedef int8_t poly8_t;
46typedef int16_t poly16_t;
47#endif
48typedef __attribute__((neon_vector_type(8))) int8_t int8x8_t;
49typedef __attribute__((neon_vector_type(16))) int8_t int8x16_t;
50typedef __attribute__((neon_vector_type(4))) int16_t int16x4_t;
51typedef __attribute__((neon_vector_type(8))) int16_t int16x8_t;
52typedef __attribute__((neon_vector_type(2))) int32_t int32x2_t;
53typedef __attribute__((neon_vector_type(4))) int32_t int32x4_t;
54typedef __attribute__((neon_vector_type(1))) int64_t int64x1_t;
55typedef __attribute__((neon_vector_type(2))) int64_t int64x2_t;
56typedef __attribute__((neon_vector_type(8))) uint8_t uint8x8_t;
57typedef __attribute__((neon_vector_type(16))) uint8_t uint8x16_t;
58typedef __attribute__((neon_vector_type(4))) uint16_t uint16x4_t;
59typedef __attribute__((neon_vector_type(8))) uint16_t uint16x8_t;
60typedef __attribute__((neon_vector_type(2))) uint32_t uint32x2_t;
61typedef __attribute__((neon_vector_type(4))) uint32_t uint32x4_t;
62typedef __attribute__((neon_vector_type(1))) uint64_t uint64x1_t;
63typedef __attribute__((neon_vector_type(2))) uint64_t uint64x2_t;
64typedef __attribute__((neon_vector_type(4))) float16_t float16x4_t;
65typedef __attribute__((neon_vector_type(8))) float16_t float16x8_t;
66typedef __attribute__((neon_vector_type(2))) float32_t float32x2_t;
67typedef __attribute__((neon_vector_type(4))) float32_t float32x4_t;
68#ifdef __aarch64__
69typedef __attribute__((neon_vector_type(1))) float64_t float64x1_t;
70typedef __attribute__((neon_vector_type(2))) float64_t float64x2_t;
71#endif
72typedef __attribute__((neon_polyvector_type(8))) poly8_t poly8x8_t;
73typedef __attribute__((neon_polyvector_type(16))) poly8_t poly8x16_t;
74typedef __attribute__((neon_polyvector_type(4))) poly16_t poly16x4_t;
75typedef __attribute__((neon_polyvector_type(8))) poly16_t poly16x8_t;
76#ifdef __aarch64__
77typedef __attribute__((neon_polyvector_type(1))) poly64_t poly64x1_t;
78typedef __attribute__((neon_polyvector_type(2))) poly64_t poly64x2_t;
79#endif
80
81typedef struct int8x8x2_t {
82  int8x8_t val[2];
83} int8x8x2_t;
84
85typedef struct int8x16x2_t {
86  int8x16_t val[2];
87} int8x16x2_t;
88
89typedef struct int16x4x2_t {
90  int16x4_t val[2];
91} int16x4x2_t;
92
93typedef struct int16x8x2_t {
94  int16x8_t val[2];
95} int16x8x2_t;
96
97typedef struct int32x2x2_t {
98  int32x2_t val[2];
99} int32x2x2_t;
100
101typedef struct int32x4x2_t {
102  int32x4_t val[2];
103} int32x4x2_t;
104
105typedef struct int64x1x2_t {
106  int64x1_t val[2];
107} int64x1x2_t;
108
109typedef struct int64x2x2_t {
110  int64x2_t val[2];
111} int64x2x2_t;
112
113typedef struct uint8x8x2_t {
114  uint8x8_t val[2];
115} uint8x8x2_t;
116
117typedef struct uint8x16x2_t {
118  uint8x16_t val[2];
119} uint8x16x2_t;
120
121typedef struct uint16x4x2_t {
122  uint16x4_t val[2];
123} uint16x4x2_t;
124
125typedef struct uint16x8x2_t {
126  uint16x8_t val[2];
127} uint16x8x2_t;
128
129typedef struct uint32x2x2_t {
130  uint32x2_t val[2];
131} uint32x2x2_t;
132
133typedef struct uint32x4x2_t {
134  uint32x4_t val[2];
135} uint32x4x2_t;
136
137typedef struct uint64x1x2_t {
138  uint64x1_t val[2];
139} uint64x1x2_t;
140
141typedef struct uint64x2x2_t {
142  uint64x2_t val[2];
143} uint64x2x2_t;
144
145typedef struct float16x4x2_t {
146  float16x4_t val[2];
147} float16x4x2_t;
148
149typedef struct float16x8x2_t {
150  float16x8_t val[2];
151} float16x8x2_t;
152
153typedef struct float32x2x2_t {
154  float32x2_t val[2];
155} float32x2x2_t;
156
157typedef struct float32x4x2_t {
158  float32x4_t val[2];
159} float32x4x2_t;
160
161#ifdef __aarch64__
162typedef struct float64x1x2_t {
163  float64x1_t val[2];
164} float64x1x2_t;
165
166typedef struct float64x2x2_t {
167  float64x2_t val[2];
168} float64x2x2_t;
169
170#endif
171typedef struct poly8x8x2_t {
172  poly8x8_t val[2];
173} poly8x8x2_t;
174
175typedef struct poly8x16x2_t {
176  poly8x16_t val[2];
177} poly8x16x2_t;
178
179typedef struct poly16x4x2_t {
180  poly16x4_t val[2];
181} poly16x4x2_t;
182
183typedef struct poly16x8x2_t {
184  poly16x8_t val[2];
185} poly16x8x2_t;
186
187#ifdef __aarch64__
188typedef struct poly64x1x2_t {
189  poly64x1_t val[2];
190} poly64x1x2_t;
191
192typedef struct poly64x2x2_t {
193  poly64x2_t val[2];
194} poly64x2x2_t;
195
196#endif
197typedef struct int8x8x3_t {
198  int8x8_t val[3];
199} int8x8x3_t;
200
201typedef struct int8x16x3_t {
202  int8x16_t val[3];
203} int8x16x3_t;
204
205typedef struct int16x4x3_t {
206  int16x4_t val[3];
207} int16x4x3_t;
208
209typedef struct int16x8x3_t {
210  int16x8_t val[3];
211} int16x8x3_t;
212
213typedef struct int32x2x3_t {
214  int32x2_t val[3];
215} int32x2x3_t;
216
217typedef struct int32x4x3_t {
218  int32x4_t val[3];
219} int32x4x3_t;
220
221typedef struct int64x1x3_t {
222  int64x1_t val[3];
223} int64x1x3_t;
224
225typedef struct int64x2x3_t {
226  int64x2_t val[3];
227} int64x2x3_t;
228
229typedef struct uint8x8x3_t {
230  uint8x8_t val[3];
231} uint8x8x3_t;
232
233typedef struct uint8x16x3_t {
234  uint8x16_t val[3];
235} uint8x16x3_t;
236
237typedef struct uint16x4x3_t {
238  uint16x4_t val[3];
239} uint16x4x3_t;
240
241typedef struct uint16x8x3_t {
242  uint16x8_t val[3];
243} uint16x8x3_t;
244
245typedef struct uint32x2x3_t {
246  uint32x2_t val[3];
247} uint32x2x3_t;
248
249typedef struct uint32x4x3_t {
250  uint32x4_t val[3];
251} uint32x4x3_t;
252
253typedef struct uint64x1x3_t {
254  uint64x1_t val[3];
255} uint64x1x3_t;
256
257typedef struct uint64x2x3_t {
258  uint64x2_t val[3];
259} uint64x2x3_t;
260
261typedef struct float16x4x3_t {
262  float16x4_t val[3];
263} float16x4x3_t;
264
265typedef struct float16x8x3_t {
266  float16x8_t val[3];
267} float16x8x3_t;
268
269typedef struct float32x2x3_t {
270  float32x2_t val[3];
271} float32x2x3_t;
272
273typedef struct float32x4x3_t {
274  float32x4_t val[3];
275} float32x4x3_t;
276
277#ifdef __aarch64__
278typedef struct float64x1x3_t {
279  float64x1_t val[3];
280} float64x1x3_t;
281
282typedef struct float64x2x3_t {
283  float64x2_t val[3];
284} float64x2x3_t;
285
286#endif
287typedef struct poly8x8x3_t {
288  poly8x8_t val[3];
289} poly8x8x3_t;
290
291typedef struct poly8x16x3_t {
292  poly8x16_t val[3];
293} poly8x16x3_t;
294
295typedef struct poly16x4x3_t {
296  poly16x4_t val[3];
297} poly16x4x3_t;
298
299typedef struct poly16x8x3_t {
300  poly16x8_t val[3];
301} poly16x8x3_t;
302
303#ifdef __aarch64__
304typedef struct poly64x1x3_t {
305  poly64x1_t val[3];
306} poly64x1x3_t;
307
308typedef struct poly64x2x3_t {
309  poly64x2_t val[3];
310} poly64x2x3_t;
311
312#endif
313typedef struct int8x8x4_t {
314  int8x8_t val[4];
315} int8x8x4_t;
316
317typedef struct int8x16x4_t {
318  int8x16_t val[4];
319} int8x16x4_t;
320
321typedef struct int16x4x4_t {
322  int16x4_t val[4];
323} int16x4x4_t;
324
325typedef struct int16x8x4_t {
326  int16x8_t val[4];
327} int16x8x4_t;
328
329typedef struct int32x2x4_t {
330  int32x2_t val[4];
331} int32x2x4_t;
332
333typedef struct int32x4x4_t {
334  int32x4_t val[4];
335} int32x4x4_t;
336
337typedef struct int64x1x4_t {
338  int64x1_t val[4];
339} int64x1x4_t;
340
341typedef struct int64x2x4_t {
342  int64x2_t val[4];
343} int64x2x4_t;
344
345typedef struct uint8x8x4_t {
346  uint8x8_t val[4];
347} uint8x8x4_t;
348
349typedef struct uint8x16x4_t {
350  uint8x16_t val[4];
351} uint8x16x4_t;
352
353typedef struct uint16x4x4_t {
354  uint16x4_t val[4];
355} uint16x4x4_t;
356
357typedef struct uint16x8x4_t {
358  uint16x8_t val[4];
359} uint16x8x4_t;
360
361typedef struct uint32x2x4_t {
362  uint32x2_t val[4];
363} uint32x2x4_t;
364
365typedef struct uint32x4x4_t {
366  uint32x4_t val[4];
367} uint32x4x4_t;
368
369typedef struct uint64x1x4_t {
370  uint64x1_t val[4];
371} uint64x1x4_t;
372
373typedef struct uint64x2x4_t {
374  uint64x2_t val[4];
375} uint64x2x4_t;
376
377typedef struct float16x4x4_t {
378  float16x4_t val[4];
379} float16x4x4_t;
380
381typedef struct float16x8x4_t {
382  float16x8_t val[4];
383} float16x8x4_t;
384
385typedef struct float32x2x4_t {
386  float32x2_t val[4];
387} float32x2x4_t;
388
389typedef struct float32x4x4_t {
390  float32x4_t val[4];
391} float32x4x4_t;
392
393#ifdef __aarch64__
394typedef struct float64x1x4_t {
395  float64x1_t val[4];
396} float64x1x4_t;
397
398typedef struct float64x2x4_t {
399  float64x2_t val[4];
400} float64x2x4_t;
401
402#endif
403typedef struct poly8x8x4_t {
404  poly8x8_t val[4];
405} poly8x8x4_t;
406
407typedef struct poly8x16x4_t {
408  poly8x16_t val[4];
409} poly8x16x4_t;
410
411typedef struct poly16x4x4_t {
412  poly16x4_t val[4];
413} poly16x4x4_t;
414
415typedef struct poly16x8x4_t {
416  poly16x8_t val[4];
417} poly16x8x4_t;
418
419#ifdef __aarch64__
420typedef struct poly64x1x4_t {
421  poly64x1_t val[4];
422} poly64x1x4_t;
423
424typedef struct poly64x2x4_t {
425  poly64x2_t val[4];
426} poly64x2x4_t;
427
428#endif
429
430#define __ai static inline __attribute__((__always_inline__, __nodebug__))
431
432#ifdef __LITTLE_ENDIAN__
433__ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
434  uint8x16_t __ret;
435  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
436  return __ret;
437}
438#else
439__ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
440  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
441  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
442  uint8x16_t __ret;
443  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
444  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
445  return __ret;
446}
447__ai uint8x16_t __noswap_vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
448  uint8x16_t __ret;
449  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
450  return __ret;
451}
452#endif
453
454#ifdef __LITTLE_ENDIAN__
455__ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
456  uint32x4_t __ret;
457  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
458  return __ret;
459}
460#else
461__ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
462  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
463  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
464  uint32x4_t __ret;
465  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
466  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
467  return __ret;
468}
469__ai uint32x4_t __noswap_vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
470  uint32x4_t __ret;
471  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
472  return __ret;
473}
474#endif
475
476#ifdef __LITTLE_ENDIAN__
477__ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
478  uint16x8_t __ret;
479  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
480  return __ret;
481}
482#else
483__ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
484  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
485  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
486  uint16x8_t __ret;
487  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
488  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
489  return __ret;
490}
491__ai uint16x8_t __noswap_vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
492  uint16x8_t __ret;
493  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
494  return __ret;
495}
496#endif
497
498#ifdef __LITTLE_ENDIAN__
499__ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
500  int8x16_t __ret;
501  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
502  return __ret;
503}
504#else
505__ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
506  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
507  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
508  int8x16_t __ret;
509  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
510  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
511  return __ret;
512}
513__ai int8x16_t __noswap_vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
514  int8x16_t __ret;
515  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
516  return __ret;
517}
518#endif
519
520#ifdef __LITTLE_ENDIAN__
521__ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1) {
522  float32x4_t __ret;
523  __ret = (float32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
524  return __ret;
525}
526#else
527__ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1) {
528  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
529  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
530  float32x4_t __ret;
531  __ret = (float32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
532  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
533  return __ret;
534}
535#endif
536
537#ifdef __LITTLE_ENDIAN__
538__ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
539  int32x4_t __ret;
540  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
541  return __ret;
542}
543#else
544__ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
545  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
546  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
547  int32x4_t __ret;
548  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
549  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
550  return __ret;
551}
552__ai int32x4_t __noswap_vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
553  int32x4_t __ret;
554  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
555  return __ret;
556}
557#endif
558
559#ifdef __LITTLE_ENDIAN__
560__ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
561  int16x8_t __ret;
562  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
563  return __ret;
564}
565#else
566__ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
567  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
568  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
569  int16x8_t __ret;
570  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
571  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
572  return __ret;
573}
574__ai int16x8_t __noswap_vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
575  int16x8_t __ret;
576  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
577  return __ret;
578}
579#endif
580
581#ifdef __LITTLE_ENDIAN__
582__ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
583  uint8x8_t __ret;
584  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
585  return __ret;
586}
587#else
588__ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
589  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
590  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
591  uint8x8_t __ret;
592  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
593  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
594  return __ret;
595}
596__ai uint8x8_t __noswap_vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
597  uint8x8_t __ret;
598  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
599  return __ret;
600}
601#endif
602
603#ifdef __LITTLE_ENDIAN__
604__ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
605  uint32x2_t __ret;
606  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
607  return __ret;
608}
609#else
610__ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
611  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
612  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
613  uint32x2_t __ret;
614  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
615  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
616  return __ret;
617}
618__ai uint32x2_t __noswap_vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
619  uint32x2_t __ret;
620  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
621  return __ret;
622}
623#endif
624
625#ifdef __LITTLE_ENDIAN__
626__ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
627  uint16x4_t __ret;
628  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
629  return __ret;
630}
631#else
632__ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
633  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
634  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
635  uint16x4_t __ret;
636  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
637  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
638  return __ret;
639}
640__ai uint16x4_t __noswap_vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
641  uint16x4_t __ret;
642  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
643  return __ret;
644}
645#endif
646
647#ifdef __LITTLE_ENDIAN__
648__ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1) {
649  int8x8_t __ret;
650  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
651  return __ret;
652}
653#else
654__ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1) {
655  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
656  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
657  int8x8_t __ret;
658  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
659  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
660  return __ret;
661}
662__ai int8x8_t __noswap_vabd_s8(int8x8_t __p0, int8x8_t __p1) {
663  int8x8_t __ret;
664  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
665  return __ret;
666}
667#endif
668
669#ifdef __LITTLE_ENDIAN__
670__ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1) {
671  float32x2_t __ret;
672  __ret = (float32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
673  return __ret;
674}
675#else
676__ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1) {
677  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
678  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
679  float32x2_t __ret;
680  __ret = (float32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
681  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
682  return __ret;
683}
684#endif
685
686#ifdef __LITTLE_ENDIAN__
687__ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1) {
688  int32x2_t __ret;
689  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
690  return __ret;
691}
692#else
693__ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1) {
694  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
695  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
696  int32x2_t __ret;
697  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
698  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
699  return __ret;
700}
701__ai int32x2_t __noswap_vabd_s32(int32x2_t __p0, int32x2_t __p1) {
702  int32x2_t __ret;
703  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
704  return __ret;
705}
706#endif
707
708#ifdef __LITTLE_ENDIAN__
709__ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1) {
710  int16x4_t __ret;
711  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
712  return __ret;
713}
714#else
715__ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1) {
716  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
717  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
718  int16x4_t __ret;
719  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
720  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
721  return __ret;
722}
723__ai int16x4_t __noswap_vabd_s16(int16x4_t __p0, int16x4_t __p1) {
724  int16x4_t __ret;
725  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
726  return __ret;
727}
728#endif
729
730#ifdef __LITTLE_ENDIAN__
731__ai int8x16_t vabsq_s8(int8x16_t __p0) {
732  int8x16_t __ret;
733  __ret = (int8x16_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 32);
734  return __ret;
735}
736#else
737__ai int8x16_t vabsq_s8(int8x16_t __p0) {
738  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
739  int8x16_t __ret;
740  __ret = (int8x16_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 32);
741  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
742  return __ret;
743}
744#endif
745
746#ifdef __LITTLE_ENDIAN__
747__ai float32x4_t vabsq_f32(float32x4_t __p0) {
748  float32x4_t __ret;
749  __ret = (float32x4_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 41);
750  return __ret;
751}
752#else
753__ai float32x4_t vabsq_f32(float32x4_t __p0) {
754  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
755  float32x4_t __ret;
756  __ret = (float32x4_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 41);
757  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
758  return __ret;
759}
760#endif
761
762#ifdef __LITTLE_ENDIAN__
763__ai int32x4_t vabsq_s32(int32x4_t __p0) {
764  int32x4_t __ret;
765  __ret = (int32x4_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 34);
766  return __ret;
767}
768#else
769__ai int32x4_t vabsq_s32(int32x4_t __p0) {
770  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
771  int32x4_t __ret;
772  __ret = (int32x4_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 34);
773  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
774  return __ret;
775}
776#endif
777
778#ifdef __LITTLE_ENDIAN__
779__ai int16x8_t vabsq_s16(int16x8_t __p0) {
780  int16x8_t __ret;
781  __ret = (int16x8_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 33);
782  return __ret;
783}
784#else
785__ai int16x8_t vabsq_s16(int16x8_t __p0) {
786  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
787  int16x8_t __ret;
788  __ret = (int16x8_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 33);
789  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
790  return __ret;
791}
792#endif
793
794#ifdef __LITTLE_ENDIAN__
795__ai int8x8_t vabs_s8(int8x8_t __p0) {
796  int8x8_t __ret;
797  __ret = (int8x8_t) __builtin_neon_vabs_v((int8x8_t)__p0, 0);
798  return __ret;
799}
800#else
801__ai int8x8_t vabs_s8(int8x8_t __p0) {
802  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
803  int8x8_t __ret;
804  __ret = (int8x8_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 0);
805  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
806  return __ret;
807}
808#endif
809
810#ifdef __LITTLE_ENDIAN__
811__ai float32x2_t vabs_f32(float32x2_t __p0) {
812  float32x2_t __ret;
813  __ret = (float32x2_t) __builtin_neon_vabs_v((int8x8_t)__p0, 9);
814  return __ret;
815}
816#else
817__ai float32x2_t vabs_f32(float32x2_t __p0) {
818  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
819  float32x2_t __ret;
820  __ret = (float32x2_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 9);
821  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
822  return __ret;
823}
824#endif
825
826#ifdef __LITTLE_ENDIAN__
827__ai int32x2_t vabs_s32(int32x2_t __p0) {
828  int32x2_t __ret;
829  __ret = (int32x2_t) __builtin_neon_vabs_v((int8x8_t)__p0, 2);
830  return __ret;
831}
832#else
833__ai int32x2_t vabs_s32(int32x2_t __p0) {
834  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
835  int32x2_t __ret;
836  __ret = (int32x2_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 2);
837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
838  return __ret;
839}
840#endif
841
842#ifdef __LITTLE_ENDIAN__
843__ai int16x4_t vabs_s16(int16x4_t __p0) {
844  int16x4_t __ret;
845  __ret = (int16x4_t) __builtin_neon_vabs_v((int8x8_t)__p0, 1);
846  return __ret;
847}
848#else
849__ai int16x4_t vabs_s16(int16x4_t __p0) {
850  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
851  int16x4_t __ret;
852  __ret = (int16x4_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 1);
853  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
854  return __ret;
855}
856#endif
857
858#ifdef __LITTLE_ENDIAN__
859__ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
860  uint8x16_t __ret;
861  __ret = __p0 + __p1;
862  return __ret;
863}
864#else
865__ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
866  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
867  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
868  uint8x16_t __ret;
869  __ret = __rev0 + __rev1;
870  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
871  return __ret;
872}
873#endif
874
875#ifdef __LITTLE_ENDIAN__
876__ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
877  uint32x4_t __ret;
878  __ret = __p0 + __p1;
879  return __ret;
880}
881#else
882__ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
883  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
884  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
885  uint32x4_t __ret;
886  __ret = __rev0 + __rev1;
887  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
888  return __ret;
889}
890#endif
891
892#ifdef __LITTLE_ENDIAN__
893__ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
894  uint64x2_t __ret;
895  __ret = __p0 + __p1;
896  return __ret;
897}
898#else
899__ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
900  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
901  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
902  uint64x2_t __ret;
903  __ret = __rev0 + __rev1;
904  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
905  return __ret;
906}
907#endif
908
909#ifdef __LITTLE_ENDIAN__
910__ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
911  uint16x8_t __ret;
912  __ret = __p0 + __p1;
913  return __ret;
914}
915#else
916__ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
917  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
918  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
919  uint16x8_t __ret;
920  __ret = __rev0 + __rev1;
921  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
922  return __ret;
923}
924#endif
925
926#ifdef __LITTLE_ENDIAN__
927__ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1) {
928  int8x16_t __ret;
929  __ret = __p0 + __p1;
930  return __ret;
931}
932#else
933__ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1) {
934  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
935  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
936  int8x16_t __ret;
937  __ret = __rev0 + __rev1;
938  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
939  return __ret;
940}
941#endif
942
943#ifdef __LITTLE_ENDIAN__
944__ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1) {
945  float32x4_t __ret;
946  __ret = __p0 + __p1;
947  return __ret;
948}
949#else
950__ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1) {
951  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
952  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
953  float32x4_t __ret;
954  __ret = __rev0 + __rev1;
955  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
956  return __ret;
957}
958#endif
959
960#ifdef __LITTLE_ENDIAN__
961__ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1) {
962  int32x4_t __ret;
963  __ret = __p0 + __p1;
964  return __ret;
965}
966#else
967__ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1) {
968  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
969  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
970  int32x4_t __ret;
971  __ret = __rev0 + __rev1;
972  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
973  return __ret;
974}
975#endif
976
977#ifdef __LITTLE_ENDIAN__
978__ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1) {
979  int64x2_t __ret;
980  __ret = __p0 + __p1;
981  return __ret;
982}
983#else
984__ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1) {
985  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
986  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
987  int64x2_t __ret;
988  __ret = __rev0 + __rev1;
989  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
990  return __ret;
991}
992#endif
993
994#ifdef __LITTLE_ENDIAN__
995__ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1) {
996  int16x8_t __ret;
997  __ret = __p0 + __p1;
998  return __ret;
999}
1000#else
1001__ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1) {
1002  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1003  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1004  int16x8_t __ret;
1005  __ret = __rev0 + __rev1;
1006  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1007  return __ret;
1008}
1009#endif
1010
1011#ifdef __LITTLE_ENDIAN__
1012__ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1013  uint8x8_t __ret;
1014  __ret = __p0 + __p1;
1015  return __ret;
1016}
1017#else
1018__ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1019  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1020  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1021  uint8x8_t __ret;
1022  __ret = __rev0 + __rev1;
1023  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1024  return __ret;
1025}
1026#endif
1027
1028#ifdef __LITTLE_ENDIAN__
1029__ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1030  uint32x2_t __ret;
1031  __ret = __p0 + __p1;
1032  return __ret;
1033}
1034#else
1035__ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1036  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1037  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1038  uint32x2_t __ret;
1039  __ret = __rev0 + __rev1;
1040  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1041  return __ret;
1042}
1043#endif
1044
1045#ifdef __LITTLE_ENDIAN__
1046__ai uint64x1_t vadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
1047  uint64x1_t __ret;
1048  __ret = __p0 + __p1;
1049  return __ret;
1050}
1051#else
1052__ai uint64x1_t vadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
1053  uint64x1_t __ret;
1054  __ret = __p0 + __p1;
1055  return __ret;
1056}
1057#endif
1058
1059#ifdef __LITTLE_ENDIAN__
1060__ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
1061  uint16x4_t __ret;
1062  __ret = __p0 + __p1;
1063  return __ret;
1064}
1065#else
1066__ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
1067  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1068  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1069  uint16x4_t __ret;
1070  __ret = __rev0 + __rev1;
1071  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1072  return __ret;
1073}
1074#endif
1075
1076#ifdef __LITTLE_ENDIAN__
1077__ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1) {
1078  int8x8_t __ret;
1079  __ret = __p0 + __p1;
1080  return __ret;
1081}
1082#else
1083__ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1) {
1084  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1085  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1086  int8x8_t __ret;
1087  __ret = __rev0 + __rev1;
1088  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1089  return __ret;
1090}
1091#endif
1092
1093#ifdef __LITTLE_ENDIAN__
1094__ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1) {
1095  float32x2_t __ret;
1096  __ret = __p0 + __p1;
1097  return __ret;
1098}
1099#else
1100__ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1) {
1101  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1102  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1103  float32x2_t __ret;
1104  __ret = __rev0 + __rev1;
1105  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1106  return __ret;
1107}
1108#endif
1109
1110#ifdef __LITTLE_ENDIAN__
1111__ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1) {
1112  int32x2_t __ret;
1113  __ret = __p0 + __p1;
1114  return __ret;
1115}
1116#else
1117__ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1) {
1118  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1119  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1120  int32x2_t __ret;
1121  __ret = __rev0 + __rev1;
1122  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1123  return __ret;
1124}
1125#endif
1126
1127#ifdef __LITTLE_ENDIAN__
1128__ai int64x1_t vadd_s64(int64x1_t __p0, int64x1_t __p1) {
1129  int64x1_t __ret;
1130  __ret = __p0 + __p1;
1131  return __ret;
1132}
1133#else
1134__ai int64x1_t vadd_s64(int64x1_t __p0, int64x1_t __p1) {
1135  int64x1_t __ret;
1136  __ret = __p0 + __p1;
1137  return __ret;
1138}
1139#endif
1140
1141#ifdef __LITTLE_ENDIAN__
1142__ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) {
1143  int16x4_t __ret;
1144  __ret = __p0 + __p1;
1145  return __ret;
1146}
1147#else
1148__ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) {
1149  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1150  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1151  int16x4_t __ret;
1152  __ret = __rev0 + __rev1;
1153  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1154  return __ret;
1155}
1156#endif
1157
1158#ifdef __LITTLE_ENDIAN__
1159__ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
1160  uint16x4_t __ret;
1161  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
1162  return __ret;
1163}
1164#else
1165__ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
1166  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1167  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1168  uint16x4_t __ret;
1169  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
1170  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1171  return __ret;
1172}
1173__ai uint16x4_t __noswap_vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
1174  uint16x4_t __ret;
1175  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
1176  return __ret;
1177}
1178#endif
1179
1180#ifdef __LITTLE_ENDIAN__
1181__ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
1182  uint32x2_t __ret;
1183  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
1184  return __ret;
1185}
1186#else
1187__ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
1188  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1189  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1190  uint32x2_t __ret;
1191  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
1192  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1193  return __ret;
1194}
1195__ai uint32x2_t __noswap_vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
1196  uint32x2_t __ret;
1197  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
1198  return __ret;
1199}
1200#endif
1201
1202#ifdef __LITTLE_ENDIAN__
1203__ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
1204  uint8x8_t __ret;
1205  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
1206  return __ret;
1207}
1208#else
1209__ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
1210  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1211  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1212  uint8x8_t __ret;
1213  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
1214  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1215  return __ret;
1216}
1217__ai uint8x8_t __noswap_vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
1218  uint8x8_t __ret;
1219  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
1220  return __ret;
1221}
1222#endif
1223
1224#ifdef __LITTLE_ENDIAN__
1225__ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
1226  int16x4_t __ret;
1227  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
1228  return __ret;
1229}
1230#else
1231__ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
1232  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1233  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1234  int16x4_t __ret;
1235  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
1236  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1237  return __ret;
1238}
1239__ai int16x4_t __noswap_vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
1240  int16x4_t __ret;
1241  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
1242  return __ret;
1243}
1244#endif
1245
1246#ifdef __LITTLE_ENDIAN__
1247__ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
1248  int32x2_t __ret;
1249  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
1250  return __ret;
1251}
1252#else
1253__ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
1254  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1255  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1256  int32x2_t __ret;
1257  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
1258  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1259  return __ret;
1260}
1261__ai int32x2_t __noswap_vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
1262  int32x2_t __ret;
1263  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
1264  return __ret;
1265}
1266#endif
1267
1268#ifdef __LITTLE_ENDIAN__
1269__ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
1270  int8x8_t __ret;
1271  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
1272  return __ret;
1273}
1274#else
1275__ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
1276  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1277  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1278  int8x8_t __ret;
1279  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
1280  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1281  return __ret;
1282}
1283__ai int8x8_t __noswap_vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
1284  int8x8_t __ret;
1285  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
1286  return __ret;
1287}
1288#endif
1289
1290#ifdef __LITTLE_ENDIAN__
1291__ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1292  uint8x16_t __ret;
1293  __ret = __p0 & __p1;
1294  return __ret;
1295}
1296#else
1297__ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1298  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1299  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1300  uint8x16_t __ret;
1301  __ret = __rev0 & __rev1;
1302  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1303  return __ret;
1304}
1305#endif
1306
1307#ifdef __LITTLE_ENDIAN__
1308__ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1309  uint32x4_t __ret;
1310  __ret = __p0 & __p1;
1311  return __ret;
1312}
1313#else
1314__ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1315  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1316  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1317  uint32x4_t __ret;
1318  __ret = __rev0 & __rev1;
1319  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1320  return __ret;
1321}
1322#endif
1323
1324#ifdef __LITTLE_ENDIAN__
1325__ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1326  uint64x2_t __ret;
1327  __ret = __p0 & __p1;
1328  return __ret;
1329}
1330#else
1331__ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1332  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1333  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1334  uint64x2_t __ret;
1335  __ret = __rev0 & __rev1;
1336  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1337  return __ret;
1338}
1339#endif
1340
1341#ifdef __LITTLE_ENDIAN__
1342__ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1343  uint16x8_t __ret;
1344  __ret = __p0 & __p1;
1345  return __ret;
1346}
1347#else
1348__ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1349  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1350  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1351  uint16x8_t __ret;
1352  __ret = __rev0 & __rev1;
1353  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1354  return __ret;
1355}
1356#endif
1357
1358#ifdef __LITTLE_ENDIAN__
1359__ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1) {
1360  int8x16_t __ret;
1361  __ret = __p0 & __p1;
1362  return __ret;
1363}
1364#else
1365__ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1) {
1366  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1367  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1368  int8x16_t __ret;
1369  __ret = __rev0 & __rev1;
1370  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1371  return __ret;
1372}
1373#endif
1374
1375#ifdef __LITTLE_ENDIAN__
1376__ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1) {
1377  int32x4_t __ret;
1378  __ret = __p0 & __p1;
1379  return __ret;
1380}
1381#else
1382__ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1) {
1383  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1384  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1385  int32x4_t __ret;
1386  __ret = __rev0 & __rev1;
1387  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1388  return __ret;
1389}
1390#endif
1391
1392#ifdef __LITTLE_ENDIAN__
1393__ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1) {
1394  int64x2_t __ret;
1395  __ret = __p0 & __p1;
1396  return __ret;
1397}
1398#else
1399__ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1) {
1400  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1401  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1402  int64x2_t __ret;
1403  __ret = __rev0 & __rev1;
1404  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1405  return __ret;
1406}
1407#endif
1408
1409#ifdef __LITTLE_ENDIAN__
1410__ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1) {
1411  int16x8_t __ret;
1412  __ret = __p0 & __p1;
1413  return __ret;
1414}
1415#else
1416__ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1) {
1417  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1418  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1419  int16x8_t __ret;
1420  __ret = __rev0 & __rev1;
1421  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1422  return __ret;
1423}
1424#endif
1425
1426#ifdef __LITTLE_ENDIAN__
1427__ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1) {
1428  uint8x8_t __ret;
1429  __ret = __p0 & __p1;
1430  return __ret;
1431}
1432#else
1433__ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1) {
1434  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1435  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1436  uint8x8_t __ret;
1437  __ret = __rev0 & __rev1;
1438  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1439  return __ret;
1440}
1441#endif
1442
1443#ifdef __LITTLE_ENDIAN__
1444__ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1) {
1445  uint32x2_t __ret;
1446  __ret = __p0 & __p1;
1447  return __ret;
1448}
1449#else
1450__ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1) {
1451  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1452  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1453  uint32x2_t __ret;
1454  __ret = __rev0 & __rev1;
1455  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1456  return __ret;
1457}
1458#endif
1459
1460#ifdef __LITTLE_ENDIAN__
1461__ai uint64x1_t vand_u64(uint64x1_t __p0, uint64x1_t __p1) {
1462  uint64x1_t __ret;
1463  __ret = __p0 & __p1;
1464  return __ret;
1465}
1466#else
1467__ai uint64x1_t vand_u64(uint64x1_t __p0, uint64x1_t __p1) {
1468  uint64x1_t __ret;
1469  __ret = __p0 & __p1;
1470  return __ret;
1471}
1472#endif
1473
1474#ifdef __LITTLE_ENDIAN__
1475__ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1) {
1476  uint16x4_t __ret;
1477  __ret = __p0 & __p1;
1478  return __ret;
1479}
1480#else
1481__ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1) {
1482  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1483  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1484  uint16x4_t __ret;
1485  __ret = __rev0 & __rev1;
1486  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1487  return __ret;
1488}
1489#endif
1490
1491#ifdef __LITTLE_ENDIAN__
1492__ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1) {
1493  int8x8_t __ret;
1494  __ret = __p0 & __p1;
1495  return __ret;
1496}
1497#else
1498__ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1) {
1499  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1500  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1501  int8x8_t __ret;
1502  __ret = __rev0 & __rev1;
1503  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1504  return __ret;
1505}
1506#endif
1507
1508#ifdef __LITTLE_ENDIAN__
1509__ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1) {
1510  int32x2_t __ret;
1511  __ret = __p0 & __p1;
1512  return __ret;
1513}
1514#else
1515__ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1) {
1516  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1517  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1518  int32x2_t __ret;
1519  __ret = __rev0 & __rev1;
1520  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1521  return __ret;
1522}
1523#endif
1524
1525#ifdef __LITTLE_ENDIAN__
1526__ai int64x1_t vand_s64(int64x1_t __p0, int64x1_t __p1) {
1527  int64x1_t __ret;
1528  __ret = __p0 & __p1;
1529  return __ret;
1530}
1531#else
1532__ai int64x1_t vand_s64(int64x1_t __p0, int64x1_t __p1) {
1533  int64x1_t __ret;
1534  __ret = __p0 & __p1;
1535  return __ret;
1536}
1537#endif
1538
1539#ifdef __LITTLE_ENDIAN__
1540__ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1) {
1541  int16x4_t __ret;
1542  __ret = __p0 & __p1;
1543  return __ret;
1544}
1545#else
1546__ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1) {
1547  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1548  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1549  int16x4_t __ret;
1550  __ret = __rev0 & __rev1;
1551  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1552  return __ret;
1553}
1554#endif
1555
1556#ifdef __LITTLE_ENDIAN__
1557__ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1558  uint8x16_t __ret;
1559  __ret = __p0 & ~__p1;
1560  return __ret;
1561}
1562#else
1563__ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1564  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1565  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1566  uint8x16_t __ret;
1567  __ret = __rev0 & ~__rev1;
1568  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1569  return __ret;
1570}
1571#endif
1572
1573#ifdef __LITTLE_ENDIAN__
1574__ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1575  uint32x4_t __ret;
1576  __ret = __p0 & ~__p1;
1577  return __ret;
1578}
1579#else
1580__ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1581  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1582  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1583  uint32x4_t __ret;
1584  __ret = __rev0 & ~__rev1;
1585  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1586  return __ret;
1587}
1588#endif
1589
1590#ifdef __LITTLE_ENDIAN__
1591__ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1592  uint64x2_t __ret;
1593  __ret = __p0 & ~__p1;
1594  return __ret;
1595}
1596#else
1597__ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1598  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1599  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1600  uint64x2_t __ret;
1601  __ret = __rev0 & ~__rev1;
1602  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1603  return __ret;
1604}
1605#endif
1606
1607#ifdef __LITTLE_ENDIAN__
1608__ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1609  uint16x8_t __ret;
1610  __ret = __p0 & ~__p1;
1611  return __ret;
1612}
1613#else
1614__ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1615  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1616  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1617  uint16x8_t __ret;
1618  __ret = __rev0 & ~__rev1;
1619  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1620  return __ret;
1621}
1622#endif
1623
1624#ifdef __LITTLE_ENDIAN__
1625__ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1) {
1626  int8x16_t __ret;
1627  __ret = __p0 & ~__p1;
1628  return __ret;
1629}
1630#else
1631__ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1) {
1632  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1633  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1634  int8x16_t __ret;
1635  __ret = __rev0 & ~__rev1;
1636  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1637  return __ret;
1638}
1639#endif
1640
1641#ifdef __LITTLE_ENDIAN__
1642__ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1) {
1643  int32x4_t __ret;
1644  __ret = __p0 & ~__p1;
1645  return __ret;
1646}
1647#else
1648__ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1) {
1649  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1650  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1651  int32x4_t __ret;
1652  __ret = __rev0 & ~__rev1;
1653  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1654  return __ret;
1655}
1656#endif
1657
1658#ifdef __LITTLE_ENDIAN__
1659__ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1) {
1660  int64x2_t __ret;
1661  __ret = __p0 & ~__p1;
1662  return __ret;
1663}
1664#else
1665__ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1) {
1666  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1667  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1668  int64x2_t __ret;
1669  __ret = __rev0 & ~__rev1;
1670  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1671  return __ret;
1672}
1673#endif
1674
1675#ifdef __LITTLE_ENDIAN__
1676__ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1) {
1677  int16x8_t __ret;
1678  __ret = __p0 & ~__p1;
1679  return __ret;
1680}
1681#else
1682__ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1) {
1683  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1684  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1685  int16x8_t __ret;
1686  __ret = __rev0 & ~__rev1;
1687  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1688  return __ret;
1689}
1690#endif
1691
1692#ifdef __LITTLE_ENDIAN__
1693__ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1) {
1694  uint8x8_t __ret;
1695  __ret = __p0 & ~__p1;
1696  return __ret;
1697}
1698#else
1699__ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1) {
1700  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1701  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1702  uint8x8_t __ret;
1703  __ret = __rev0 & ~__rev1;
1704  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1705  return __ret;
1706}
1707#endif
1708
1709#ifdef __LITTLE_ENDIAN__
1710__ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1) {
1711  uint32x2_t __ret;
1712  __ret = __p0 & ~__p1;
1713  return __ret;
1714}
1715#else
1716__ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1) {
1717  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1718  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1719  uint32x2_t __ret;
1720  __ret = __rev0 & ~__rev1;
1721  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1722  return __ret;
1723}
1724#endif
1725
1726#ifdef __LITTLE_ENDIAN__
1727__ai uint64x1_t vbic_u64(uint64x1_t __p0, uint64x1_t __p1) {
1728  uint64x1_t __ret;
1729  __ret = __p0 & ~__p1;
1730  return __ret;
1731}
1732#else
1733__ai uint64x1_t vbic_u64(uint64x1_t __p0, uint64x1_t __p1) {
1734  uint64x1_t __ret;
1735  __ret = __p0 & ~__p1;
1736  return __ret;
1737}
1738#endif
1739
1740#ifdef __LITTLE_ENDIAN__
1741__ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1) {
1742  uint16x4_t __ret;
1743  __ret = __p0 & ~__p1;
1744  return __ret;
1745}
1746#else
1747__ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1) {
1748  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1749  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1750  uint16x4_t __ret;
1751  __ret = __rev0 & ~__rev1;
1752  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1753  return __ret;
1754}
1755#endif
1756
1757#ifdef __LITTLE_ENDIAN__
1758__ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1) {
1759  int8x8_t __ret;
1760  __ret = __p0 & ~__p1;
1761  return __ret;
1762}
1763#else
1764__ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1) {
1765  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1766  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1767  int8x8_t __ret;
1768  __ret = __rev0 & ~__rev1;
1769  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1770  return __ret;
1771}
1772#endif
1773
1774#ifdef __LITTLE_ENDIAN__
1775__ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1) {
1776  int32x2_t __ret;
1777  __ret = __p0 & ~__p1;
1778  return __ret;
1779}
1780#else
1781__ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1) {
1782  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1783  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1784  int32x2_t __ret;
1785  __ret = __rev0 & ~__rev1;
1786  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1787  return __ret;
1788}
1789#endif
1790
1791#ifdef __LITTLE_ENDIAN__
1792__ai int64x1_t vbic_s64(int64x1_t __p0, int64x1_t __p1) {
1793  int64x1_t __ret;
1794  __ret = __p0 & ~__p1;
1795  return __ret;
1796}
1797#else
1798__ai int64x1_t vbic_s64(int64x1_t __p0, int64x1_t __p1) {
1799  int64x1_t __ret;
1800  __ret = __p0 & ~__p1;
1801  return __ret;
1802}
1803#endif
1804
1805#ifdef __LITTLE_ENDIAN__
1806__ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1) {
1807  int16x4_t __ret;
1808  __ret = __p0 & ~__p1;
1809  return __ret;
1810}
1811#else
1812__ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1) {
1813  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1814  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1815  int16x4_t __ret;
1816  __ret = __rev0 & ~__rev1;
1817  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1818  return __ret;
1819}
1820#endif
1821
1822#ifdef __LITTLE_ENDIAN__
1823__ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2) {
1824  poly8x8_t __ret;
1825  __ret = (poly8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 4);
1826  return __ret;
1827}
1828#else
1829__ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2) {
1830  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1831  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1832  poly8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
1833  poly8x8_t __ret;
1834  __ret = (poly8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 4);
1835  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1836  return __ret;
1837}
1838#endif
1839
1840#ifdef __LITTLE_ENDIAN__
1841__ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2) {
1842  poly16x4_t __ret;
1843  __ret = (poly16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 5);
1844  return __ret;
1845}
1846#else
1847__ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2) {
1848  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1849  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1850  poly16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
1851  poly16x4_t __ret;
1852  __ret = (poly16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 5);
1853  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1854  return __ret;
1855}
1856#endif
1857
1858#ifdef __LITTLE_ENDIAN__
1859__ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2) {
1860  poly8x16_t __ret;
1861  __ret = (poly8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 36);
1862  return __ret;
1863}
1864#else
1865__ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2) {
1866  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1867  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1868  poly8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1869  poly8x16_t __ret;
1870  __ret = (poly8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 36);
1871  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1872  return __ret;
1873}
1874#endif
1875
1876#ifdef __LITTLE_ENDIAN__
1877__ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2) {
1878  poly16x8_t __ret;
1879  __ret = (poly16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 37);
1880  return __ret;
1881}
1882#else
1883__ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2) {
1884  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1885  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1886  poly16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
1887  poly16x8_t __ret;
1888  __ret = (poly16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 37);
1889  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1890  return __ret;
1891}
1892#endif
1893
1894#ifdef __LITTLE_ENDIAN__
1895__ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
1896  uint8x16_t __ret;
1897  __ret = (uint8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
1898  return __ret;
1899}
1900#else
1901__ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
1902  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1903  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1904  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1905  uint8x16_t __ret;
1906  __ret = (uint8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
1907  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1908  return __ret;
1909}
1910#endif
1911
1912#ifdef __LITTLE_ENDIAN__
1913__ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
1914  uint32x4_t __ret;
1915  __ret = (uint32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
1916  return __ret;
1917}
1918#else
1919__ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
1920  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1921  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1922  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
1923  uint32x4_t __ret;
1924  __ret = (uint32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
1925  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1926  return __ret;
1927}
1928#endif
1929
1930#ifdef __LITTLE_ENDIAN__
1931__ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
1932  uint64x2_t __ret;
1933  __ret = (uint64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51);
1934  return __ret;
1935}
1936#else
1937__ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
1938  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1939  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1940  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
1941  uint64x2_t __ret;
1942  __ret = (uint64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51);
1943  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1944  return __ret;
1945}
1946#endif
1947
1948#ifdef __LITTLE_ENDIAN__
1949__ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
1950  uint16x8_t __ret;
1951  __ret = (uint16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 49);
1952  return __ret;
1953}
1954#else
1955__ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
1956  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1957  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1958  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
1959  uint16x8_t __ret;
1960  __ret = (uint16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 49);
1961  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1962  return __ret;
1963}
1964#endif
1965
1966#ifdef __LITTLE_ENDIAN__
1967__ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
1968  int8x16_t __ret;
1969  __ret = (int8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
1970  return __ret;
1971}
1972#else
1973__ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
1974  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1975  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1976  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1977  int8x16_t __ret;
1978  __ret = (int8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
1979  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1980  return __ret;
1981}
1982#endif
1983
1984#ifdef __LITTLE_ENDIAN__
1985__ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
1986  float32x4_t __ret;
1987  __ret = (float32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
1988  return __ret;
1989}
1990#else
1991__ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
1992  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1993  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1994  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
1995  float32x4_t __ret;
1996  __ret = (float32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
1997  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1998  return __ret;
1999}
2000#endif
2001
2002#ifdef __LITTLE_ENDIAN__
2003__ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
2004  int32x4_t __ret;
2005  __ret = (int32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
2006  return __ret;
2007}
2008#else
2009__ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
2010  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2011  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2012  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2013  int32x4_t __ret;
2014  __ret = (int32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
2015  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2016  return __ret;
2017}
2018#endif
2019
2020#ifdef __LITTLE_ENDIAN__
2021__ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
2022  int64x2_t __ret;
2023  __ret = (int64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 35);
2024  return __ret;
2025}
2026#else
2027__ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
2028  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2029  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2030  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2031  int64x2_t __ret;
2032  __ret = (int64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 35);
2033  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2034  return __ret;
2035}
2036#endif
2037
2038#ifdef __LITTLE_ENDIAN__
2039__ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
2040  int16x8_t __ret;
2041  __ret = (int16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33);
2042  return __ret;
2043}
2044#else
2045__ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
2046  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2047  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2048  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2049  int16x8_t __ret;
2050  __ret = (int16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33);
2051  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2052  return __ret;
2053}
2054#endif
2055
2056#ifdef __LITTLE_ENDIAN__
2057__ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
2058  uint8x8_t __ret;
2059  __ret = (uint8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 16);
2060  return __ret;
2061}
2062#else
2063__ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
2064  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2065  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2066  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2067  uint8x8_t __ret;
2068  __ret = (uint8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 16);
2069  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2070  return __ret;
2071}
2072#endif
2073
2074#ifdef __LITTLE_ENDIAN__
2075__ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
2076  uint32x2_t __ret;
2077  __ret = (uint32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 18);
2078  return __ret;
2079}
2080#else
2081__ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
2082  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2083  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2084  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2085  uint32x2_t __ret;
2086  __ret = (uint32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 18);
2087  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2088  return __ret;
2089}
2090#endif
2091
2092#ifdef __LITTLE_ENDIAN__
2093__ai uint64x1_t vbsl_u64(uint64x1_t __p0, uint64x1_t __p1, uint64x1_t __p2) {
2094  uint64x1_t __ret;
2095  __ret = (uint64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 19);
2096  return __ret;
2097}
2098#else
2099__ai uint64x1_t vbsl_u64(uint64x1_t __p0, uint64x1_t __p1, uint64x1_t __p2) {
2100  uint64x1_t __ret;
2101  __ret = (uint64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 19);
2102  return __ret;
2103}
2104#endif
2105
2106#ifdef __LITTLE_ENDIAN__
2107__ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
2108  uint16x4_t __ret;
2109  __ret = (uint16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 17);
2110  return __ret;
2111}
2112#else
2113__ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
2114  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2115  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2116  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2117  uint16x4_t __ret;
2118  __ret = (uint16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 17);
2119  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2120  return __ret;
2121}
2122#endif
2123
2124#ifdef __LITTLE_ENDIAN__
2125__ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
2126  int8x8_t __ret;
2127  __ret = (int8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 0);
2128  return __ret;
2129}
2130#else
2131__ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
2132  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2133  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2134  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2135  int8x8_t __ret;
2136  __ret = (int8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 0);
2137  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2138  return __ret;
2139}
2140#endif
2141
2142#ifdef __LITTLE_ENDIAN__
2143__ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
2144  float32x2_t __ret;
2145  __ret = (float32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
2146  return __ret;
2147}
2148#else
2149__ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
2150  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2151  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2152  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2153  float32x2_t __ret;
2154  __ret = (float32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
2155  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2156  return __ret;
2157}
2158#endif
2159
2160#ifdef __LITTLE_ENDIAN__
2161__ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
2162  int32x2_t __ret;
2163  __ret = (int32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
2164  return __ret;
2165}
2166#else
2167__ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
2168  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2169  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2170  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2171  int32x2_t __ret;
2172  __ret = (int32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 2);
2173  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2174  return __ret;
2175}
2176#endif
2177
2178#ifdef __LITTLE_ENDIAN__
2179__ai int64x1_t vbsl_s64(uint64x1_t __p0, int64x1_t __p1, int64x1_t __p2) {
2180  int64x1_t __ret;
2181  __ret = (int64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 3);
2182  return __ret;
2183}
2184#else
2185__ai int64x1_t vbsl_s64(uint64x1_t __p0, int64x1_t __p1, int64x1_t __p2) {
2186  int64x1_t __ret;
2187  __ret = (int64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 3);
2188  return __ret;
2189}
2190#endif
2191
2192#ifdef __LITTLE_ENDIAN__
2193__ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
2194  int16x4_t __ret;
2195  __ret = (int16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 1);
2196  return __ret;
2197}
2198#else
2199__ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
2200  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2201  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2202  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2203  int16x4_t __ret;
2204  __ret = (int16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 1);
2205  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2206  return __ret;
2207}
2208#endif
2209
2210#ifdef __LITTLE_ENDIAN__
2211__ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1) {
2212  uint32x4_t __ret;
2213  __ret = (uint32x4_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2214  return __ret;
2215}
2216#else
2217__ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1) {
2218  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2219  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2220  uint32x4_t __ret;
2221  __ret = (uint32x4_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2222  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2223  return __ret;
2224}
2225#endif
2226
2227#ifdef __LITTLE_ENDIAN__
2228__ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1) {
2229  uint32x2_t __ret;
2230  __ret = (uint32x2_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2231  return __ret;
2232}
2233#else
2234__ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1) {
2235  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2236  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2237  uint32x2_t __ret;
2238  __ret = (uint32x2_t) __builtin_neon_vcage_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2239  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2240  return __ret;
2241}
2242#endif
2243
2244#ifdef __LITTLE_ENDIAN__
2245__ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1) {
2246  uint32x4_t __ret;
2247  __ret = (uint32x4_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2248  return __ret;
2249}
2250#else
2251__ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1) {
2252  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2253  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2254  uint32x4_t __ret;
2255  __ret = (uint32x4_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2256  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2257  return __ret;
2258}
2259#endif
2260
2261#ifdef __LITTLE_ENDIAN__
2262__ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1) {
2263  uint32x2_t __ret;
2264  __ret = (uint32x2_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2265  return __ret;
2266}
2267#else
2268__ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1) {
2269  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2270  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2271  uint32x2_t __ret;
2272  __ret = (uint32x2_t) __builtin_neon_vcagt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2273  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2274  return __ret;
2275}
2276#endif
2277
2278#ifdef __LITTLE_ENDIAN__
2279__ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1) {
2280  uint32x4_t __ret;
2281  __ret = (uint32x4_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2282  return __ret;
2283}
2284#else
2285__ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1) {
2286  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2287  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2288  uint32x4_t __ret;
2289  __ret = (uint32x4_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2290  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2291  return __ret;
2292}
2293#endif
2294
2295#ifdef __LITTLE_ENDIAN__
2296__ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1) {
2297  uint32x2_t __ret;
2298  __ret = (uint32x2_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2299  return __ret;
2300}
2301#else
2302__ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1) {
2303  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2304  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2305  uint32x2_t __ret;
2306  __ret = (uint32x2_t) __builtin_neon_vcale_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2307  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2308  return __ret;
2309}
2310#endif
2311
2312#ifdef __LITTLE_ENDIAN__
2313__ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1) {
2314  uint32x4_t __ret;
2315  __ret = (uint32x4_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2316  return __ret;
2317}
2318#else
2319__ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1) {
2320  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2321  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2322  uint32x4_t __ret;
2323  __ret = (uint32x4_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2324  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2325  return __ret;
2326}
2327#endif
2328
2329#ifdef __LITTLE_ENDIAN__
2330__ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1) {
2331  uint32x2_t __ret;
2332  __ret = (uint32x2_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2333  return __ret;
2334}
2335#else
2336__ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1) {
2337  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2338  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2339  uint32x2_t __ret;
2340  __ret = (uint32x2_t) __builtin_neon_vcalt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2341  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2342  return __ret;
2343}
2344#endif
2345
2346#ifdef __LITTLE_ENDIAN__
2347__ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1) {
2348  uint8x8_t __ret;
2349  __ret = (uint8x8_t)(__p0 == __p1);
2350  return __ret;
2351}
2352#else
2353__ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1) {
2354  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2355  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2356  uint8x8_t __ret;
2357  __ret = (uint8x8_t)(__rev0 == __rev1);
2358  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2359  return __ret;
2360}
2361#endif
2362
2363#ifdef __LITTLE_ENDIAN__
2364__ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1) {
2365  uint8x16_t __ret;
2366  __ret = (uint8x16_t)(__p0 == __p1);
2367  return __ret;
2368}
2369#else
2370__ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1) {
2371  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2372  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2373  uint8x16_t __ret;
2374  __ret = (uint8x16_t)(__rev0 == __rev1);
2375  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2376  return __ret;
2377}
2378#endif
2379
2380#ifdef __LITTLE_ENDIAN__
2381__ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2382  uint8x16_t __ret;
2383  __ret = (uint8x16_t)(__p0 == __p1);
2384  return __ret;
2385}
2386#else
2387__ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2388  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2389  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2390  uint8x16_t __ret;
2391  __ret = (uint8x16_t)(__rev0 == __rev1);
2392  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2393  return __ret;
2394}
2395#endif
2396
2397#ifdef __LITTLE_ENDIAN__
2398__ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2399  uint32x4_t __ret;
2400  __ret = (uint32x4_t)(__p0 == __p1);
2401  return __ret;
2402}
2403#else
2404__ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2405  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2406  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2407  uint32x4_t __ret;
2408  __ret = (uint32x4_t)(__rev0 == __rev1);
2409  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2410  return __ret;
2411}
2412#endif
2413
2414#ifdef __LITTLE_ENDIAN__
2415__ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2416  uint16x8_t __ret;
2417  __ret = (uint16x8_t)(__p0 == __p1);
2418  return __ret;
2419}
2420#else
2421__ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2422  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2423  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2424  uint16x8_t __ret;
2425  __ret = (uint16x8_t)(__rev0 == __rev1);
2426  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2427  return __ret;
2428}
2429#endif
2430
2431#ifdef __LITTLE_ENDIAN__
2432__ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1) {
2433  uint8x16_t __ret;
2434  __ret = (uint8x16_t)(__p0 == __p1);
2435  return __ret;
2436}
2437#else
2438__ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1) {
2439  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2440  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2441  uint8x16_t __ret;
2442  __ret = (uint8x16_t)(__rev0 == __rev1);
2443  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2444  return __ret;
2445}
2446#endif
2447
2448#ifdef __LITTLE_ENDIAN__
2449__ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1) {
2450  uint32x4_t __ret;
2451  __ret = (uint32x4_t)(__p0 == __p1);
2452  return __ret;
2453}
2454#else
2455__ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1) {
2456  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2457  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2458  uint32x4_t __ret;
2459  __ret = (uint32x4_t)(__rev0 == __rev1);
2460  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2461  return __ret;
2462}
2463#endif
2464
2465#ifdef __LITTLE_ENDIAN__
2466__ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1) {
2467  uint32x4_t __ret;
2468  __ret = (uint32x4_t)(__p0 == __p1);
2469  return __ret;
2470}
2471#else
2472__ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1) {
2473  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2474  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2475  uint32x4_t __ret;
2476  __ret = (uint32x4_t)(__rev0 == __rev1);
2477  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2478  return __ret;
2479}
2480#endif
2481
2482#ifdef __LITTLE_ENDIAN__
2483__ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1) {
2484  uint16x8_t __ret;
2485  __ret = (uint16x8_t)(__p0 == __p1);
2486  return __ret;
2487}
2488#else
2489__ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1) {
2490  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2491  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2492  uint16x8_t __ret;
2493  __ret = (uint16x8_t)(__rev0 == __rev1);
2494  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2495  return __ret;
2496}
2497#endif
2498
2499#ifdef __LITTLE_ENDIAN__
2500__ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1) {
2501  uint8x8_t __ret;
2502  __ret = (uint8x8_t)(__p0 == __p1);
2503  return __ret;
2504}
2505#else
2506__ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1) {
2507  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2508  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2509  uint8x8_t __ret;
2510  __ret = (uint8x8_t)(__rev0 == __rev1);
2511  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2512  return __ret;
2513}
2514#endif
2515
2516#ifdef __LITTLE_ENDIAN__
2517__ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1) {
2518  uint32x2_t __ret;
2519  __ret = (uint32x2_t)(__p0 == __p1);
2520  return __ret;
2521}
2522#else
2523__ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1) {
2524  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2525  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2526  uint32x2_t __ret;
2527  __ret = (uint32x2_t)(__rev0 == __rev1);
2528  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2529  return __ret;
2530}
2531#endif
2532
2533#ifdef __LITTLE_ENDIAN__
2534__ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1) {
2535  uint16x4_t __ret;
2536  __ret = (uint16x4_t)(__p0 == __p1);
2537  return __ret;
2538}
2539#else
2540__ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1) {
2541  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2542  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2543  uint16x4_t __ret;
2544  __ret = (uint16x4_t)(__rev0 == __rev1);
2545  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2546  return __ret;
2547}
2548#endif
2549
2550#ifdef __LITTLE_ENDIAN__
2551__ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1) {
2552  uint8x8_t __ret;
2553  __ret = (uint8x8_t)(__p0 == __p1);
2554  return __ret;
2555}
2556#else
2557__ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1) {
2558  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2559  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2560  uint8x8_t __ret;
2561  __ret = (uint8x8_t)(__rev0 == __rev1);
2562  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2563  return __ret;
2564}
2565#endif
2566
2567#ifdef __LITTLE_ENDIAN__
2568__ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1) {
2569  uint32x2_t __ret;
2570  __ret = (uint32x2_t)(__p0 == __p1);
2571  return __ret;
2572}
2573#else
2574__ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1) {
2575  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2576  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2577  uint32x2_t __ret;
2578  __ret = (uint32x2_t)(__rev0 == __rev1);
2579  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2580  return __ret;
2581}
2582#endif
2583
2584#ifdef __LITTLE_ENDIAN__
2585__ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1) {
2586  uint32x2_t __ret;
2587  __ret = (uint32x2_t)(__p0 == __p1);
2588  return __ret;
2589}
2590#else
2591__ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1) {
2592  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2593  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2594  uint32x2_t __ret;
2595  __ret = (uint32x2_t)(__rev0 == __rev1);
2596  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2597  return __ret;
2598}
2599#endif
2600
2601#ifdef __LITTLE_ENDIAN__
2602__ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1) {
2603  uint16x4_t __ret;
2604  __ret = (uint16x4_t)(__p0 == __p1);
2605  return __ret;
2606}
2607#else
2608__ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1) {
2609  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2610  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2611  uint16x4_t __ret;
2612  __ret = (uint16x4_t)(__rev0 == __rev1);
2613  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2614  return __ret;
2615}
2616#endif
2617
2618#ifdef __LITTLE_ENDIAN__
2619__ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2620  uint8x16_t __ret;
2621  __ret = (uint8x16_t)(__p0 >= __p1);
2622  return __ret;
2623}
2624#else
2625__ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2626  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2627  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2628  uint8x16_t __ret;
2629  __ret = (uint8x16_t)(__rev0 >= __rev1);
2630  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2631  return __ret;
2632}
2633#endif
2634
2635#ifdef __LITTLE_ENDIAN__
2636__ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2637  uint32x4_t __ret;
2638  __ret = (uint32x4_t)(__p0 >= __p1);
2639  return __ret;
2640}
2641#else
2642__ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2643  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2644  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2645  uint32x4_t __ret;
2646  __ret = (uint32x4_t)(__rev0 >= __rev1);
2647  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2648  return __ret;
2649}
2650#endif
2651
2652#ifdef __LITTLE_ENDIAN__
2653__ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2654  uint16x8_t __ret;
2655  __ret = (uint16x8_t)(__p0 >= __p1);
2656  return __ret;
2657}
2658#else
2659__ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2660  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2661  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2662  uint16x8_t __ret;
2663  __ret = (uint16x8_t)(__rev0 >= __rev1);
2664  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2665  return __ret;
2666}
2667#endif
2668
2669#ifdef __LITTLE_ENDIAN__
2670__ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1) {
2671  uint8x16_t __ret;
2672  __ret = (uint8x16_t)(__p0 >= __p1);
2673  return __ret;
2674}
2675#else
2676__ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1) {
2677  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2678  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2679  uint8x16_t __ret;
2680  __ret = (uint8x16_t)(__rev0 >= __rev1);
2681  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2682  return __ret;
2683}
2684#endif
2685
2686#ifdef __LITTLE_ENDIAN__
2687__ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1) {
2688  uint32x4_t __ret;
2689  __ret = (uint32x4_t)(__p0 >= __p1);
2690  return __ret;
2691}
2692#else
2693__ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1) {
2694  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2695  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2696  uint32x4_t __ret;
2697  __ret = (uint32x4_t)(__rev0 >= __rev1);
2698  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2699  return __ret;
2700}
2701#endif
2702
2703#ifdef __LITTLE_ENDIAN__
2704__ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1) {
2705  uint32x4_t __ret;
2706  __ret = (uint32x4_t)(__p0 >= __p1);
2707  return __ret;
2708}
2709#else
2710__ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1) {
2711  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2712  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2713  uint32x4_t __ret;
2714  __ret = (uint32x4_t)(__rev0 >= __rev1);
2715  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2716  return __ret;
2717}
2718#endif
2719
2720#ifdef __LITTLE_ENDIAN__
2721__ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1) {
2722  uint16x8_t __ret;
2723  __ret = (uint16x8_t)(__p0 >= __p1);
2724  return __ret;
2725}
2726#else
2727__ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1) {
2728  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2729  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2730  uint16x8_t __ret;
2731  __ret = (uint16x8_t)(__rev0 >= __rev1);
2732  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2733  return __ret;
2734}
2735#endif
2736
2737#ifdef __LITTLE_ENDIAN__
2738__ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1) {
2739  uint8x8_t __ret;
2740  __ret = (uint8x8_t)(__p0 >= __p1);
2741  return __ret;
2742}
2743#else
2744__ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1) {
2745  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2746  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2747  uint8x8_t __ret;
2748  __ret = (uint8x8_t)(__rev0 >= __rev1);
2749  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2750  return __ret;
2751}
2752#endif
2753
2754#ifdef __LITTLE_ENDIAN__
2755__ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1) {
2756  uint32x2_t __ret;
2757  __ret = (uint32x2_t)(__p0 >= __p1);
2758  return __ret;
2759}
2760#else
2761__ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1) {
2762  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2763  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2764  uint32x2_t __ret;
2765  __ret = (uint32x2_t)(__rev0 >= __rev1);
2766  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2767  return __ret;
2768}
2769#endif
2770
2771#ifdef __LITTLE_ENDIAN__
2772__ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1) {
2773  uint16x4_t __ret;
2774  __ret = (uint16x4_t)(__p0 >= __p1);
2775  return __ret;
2776}
2777#else
2778__ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1) {
2779  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2780  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2781  uint16x4_t __ret;
2782  __ret = (uint16x4_t)(__rev0 >= __rev1);
2783  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2784  return __ret;
2785}
2786#endif
2787
2788#ifdef __LITTLE_ENDIAN__
2789__ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1) {
2790  uint8x8_t __ret;
2791  __ret = (uint8x8_t)(__p0 >= __p1);
2792  return __ret;
2793}
2794#else
2795__ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1) {
2796  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2797  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2798  uint8x8_t __ret;
2799  __ret = (uint8x8_t)(__rev0 >= __rev1);
2800  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2801  return __ret;
2802}
2803#endif
2804
2805#ifdef __LITTLE_ENDIAN__
2806__ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1) {
2807  uint32x2_t __ret;
2808  __ret = (uint32x2_t)(__p0 >= __p1);
2809  return __ret;
2810}
2811#else
2812__ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1) {
2813  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2814  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2815  uint32x2_t __ret;
2816  __ret = (uint32x2_t)(__rev0 >= __rev1);
2817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2818  return __ret;
2819}
2820#endif
2821
2822#ifdef __LITTLE_ENDIAN__
2823__ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1) {
2824  uint32x2_t __ret;
2825  __ret = (uint32x2_t)(__p0 >= __p1);
2826  return __ret;
2827}
2828#else
2829__ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1) {
2830  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2831  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2832  uint32x2_t __ret;
2833  __ret = (uint32x2_t)(__rev0 >= __rev1);
2834  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2835  return __ret;
2836}
2837#endif
2838
2839#ifdef __LITTLE_ENDIAN__
2840__ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1) {
2841  uint16x4_t __ret;
2842  __ret = (uint16x4_t)(__p0 >= __p1);
2843  return __ret;
2844}
2845#else
2846__ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1) {
2847  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2848  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2849  uint16x4_t __ret;
2850  __ret = (uint16x4_t)(__rev0 >= __rev1);
2851  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2852  return __ret;
2853}
2854#endif
2855
2856#ifdef __LITTLE_ENDIAN__
2857__ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2858  uint8x16_t __ret;
2859  __ret = (uint8x16_t)(__p0 > __p1);
2860  return __ret;
2861}
2862#else
2863__ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2864  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2865  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2866  uint8x16_t __ret;
2867  __ret = (uint8x16_t)(__rev0 > __rev1);
2868  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2869  return __ret;
2870}
2871#endif
2872
2873#ifdef __LITTLE_ENDIAN__
2874__ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2875  uint32x4_t __ret;
2876  __ret = (uint32x4_t)(__p0 > __p1);
2877  return __ret;
2878}
2879#else
2880__ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2881  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2882  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2883  uint32x4_t __ret;
2884  __ret = (uint32x4_t)(__rev0 > __rev1);
2885  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2886  return __ret;
2887}
2888#endif
2889
2890#ifdef __LITTLE_ENDIAN__
2891__ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2892  uint16x8_t __ret;
2893  __ret = (uint16x8_t)(__p0 > __p1);
2894  return __ret;
2895}
2896#else
2897__ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2898  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2899  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2900  uint16x8_t __ret;
2901  __ret = (uint16x8_t)(__rev0 > __rev1);
2902  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2903  return __ret;
2904}
2905#endif
2906
2907#ifdef __LITTLE_ENDIAN__
2908__ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1) {
2909  uint8x16_t __ret;
2910  __ret = (uint8x16_t)(__p0 > __p1);
2911  return __ret;
2912}
2913#else
2914__ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1) {
2915  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2916  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2917  uint8x16_t __ret;
2918  __ret = (uint8x16_t)(__rev0 > __rev1);
2919  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2920  return __ret;
2921}
2922#endif
2923
2924#ifdef __LITTLE_ENDIAN__
2925__ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1) {
2926  uint32x4_t __ret;
2927  __ret = (uint32x4_t)(__p0 > __p1);
2928  return __ret;
2929}
2930#else
2931__ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1) {
2932  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2933  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2934  uint32x4_t __ret;
2935  __ret = (uint32x4_t)(__rev0 > __rev1);
2936  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2937  return __ret;
2938}
2939#endif
2940
2941#ifdef __LITTLE_ENDIAN__
2942__ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1) {
2943  uint32x4_t __ret;
2944  __ret = (uint32x4_t)(__p0 > __p1);
2945  return __ret;
2946}
2947#else
2948__ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1) {
2949  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2950  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2951  uint32x4_t __ret;
2952  __ret = (uint32x4_t)(__rev0 > __rev1);
2953  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2954  return __ret;
2955}
2956#endif
2957
2958#ifdef __LITTLE_ENDIAN__
2959__ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1) {
2960  uint16x8_t __ret;
2961  __ret = (uint16x8_t)(__p0 > __p1);
2962  return __ret;
2963}
2964#else
2965__ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1) {
2966  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2967  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2968  uint16x8_t __ret;
2969  __ret = (uint16x8_t)(__rev0 > __rev1);
2970  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2971  return __ret;
2972}
2973#endif
2974
2975#ifdef __LITTLE_ENDIAN__
2976__ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1) {
2977  uint8x8_t __ret;
2978  __ret = (uint8x8_t)(__p0 > __p1);
2979  return __ret;
2980}
2981#else
2982__ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1) {
2983  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2984  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2985  uint8x8_t __ret;
2986  __ret = (uint8x8_t)(__rev0 > __rev1);
2987  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2988  return __ret;
2989}
2990#endif
2991
2992#ifdef __LITTLE_ENDIAN__
2993__ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1) {
2994  uint32x2_t __ret;
2995  __ret = (uint32x2_t)(__p0 > __p1);
2996  return __ret;
2997}
2998#else
2999__ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1) {
3000  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3001  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3002  uint32x2_t __ret;
3003  __ret = (uint32x2_t)(__rev0 > __rev1);
3004  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3005  return __ret;
3006}
3007#endif
3008
3009#ifdef __LITTLE_ENDIAN__
3010__ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3011  uint16x4_t __ret;
3012  __ret = (uint16x4_t)(__p0 > __p1);
3013  return __ret;
3014}
3015#else
3016__ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3017  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3018  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3019  uint16x4_t __ret;
3020  __ret = (uint16x4_t)(__rev0 > __rev1);
3021  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3022  return __ret;
3023}
3024#endif
3025
3026#ifdef __LITTLE_ENDIAN__
3027__ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1) {
3028  uint8x8_t __ret;
3029  __ret = (uint8x8_t)(__p0 > __p1);
3030  return __ret;
3031}
3032#else
3033__ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1) {
3034  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3035  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3036  uint8x8_t __ret;
3037  __ret = (uint8x8_t)(__rev0 > __rev1);
3038  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3039  return __ret;
3040}
3041#endif
3042
3043#ifdef __LITTLE_ENDIAN__
3044__ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1) {
3045  uint32x2_t __ret;
3046  __ret = (uint32x2_t)(__p0 > __p1);
3047  return __ret;
3048}
3049#else
3050__ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1) {
3051  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3052  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3053  uint32x2_t __ret;
3054  __ret = (uint32x2_t)(__rev0 > __rev1);
3055  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3056  return __ret;
3057}
3058#endif
3059
3060#ifdef __LITTLE_ENDIAN__
3061__ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1) {
3062  uint32x2_t __ret;
3063  __ret = (uint32x2_t)(__p0 > __p1);
3064  return __ret;
3065}
3066#else
3067__ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1) {
3068  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3069  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3070  uint32x2_t __ret;
3071  __ret = (uint32x2_t)(__rev0 > __rev1);
3072  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3073  return __ret;
3074}
3075#endif
3076
3077#ifdef __LITTLE_ENDIAN__
3078__ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1) {
3079  uint16x4_t __ret;
3080  __ret = (uint16x4_t)(__p0 > __p1);
3081  return __ret;
3082}
3083#else
3084__ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1) {
3085  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3086  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3087  uint16x4_t __ret;
3088  __ret = (uint16x4_t)(__rev0 > __rev1);
3089  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3090  return __ret;
3091}
3092#endif
3093
3094#ifdef __LITTLE_ENDIAN__
3095__ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3096  uint8x16_t __ret;
3097  __ret = (uint8x16_t)(__p0 <= __p1);
3098  return __ret;
3099}
3100#else
3101__ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3102  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3103  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3104  uint8x16_t __ret;
3105  __ret = (uint8x16_t)(__rev0 <= __rev1);
3106  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3107  return __ret;
3108}
3109#endif
3110
3111#ifdef __LITTLE_ENDIAN__
3112__ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3113  uint32x4_t __ret;
3114  __ret = (uint32x4_t)(__p0 <= __p1);
3115  return __ret;
3116}
3117#else
3118__ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3119  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3120  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3121  uint32x4_t __ret;
3122  __ret = (uint32x4_t)(__rev0 <= __rev1);
3123  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3124  return __ret;
3125}
3126#endif
3127
3128#ifdef __LITTLE_ENDIAN__
3129__ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3130  uint16x8_t __ret;
3131  __ret = (uint16x8_t)(__p0 <= __p1);
3132  return __ret;
3133}
3134#else
3135__ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3136  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3137  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3138  uint16x8_t __ret;
3139  __ret = (uint16x8_t)(__rev0 <= __rev1);
3140  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3141  return __ret;
3142}
3143#endif
3144
3145#ifdef __LITTLE_ENDIAN__
3146__ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1) {
3147  uint8x16_t __ret;
3148  __ret = (uint8x16_t)(__p0 <= __p1);
3149  return __ret;
3150}
3151#else
3152__ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1) {
3153  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3154  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3155  uint8x16_t __ret;
3156  __ret = (uint8x16_t)(__rev0 <= __rev1);
3157  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3158  return __ret;
3159}
3160#endif
3161
3162#ifdef __LITTLE_ENDIAN__
3163__ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1) {
3164  uint32x4_t __ret;
3165  __ret = (uint32x4_t)(__p0 <= __p1);
3166  return __ret;
3167}
3168#else
3169__ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1) {
3170  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3171  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3172  uint32x4_t __ret;
3173  __ret = (uint32x4_t)(__rev0 <= __rev1);
3174  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3175  return __ret;
3176}
3177#endif
3178
3179#ifdef __LITTLE_ENDIAN__
3180__ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1) {
3181  uint32x4_t __ret;
3182  __ret = (uint32x4_t)(__p0 <= __p1);
3183  return __ret;
3184}
3185#else
3186__ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1) {
3187  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3188  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3189  uint32x4_t __ret;
3190  __ret = (uint32x4_t)(__rev0 <= __rev1);
3191  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3192  return __ret;
3193}
3194#endif
3195
3196#ifdef __LITTLE_ENDIAN__
3197__ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1) {
3198  uint16x8_t __ret;
3199  __ret = (uint16x8_t)(__p0 <= __p1);
3200  return __ret;
3201}
3202#else
3203__ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1) {
3204  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3205  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3206  uint16x8_t __ret;
3207  __ret = (uint16x8_t)(__rev0 <= __rev1);
3208  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3209  return __ret;
3210}
3211#endif
3212
3213#ifdef __LITTLE_ENDIAN__
3214__ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1) {
3215  uint8x8_t __ret;
3216  __ret = (uint8x8_t)(__p0 <= __p1);
3217  return __ret;
3218}
3219#else
3220__ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1) {
3221  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3222  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3223  uint8x8_t __ret;
3224  __ret = (uint8x8_t)(__rev0 <= __rev1);
3225  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3226  return __ret;
3227}
3228#endif
3229
3230#ifdef __LITTLE_ENDIAN__
3231__ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1) {
3232  uint32x2_t __ret;
3233  __ret = (uint32x2_t)(__p0 <= __p1);
3234  return __ret;
3235}
3236#else
3237__ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1) {
3238  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3239  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3240  uint32x2_t __ret;
3241  __ret = (uint32x2_t)(__rev0 <= __rev1);
3242  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3243  return __ret;
3244}
3245#endif
3246
3247#ifdef __LITTLE_ENDIAN__
3248__ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1) {
3249  uint16x4_t __ret;
3250  __ret = (uint16x4_t)(__p0 <= __p1);
3251  return __ret;
3252}
3253#else
3254__ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1) {
3255  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3256  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3257  uint16x4_t __ret;
3258  __ret = (uint16x4_t)(__rev0 <= __rev1);
3259  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3260  return __ret;
3261}
3262#endif
3263
3264#ifdef __LITTLE_ENDIAN__
3265__ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1) {
3266  uint8x8_t __ret;
3267  __ret = (uint8x8_t)(__p0 <= __p1);
3268  return __ret;
3269}
3270#else
3271__ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1) {
3272  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3273  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3274  uint8x8_t __ret;
3275  __ret = (uint8x8_t)(__rev0 <= __rev1);
3276  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3277  return __ret;
3278}
3279#endif
3280
3281#ifdef __LITTLE_ENDIAN__
3282__ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1) {
3283  uint32x2_t __ret;
3284  __ret = (uint32x2_t)(__p0 <= __p1);
3285  return __ret;
3286}
3287#else
3288__ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1) {
3289  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3290  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3291  uint32x2_t __ret;
3292  __ret = (uint32x2_t)(__rev0 <= __rev1);
3293  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3294  return __ret;
3295}
3296#endif
3297
3298#ifdef __LITTLE_ENDIAN__
3299__ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1) {
3300  uint32x2_t __ret;
3301  __ret = (uint32x2_t)(__p0 <= __p1);
3302  return __ret;
3303}
3304#else
3305__ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1) {
3306  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3307  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3308  uint32x2_t __ret;
3309  __ret = (uint32x2_t)(__rev0 <= __rev1);
3310  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3311  return __ret;
3312}
3313#endif
3314
3315#ifdef __LITTLE_ENDIAN__
3316__ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) {
3317  uint16x4_t __ret;
3318  __ret = (uint16x4_t)(__p0 <= __p1);
3319  return __ret;
3320}
3321#else
3322__ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) {
3323  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3324  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3325  uint16x4_t __ret;
3326  __ret = (uint16x4_t)(__rev0 <= __rev1);
3327  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3328  return __ret;
3329}
3330#endif
3331
3332#ifdef __LITTLE_ENDIAN__
3333__ai int8x16_t vclsq_s8(int8x16_t __p0) {
3334  int8x16_t __ret;
3335  __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 32);
3336  return __ret;
3337}
3338#else
3339__ai int8x16_t vclsq_s8(int8x16_t __p0) {
3340  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3341  int8x16_t __ret;
3342  __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 32);
3343  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3344  return __ret;
3345}
3346#endif
3347
3348#ifdef __LITTLE_ENDIAN__
3349__ai int32x4_t vclsq_s32(int32x4_t __p0) {
3350  int32x4_t __ret;
3351  __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 34);
3352  return __ret;
3353}
3354#else
3355__ai int32x4_t vclsq_s32(int32x4_t __p0) {
3356  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3357  int32x4_t __ret;
3358  __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 34);
3359  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3360  return __ret;
3361}
3362#endif
3363
3364#ifdef __LITTLE_ENDIAN__
3365__ai int16x8_t vclsq_s16(int16x8_t __p0) {
3366  int16x8_t __ret;
3367  __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 33);
3368  return __ret;
3369}
3370#else
3371__ai int16x8_t vclsq_s16(int16x8_t __p0) {
3372  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3373  int16x8_t __ret;
3374  __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 33);
3375  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3376  return __ret;
3377}
3378#endif
3379
3380#ifdef __LITTLE_ENDIAN__
3381__ai int8x8_t vcls_s8(int8x8_t __p0) {
3382  int8x8_t __ret;
3383  __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__p0, 0);
3384  return __ret;
3385}
3386#else
3387__ai int8x8_t vcls_s8(int8x8_t __p0) {
3388  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3389  int8x8_t __ret;
3390  __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 0);
3391  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3392  return __ret;
3393}
3394#endif
3395
3396#ifdef __LITTLE_ENDIAN__
3397__ai int32x2_t vcls_s32(int32x2_t __p0) {
3398  int32x2_t __ret;
3399  __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__p0, 2);
3400  return __ret;
3401}
3402#else
3403__ai int32x2_t vcls_s32(int32x2_t __p0) {
3404  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3405  int32x2_t __ret;
3406  __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 2);
3407  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3408  return __ret;
3409}
3410#endif
3411
3412#ifdef __LITTLE_ENDIAN__
3413__ai int16x4_t vcls_s16(int16x4_t __p0) {
3414  int16x4_t __ret;
3415  __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__p0, 1);
3416  return __ret;
3417}
3418#else
3419__ai int16x4_t vcls_s16(int16x4_t __p0) {
3420  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3421  int16x4_t __ret;
3422  __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 1);
3423  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3424  return __ret;
3425}
3426#endif
3427
3428#ifdef __LITTLE_ENDIAN__
3429__ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3430  uint8x16_t __ret;
3431  __ret = (uint8x16_t)(__p0 < __p1);
3432  return __ret;
3433}
3434#else
3435__ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3436  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3437  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3438  uint8x16_t __ret;
3439  __ret = (uint8x16_t)(__rev0 < __rev1);
3440  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3441  return __ret;
3442}
3443#endif
3444
3445#ifdef __LITTLE_ENDIAN__
3446__ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3447  uint32x4_t __ret;
3448  __ret = (uint32x4_t)(__p0 < __p1);
3449  return __ret;
3450}
3451#else
3452__ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3453  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3454  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3455  uint32x4_t __ret;
3456  __ret = (uint32x4_t)(__rev0 < __rev1);
3457  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3458  return __ret;
3459}
3460#endif
3461
3462#ifdef __LITTLE_ENDIAN__
3463__ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3464  uint16x8_t __ret;
3465  __ret = (uint16x8_t)(__p0 < __p1);
3466  return __ret;
3467}
3468#else
3469__ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3470  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3471  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3472  uint16x8_t __ret;
3473  __ret = (uint16x8_t)(__rev0 < __rev1);
3474  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3475  return __ret;
3476}
3477#endif
3478
3479#ifdef __LITTLE_ENDIAN__
3480__ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1) {
3481  uint8x16_t __ret;
3482  __ret = (uint8x16_t)(__p0 < __p1);
3483  return __ret;
3484}
3485#else
3486__ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1) {
3487  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3488  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3489  uint8x16_t __ret;
3490  __ret = (uint8x16_t)(__rev0 < __rev1);
3491  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3492  return __ret;
3493}
3494#endif
3495
3496#ifdef __LITTLE_ENDIAN__
3497__ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1) {
3498  uint32x4_t __ret;
3499  __ret = (uint32x4_t)(__p0 < __p1);
3500  return __ret;
3501}
3502#else
3503__ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1) {
3504  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3505  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3506  uint32x4_t __ret;
3507  __ret = (uint32x4_t)(__rev0 < __rev1);
3508  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3509  return __ret;
3510}
3511#endif
3512
3513#ifdef __LITTLE_ENDIAN__
3514__ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1) {
3515  uint32x4_t __ret;
3516  __ret = (uint32x4_t)(__p0 < __p1);
3517  return __ret;
3518}
3519#else
3520__ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1) {
3521  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3522  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3523  uint32x4_t __ret;
3524  __ret = (uint32x4_t)(__rev0 < __rev1);
3525  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3526  return __ret;
3527}
3528#endif
3529
3530#ifdef __LITTLE_ENDIAN__
3531__ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1) {
3532  uint16x8_t __ret;
3533  __ret = (uint16x8_t)(__p0 < __p1);
3534  return __ret;
3535}
3536#else
3537__ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1) {
3538  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3539  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3540  uint16x8_t __ret;
3541  __ret = (uint16x8_t)(__rev0 < __rev1);
3542  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3543  return __ret;
3544}
3545#endif
3546
3547#ifdef __LITTLE_ENDIAN__
3548__ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1) {
3549  uint8x8_t __ret;
3550  __ret = (uint8x8_t)(__p0 < __p1);
3551  return __ret;
3552}
3553#else
3554__ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1) {
3555  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3556  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3557  uint8x8_t __ret;
3558  __ret = (uint8x8_t)(__rev0 < __rev1);
3559  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3560  return __ret;
3561}
3562#endif
3563
3564#ifdef __LITTLE_ENDIAN__
3565__ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1) {
3566  uint32x2_t __ret;
3567  __ret = (uint32x2_t)(__p0 < __p1);
3568  return __ret;
3569}
3570#else
3571__ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1) {
3572  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3573  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3574  uint32x2_t __ret;
3575  __ret = (uint32x2_t)(__rev0 < __rev1);
3576  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3577  return __ret;
3578}
3579#endif
3580
3581#ifdef __LITTLE_ENDIAN__
3582__ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3583  uint16x4_t __ret;
3584  __ret = (uint16x4_t)(__p0 < __p1);
3585  return __ret;
3586}
3587#else
3588__ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3589  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3590  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3591  uint16x4_t __ret;
3592  __ret = (uint16x4_t)(__rev0 < __rev1);
3593  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3594  return __ret;
3595}
3596#endif
3597
3598#ifdef __LITTLE_ENDIAN__
3599__ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1) {
3600  uint8x8_t __ret;
3601  __ret = (uint8x8_t)(__p0 < __p1);
3602  return __ret;
3603}
3604#else
3605__ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1) {
3606  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3607  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3608  uint8x8_t __ret;
3609  __ret = (uint8x8_t)(__rev0 < __rev1);
3610  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3611  return __ret;
3612}
3613#endif
3614
3615#ifdef __LITTLE_ENDIAN__
3616__ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1) {
3617  uint32x2_t __ret;
3618  __ret = (uint32x2_t)(__p0 < __p1);
3619  return __ret;
3620}
3621#else
3622__ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1) {
3623  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3624  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3625  uint32x2_t __ret;
3626  __ret = (uint32x2_t)(__rev0 < __rev1);
3627  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3628  return __ret;
3629}
3630#endif
3631
3632#ifdef __LITTLE_ENDIAN__
3633__ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1) {
3634  uint32x2_t __ret;
3635  __ret = (uint32x2_t)(__p0 < __p1);
3636  return __ret;
3637}
3638#else
3639__ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1) {
3640  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3641  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3642  uint32x2_t __ret;
3643  __ret = (uint32x2_t)(__rev0 < __rev1);
3644  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3645  return __ret;
3646}
3647#endif
3648
3649#ifdef __LITTLE_ENDIAN__
3650__ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1) {
3651  uint16x4_t __ret;
3652  __ret = (uint16x4_t)(__p0 < __p1);
3653  return __ret;
3654}
3655#else
3656__ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1) {
3657  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3658  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3659  uint16x4_t __ret;
3660  __ret = (uint16x4_t)(__rev0 < __rev1);
3661  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3662  return __ret;
3663}
3664#endif
3665
3666#ifdef __LITTLE_ENDIAN__
3667__ai uint8x16_t vclzq_u8(uint8x16_t __p0) {
3668  uint8x16_t __ret;
3669  __ret = (uint8x16_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 48);
3670  return __ret;
3671}
3672#else
3673__ai uint8x16_t vclzq_u8(uint8x16_t __p0) {
3674  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3675  uint8x16_t __ret;
3676  __ret = (uint8x16_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 48);
3677  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3678  return __ret;
3679}
3680#endif
3681
3682#ifdef __LITTLE_ENDIAN__
3683__ai uint32x4_t vclzq_u32(uint32x4_t __p0) {
3684  uint32x4_t __ret;
3685  __ret = (uint32x4_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 50);
3686  return __ret;
3687}
3688#else
3689__ai uint32x4_t vclzq_u32(uint32x4_t __p0) {
3690  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3691  uint32x4_t __ret;
3692  __ret = (uint32x4_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 50);
3693  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3694  return __ret;
3695}
3696#endif
3697
3698#ifdef __LITTLE_ENDIAN__
3699__ai uint16x8_t vclzq_u16(uint16x8_t __p0) {
3700  uint16x8_t __ret;
3701  __ret = (uint16x8_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 49);
3702  return __ret;
3703}
3704#else
3705__ai uint16x8_t vclzq_u16(uint16x8_t __p0) {
3706  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3707  uint16x8_t __ret;
3708  __ret = (uint16x8_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 49);
3709  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3710  return __ret;
3711}
3712#endif
3713
3714#ifdef __LITTLE_ENDIAN__
3715__ai int8x16_t vclzq_s8(int8x16_t __p0) {
3716  int8x16_t __ret;
3717  __ret = (int8x16_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 32);
3718  return __ret;
3719}
3720#else
3721__ai int8x16_t vclzq_s8(int8x16_t __p0) {
3722  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3723  int8x16_t __ret;
3724  __ret = (int8x16_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 32);
3725  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3726  return __ret;
3727}
3728#endif
3729
3730#ifdef __LITTLE_ENDIAN__
3731__ai int32x4_t vclzq_s32(int32x4_t __p0) {
3732  int32x4_t __ret;
3733  __ret = (int32x4_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 34);
3734  return __ret;
3735}
3736#else
3737__ai int32x4_t vclzq_s32(int32x4_t __p0) {
3738  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3739  int32x4_t __ret;
3740  __ret = (int32x4_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 34);
3741  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3742  return __ret;
3743}
3744#endif
3745
3746#ifdef __LITTLE_ENDIAN__
3747__ai int16x8_t vclzq_s16(int16x8_t __p0) {
3748  int16x8_t __ret;
3749  __ret = (int16x8_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 33);
3750  return __ret;
3751}
3752#else
3753__ai int16x8_t vclzq_s16(int16x8_t __p0) {
3754  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3755  int16x8_t __ret;
3756  __ret = (int16x8_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 33);
3757  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3758  return __ret;
3759}
3760#endif
3761
3762#ifdef __LITTLE_ENDIAN__
3763__ai uint8x8_t vclz_u8(uint8x8_t __p0) {
3764  uint8x8_t __ret;
3765  __ret = (uint8x8_t) __builtin_neon_vclz_v((int8x8_t)__p0, 16);
3766  return __ret;
3767}
3768#else
3769__ai uint8x8_t vclz_u8(uint8x8_t __p0) {
3770  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3771  uint8x8_t __ret;
3772  __ret = (uint8x8_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 16);
3773  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3774  return __ret;
3775}
3776#endif
3777
3778#ifdef __LITTLE_ENDIAN__
3779__ai uint32x2_t vclz_u32(uint32x2_t __p0) {
3780  uint32x2_t __ret;
3781  __ret = (uint32x2_t) __builtin_neon_vclz_v((int8x8_t)__p0, 18);
3782  return __ret;
3783}
3784#else
3785__ai uint32x2_t vclz_u32(uint32x2_t __p0) {
3786  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3787  uint32x2_t __ret;
3788  __ret = (uint32x2_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 18);
3789  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3790  return __ret;
3791}
3792#endif
3793
3794#ifdef __LITTLE_ENDIAN__
3795__ai uint16x4_t vclz_u16(uint16x4_t __p0) {
3796  uint16x4_t __ret;
3797  __ret = (uint16x4_t) __builtin_neon_vclz_v((int8x8_t)__p0, 17);
3798  return __ret;
3799}
3800#else
3801__ai uint16x4_t vclz_u16(uint16x4_t __p0) {
3802  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3803  uint16x4_t __ret;
3804  __ret = (uint16x4_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 17);
3805  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3806  return __ret;
3807}
3808#endif
3809
3810#ifdef __LITTLE_ENDIAN__
3811__ai int8x8_t vclz_s8(int8x8_t __p0) {
3812  int8x8_t __ret;
3813  __ret = (int8x8_t) __builtin_neon_vclz_v((int8x8_t)__p0, 0);
3814  return __ret;
3815}
3816#else
3817__ai int8x8_t vclz_s8(int8x8_t __p0) {
3818  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3819  int8x8_t __ret;
3820  __ret = (int8x8_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 0);
3821  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3822  return __ret;
3823}
3824#endif
3825
3826#ifdef __LITTLE_ENDIAN__
3827__ai int32x2_t vclz_s32(int32x2_t __p0) {
3828  int32x2_t __ret;
3829  __ret = (int32x2_t) __builtin_neon_vclz_v((int8x8_t)__p0, 2);
3830  return __ret;
3831}
3832#else
3833__ai int32x2_t vclz_s32(int32x2_t __p0) {
3834  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3835  int32x2_t __ret;
3836  __ret = (int32x2_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 2);
3837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3838  return __ret;
3839}
3840#endif
3841
3842#ifdef __LITTLE_ENDIAN__
3843__ai int16x4_t vclz_s16(int16x4_t __p0) {
3844  int16x4_t __ret;
3845  __ret = (int16x4_t) __builtin_neon_vclz_v((int8x8_t)__p0, 1);
3846  return __ret;
3847}
3848#else
3849__ai int16x4_t vclz_s16(int16x4_t __p0) {
3850  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3851  int16x4_t __ret;
3852  __ret = (int16x4_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 1);
3853  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3854  return __ret;
3855}
3856#endif
3857
3858#ifdef __LITTLE_ENDIAN__
3859__ai poly8x8_t vcnt_p8(poly8x8_t __p0) {
3860  poly8x8_t __ret;
3861  __ret = (poly8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 4);
3862  return __ret;
3863}
3864#else
3865__ai poly8x8_t vcnt_p8(poly8x8_t __p0) {
3866  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3867  poly8x8_t __ret;
3868  __ret = (poly8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 4);
3869  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3870  return __ret;
3871}
3872#endif
3873
3874#ifdef __LITTLE_ENDIAN__
3875__ai poly8x16_t vcntq_p8(poly8x16_t __p0) {
3876  poly8x16_t __ret;
3877  __ret = (poly8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 36);
3878  return __ret;
3879}
3880#else
3881__ai poly8x16_t vcntq_p8(poly8x16_t __p0) {
3882  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3883  poly8x16_t __ret;
3884  __ret = (poly8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 36);
3885  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3886  return __ret;
3887}
3888#endif
3889
3890#ifdef __LITTLE_ENDIAN__
3891__ai uint8x16_t vcntq_u8(uint8x16_t __p0) {
3892  uint8x16_t __ret;
3893  __ret = (uint8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 48);
3894  return __ret;
3895}
3896#else
3897__ai uint8x16_t vcntq_u8(uint8x16_t __p0) {
3898  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3899  uint8x16_t __ret;
3900  __ret = (uint8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 48);
3901  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3902  return __ret;
3903}
3904#endif
3905
3906#ifdef __LITTLE_ENDIAN__
3907__ai int8x16_t vcntq_s8(int8x16_t __p0) {
3908  int8x16_t __ret;
3909  __ret = (int8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 32);
3910  return __ret;
3911}
3912#else
3913__ai int8x16_t vcntq_s8(int8x16_t __p0) {
3914  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3915  int8x16_t __ret;
3916  __ret = (int8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 32);
3917  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3918  return __ret;
3919}
3920#endif
3921
3922#ifdef __LITTLE_ENDIAN__
3923__ai uint8x8_t vcnt_u8(uint8x8_t __p0) {
3924  uint8x8_t __ret;
3925  __ret = (uint8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 16);
3926  return __ret;
3927}
3928#else
3929__ai uint8x8_t vcnt_u8(uint8x8_t __p0) {
3930  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3931  uint8x8_t __ret;
3932  __ret = (uint8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 16);
3933  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3934  return __ret;
3935}
3936#endif
3937
3938#ifdef __LITTLE_ENDIAN__
3939__ai int8x8_t vcnt_s8(int8x8_t __p0) {
3940  int8x8_t __ret;
3941  __ret = (int8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 0);
3942  return __ret;
3943}
3944#else
3945__ai int8x8_t vcnt_s8(int8x8_t __p0) {
3946  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3947  int8x8_t __ret;
3948  __ret = (int8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 0);
3949  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3950  return __ret;
3951}
3952#endif
3953
3954#ifdef __LITTLE_ENDIAN__
3955__ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1) {
3956  poly8x16_t __ret;
3957  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
3958  return __ret;
3959}
3960#else
3961__ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1) {
3962  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3963  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3964  poly8x16_t __ret;
3965  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
3966  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3967  return __ret;
3968}
3969#endif
3970
3971#ifdef __LITTLE_ENDIAN__
3972__ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1) {
3973  poly16x8_t __ret;
3974  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
3975  return __ret;
3976}
3977#else
3978__ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1) {
3979  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3980  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3981  poly16x8_t __ret;
3982  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
3983  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3984  return __ret;
3985}
3986#endif
3987
3988#ifdef __LITTLE_ENDIAN__
3989__ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
3990  uint8x16_t __ret;
3991  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
3992  return __ret;
3993}
3994#else
3995__ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
3996  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3997  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3998  uint8x16_t __ret;
3999  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4000  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4001  return __ret;
4002}
4003__ai uint8x16_t __noswap_vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
4004  uint8x16_t __ret;
4005  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4006  return __ret;
4007}
4008#endif
4009
4010#ifdef __LITTLE_ENDIAN__
4011__ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
4012  uint32x4_t __ret;
4013  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4014  return __ret;
4015}
4016#else
4017__ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
4018  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4019  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4020  uint32x4_t __ret;
4021  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
4022  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4023  return __ret;
4024}
4025__ai uint32x4_t __noswap_vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
4026  uint32x4_t __ret;
4027  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4028  return __ret;
4029}
4030#endif
4031
4032#ifdef __LITTLE_ENDIAN__
4033__ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1) {
4034  uint64x2_t __ret;
4035  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4036  return __ret;
4037}
4038#else
4039__ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1) {
4040  uint64x2_t __ret;
4041  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4042  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4043  return __ret;
4044}
4045#endif
4046
4047#ifdef __LITTLE_ENDIAN__
4048__ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
4049  uint16x8_t __ret;
4050  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4051  return __ret;
4052}
4053#else
4054__ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
4055  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4056  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4057  uint16x8_t __ret;
4058  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
4059  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4060  return __ret;
4061}
4062__ai uint16x8_t __noswap_vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
4063  uint16x8_t __ret;
4064  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4065  return __ret;
4066}
4067#endif
4068
4069#ifdef __LITTLE_ENDIAN__
4070__ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
4071  int8x16_t __ret;
4072  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4073  return __ret;
4074}
4075#else
4076__ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
4077  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4078  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4079  int8x16_t __ret;
4080  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4081  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4082  return __ret;
4083}
4084__ai int8x16_t __noswap_vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
4085  int8x16_t __ret;
4086  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4087  return __ret;
4088}
4089#endif
4090
4091#ifdef __LITTLE_ENDIAN__
4092__ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
4093  float32x4_t __ret;
4094  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4095  return __ret;
4096}
4097#else
4098__ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
4099  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4100  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4101  float32x4_t __ret;
4102  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
4103  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4104  return __ret;
4105}
4106__ai float32x4_t __noswap_vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
4107  float32x4_t __ret;
4108  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4109  return __ret;
4110}
4111#endif
4112
4113#ifdef __LITTLE_ENDIAN__
4114__ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
4115  float16x8_t __ret;
4116  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4117  return __ret;
4118}
4119#else
4120__ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
4121  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4122  float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4123  float16x8_t __ret;
4124  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
4125  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4126  return __ret;
4127}
4128__ai float16x8_t __noswap_vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
4129  float16x8_t __ret;
4130  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4131  return __ret;
4132}
4133#endif
4134
4135#ifdef __LITTLE_ENDIAN__
4136__ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
4137  int32x4_t __ret;
4138  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4139  return __ret;
4140}
4141#else
4142__ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
4143  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4144  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4145  int32x4_t __ret;
4146  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
4147  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4148  return __ret;
4149}
4150__ai int32x4_t __noswap_vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
4151  int32x4_t __ret;
4152  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4153  return __ret;
4154}
4155#endif
4156
4157#ifdef __LITTLE_ENDIAN__
4158__ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1) {
4159  int64x2_t __ret;
4160  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4161  return __ret;
4162}
4163#else
4164__ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1) {
4165  int64x2_t __ret;
4166  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4167  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4168  return __ret;
4169}
4170#endif
4171
4172#ifdef __LITTLE_ENDIAN__
4173__ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
4174  int16x8_t __ret;
4175  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4176  return __ret;
4177}
4178#else
4179__ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
4180  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4181  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4182  int16x8_t __ret;
4183  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
4184  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4185  return __ret;
4186}
4187__ai int16x8_t __noswap_vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
4188  int16x8_t __ret;
4189  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4190  return __ret;
4191}
4192#endif
4193
4194#ifdef __LITTLE_ENDIAN__
4195__ai poly8x8_t vcreate_p8(uint64_t __p0) {
4196  poly8x8_t __ret;
4197  __ret = (poly8x8_t)(__p0);
4198  return __ret;
4199}
4200#else
4201__ai poly8x8_t vcreate_p8(uint64_t __p0) {
4202  poly8x8_t __ret;
4203  __ret = (poly8x8_t)(__p0);
4204  return __ret;
4205}
4206#endif
4207
4208#ifdef __LITTLE_ENDIAN__
4209__ai poly16x4_t vcreate_p16(uint64_t __p0) {
4210  poly16x4_t __ret;
4211  __ret = (poly16x4_t)(__p0);
4212  return __ret;
4213}
4214#else
4215__ai poly16x4_t vcreate_p16(uint64_t __p0) {
4216  poly16x4_t __ret;
4217  __ret = (poly16x4_t)(__p0);
4218  return __ret;
4219}
4220#endif
4221
4222#ifdef __LITTLE_ENDIAN__
4223__ai uint8x8_t vcreate_u8(uint64_t __p0) {
4224  uint8x8_t __ret;
4225  __ret = (uint8x8_t)(__p0);
4226  return __ret;
4227}
4228#else
4229__ai uint8x8_t vcreate_u8(uint64_t __p0) {
4230  uint8x8_t __ret;
4231  __ret = (uint8x8_t)(__p0);
4232  return __ret;
4233}
4234#endif
4235
4236#ifdef __LITTLE_ENDIAN__
4237__ai uint32x2_t vcreate_u32(uint64_t __p0) {
4238  uint32x2_t __ret;
4239  __ret = (uint32x2_t)(__p0);
4240  return __ret;
4241}
4242#else
4243__ai uint32x2_t vcreate_u32(uint64_t __p0) {
4244  uint32x2_t __ret;
4245  __ret = (uint32x2_t)(__p0);
4246  return __ret;
4247}
4248#endif
4249
4250#ifdef __LITTLE_ENDIAN__
4251__ai uint64x1_t vcreate_u64(uint64_t __p0) {
4252  uint64x1_t __ret;
4253  __ret = (uint64x1_t)(__p0);
4254  return __ret;
4255}
4256#else
4257__ai uint64x1_t vcreate_u64(uint64_t __p0) {
4258  uint64x1_t __ret;
4259  __ret = (uint64x1_t)(__p0);
4260  return __ret;
4261}
4262#endif
4263
4264#ifdef __LITTLE_ENDIAN__
4265__ai uint16x4_t vcreate_u16(uint64_t __p0) {
4266  uint16x4_t __ret;
4267  __ret = (uint16x4_t)(__p0);
4268  return __ret;
4269}
4270#else
4271__ai uint16x4_t vcreate_u16(uint64_t __p0) {
4272  uint16x4_t __ret;
4273  __ret = (uint16x4_t)(__p0);
4274  return __ret;
4275}
4276#endif
4277
4278#ifdef __LITTLE_ENDIAN__
4279__ai int8x8_t vcreate_s8(uint64_t __p0) {
4280  int8x8_t __ret;
4281  __ret = (int8x8_t)(__p0);
4282  return __ret;
4283}
4284#else
4285__ai int8x8_t vcreate_s8(uint64_t __p0) {
4286  int8x8_t __ret;
4287  __ret = (int8x8_t)(__p0);
4288  return __ret;
4289}
4290#endif
4291
4292#ifdef __LITTLE_ENDIAN__
4293__ai float32x2_t vcreate_f32(uint64_t __p0) {
4294  float32x2_t __ret;
4295  __ret = (float32x2_t)(__p0);
4296  return __ret;
4297}
4298#else
4299__ai float32x2_t vcreate_f32(uint64_t __p0) {
4300  float32x2_t __ret;
4301  __ret = (float32x2_t)(__p0);
4302  return __ret;
4303}
4304#endif
4305
4306#ifdef __LITTLE_ENDIAN__
4307__ai float16x4_t vcreate_f16(uint64_t __p0) {
4308  float16x4_t __ret;
4309  __ret = (float16x4_t)(__p0);
4310  return __ret;
4311}
4312#else
4313__ai float16x4_t vcreate_f16(uint64_t __p0) {
4314  float16x4_t __ret;
4315  __ret = (float16x4_t)(__p0);
4316  return __ret;
4317}
4318#endif
4319
4320#ifdef __LITTLE_ENDIAN__
4321__ai int32x2_t vcreate_s32(uint64_t __p0) {
4322  int32x2_t __ret;
4323  __ret = (int32x2_t)(__p0);
4324  return __ret;
4325}
4326#else
4327__ai int32x2_t vcreate_s32(uint64_t __p0) {
4328  int32x2_t __ret;
4329  __ret = (int32x2_t)(__p0);
4330  return __ret;
4331}
4332#endif
4333
4334#ifdef __LITTLE_ENDIAN__
4335__ai int64x1_t vcreate_s64(uint64_t __p0) {
4336  int64x1_t __ret;
4337  __ret = (int64x1_t)(__p0);
4338  return __ret;
4339}
4340#else
4341__ai int64x1_t vcreate_s64(uint64_t __p0) {
4342  int64x1_t __ret;
4343  __ret = (int64x1_t)(__p0);
4344  return __ret;
4345}
4346#endif
4347
4348#ifdef __LITTLE_ENDIAN__
4349__ai int16x4_t vcreate_s16(uint64_t __p0) {
4350  int16x4_t __ret;
4351  __ret = (int16x4_t)(__p0);
4352  return __ret;
4353}
4354#else
4355__ai int16x4_t vcreate_s16(uint64_t __p0) {
4356  int16x4_t __ret;
4357  __ret = (int16x4_t)(__p0);
4358  return __ret;
4359}
4360#endif
4361
4362#ifdef __LITTLE_ENDIAN__
4363__ai float16x4_t vcvt_f16_f32(float32x4_t __p0) {
4364  float16x4_t __ret;
4365  __ret = (float16x4_t) __builtin_neon_vcvt_f16_v((int8x16_t)__p0, 8);
4366  return __ret;
4367}
4368#else
4369__ai float16x4_t vcvt_f16_f32(float32x4_t __p0) {
4370  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4371  float16x4_t __ret;
4372  __ret = (float16x4_t) __builtin_neon_vcvt_f16_v((int8x16_t)__rev0, 8);
4373  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4374  return __ret;
4375}
4376__ai float16x4_t __noswap_vcvt_f16_f32(float32x4_t __p0) {
4377  float16x4_t __ret;
4378  __ret = (float16x4_t) __builtin_neon_vcvt_f16_v((int8x16_t)__p0, 8);
4379  return __ret;
4380}
4381#endif
4382
4383#ifdef __LITTLE_ENDIAN__
4384__ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0) {
4385  float32x4_t __ret;
4386  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__p0, 50);
4387  return __ret;
4388}
4389#else
4390__ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0) {
4391  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4392  float32x4_t __ret;
4393  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__rev0, 50);
4394  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4395  return __ret;
4396}
4397#endif
4398
4399#ifdef __LITTLE_ENDIAN__
4400__ai float32x4_t vcvtq_f32_s32(int32x4_t __p0) {
4401  float32x4_t __ret;
4402  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__p0, 34);
4403  return __ret;
4404}
4405#else
4406__ai float32x4_t vcvtq_f32_s32(int32x4_t __p0) {
4407  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4408  float32x4_t __ret;
4409  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__rev0, 34);
4410  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4411  return __ret;
4412}
4413#endif
4414
4415#ifdef __LITTLE_ENDIAN__
4416__ai float32x2_t vcvt_f32_u32(uint32x2_t __p0) {
4417  float32x2_t __ret;
4418  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__p0, 18);
4419  return __ret;
4420}
4421#else
4422__ai float32x2_t vcvt_f32_u32(uint32x2_t __p0) {
4423  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4424  float32x2_t __ret;
4425  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__rev0, 18);
4426  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4427  return __ret;
4428}
4429#endif
4430
4431#ifdef __LITTLE_ENDIAN__
4432__ai float32x2_t vcvt_f32_s32(int32x2_t __p0) {
4433  float32x2_t __ret;
4434  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__p0, 2);
4435  return __ret;
4436}
4437#else
4438__ai float32x2_t vcvt_f32_s32(int32x2_t __p0) {
4439  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4440  float32x2_t __ret;
4441  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__rev0, 2);
4442  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4443  return __ret;
4444}
4445#endif
4446
4447#ifdef __LITTLE_ENDIAN__
4448__ai float32x4_t vcvt_f32_f16(float16x4_t __p0) {
4449  float32x4_t __ret;
4450  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__p0, 8);
4451  return __ret;
4452}
4453#else
4454__ai float32x4_t vcvt_f32_f16(float16x4_t __p0) {
4455  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4456  float32x4_t __ret;
4457  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__rev0, 8);
4458  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4459  return __ret;
4460}
4461__ai float32x4_t __noswap_vcvt_f32_f16(float16x4_t __p0) {
4462  float32x4_t __ret;
4463  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__p0, 8);
4464  return __ret;
4465}
4466#endif
4467
4468#ifdef __LITTLE_ENDIAN__
4469#define vcvtq_n_f32_u32(__p0, __p1) __extension__ ({ \
4470  uint32x4_t __s0 = __p0; \
4471  float32x4_t __ret; \
4472  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__s0, __p1, 50); \
4473  __ret; \
4474})
4475#else
4476#define vcvtq_n_f32_u32(__p0, __p1) __extension__ ({ \
4477  uint32x4_t __s0 = __p0; \
4478  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4479  float32x4_t __ret; \
4480  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__rev0, __p1, 50); \
4481  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4482  __ret; \
4483})
4484#endif
4485
4486#ifdef __LITTLE_ENDIAN__
4487#define vcvtq_n_f32_s32(__p0, __p1) __extension__ ({ \
4488  int32x4_t __s0 = __p0; \
4489  float32x4_t __ret; \
4490  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__s0, __p1, 34); \
4491  __ret; \
4492})
4493#else
4494#define vcvtq_n_f32_s32(__p0, __p1) __extension__ ({ \
4495  int32x4_t __s0 = __p0; \
4496  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4497  float32x4_t __ret; \
4498  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__rev0, __p1, 34); \
4499  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4500  __ret; \
4501})
4502#endif
4503
4504#ifdef __LITTLE_ENDIAN__
4505#define vcvt_n_f32_u32(__p0, __p1) __extension__ ({ \
4506  uint32x2_t __s0 = __p0; \
4507  float32x2_t __ret; \
4508  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__s0, __p1, 18); \
4509  __ret; \
4510})
4511#else
4512#define vcvt_n_f32_u32(__p0, __p1) __extension__ ({ \
4513  uint32x2_t __s0 = __p0; \
4514  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4515  float32x2_t __ret; \
4516  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__rev0, __p1, 18); \
4517  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4518  __ret; \
4519})
4520#endif
4521
4522#ifdef __LITTLE_ENDIAN__
4523#define vcvt_n_f32_s32(__p0, __p1) __extension__ ({ \
4524  int32x2_t __s0 = __p0; \
4525  float32x2_t __ret; \
4526  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__s0, __p1, 2); \
4527  __ret; \
4528})
4529#else
4530#define vcvt_n_f32_s32(__p0, __p1) __extension__ ({ \
4531  int32x2_t __s0 = __p0; \
4532  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4533  float32x2_t __ret; \
4534  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__rev0, __p1, 2); \
4535  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4536  __ret; \
4537})
4538#endif
4539
4540#ifdef __LITTLE_ENDIAN__
4541#define vcvtq_n_s32_f32(__p0, __p1) __extension__ ({ \
4542  float32x4_t __s0 = __p0; \
4543  int32x4_t __ret; \
4544  __ret = (int32x4_t) __builtin_neon_vcvtq_n_s32_v((int8x16_t)__s0, __p1, 34); \
4545  __ret; \
4546})
4547#else
4548#define vcvtq_n_s32_f32(__p0, __p1) __extension__ ({ \
4549  float32x4_t __s0 = __p0; \
4550  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4551  int32x4_t __ret; \
4552  __ret = (int32x4_t) __builtin_neon_vcvtq_n_s32_v((int8x16_t)__rev0, __p1, 34); \
4553  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4554  __ret; \
4555})
4556#endif
4557
4558#ifdef __LITTLE_ENDIAN__
4559#define vcvt_n_s32_f32(__p0, __p1) __extension__ ({ \
4560  float32x2_t __s0 = __p0; \
4561  int32x2_t __ret; \
4562  __ret = (int32x2_t) __builtin_neon_vcvt_n_s32_v((int8x8_t)__s0, __p1, 2); \
4563  __ret; \
4564})
4565#else
4566#define vcvt_n_s32_f32(__p0, __p1) __extension__ ({ \
4567  float32x2_t __s0 = __p0; \
4568  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4569  int32x2_t __ret; \
4570  __ret = (int32x2_t) __builtin_neon_vcvt_n_s32_v((int8x8_t)__rev0, __p1, 2); \
4571  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4572  __ret; \
4573})
4574#endif
4575
4576#ifdef __LITTLE_ENDIAN__
4577#define vcvtq_n_u32_f32(__p0, __p1) __extension__ ({ \
4578  float32x4_t __s0 = __p0; \
4579  uint32x4_t __ret; \
4580  __ret = (uint32x4_t) __builtin_neon_vcvtq_n_u32_v((int8x16_t)__s0, __p1, 50); \
4581  __ret; \
4582})
4583#else
4584#define vcvtq_n_u32_f32(__p0, __p1) __extension__ ({ \
4585  float32x4_t __s0 = __p0; \
4586  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4587  uint32x4_t __ret; \
4588  __ret = (uint32x4_t) __builtin_neon_vcvtq_n_u32_v((int8x16_t)__rev0, __p1, 50); \
4589  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4590  __ret; \
4591})
4592#endif
4593
4594#ifdef __LITTLE_ENDIAN__
4595#define vcvt_n_u32_f32(__p0, __p1) __extension__ ({ \
4596  float32x2_t __s0 = __p0; \
4597  uint32x2_t __ret; \
4598  __ret = (uint32x2_t) __builtin_neon_vcvt_n_u32_v((int8x8_t)__s0, __p1, 18); \
4599  __ret; \
4600})
4601#else
4602#define vcvt_n_u32_f32(__p0, __p1) __extension__ ({ \
4603  float32x2_t __s0 = __p0; \
4604  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4605  uint32x2_t __ret; \
4606  __ret = (uint32x2_t) __builtin_neon_vcvt_n_u32_v((int8x8_t)__rev0, __p1, 18); \
4607  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4608  __ret; \
4609})
4610#endif
4611
4612#ifdef __LITTLE_ENDIAN__
4613__ai int32x4_t vcvtq_s32_f32(float32x4_t __p0) {
4614  int32x4_t __ret;
4615  __ret = (int32x4_t) __builtin_neon_vcvtq_s32_v((int8x16_t)__p0, 34);
4616  return __ret;
4617}
4618#else
4619__ai int32x4_t vcvtq_s32_f32(float32x4_t __p0) {
4620  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4621  int32x4_t __ret;
4622  __ret = (int32x4_t) __builtin_neon_vcvtq_s32_v((int8x16_t)__rev0, 34);
4623  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4624  return __ret;
4625}
4626#endif
4627
4628#ifdef __LITTLE_ENDIAN__
4629__ai int32x2_t vcvt_s32_f32(float32x2_t __p0) {
4630  int32x2_t __ret;
4631  __ret = (int32x2_t) __builtin_neon_vcvt_s32_v((int8x8_t)__p0, 2);
4632  return __ret;
4633}
4634#else
4635__ai int32x2_t vcvt_s32_f32(float32x2_t __p0) {
4636  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4637  int32x2_t __ret;
4638  __ret = (int32x2_t) __builtin_neon_vcvt_s32_v((int8x8_t)__rev0, 2);
4639  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4640  return __ret;
4641}
4642#endif
4643
4644#ifdef __LITTLE_ENDIAN__
4645__ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0) {
4646  uint32x4_t __ret;
4647  __ret = (uint32x4_t) __builtin_neon_vcvtq_u32_v((int8x16_t)__p0, 50);
4648  return __ret;
4649}
4650#else
4651__ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0) {
4652  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4653  uint32x4_t __ret;
4654  __ret = (uint32x4_t) __builtin_neon_vcvtq_u32_v((int8x16_t)__rev0, 50);
4655  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4656  return __ret;
4657}
4658#endif
4659
4660#ifdef __LITTLE_ENDIAN__
4661__ai uint32x2_t vcvt_u32_f32(float32x2_t __p0) {
4662  uint32x2_t __ret;
4663  __ret = (uint32x2_t) __builtin_neon_vcvt_u32_v((int8x8_t)__p0, 18);
4664  return __ret;
4665}
4666#else
4667__ai uint32x2_t vcvt_u32_f32(float32x2_t __p0) {
4668  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4669  uint32x2_t __ret;
4670  __ret = (uint32x2_t) __builtin_neon_vcvt_u32_v((int8x8_t)__rev0, 18);
4671  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4672  return __ret;
4673}
4674#endif
4675
4676#ifdef __LITTLE_ENDIAN__
4677#define vdup_lane_p8(__p0, __p1) __extension__ ({ \
4678  poly8x8_t __s0 = __p0; \
4679  poly8x8_t __ret; \
4680  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4681  __ret; \
4682})
4683#else
4684#define vdup_lane_p8(__p0, __p1) __extension__ ({ \
4685  poly8x8_t __s0 = __p0; \
4686  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4687  poly8x8_t __ret; \
4688  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4689  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4690  __ret; \
4691})
4692#endif
4693
4694#ifdef __LITTLE_ENDIAN__
4695#define vdup_lane_p16(__p0, __p1) __extension__ ({ \
4696  poly16x4_t __s0 = __p0; \
4697  poly16x4_t __ret; \
4698  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4699  __ret; \
4700})
4701#else
4702#define vdup_lane_p16(__p0, __p1) __extension__ ({ \
4703  poly16x4_t __s0 = __p0; \
4704  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4705  poly16x4_t __ret; \
4706  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4707  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4708  __ret; \
4709})
4710#endif
4711
4712#ifdef __LITTLE_ENDIAN__
4713#define vdupq_lane_p8(__p0, __p1) __extension__ ({ \
4714  poly8x8_t __s0 = __p0; \
4715  poly8x16_t __ret; \
4716  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4717  __ret; \
4718})
4719#else
4720#define vdupq_lane_p8(__p0, __p1) __extension__ ({ \
4721  poly8x8_t __s0 = __p0; \
4722  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4723  poly8x16_t __ret; \
4724  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4725  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
4726  __ret; \
4727})
4728#endif
4729
4730#ifdef __LITTLE_ENDIAN__
4731#define vdupq_lane_p16(__p0, __p1) __extension__ ({ \
4732  poly16x4_t __s0 = __p0; \
4733  poly16x8_t __ret; \
4734  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4735  __ret; \
4736})
4737#else
4738#define vdupq_lane_p16(__p0, __p1) __extension__ ({ \
4739  poly16x4_t __s0 = __p0; \
4740  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4741  poly16x8_t __ret; \
4742  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4743  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4744  __ret; \
4745})
4746#endif
4747
4748#ifdef __LITTLE_ENDIAN__
4749#define vdupq_lane_u8(__p0, __p1) __extension__ ({ \
4750  uint8x8_t __s0 = __p0; \
4751  uint8x16_t __ret; \
4752  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4753  __ret; \
4754})
4755#else
4756#define vdupq_lane_u8(__p0, __p1) __extension__ ({ \
4757  uint8x8_t __s0 = __p0; \
4758  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4759  uint8x16_t __ret; \
4760  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4761  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
4762  __ret; \
4763})
4764#endif
4765
4766#ifdef __LITTLE_ENDIAN__
4767#define vdupq_lane_u32(__p0, __p1) __extension__ ({ \
4768  uint32x2_t __s0 = __p0; \
4769  uint32x4_t __ret; \
4770  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4771  __ret; \
4772})
4773#else
4774#define vdupq_lane_u32(__p0, __p1) __extension__ ({ \
4775  uint32x2_t __s0 = __p0; \
4776  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4777  uint32x4_t __ret; \
4778  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4779  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4780  __ret; \
4781})
4782#endif
4783
4784#ifdef __LITTLE_ENDIAN__
4785#define vdupq_lane_u64(__p0, __p1) __extension__ ({ \
4786  uint64x1_t __s0 = __p0; \
4787  uint64x2_t __ret; \
4788  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4789  __ret; \
4790})
4791#else
4792#define vdupq_lane_u64(__p0, __p1) __extension__ ({ \
4793  uint64x1_t __s0 = __p0; \
4794  uint64x2_t __ret; \
4795  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4796  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4797  __ret; \
4798})
4799#endif
4800
4801#ifdef __LITTLE_ENDIAN__
4802#define vdupq_lane_u16(__p0, __p1) __extension__ ({ \
4803  uint16x4_t __s0 = __p0; \
4804  uint16x8_t __ret; \
4805  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4806  __ret; \
4807})
4808#else
4809#define vdupq_lane_u16(__p0, __p1) __extension__ ({ \
4810  uint16x4_t __s0 = __p0; \
4811  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4812  uint16x8_t __ret; \
4813  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4814  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4815  __ret; \
4816})
4817#endif
4818
4819#ifdef __LITTLE_ENDIAN__
4820#define vdupq_lane_s8(__p0, __p1) __extension__ ({ \
4821  int8x8_t __s0 = __p0; \
4822  int8x16_t __ret; \
4823  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4824  __ret; \
4825})
4826#else
4827#define vdupq_lane_s8(__p0, __p1) __extension__ ({ \
4828  int8x8_t __s0 = __p0; \
4829  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4830  int8x16_t __ret; \
4831  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4832  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
4833  __ret; \
4834})
4835#endif
4836
4837#ifdef __LITTLE_ENDIAN__
4838#define vdupq_lane_f32(__p0, __p1) __extension__ ({ \
4839  float32x2_t __s0 = __p0; \
4840  float32x4_t __ret; \
4841  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4842  __ret; \
4843})
4844#else
4845#define vdupq_lane_f32(__p0, __p1) __extension__ ({ \
4846  float32x2_t __s0 = __p0; \
4847  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4848  float32x4_t __ret; \
4849  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4850  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4851  __ret; \
4852})
4853#endif
4854
4855#ifdef __LITTLE_ENDIAN__
4856#define vdupq_lane_s32(__p0, __p1) __extension__ ({ \
4857  int32x2_t __s0 = __p0; \
4858  int32x4_t __ret; \
4859  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4860  __ret; \
4861})
4862#else
4863#define vdupq_lane_s32(__p0, __p1) __extension__ ({ \
4864  int32x2_t __s0 = __p0; \
4865  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4866  int32x4_t __ret; \
4867  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4868  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4869  __ret; \
4870})
4871#endif
4872
4873#ifdef __LITTLE_ENDIAN__
4874#define vdupq_lane_s64(__p0, __p1) __extension__ ({ \
4875  int64x1_t __s0 = __p0; \
4876  int64x2_t __ret; \
4877  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4878  __ret; \
4879})
4880#else
4881#define vdupq_lane_s64(__p0, __p1) __extension__ ({ \
4882  int64x1_t __s0 = __p0; \
4883  int64x2_t __ret; \
4884  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4885  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4886  __ret; \
4887})
4888#endif
4889
4890#ifdef __LITTLE_ENDIAN__
4891#define vdupq_lane_s16(__p0, __p1) __extension__ ({ \
4892  int16x4_t __s0 = __p0; \
4893  int16x8_t __ret; \
4894  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4895  __ret; \
4896})
4897#else
4898#define vdupq_lane_s16(__p0, __p1) __extension__ ({ \
4899  int16x4_t __s0 = __p0; \
4900  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4901  int16x8_t __ret; \
4902  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4903  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4904  __ret; \
4905})
4906#endif
4907
4908#ifdef __LITTLE_ENDIAN__
4909#define vdup_lane_u8(__p0, __p1) __extension__ ({ \
4910  uint8x8_t __s0 = __p0; \
4911  uint8x8_t __ret; \
4912  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4913  __ret; \
4914})
4915#else
4916#define vdup_lane_u8(__p0, __p1) __extension__ ({ \
4917  uint8x8_t __s0 = __p0; \
4918  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4919  uint8x8_t __ret; \
4920  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4921  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4922  __ret; \
4923})
4924#endif
4925
4926#ifdef __LITTLE_ENDIAN__
4927#define vdup_lane_u32(__p0, __p1) __extension__ ({ \
4928  uint32x2_t __s0 = __p0; \
4929  uint32x2_t __ret; \
4930  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4931  __ret; \
4932})
4933#else
4934#define vdup_lane_u32(__p0, __p1) __extension__ ({ \
4935  uint32x2_t __s0 = __p0; \
4936  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4937  uint32x2_t __ret; \
4938  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
4939  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4940  __ret; \
4941})
4942#endif
4943
4944#ifdef __LITTLE_ENDIAN__
4945#define vdup_lane_u64(__p0, __p1) __extension__ ({ \
4946  uint64x1_t __s0 = __p0; \
4947  uint64x1_t __ret; \
4948  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
4949  __ret; \
4950})
4951#else
4952#define vdup_lane_u64(__p0, __p1) __extension__ ({ \
4953  uint64x1_t __s0 = __p0; \
4954  uint64x1_t __ret; \
4955  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
4956  __ret; \
4957})
4958#endif
4959
4960#ifdef __LITTLE_ENDIAN__
4961#define vdup_lane_u16(__p0, __p1) __extension__ ({ \
4962  uint16x4_t __s0 = __p0; \
4963  uint16x4_t __ret; \
4964  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4965  __ret; \
4966})
4967#else
4968#define vdup_lane_u16(__p0, __p1) __extension__ ({ \
4969  uint16x4_t __s0 = __p0; \
4970  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4971  uint16x4_t __ret; \
4972  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4973  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4974  __ret; \
4975})
4976#endif
4977
4978#ifdef __LITTLE_ENDIAN__
4979#define vdup_lane_s8(__p0, __p1) __extension__ ({ \
4980  int8x8_t __s0 = __p0; \
4981  int8x8_t __ret; \
4982  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4983  __ret; \
4984})
4985#else
4986#define vdup_lane_s8(__p0, __p1) __extension__ ({ \
4987  int8x8_t __s0 = __p0; \
4988  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4989  int8x8_t __ret; \
4990  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4991  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4992  __ret; \
4993})
4994#endif
4995
4996#ifdef __LITTLE_ENDIAN__
4997#define vdup_lane_f32(__p0, __p1) __extension__ ({ \
4998  float32x2_t __s0 = __p0; \
4999  float32x2_t __ret; \
5000  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
5001  __ret; \
5002})
5003#else
5004#define vdup_lane_f32(__p0, __p1) __extension__ ({ \
5005  float32x2_t __s0 = __p0; \
5006  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5007  float32x2_t __ret; \
5008  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
5009  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5010  __ret; \
5011})
5012#endif
5013
5014#ifdef __LITTLE_ENDIAN__
5015#define vdup_lane_s32(__p0, __p1) __extension__ ({ \
5016  int32x2_t __s0 = __p0; \
5017  int32x2_t __ret; \
5018  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
5019  __ret; \
5020})
5021#else
5022#define vdup_lane_s32(__p0, __p1) __extension__ ({ \
5023  int32x2_t __s0 = __p0; \
5024  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5025  int32x2_t __ret; \
5026  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
5027  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5028  __ret; \
5029})
5030#endif
5031
5032#ifdef __LITTLE_ENDIAN__
5033#define vdup_lane_s64(__p0, __p1) __extension__ ({ \
5034  int64x1_t __s0 = __p0; \
5035  int64x1_t __ret; \
5036  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
5037  __ret; \
5038})
5039#else
5040#define vdup_lane_s64(__p0, __p1) __extension__ ({ \
5041  int64x1_t __s0 = __p0; \
5042  int64x1_t __ret; \
5043  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
5044  __ret; \
5045})
5046#endif
5047
5048#ifdef __LITTLE_ENDIAN__
5049#define vdup_lane_s16(__p0, __p1) __extension__ ({ \
5050  int16x4_t __s0 = __p0; \
5051  int16x4_t __ret; \
5052  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
5053  __ret; \
5054})
5055#else
5056#define vdup_lane_s16(__p0, __p1) __extension__ ({ \
5057  int16x4_t __s0 = __p0; \
5058  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5059  int16x4_t __ret; \
5060  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
5061  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5062  __ret; \
5063})
5064#endif
5065
5066#ifdef __LITTLE_ENDIAN__
5067__ai poly8x8_t vdup_n_p8(poly8_t __p0) {
5068  poly8x8_t __ret;
5069  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5070  return __ret;
5071}
5072#else
5073__ai poly8x8_t vdup_n_p8(poly8_t __p0) {
5074  poly8x8_t __ret;
5075  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5076  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5077  return __ret;
5078}
5079#endif
5080
5081#ifdef __LITTLE_ENDIAN__
5082__ai poly16x4_t vdup_n_p16(poly16_t __p0) {
5083  poly16x4_t __ret;
5084  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
5085  return __ret;
5086}
5087#else
5088__ai poly16x4_t vdup_n_p16(poly16_t __p0) {
5089  poly16x4_t __ret;
5090  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
5091  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5092  return __ret;
5093}
5094#endif
5095
5096#ifdef __LITTLE_ENDIAN__
5097__ai poly8x16_t vdupq_n_p8(poly8_t __p0) {
5098  poly8x16_t __ret;
5099  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5100  return __ret;
5101}
5102#else
5103__ai poly8x16_t vdupq_n_p8(poly8_t __p0) {
5104  poly8x16_t __ret;
5105  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5106  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5107  return __ret;
5108}
5109#endif
5110
5111#ifdef __LITTLE_ENDIAN__
5112__ai poly16x8_t vdupq_n_p16(poly16_t __p0) {
5113  poly16x8_t __ret;
5114  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5115  return __ret;
5116}
5117#else
5118__ai poly16x8_t vdupq_n_p16(poly16_t __p0) {
5119  poly16x8_t __ret;
5120  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5121  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5122  return __ret;
5123}
5124#endif
5125
5126#ifdef __LITTLE_ENDIAN__
5127__ai uint8x16_t vdupq_n_u8(uint8_t __p0) {
5128  uint8x16_t __ret;
5129  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5130  return __ret;
5131}
5132#else
5133__ai uint8x16_t vdupq_n_u8(uint8_t __p0) {
5134  uint8x16_t __ret;
5135  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5136  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5137  return __ret;
5138}
5139#endif
5140
5141#ifdef __LITTLE_ENDIAN__
5142__ai uint32x4_t vdupq_n_u32(uint32_t __p0) {
5143  uint32x4_t __ret;
5144  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
5145  return __ret;
5146}
5147#else
5148__ai uint32x4_t vdupq_n_u32(uint32_t __p0) {
5149  uint32x4_t __ret;
5150  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
5151  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5152  return __ret;
5153}
5154#endif
5155
5156#ifdef __LITTLE_ENDIAN__
5157__ai uint64x2_t vdupq_n_u64(uint64_t __p0) {
5158  uint64x2_t __ret;
5159  __ret = (uint64x2_t) {__p0, __p0};
5160  return __ret;
5161}
5162#else
5163__ai uint64x2_t vdupq_n_u64(uint64_t __p0) {
5164  uint64x2_t __ret;
5165  __ret = (uint64x2_t) {__p0, __p0};
5166  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5167  return __ret;
5168}
5169#endif
5170
5171#ifdef __LITTLE_ENDIAN__
5172__ai uint16x8_t vdupq_n_u16(uint16_t __p0) {
5173  uint16x8_t __ret;
5174  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5175  return __ret;
5176}
5177#else
5178__ai uint16x8_t vdupq_n_u16(uint16_t __p0) {
5179  uint16x8_t __ret;
5180  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5181  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5182  return __ret;
5183}
5184#endif
5185
5186#ifdef __LITTLE_ENDIAN__
5187__ai int8x16_t vdupq_n_s8(int8_t __p0) {
5188  int8x16_t __ret;
5189  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5190  return __ret;
5191}
5192#else
5193__ai int8x16_t vdupq_n_s8(int8_t __p0) {
5194  int8x16_t __ret;
5195  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5196  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5197  return __ret;
5198}
5199#endif
5200
5201#ifdef __LITTLE_ENDIAN__
5202__ai float32x4_t vdupq_n_f32(float32_t __p0) {
5203  float32x4_t __ret;
5204  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
5205  return __ret;
5206}
5207#else
5208__ai float32x4_t vdupq_n_f32(float32_t __p0) {
5209  float32x4_t __ret;
5210  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
5211  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5212  return __ret;
5213}
5214#endif
5215
5216#ifdef __LITTLE_ENDIAN__
5217#define vdupq_n_f16(__p0) __extension__ ({ \
5218  float16_t __s0 = __p0; \
5219  float16x8_t __ret; \
5220  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
5221  __ret; \
5222})
5223#else
5224#define vdupq_n_f16(__p0) __extension__ ({ \
5225  float16_t __s0 = __p0; \
5226  float16x8_t __ret; \
5227  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
5228  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5229  __ret; \
5230})
5231#endif
5232
5233#ifdef __LITTLE_ENDIAN__
5234__ai int32x4_t vdupq_n_s32(int32_t __p0) {
5235  int32x4_t __ret;
5236  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
5237  return __ret;
5238}
5239#else
5240__ai int32x4_t vdupq_n_s32(int32_t __p0) {
5241  int32x4_t __ret;
5242  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
5243  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5244  return __ret;
5245}
5246#endif
5247
5248#ifdef __LITTLE_ENDIAN__
5249__ai int64x2_t vdupq_n_s64(int64_t __p0) {
5250  int64x2_t __ret;
5251  __ret = (int64x2_t) {__p0, __p0};
5252  return __ret;
5253}
5254#else
5255__ai int64x2_t vdupq_n_s64(int64_t __p0) {
5256  int64x2_t __ret;
5257  __ret = (int64x2_t) {__p0, __p0};
5258  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5259  return __ret;
5260}
5261#endif
5262
5263#ifdef __LITTLE_ENDIAN__
5264__ai int16x8_t vdupq_n_s16(int16_t __p0) {
5265  int16x8_t __ret;
5266  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5267  return __ret;
5268}
5269#else
5270__ai int16x8_t vdupq_n_s16(int16_t __p0) {
5271  int16x8_t __ret;
5272  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5273  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5274  return __ret;
5275}
5276#endif
5277
5278#ifdef __LITTLE_ENDIAN__
5279__ai uint8x8_t vdup_n_u8(uint8_t __p0) {
5280  uint8x8_t __ret;
5281  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5282  return __ret;
5283}
5284#else
5285__ai uint8x8_t vdup_n_u8(uint8_t __p0) {
5286  uint8x8_t __ret;
5287  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5288  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5289  return __ret;
5290}
5291#endif
5292
5293#ifdef __LITTLE_ENDIAN__
5294__ai uint32x2_t vdup_n_u32(uint32_t __p0) {
5295  uint32x2_t __ret;
5296  __ret = (uint32x2_t) {__p0, __p0};
5297  return __ret;
5298}
5299#else
5300__ai uint32x2_t vdup_n_u32(uint32_t __p0) {
5301  uint32x2_t __ret;
5302  __ret = (uint32x2_t) {__p0, __p0};
5303  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5304  return __ret;
5305}
5306#endif
5307
5308#ifdef __LITTLE_ENDIAN__
5309__ai uint64x1_t vdup_n_u64(uint64_t __p0) {
5310  uint64x1_t __ret;
5311  __ret = (uint64x1_t) {__p0};
5312  return __ret;
5313}
5314#else
5315__ai uint64x1_t vdup_n_u64(uint64_t __p0) {
5316  uint64x1_t __ret;
5317  __ret = (uint64x1_t) {__p0};
5318  return __ret;
5319}
5320#endif
5321
5322#ifdef __LITTLE_ENDIAN__
5323__ai uint16x4_t vdup_n_u16(uint16_t __p0) {
5324  uint16x4_t __ret;
5325  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
5326  return __ret;
5327}
5328#else
5329__ai uint16x4_t vdup_n_u16(uint16_t __p0) {
5330  uint16x4_t __ret;
5331  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
5332  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5333  return __ret;
5334}
5335#endif
5336
5337#ifdef __LITTLE_ENDIAN__
5338__ai int8x8_t vdup_n_s8(int8_t __p0) {
5339  int8x8_t __ret;
5340  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5341  return __ret;
5342}
5343#else
5344__ai int8x8_t vdup_n_s8(int8_t __p0) {
5345  int8x8_t __ret;
5346  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5347  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5348  return __ret;
5349}
5350#endif
5351
5352#ifdef __LITTLE_ENDIAN__
5353__ai float32x2_t vdup_n_f32(float32_t __p0) {
5354  float32x2_t __ret;
5355  __ret = (float32x2_t) {__p0, __p0};
5356  return __ret;
5357}
5358#else
5359__ai float32x2_t vdup_n_f32(float32_t __p0) {
5360  float32x2_t __ret;
5361  __ret = (float32x2_t) {__p0, __p0};
5362  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5363  return __ret;
5364}
5365#endif
5366
5367#ifdef __LITTLE_ENDIAN__
5368#define vdup_n_f16(__p0) __extension__ ({ \
5369  float16_t __s0 = __p0; \
5370  float16x4_t __ret; \
5371  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
5372  __ret; \
5373})
5374#else
5375#define vdup_n_f16(__p0) __extension__ ({ \
5376  float16_t __s0 = __p0; \
5377  float16x4_t __ret; \
5378  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
5379  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5380  __ret; \
5381})
5382#endif
5383
5384#ifdef __LITTLE_ENDIAN__
5385__ai int32x2_t vdup_n_s32(int32_t __p0) {
5386  int32x2_t __ret;
5387  __ret = (int32x2_t) {__p0, __p0};
5388  return __ret;
5389}
5390#else
5391__ai int32x2_t vdup_n_s32(int32_t __p0) {
5392  int32x2_t __ret;
5393  __ret = (int32x2_t) {__p0, __p0};
5394  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5395  return __ret;
5396}
5397#endif
5398
5399#ifdef __LITTLE_ENDIAN__
5400__ai int64x1_t vdup_n_s64(int64_t __p0) {
5401  int64x1_t __ret;
5402  __ret = (int64x1_t) {__p0};
5403  return __ret;
5404}
5405#else
5406__ai int64x1_t vdup_n_s64(int64_t __p0) {
5407  int64x1_t __ret;
5408  __ret = (int64x1_t) {__p0};
5409  return __ret;
5410}
5411#endif
5412
5413#ifdef __LITTLE_ENDIAN__
5414__ai int16x4_t vdup_n_s16(int16_t __p0) {
5415  int16x4_t __ret;
5416  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
5417  return __ret;
5418}
5419#else
5420__ai int16x4_t vdup_n_s16(int16_t __p0) {
5421  int16x4_t __ret;
5422  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
5423  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5424  return __ret;
5425}
5426#endif
5427
5428#ifdef __LITTLE_ENDIAN__
5429__ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1) {
5430  uint8x16_t __ret;
5431  __ret = __p0 ^ __p1;
5432  return __ret;
5433}
5434#else
5435__ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1) {
5436  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5437  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5438  uint8x16_t __ret;
5439  __ret = __rev0 ^ __rev1;
5440  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5441  return __ret;
5442}
5443#endif
5444
5445#ifdef __LITTLE_ENDIAN__
5446__ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1) {
5447  uint32x4_t __ret;
5448  __ret = __p0 ^ __p1;
5449  return __ret;
5450}
5451#else
5452__ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1) {
5453  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5454  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5455  uint32x4_t __ret;
5456  __ret = __rev0 ^ __rev1;
5457  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5458  return __ret;
5459}
5460#endif
5461
5462#ifdef __LITTLE_ENDIAN__
5463__ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1) {
5464  uint64x2_t __ret;
5465  __ret = __p0 ^ __p1;
5466  return __ret;
5467}
5468#else
5469__ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1) {
5470  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5471  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5472  uint64x2_t __ret;
5473  __ret = __rev0 ^ __rev1;
5474  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5475  return __ret;
5476}
5477#endif
5478
5479#ifdef __LITTLE_ENDIAN__
5480__ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1) {
5481  uint16x8_t __ret;
5482  __ret = __p0 ^ __p1;
5483  return __ret;
5484}
5485#else
5486__ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1) {
5487  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5488  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5489  uint16x8_t __ret;
5490  __ret = __rev0 ^ __rev1;
5491  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5492  return __ret;
5493}
5494#endif
5495
5496#ifdef __LITTLE_ENDIAN__
5497__ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1) {
5498  int8x16_t __ret;
5499  __ret = __p0 ^ __p1;
5500  return __ret;
5501}
5502#else
5503__ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1) {
5504  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5505  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5506  int8x16_t __ret;
5507  __ret = __rev0 ^ __rev1;
5508  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5509  return __ret;
5510}
5511#endif
5512
5513#ifdef __LITTLE_ENDIAN__
5514__ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1) {
5515  int32x4_t __ret;
5516  __ret = __p0 ^ __p1;
5517  return __ret;
5518}
5519#else
5520__ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1) {
5521  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5522  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5523  int32x4_t __ret;
5524  __ret = __rev0 ^ __rev1;
5525  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5526  return __ret;
5527}
5528#endif
5529
5530#ifdef __LITTLE_ENDIAN__
5531__ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1) {
5532  int64x2_t __ret;
5533  __ret = __p0 ^ __p1;
5534  return __ret;
5535}
5536#else
5537__ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1) {
5538  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5539  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5540  int64x2_t __ret;
5541  __ret = __rev0 ^ __rev1;
5542  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5543  return __ret;
5544}
5545#endif
5546
5547#ifdef __LITTLE_ENDIAN__
5548__ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1) {
5549  int16x8_t __ret;
5550  __ret = __p0 ^ __p1;
5551  return __ret;
5552}
5553#else
5554__ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1) {
5555  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5556  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5557  int16x8_t __ret;
5558  __ret = __rev0 ^ __rev1;
5559  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5560  return __ret;
5561}
5562#endif
5563
5564#ifdef __LITTLE_ENDIAN__
5565__ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1) {
5566  uint8x8_t __ret;
5567  __ret = __p0 ^ __p1;
5568  return __ret;
5569}
5570#else
5571__ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1) {
5572  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5573  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5574  uint8x8_t __ret;
5575  __ret = __rev0 ^ __rev1;
5576  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5577  return __ret;
5578}
5579#endif
5580
5581#ifdef __LITTLE_ENDIAN__
5582__ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1) {
5583  uint32x2_t __ret;
5584  __ret = __p0 ^ __p1;
5585  return __ret;
5586}
5587#else
5588__ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1) {
5589  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5590  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5591  uint32x2_t __ret;
5592  __ret = __rev0 ^ __rev1;
5593  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5594  return __ret;
5595}
5596#endif
5597
5598#ifdef __LITTLE_ENDIAN__
5599__ai uint64x1_t veor_u64(uint64x1_t __p0, uint64x1_t __p1) {
5600  uint64x1_t __ret;
5601  __ret = __p0 ^ __p1;
5602  return __ret;
5603}
5604#else
5605__ai uint64x1_t veor_u64(uint64x1_t __p0, uint64x1_t __p1) {
5606  uint64x1_t __ret;
5607  __ret = __p0 ^ __p1;
5608  return __ret;
5609}
5610#endif
5611
5612#ifdef __LITTLE_ENDIAN__
5613__ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1) {
5614  uint16x4_t __ret;
5615  __ret = __p0 ^ __p1;
5616  return __ret;
5617}
5618#else
5619__ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1) {
5620  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5621  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5622  uint16x4_t __ret;
5623  __ret = __rev0 ^ __rev1;
5624  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5625  return __ret;
5626}
5627#endif
5628
5629#ifdef __LITTLE_ENDIAN__
5630__ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1) {
5631  int8x8_t __ret;
5632  __ret = __p0 ^ __p1;
5633  return __ret;
5634}
5635#else
5636__ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1) {
5637  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5638  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5639  int8x8_t __ret;
5640  __ret = __rev0 ^ __rev1;
5641  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5642  return __ret;
5643}
5644#endif
5645
5646#ifdef __LITTLE_ENDIAN__
5647__ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1) {
5648  int32x2_t __ret;
5649  __ret = __p0 ^ __p1;
5650  return __ret;
5651}
5652#else
5653__ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1) {
5654  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5655  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5656  int32x2_t __ret;
5657  __ret = __rev0 ^ __rev1;
5658  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5659  return __ret;
5660}
5661#endif
5662
5663#ifdef __LITTLE_ENDIAN__
5664__ai int64x1_t veor_s64(int64x1_t __p0, int64x1_t __p1) {
5665  int64x1_t __ret;
5666  __ret = __p0 ^ __p1;
5667  return __ret;
5668}
5669#else
5670__ai int64x1_t veor_s64(int64x1_t __p0, int64x1_t __p1) {
5671  int64x1_t __ret;
5672  __ret = __p0 ^ __p1;
5673  return __ret;
5674}
5675#endif
5676
5677#ifdef __LITTLE_ENDIAN__
5678__ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1) {
5679  int16x4_t __ret;
5680  __ret = __p0 ^ __p1;
5681  return __ret;
5682}
5683#else
5684__ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1) {
5685  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5686  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5687  int16x4_t __ret;
5688  __ret = __rev0 ^ __rev1;
5689  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5690  return __ret;
5691}
5692#endif
5693
5694#ifdef __LITTLE_ENDIAN__
5695#define vext_p8(__p0, __p1, __p2) __extension__ ({ \
5696  poly8x8_t __s0 = __p0; \
5697  poly8x8_t __s1 = __p1; \
5698  poly8x8_t __ret; \
5699  __ret = (poly8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
5700  __ret; \
5701})
5702#else
5703#define vext_p8(__p0, __p1, __p2) __extension__ ({ \
5704  poly8x8_t __s0 = __p0; \
5705  poly8x8_t __s1 = __p1; \
5706  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5707  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5708  poly8x8_t __ret; \
5709  __ret = (poly8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
5710  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5711  __ret; \
5712})
5713#endif
5714
5715#ifdef __LITTLE_ENDIAN__
5716#define vext_p16(__p0, __p1, __p2) __extension__ ({ \
5717  poly16x4_t __s0 = __p0; \
5718  poly16x4_t __s1 = __p1; \
5719  poly16x4_t __ret; \
5720  __ret = (poly16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
5721  __ret; \
5722})
5723#else
5724#define vext_p16(__p0, __p1, __p2) __extension__ ({ \
5725  poly16x4_t __s0 = __p0; \
5726  poly16x4_t __s1 = __p1; \
5727  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5728  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5729  poly16x4_t __ret; \
5730  __ret = (poly16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
5731  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5732  __ret; \
5733})
5734#endif
5735
5736#ifdef __LITTLE_ENDIAN__
5737#define vextq_p8(__p0, __p1, __p2) __extension__ ({ \
5738  poly8x16_t __s0 = __p0; \
5739  poly8x16_t __s1 = __p1; \
5740  poly8x16_t __ret; \
5741  __ret = (poly8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
5742  __ret; \
5743})
5744#else
5745#define vextq_p8(__p0, __p1, __p2) __extension__ ({ \
5746  poly8x16_t __s0 = __p0; \
5747  poly8x16_t __s1 = __p1; \
5748  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5749  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5750  poly8x16_t __ret; \
5751  __ret = (poly8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
5752  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5753  __ret; \
5754})
5755#endif
5756
5757#ifdef __LITTLE_ENDIAN__
5758#define vextq_p16(__p0, __p1, __p2) __extension__ ({ \
5759  poly16x8_t __s0 = __p0; \
5760  poly16x8_t __s1 = __p1; \
5761  poly16x8_t __ret; \
5762  __ret = (poly16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
5763  __ret; \
5764})
5765#else
5766#define vextq_p16(__p0, __p1, __p2) __extension__ ({ \
5767  poly16x8_t __s0 = __p0; \
5768  poly16x8_t __s1 = __p1; \
5769  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5770  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5771  poly16x8_t __ret; \
5772  __ret = (poly16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
5773  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5774  __ret; \
5775})
5776#endif
5777
5778#ifdef __LITTLE_ENDIAN__
5779#define vextq_u8(__p0, __p1, __p2) __extension__ ({ \
5780  uint8x16_t __s0 = __p0; \
5781  uint8x16_t __s1 = __p1; \
5782  uint8x16_t __ret; \
5783  __ret = (uint8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
5784  __ret; \
5785})
5786#else
5787#define vextq_u8(__p0, __p1, __p2) __extension__ ({ \
5788  uint8x16_t __s0 = __p0; \
5789  uint8x16_t __s1 = __p1; \
5790  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5791  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5792  uint8x16_t __ret; \
5793  __ret = (uint8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
5794  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5795  __ret; \
5796})
5797#endif
5798
5799#ifdef __LITTLE_ENDIAN__
5800#define vextq_u32(__p0, __p1, __p2) __extension__ ({ \
5801  uint32x4_t __s0 = __p0; \
5802  uint32x4_t __s1 = __p1; \
5803  uint32x4_t __ret; \
5804  __ret = (uint32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
5805  __ret; \
5806})
5807#else
5808#define vextq_u32(__p0, __p1, __p2) __extension__ ({ \
5809  uint32x4_t __s0 = __p0; \
5810  uint32x4_t __s1 = __p1; \
5811  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5812  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5813  uint32x4_t __ret; \
5814  __ret = (uint32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
5815  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5816  __ret; \
5817})
5818#endif
5819
5820#ifdef __LITTLE_ENDIAN__
5821#define vextq_u64(__p0, __p1, __p2) __extension__ ({ \
5822  uint64x2_t __s0 = __p0; \
5823  uint64x2_t __s1 = __p1; \
5824  uint64x2_t __ret; \
5825  __ret = (uint64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
5826  __ret; \
5827})
5828#else
5829#define vextq_u64(__p0, __p1, __p2) __extension__ ({ \
5830  uint64x2_t __s0 = __p0; \
5831  uint64x2_t __s1 = __p1; \
5832  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5833  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
5834  uint64x2_t __ret; \
5835  __ret = (uint64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
5836  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5837  __ret; \
5838})
5839#endif
5840
5841#ifdef __LITTLE_ENDIAN__
5842#define vextq_u16(__p0, __p1, __p2) __extension__ ({ \
5843  uint16x8_t __s0 = __p0; \
5844  uint16x8_t __s1 = __p1; \
5845  uint16x8_t __ret; \
5846  __ret = (uint16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
5847  __ret; \
5848})
5849#else
5850#define vextq_u16(__p0, __p1, __p2) __extension__ ({ \
5851  uint16x8_t __s0 = __p0; \
5852  uint16x8_t __s1 = __p1; \
5853  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5854  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5855  uint16x8_t __ret; \
5856  __ret = (uint16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
5857  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5858  __ret; \
5859})
5860#endif
5861
5862#ifdef __LITTLE_ENDIAN__
5863#define vextq_s8(__p0, __p1, __p2) __extension__ ({ \
5864  int8x16_t __s0 = __p0; \
5865  int8x16_t __s1 = __p1; \
5866  int8x16_t __ret; \
5867  __ret = (int8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
5868  __ret; \
5869})
5870#else
5871#define vextq_s8(__p0, __p1, __p2) __extension__ ({ \
5872  int8x16_t __s0 = __p0; \
5873  int8x16_t __s1 = __p1; \
5874  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5875  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5876  int8x16_t __ret; \
5877  __ret = (int8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
5878  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5879  __ret; \
5880})
5881#endif
5882
5883#ifdef __LITTLE_ENDIAN__
5884#define vextq_f32(__p0, __p1, __p2) __extension__ ({ \
5885  float32x4_t __s0 = __p0; \
5886  float32x4_t __s1 = __p1; \
5887  float32x4_t __ret; \
5888  __ret = (float32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 41); \
5889  __ret; \
5890})
5891#else
5892#define vextq_f32(__p0, __p1, __p2) __extension__ ({ \
5893  float32x4_t __s0 = __p0; \
5894  float32x4_t __s1 = __p1; \
5895  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5896  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5897  float32x4_t __ret; \
5898  __ret = (float32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 41); \
5899  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5900  __ret; \
5901})
5902#endif
5903
5904#ifdef __LITTLE_ENDIAN__
5905#define vextq_s32(__p0, __p1, __p2) __extension__ ({ \
5906  int32x4_t __s0 = __p0; \
5907  int32x4_t __s1 = __p1; \
5908  int32x4_t __ret; \
5909  __ret = (int32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
5910  __ret; \
5911})
5912#else
5913#define vextq_s32(__p0, __p1, __p2) __extension__ ({ \
5914  int32x4_t __s0 = __p0; \
5915  int32x4_t __s1 = __p1; \
5916  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5917  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5918  int32x4_t __ret; \
5919  __ret = (int32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
5920  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5921  __ret; \
5922})
5923#endif
5924
5925#ifdef __LITTLE_ENDIAN__
5926#define vextq_s64(__p0, __p1, __p2) __extension__ ({ \
5927  int64x2_t __s0 = __p0; \
5928  int64x2_t __s1 = __p1; \
5929  int64x2_t __ret; \
5930  __ret = (int64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
5931  __ret; \
5932})
5933#else
5934#define vextq_s64(__p0, __p1, __p2) __extension__ ({ \
5935  int64x2_t __s0 = __p0; \
5936  int64x2_t __s1 = __p1; \
5937  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5938  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
5939  int64x2_t __ret; \
5940  __ret = (int64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
5941  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5942  __ret; \
5943})
5944#endif
5945
5946#ifdef __LITTLE_ENDIAN__
5947#define vextq_s16(__p0, __p1, __p2) __extension__ ({ \
5948  int16x8_t __s0 = __p0; \
5949  int16x8_t __s1 = __p1; \
5950  int16x8_t __ret; \
5951  __ret = (int16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
5952  __ret; \
5953})
5954#else
5955#define vextq_s16(__p0, __p1, __p2) __extension__ ({ \
5956  int16x8_t __s0 = __p0; \
5957  int16x8_t __s1 = __p1; \
5958  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5959  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5960  int16x8_t __ret; \
5961  __ret = (int16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
5962  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5963  __ret; \
5964})
5965#endif
5966
5967#ifdef __LITTLE_ENDIAN__
5968#define vext_u8(__p0, __p1, __p2) __extension__ ({ \
5969  uint8x8_t __s0 = __p0; \
5970  uint8x8_t __s1 = __p1; \
5971  uint8x8_t __ret; \
5972  __ret = (uint8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
5973  __ret; \
5974})
5975#else
5976#define vext_u8(__p0, __p1, __p2) __extension__ ({ \
5977  uint8x8_t __s0 = __p0; \
5978  uint8x8_t __s1 = __p1; \
5979  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5980  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5981  uint8x8_t __ret; \
5982  __ret = (uint8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
5983  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5984  __ret; \
5985})
5986#endif
5987
5988#ifdef __LITTLE_ENDIAN__
5989#define vext_u32(__p0, __p1, __p2) __extension__ ({ \
5990  uint32x2_t __s0 = __p0; \
5991  uint32x2_t __s1 = __p1; \
5992  uint32x2_t __ret; \
5993  __ret = (uint32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
5994  __ret; \
5995})
5996#else
5997#define vext_u32(__p0, __p1, __p2) __extension__ ({ \
5998  uint32x2_t __s0 = __p0; \
5999  uint32x2_t __s1 = __p1; \
6000  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6001  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6002  uint32x2_t __ret; \
6003  __ret = (uint32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
6004  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6005  __ret; \
6006})
6007#endif
6008
6009#ifdef __LITTLE_ENDIAN__
6010#define vext_u64(__p0, __p1, __p2) __extension__ ({ \
6011  uint64x1_t __s0 = __p0; \
6012  uint64x1_t __s1 = __p1; \
6013  uint64x1_t __ret; \
6014  __ret = (uint64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
6015  __ret; \
6016})
6017#else
6018#define vext_u64(__p0, __p1, __p2) __extension__ ({ \
6019  uint64x1_t __s0 = __p0; \
6020  uint64x1_t __s1 = __p1; \
6021  uint64x1_t __ret; \
6022  __ret = (uint64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
6023  __ret; \
6024})
6025#endif
6026
6027#ifdef __LITTLE_ENDIAN__
6028#define vext_u16(__p0, __p1, __p2) __extension__ ({ \
6029  uint16x4_t __s0 = __p0; \
6030  uint16x4_t __s1 = __p1; \
6031  uint16x4_t __ret; \
6032  __ret = (uint16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
6033  __ret; \
6034})
6035#else
6036#define vext_u16(__p0, __p1, __p2) __extension__ ({ \
6037  uint16x4_t __s0 = __p0; \
6038  uint16x4_t __s1 = __p1; \
6039  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6040  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6041  uint16x4_t __ret; \
6042  __ret = (uint16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
6043  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6044  __ret; \
6045})
6046#endif
6047
6048#ifdef __LITTLE_ENDIAN__
6049#define vext_s8(__p0, __p1, __p2) __extension__ ({ \
6050  int8x8_t __s0 = __p0; \
6051  int8x8_t __s1 = __p1; \
6052  int8x8_t __ret; \
6053  __ret = (int8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
6054  __ret; \
6055})
6056#else
6057#define vext_s8(__p0, __p1, __p2) __extension__ ({ \
6058  int8x8_t __s0 = __p0; \
6059  int8x8_t __s1 = __p1; \
6060  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6061  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
6062  int8x8_t __ret; \
6063  __ret = (int8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
6064  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6065  __ret; \
6066})
6067#endif
6068
6069#ifdef __LITTLE_ENDIAN__
6070#define vext_f32(__p0, __p1, __p2) __extension__ ({ \
6071  float32x2_t __s0 = __p0; \
6072  float32x2_t __s1 = __p1; \
6073  float32x2_t __ret; \
6074  __ret = (float32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 9); \
6075  __ret; \
6076})
6077#else
6078#define vext_f32(__p0, __p1, __p2) __extension__ ({ \
6079  float32x2_t __s0 = __p0; \
6080  float32x2_t __s1 = __p1; \
6081  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6082  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6083  float32x2_t __ret; \
6084  __ret = (float32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 9); \
6085  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6086  __ret; \
6087})
6088#endif
6089
6090#ifdef __LITTLE_ENDIAN__
6091#define vext_s32(__p0, __p1, __p2) __extension__ ({ \
6092  int32x2_t __s0 = __p0; \
6093  int32x2_t __s1 = __p1; \
6094  int32x2_t __ret; \
6095  __ret = (int32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
6096  __ret; \
6097})
6098#else
6099#define vext_s32(__p0, __p1, __p2) __extension__ ({ \
6100  int32x2_t __s0 = __p0; \
6101  int32x2_t __s1 = __p1; \
6102  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6103  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6104  int32x2_t __ret; \
6105  __ret = (int32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
6106  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6107  __ret; \
6108})
6109#endif
6110
6111#ifdef __LITTLE_ENDIAN__
6112#define vext_s64(__p0, __p1, __p2) __extension__ ({ \
6113  int64x1_t __s0 = __p0; \
6114  int64x1_t __s1 = __p1; \
6115  int64x1_t __ret; \
6116  __ret = (int64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
6117  __ret; \
6118})
6119#else
6120#define vext_s64(__p0, __p1, __p2) __extension__ ({ \
6121  int64x1_t __s0 = __p0; \
6122  int64x1_t __s1 = __p1; \
6123  int64x1_t __ret; \
6124  __ret = (int64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
6125  __ret; \
6126})
6127#endif
6128
6129#ifdef __LITTLE_ENDIAN__
6130#define vext_s16(__p0, __p1, __p2) __extension__ ({ \
6131  int16x4_t __s0 = __p0; \
6132  int16x4_t __s1 = __p1; \
6133  int16x4_t __ret; \
6134  __ret = (int16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
6135  __ret; \
6136})
6137#else
6138#define vext_s16(__p0, __p1, __p2) __extension__ ({ \
6139  int16x4_t __s0 = __p0; \
6140  int16x4_t __s1 = __p1; \
6141  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6142  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6143  int16x4_t __ret; \
6144  __ret = (int16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
6145  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6146  __ret; \
6147})
6148#endif
6149
6150#ifdef __LITTLE_ENDIAN__
6151__ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
6152  float32x4_t __ret;
6153  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
6154  return __ret;
6155}
6156#else
6157__ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
6158  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6159  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
6160  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
6161  float32x4_t __ret;
6162  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
6163  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6164  return __ret;
6165}
6166__ai float32x4_t __noswap_vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
6167  float32x4_t __ret;
6168  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
6169  return __ret;
6170}
6171#endif
6172
6173#ifdef __LITTLE_ENDIAN__
6174__ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
6175  float32x2_t __ret;
6176  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
6177  return __ret;
6178}
6179#else
6180__ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
6181  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6182  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
6183  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
6184  float32x2_t __ret;
6185  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
6186  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6187  return __ret;
6188}
6189__ai float32x2_t __noswap_vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
6190  float32x2_t __ret;
6191  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
6192  return __ret;
6193}
6194#endif
6195
6196#ifdef __LITTLE_ENDIAN__
6197__ai poly8x8_t vget_high_p8(poly8x16_t __p0) {
6198  poly8x8_t __ret;
6199  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6200  return __ret;
6201}
6202#else
6203__ai poly8x8_t vget_high_p8(poly8x16_t __p0) {
6204  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6205  poly8x8_t __ret;
6206  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
6207  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6208  return __ret;
6209}
6210__ai poly8x8_t __noswap_vget_high_p8(poly8x16_t __p0) {
6211  poly8x8_t __ret;
6212  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6213  return __ret;
6214}
6215#endif
6216
6217#ifdef __LITTLE_ENDIAN__
6218__ai poly16x4_t vget_high_p16(poly16x8_t __p0) {
6219  poly16x4_t __ret;
6220  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6221  return __ret;
6222}
6223#else
6224__ai poly16x4_t vget_high_p16(poly16x8_t __p0) {
6225  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6226  poly16x4_t __ret;
6227  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6228  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6229  return __ret;
6230}
6231#endif
6232
6233#ifdef __LITTLE_ENDIAN__
6234__ai uint8x8_t vget_high_u8(uint8x16_t __p0) {
6235  uint8x8_t __ret;
6236  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6237  return __ret;
6238}
6239#else
6240__ai uint8x8_t vget_high_u8(uint8x16_t __p0) {
6241  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6242  uint8x8_t __ret;
6243  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
6244  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6245  return __ret;
6246}
6247__ai uint8x8_t __noswap_vget_high_u8(uint8x16_t __p0) {
6248  uint8x8_t __ret;
6249  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6250  return __ret;
6251}
6252#endif
6253
6254#ifdef __LITTLE_ENDIAN__
6255__ai uint32x2_t vget_high_u32(uint32x4_t __p0) {
6256  uint32x2_t __ret;
6257  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6258  return __ret;
6259}
6260#else
6261__ai uint32x2_t vget_high_u32(uint32x4_t __p0) {
6262  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6263  uint32x2_t __ret;
6264  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
6265  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6266  return __ret;
6267}
6268__ai uint32x2_t __noswap_vget_high_u32(uint32x4_t __p0) {
6269  uint32x2_t __ret;
6270  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6271  return __ret;
6272}
6273#endif
6274
6275#ifdef __LITTLE_ENDIAN__
6276__ai uint64x1_t vget_high_u64(uint64x2_t __p0) {
6277  uint64x1_t __ret;
6278  __ret = __builtin_shufflevector(__p0, __p0, 1);
6279  return __ret;
6280}
6281#else
6282__ai uint64x1_t vget_high_u64(uint64x2_t __p0) {
6283  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6284  uint64x1_t __ret;
6285  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
6286  return __ret;
6287}
6288#endif
6289
6290#ifdef __LITTLE_ENDIAN__
6291__ai uint16x4_t vget_high_u16(uint16x8_t __p0) {
6292  uint16x4_t __ret;
6293  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6294  return __ret;
6295}
6296#else
6297__ai uint16x4_t vget_high_u16(uint16x8_t __p0) {
6298  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6299  uint16x4_t __ret;
6300  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6301  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6302  return __ret;
6303}
6304__ai uint16x4_t __noswap_vget_high_u16(uint16x8_t __p0) {
6305  uint16x4_t __ret;
6306  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6307  return __ret;
6308}
6309#endif
6310
6311#ifdef __LITTLE_ENDIAN__
6312__ai int8x8_t vget_high_s8(int8x16_t __p0) {
6313  int8x8_t __ret;
6314  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6315  return __ret;
6316}
6317#else
6318__ai int8x8_t vget_high_s8(int8x16_t __p0) {
6319  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6320  int8x8_t __ret;
6321  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
6322  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6323  return __ret;
6324}
6325__ai int8x8_t __noswap_vget_high_s8(int8x16_t __p0) {
6326  int8x8_t __ret;
6327  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6328  return __ret;
6329}
6330#endif
6331
6332#ifdef __LITTLE_ENDIAN__
6333__ai float32x2_t vget_high_f32(float32x4_t __p0) {
6334  float32x2_t __ret;
6335  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6336  return __ret;
6337}
6338#else
6339__ai float32x2_t vget_high_f32(float32x4_t __p0) {
6340  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6341  float32x2_t __ret;
6342  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
6343  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6344  return __ret;
6345}
6346__ai float32x2_t __noswap_vget_high_f32(float32x4_t __p0) {
6347  float32x2_t __ret;
6348  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6349  return __ret;
6350}
6351#endif
6352
6353#ifdef __LITTLE_ENDIAN__
6354__ai float16x4_t vget_high_f16(float16x8_t __p0) {
6355  float16x4_t __ret;
6356  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6357  return __ret;
6358}
6359#else
6360__ai float16x4_t vget_high_f16(float16x8_t __p0) {
6361  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6362  float16x4_t __ret;
6363  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6364  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6365  return __ret;
6366}
6367__ai float16x4_t __noswap_vget_high_f16(float16x8_t __p0) {
6368  float16x4_t __ret;
6369  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6370  return __ret;
6371}
6372#endif
6373
6374#ifdef __LITTLE_ENDIAN__
6375__ai int32x2_t vget_high_s32(int32x4_t __p0) {
6376  int32x2_t __ret;
6377  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6378  return __ret;
6379}
6380#else
6381__ai int32x2_t vget_high_s32(int32x4_t __p0) {
6382  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6383  int32x2_t __ret;
6384  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
6385  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6386  return __ret;
6387}
6388__ai int32x2_t __noswap_vget_high_s32(int32x4_t __p0) {
6389  int32x2_t __ret;
6390  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6391  return __ret;
6392}
6393#endif
6394
6395#ifdef __LITTLE_ENDIAN__
6396__ai int64x1_t vget_high_s64(int64x2_t __p0) {
6397  int64x1_t __ret;
6398  __ret = __builtin_shufflevector(__p0, __p0, 1);
6399  return __ret;
6400}
6401#else
6402__ai int64x1_t vget_high_s64(int64x2_t __p0) {
6403  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6404  int64x1_t __ret;
6405  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
6406  return __ret;
6407}
6408#endif
6409
6410#ifdef __LITTLE_ENDIAN__
6411__ai int16x4_t vget_high_s16(int16x8_t __p0) {
6412  int16x4_t __ret;
6413  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6414  return __ret;
6415}
6416#else
6417__ai int16x4_t vget_high_s16(int16x8_t __p0) {
6418  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6419  int16x4_t __ret;
6420  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6421  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6422  return __ret;
6423}
6424__ai int16x4_t __noswap_vget_high_s16(int16x8_t __p0) {
6425  int16x4_t __ret;
6426  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6427  return __ret;
6428}
6429#endif
6430
6431#ifdef __LITTLE_ENDIAN__
6432#define vget_lane_p8(__p0, __p1) __extension__ ({ \
6433  poly8x8_t __s0 = __p0; \
6434  poly8_t __ret; \
6435  __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6436  __ret; \
6437})
6438#else
6439#define vget_lane_p8(__p0, __p1) __extension__ ({ \
6440  poly8x8_t __s0 = __p0; \
6441  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6442  poly8_t __ret; \
6443  __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
6444  __ret; \
6445})
6446#define __noswap_vget_lane_p8(__p0, __p1) __extension__ ({ \
6447  poly8x8_t __s0 = __p0; \
6448  poly8_t __ret; \
6449  __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6450  __ret; \
6451})
6452#endif
6453
6454#ifdef __LITTLE_ENDIAN__
6455#define vget_lane_p16(__p0, __p1) __extension__ ({ \
6456  poly16x4_t __s0 = __p0; \
6457  poly16_t __ret; \
6458  __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6459  __ret; \
6460})
6461#else
6462#define vget_lane_p16(__p0, __p1) __extension__ ({ \
6463  poly16x4_t __s0 = __p0; \
6464  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6465  poly16_t __ret; \
6466  __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
6467  __ret; \
6468})
6469#define __noswap_vget_lane_p16(__p0, __p1) __extension__ ({ \
6470  poly16x4_t __s0 = __p0; \
6471  poly16_t __ret; \
6472  __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6473  __ret; \
6474})
6475#endif
6476
6477#ifdef __LITTLE_ENDIAN__
6478#define vgetq_lane_p8(__p0, __p1) __extension__ ({ \
6479  poly8x16_t __s0 = __p0; \
6480  poly8_t __ret; \
6481  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6482  __ret; \
6483})
6484#else
6485#define vgetq_lane_p8(__p0, __p1) __extension__ ({ \
6486  poly8x16_t __s0 = __p0; \
6487  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6488  poly8_t __ret; \
6489  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
6490  __ret; \
6491})
6492#define __noswap_vgetq_lane_p8(__p0, __p1) __extension__ ({ \
6493  poly8x16_t __s0 = __p0; \
6494  poly8_t __ret; \
6495  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6496  __ret; \
6497})
6498#endif
6499
6500#ifdef __LITTLE_ENDIAN__
6501#define vgetq_lane_p16(__p0, __p1) __extension__ ({ \
6502  poly16x8_t __s0 = __p0; \
6503  poly16_t __ret; \
6504  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6505  __ret; \
6506})
6507#else
6508#define vgetq_lane_p16(__p0, __p1) __extension__ ({ \
6509  poly16x8_t __s0 = __p0; \
6510  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6511  poly16_t __ret; \
6512  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
6513  __ret; \
6514})
6515#define __noswap_vgetq_lane_p16(__p0, __p1) __extension__ ({ \
6516  poly16x8_t __s0 = __p0; \
6517  poly16_t __ret; \
6518  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6519  __ret; \
6520})
6521#endif
6522
6523#ifdef __LITTLE_ENDIAN__
6524#define vgetq_lane_u8(__p0, __p1) __extension__ ({ \
6525  uint8x16_t __s0 = __p0; \
6526  uint8_t __ret; \
6527  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6528  __ret; \
6529})
6530#else
6531#define vgetq_lane_u8(__p0, __p1) __extension__ ({ \
6532  uint8x16_t __s0 = __p0; \
6533  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6534  uint8_t __ret; \
6535  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
6536  __ret; \
6537})
6538#define __noswap_vgetq_lane_u8(__p0, __p1) __extension__ ({ \
6539  uint8x16_t __s0 = __p0; \
6540  uint8_t __ret; \
6541  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6542  __ret; \
6543})
6544#endif
6545
6546#ifdef __LITTLE_ENDIAN__
6547#define vgetq_lane_u32(__p0, __p1) __extension__ ({ \
6548  uint32x4_t __s0 = __p0; \
6549  uint32_t __ret; \
6550  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6551  __ret; \
6552})
6553#else
6554#define vgetq_lane_u32(__p0, __p1) __extension__ ({ \
6555  uint32x4_t __s0 = __p0; \
6556  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6557  uint32_t __ret; \
6558  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__rev0, __p1); \
6559  __ret; \
6560})
6561#define __noswap_vgetq_lane_u32(__p0, __p1) __extension__ ({ \
6562  uint32x4_t __s0 = __p0; \
6563  uint32_t __ret; \
6564  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6565  __ret; \
6566})
6567#endif
6568
6569#ifdef __LITTLE_ENDIAN__
6570#define vgetq_lane_u64(__p0, __p1) __extension__ ({ \
6571  uint64x2_t __s0 = __p0; \
6572  uint64_t __ret; \
6573  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6574  __ret; \
6575})
6576#else
6577#define vgetq_lane_u64(__p0, __p1) __extension__ ({ \
6578  uint64x2_t __s0 = __p0; \
6579  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6580  uint64_t __ret; \
6581  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
6582  __ret; \
6583})
6584#define __noswap_vgetq_lane_u64(__p0, __p1) __extension__ ({ \
6585  uint64x2_t __s0 = __p0; \
6586  uint64_t __ret; \
6587  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6588  __ret; \
6589})
6590#endif
6591
6592#ifdef __LITTLE_ENDIAN__
6593#define vgetq_lane_u16(__p0, __p1) __extension__ ({ \
6594  uint16x8_t __s0 = __p0; \
6595  uint16_t __ret; \
6596  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6597  __ret; \
6598})
6599#else
6600#define vgetq_lane_u16(__p0, __p1) __extension__ ({ \
6601  uint16x8_t __s0 = __p0; \
6602  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6603  uint16_t __ret; \
6604  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
6605  __ret; \
6606})
6607#define __noswap_vgetq_lane_u16(__p0, __p1) __extension__ ({ \
6608  uint16x8_t __s0 = __p0; \
6609  uint16_t __ret; \
6610  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6611  __ret; \
6612})
6613#endif
6614
6615#ifdef __LITTLE_ENDIAN__
6616#define vgetq_lane_s8(__p0, __p1) __extension__ ({ \
6617  int8x16_t __s0 = __p0; \
6618  int8_t __ret; \
6619  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6620  __ret; \
6621})
6622#else
6623#define vgetq_lane_s8(__p0, __p1) __extension__ ({ \
6624  int8x16_t __s0 = __p0; \
6625  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6626  int8_t __ret; \
6627  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
6628  __ret; \
6629})
6630#define __noswap_vgetq_lane_s8(__p0, __p1) __extension__ ({ \
6631  int8x16_t __s0 = __p0; \
6632  int8_t __ret; \
6633  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6634  __ret; \
6635})
6636#endif
6637
6638#ifdef __LITTLE_ENDIAN__
6639#define vgetq_lane_f32(__p0, __p1) __extension__ ({ \
6640  float32x4_t __s0 = __p0; \
6641  float32_t __ret; \
6642  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__s0, __p1); \
6643  __ret; \
6644})
6645#else
6646#define vgetq_lane_f32(__p0, __p1) __extension__ ({ \
6647  float32x4_t __s0 = __p0; \
6648  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6649  float32_t __ret; \
6650  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__rev0, __p1); \
6651  __ret; \
6652})
6653#define __noswap_vgetq_lane_f32(__p0, __p1) __extension__ ({ \
6654  float32x4_t __s0 = __p0; \
6655  float32_t __ret; \
6656  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__s0, __p1); \
6657  __ret; \
6658})
6659#endif
6660
6661#ifdef __LITTLE_ENDIAN__
6662#define vgetq_lane_s32(__p0, __p1) __extension__ ({ \
6663  int32x4_t __s0 = __p0; \
6664  int32_t __ret; \
6665  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6666  __ret; \
6667})
6668#else
6669#define vgetq_lane_s32(__p0, __p1) __extension__ ({ \
6670  int32x4_t __s0 = __p0; \
6671  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6672  int32_t __ret; \
6673  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__rev0, __p1); \
6674  __ret; \
6675})
6676#define __noswap_vgetq_lane_s32(__p0, __p1) __extension__ ({ \
6677  int32x4_t __s0 = __p0; \
6678  int32_t __ret; \
6679  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6680  __ret; \
6681})
6682#endif
6683
6684#ifdef __LITTLE_ENDIAN__
6685#define vgetq_lane_s64(__p0, __p1) __extension__ ({ \
6686  int64x2_t __s0 = __p0; \
6687  int64_t __ret; \
6688  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6689  __ret; \
6690})
6691#else
6692#define vgetq_lane_s64(__p0, __p1) __extension__ ({ \
6693  int64x2_t __s0 = __p0; \
6694  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6695  int64_t __ret; \
6696  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
6697  __ret; \
6698})
6699#define __noswap_vgetq_lane_s64(__p0, __p1) __extension__ ({ \
6700  int64x2_t __s0 = __p0; \
6701  int64_t __ret; \
6702  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6703  __ret; \
6704})
6705#endif
6706
6707#ifdef __LITTLE_ENDIAN__
6708#define vgetq_lane_s16(__p0, __p1) __extension__ ({ \
6709  int16x8_t __s0 = __p0; \
6710  int16_t __ret; \
6711  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6712  __ret; \
6713})
6714#else
6715#define vgetq_lane_s16(__p0, __p1) __extension__ ({ \
6716  int16x8_t __s0 = __p0; \
6717  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6718  int16_t __ret; \
6719  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
6720  __ret; \
6721})
6722#define __noswap_vgetq_lane_s16(__p0, __p1) __extension__ ({ \
6723  int16x8_t __s0 = __p0; \
6724  int16_t __ret; \
6725  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6726  __ret; \
6727})
6728#endif
6729
6730#ifdef __LITTLE_ENDIAN__
6731#define vget_lane_u8(__p0, __p1) __extension__ ({ \
6732  uint8x8_t __s0 = __p0; \
6733  uint8_t __ret; \
6734  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6735  __ret; \
6736})
6737#else
6738#define vget_lane_u8(__p0, __p1) __extension__ ({ \
6739  uint8x8_t __s0 = __p0; \
6740  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6741  uint8_t __ret; \
6742  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
6743  __ret; \
6744})
6745#define __noswap_vget_lane_u8(__p0, __p1) __extension__ ({ \
6746  uint8x8_t __s0 = __p0; \
6747  uint8_t __ret; \
6748  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6749  __ret; \
6750})
6751#endif
6752
6753#ifdef __LITTLE_ENDIAN__
6754#define vget_lane_u32(__p0, __p1) __extension__ ({ \
6755  uint32x2_t __s0 = __p0; \
6756  uint32_t __ret; \
6757  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6758  __ret; \
6759})
6760#else
6761#define vget_lane_u32(__p0, __p1) __extension__ ({ \
6762  uint32x2_t __s0 = __p0; \
6763  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6764  uint32_t __ret; \
6765  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__rev0, __p1); \
6766  __ret; \
6767})
6768#define __noswap_vget_lane_u32(__p0, __p1) __extension__ ({ \
6769  uint32x2_t __s0 = __p0; \
6770  uint32_t __ret; \
6771  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6772  __ret; \
6773})
6774#endif
6775
6776#ifdef __LITTLE_ENDIAN__
6777#define vget_lane_u64(__p0, __p1) __extension__ ({ \
6778  uint64x1_t __s0 = __p0; \
6779  uint64_t __ret; \
6780  __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6781  __ret; \
6782})
6783#else
6784#define vget_lane_u64(__p0, __p1) __extension__ ({ \
6785  uint64x1_t __s0 = __p0; \
6786  uint64_t __ret; \
6787  __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6788  __ret; \
6789})
6790#define __noswap_vget_lane_u64(__p0, __p1) __extension__ ({ \
6791  uint64x1_t __s0 = __p0; \
6792  uint64_t __ret; \
6793  __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6794  __ret; \
6795})
6796#endif
6797
6798#ifdef __LITTLE_ENDIAN__
6799#define vget_lane_u16(__p0, __p1) __extension__ ({ \
6800  uint16x4_t __s0 = __p0; \
6801  uint16_t __ret; \
6802  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6803  __ret; \
6804})
6805#else
6806#define vget_lane_u16(__p0, __p1) __extension__ ({ \
6807  uint16x4_t __s0 = __p0; \
6808  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6809  uint16_t __ret; \
6810  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
6811  __ret; \
6812})
6813#define __noswap_vget_lane_u16(__p0, __p1) __extension__ ({ \
6814  uint16x4_t __s0 = __p0; \
6815  uint16_t __ret; \
6816  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6817  __ret; \
6818})
6819#endif
6820
6821#ifdef __LITTLE_ENDIAN__
6822#define vget_lane_s8(__p0, __p1) __extension__ ({ \
6823  int8x8_t __s0 = __p0; \
6824  int8_t __ret; \
6825  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6826  __ret; \
6827})
6828#else
6829#define vget_lane_s8(__p0, __p1) __extension__ ({ \
6830  int8x8_t __s0 = __p0; \
6831  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6832  int8_t __ret; \
6833  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
6834  __ret; \
6835})
6836#define __noswap_vget_lane_s8(__p0, __p1) __extension__ ({ \
6837  int8x8_t __s0 = __p0; \
6838  int8_t __ret; \
6839  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6840  __ret; \
6841})
6842#endif
6843
6844#ifdef __LITTLE_ENDIAN__
6845#define vget_lane_f32(__p0, __p1) __extension__ ({ \
6846  float32x2_t __s0 = __p0; \
6847  float32_t __ret; \
6848  __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__s0, __p1); \
6849  __ret; \
6850})
6851#else
6852#define vget_lane_f32(__p0, __p1) __extension__ ({ \
6853  float32x2_t __s0 = __p0; \
6854  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6855  float32_t __ret; \
6856  __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__rev0, __p1); \
6857  __ret; \
6858})
6859#define __noswap_vget_lane_f32(__p0, __p1) __extension__ ({ \
6860  float32x2_t __s0 = __p0; \
6861  float32_t __ret; \
6862  __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__s0, __p1); \
6863  __ret; \
6864})
6865#endif
6866
6867#ifdef __LITTLE_ENDIAN__
6868#define vget_lane_s32(__p0, __p1) __extension__ ({ \
6869  int32x2_t __s0 = __p0; \
6870  int32_t __ret; \
6871  __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6872  __ret; \
6873})
6874#else
6875#define vget_lane_s32(__p0, __p1) __extension__ ({ \
6876  int32x2_t __s0 = __p0; \
6877  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6878  int32_t __ret; \
6879  __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__rev0, __p1); \
6880  __ret; \
6881})
6882#define __noswap_vget_lane_s32(__p0, __p1) __extension__ ({ \
6883  int32x2_t __s0 = __p0; \
6884  int32_t __ret; \
6885  __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6886  __ret; \
6887})
6888#endif
6889
6890#ifdef __LITTLE_ENDIAN__
6891#define vget_lane_s64(__p0, __p1) __extension__ ({ \
6892  int64x1_t __s0 = __p0; \
6893  int64_t __ret; \
6894  __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6895  __ret; \
6896})
6897#else
6898#define vget_lane_s64(__p0, __p1) __extension__ ({ \
6899  int64x1_t __s0 = __p0; \
6900  int64_t __ret; \
6901  __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6902  __ret; \
6903})
6904#define __noswap_vget_lane_s64(__p0, __p1) __extension__ ({ \
6905  int64x1_t __s0 = __p0; \
6906  int64_t __ret; \
6907  __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6908  __ret; \
6909})
6910#endif
6911
6912#ifdef __LITTLE_ENDIAN__
6913#define vget_lane_s16(__p0, __p1) __extension__ ({ \
6914  int16x4_t __s0 = __p0; \
6915  int16_t __ret; \
6916  __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6917  __ret; \
6918})
6919#else
6920#define vget_lane_s16(__p0, __p1) __extension__ ({ \
6921  int16x4_t __s0 = __p0; \
6922  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6923  int16_t __ret; \
6924  __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
6925  __ret; \
6926})
6927#define __noswap_vget_lane_s16(__p0, __p1) __extension__ ({ \
6928  int16x4_t __s0 = __p0; \
6929  int16_t __ret; \
6930  __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6931  __ret; \
6932})
6933#endif
6934
6935#ifdef __LITTLE_ENDIAN__
6936__ai poly8x8_t vget_low_p8(poly8x16_t __p0) {
6937  poly8x8_t __ret;
6938  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
6939  return __ret;
6940}
6941#else
6942__ai poly8x8_t vget_low_p8(poly8x16_t __p0) {
6943  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6944  poly8x8_t __ret;
6945  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
6946  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6947  return __ret;
6948}
6949#endif
6950
6951#ifdef __LITTLE_ENDIAN__
6952__ai poly16x4_t vget_low_p16(poly16x8_t __p0) {
6953  poly16x4_t __ret;
6954  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
6955  return __ret;
6956}
6957#else
6958__ai poly16x4_t vget_low_p16(poly16x8_t __p0) {
6959  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6960  poly16x4_t __ret;
6961  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
6962  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6963  return __ret;
6964}
6965#endif
6966
6967#ifdef __LITTLE_ENDIAN__
6968__ai uint8x8_t vget_low_u8(uint8x16_t __p0) {
6969  uint8x8_t __ret;
6970  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
6971  return __ret;
6972}
6973#else
6974__ai uint8x8_t vget_low_u8(uint8x16_t __p0) {
6975  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6976  uint8x8_t __ret;
6977  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
6978  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6979  return __ret;
6980}
6981#endif
6982
6983#ifdef __LITTLE_ENDIAN__
6984__ai uint32x2_t vget_low_u32(uint32x4_t __p0) {
6985  uint32x2_t __ret;
6986  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
6987  return __ret;
6988}
6989#else
6990__ai uint32x2_t vget_low_u32(uint32x4_t __p0) {
6991  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6992  uint32x2_t __ret;
6993  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
6994  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6995  return __ret;
6996}
6997#endif
6998
6999#ifdef __LITTLE_ENDIAN__
7000__ai uint64x1_t vget_low_u64(uint64x2_t __p0) {
7001  uint64x1_t __ret;
7002  __ret = __builtin_shufflevector(__p0, __p0, 0);
7003  return __ret;
7004}
7005#else
7006__ai uint64x1_t vget_low_u64(uint64x2_t __p0) {
7007  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7008  uint64x1_t __ret;
7009  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
7010  return __ret;
7011}
7012#endif
7013
7014#ifdef __LITTLE_ENDIAN__
7015__ai uint16x4_t vget_low_u16(uint16x8_t __p0) {
7016  uint16x4_t __ret;
7017  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7018  return __ret;
7019}
7020#else
7021__ai uint16x4_t vget_low_u16(uint16x8_t __p0) {
7022  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7023  uint16x4_t __ret;
7024  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7025  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7026  return __ret;
7027}
7028#endif
7029
7030#ifdef __LITTLE_ENDIAN__
7031__ai int8x8_t vget_low_s8(int8x16_t __p0) {
7032  int8x8_t __ret;
7033  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
7034  return __ret;
7035}
7036#else
7037__ai int8x8_t vget_low_s8(int8x16_t __p0) {
7038  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7039  int8x8_t __ret;
7040  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
7041  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7042  return __ret;
7043}
7044#endif
7045
7046#ifdef __LITTLE_ENDIAN__
7047__ai float32x2_t vget_low_f32(float32x4_t __p0) {
7048  float32x2_t __ret;
7049  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
7050  return __ret;
7051}
7052#else
7053__ai float32x2_t vget_low_f32(float32x4_t __p0) {
7054  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7055  float32x2_t __ret;
7056  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
7057  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7058  return __ret;
7059}
7060#endif
7061
7062#ifdef __LITTLE_ENDIAN__
7063__ai float16x4_t vget_low_f16(float16x8_t __p0) {
7064  float16x4_t __ret;
7065  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7066  return __ret;
7067}
7068#else
7069__ai float16x4_t vget_low_f16(float16x8_t __p0) {
7070  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7071  float16x4_t __ret;
7072  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7073  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7074  return __ret;
7075}
7076#endif
7077
7078#ifdef __LITTLE_ENDIAN__
7079__ai int32x2_t vget_low_s32(int32x4_t __p0) {
7080  int32x2_t __ret;
7081  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
7082  return __ret;
7083}
7084#else
7085__ai int32x2_t vget_low_s32(int32x4_t __p0) {
7086  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7087  int32x2_t __ret;
7088  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
7089  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7090  return __ret;
7091}
7092#endif
7093
7094#ifdef __LITTLE_ENDIAN__
7095__ai int64x1_t vget_low_s64(int64x2_t __p0) {
7096  int64x1_t __ret;
7097  __ret = __builtin_shufflevector(__p0, __p0, 0);
7098  return __ret;
7099}
7100#else
7101__ai int64x1_t vget_low_s64(int64x2_t __p0) {
7102  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7103  int64x1_t __ret;
7104  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
7105  return __ret;
7106}
7107#endif
7108
7109#ifdef __LITTLE_ENDIAN__
7110__ai int16x4_t vget_low_s16(int16x8_t __p0) {
7111  int16x4_t __ret;
7112  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7113  return __ret;
7114}
7115#else
7116__ai int16x4_t vget_low_s16(int16x8_t __p0) {
7117  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7118  int16x4_t __ret;
7119  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7120  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7121  return __ret;
7122}
7123#endif
7124
7125#ifdef __LITTLE_ENDIAN__
7126__ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7127  uint8x16_t __ret;
7128  __ret = (uint8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
7129  return __ret;
7130}
7131#else
7132__ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7133  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7134  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7135  uint8x16_t __ret;
7136  __ret = (uint8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
7137  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7138  return __ret;
7139}
7140#endif
7141
7142#ifdef __LITTLE_ENDIAN__
7143__ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7144  uint32x4_t __ret;
7145  __ret = (uint32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
7146  return __ret;
7147}
7148#else
7149__ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7150  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7151  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7152  uint32x4_t __ret;
7153  __ret = (uint32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
7154  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7155  return __ret;
7156}
7157#endif
7158
7159#ifdef __LITTLE_ENDIAN__
7160__ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7161  uint16x8_t __ret;
7162  __ret = (uint16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
7163  return __ret;
7164}
7165#else
7166__ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7167  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7168  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7169  uint16x8_t __ret;
7170  __ret = (uint16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
7171  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7172  return __ret;
7173}
7174#endif
7175
7176#ifdef __LITTLE_ENDIAN__
7177__ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
7178  int8x16_t __ret;
7179  __ret = (int8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
7180  return __ret;
7181}
7182#else
7183__ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
7184  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7185  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7186  int8x16_t __ret;
7187  __ret = (int8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
7188  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7189  return __ret;
7190}
7191#endif
7192
7193#ifdef __LITTLE_ENDIAN__
7194__ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
7195  int32x4_t __ret;
7196  __ret = (int32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
7197  return __ret;
7198}
7199#else
7200__ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
7201  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7202  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7203  int32x4_t __ret;
7204  __ret = (int32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
7205  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7206  return __ret;
7207}
7208#endif
7209
7210#ifdef __LITTLE_ENDIAN__
7211__ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
7212  int16x8_t __ret;
7213  __ret = (int16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
7214  return __ret;
7215}
7216#else
7217__ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
7218  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7219  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7220  int16x8_t __ret;
7221  __ret = (int16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
7222  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7223  return __ret;
7224}
7225#endif
7226
7227#ifdef __LITTLE_ENDIAN__
7228__ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
7229  uint8x8_t __ret;
7230  __ret = (uint8x8_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
7231  return __ret;
7232}
7233#else
7234__ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
7235  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7236  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7237  uint8x8_t __ret;
7238  __ret = (uint8x8_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
7239  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7240  return __ret;
7241}
7242#endif
7243
7244#ifdef __LITTLE_ENDIAN__
7245__ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
7246  uint32x2_t __ret;
7247  __ret = (uint32x2_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
7248  return __ret;
7249}
7250#else
7251__ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
7252  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7253  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7254  uint32x2_t __ret;
7255  __ret = (uint32x2_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
7256  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7257  return __ret;
7258}
7259#endif
7260
7261#ifdef __LITTLE_ENDIAN__
7262__ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
7263  uint16x4_t __ret;
7264  __ret = (uint16x4_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
7265  return __ret;
7266}
7267#else
7268__ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
7269  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7270  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7271  uint16x4_t __ret;
7272  __ret = (uint16x4_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
7273  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7274  return __ret;
7275}
7276#endif
7277
7278#ifdef __LITTLE_ENDIAN__
7279__ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1) {
7280  int8x8_t __ret;
7281  __ret = (int8x8_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
7282  return __ret;
7283}
7284#else
7285__ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1) {
7286  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7287  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7288  int8x8_t __ret;
7289  __ret = (int8x8_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
7290  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7291  return __ret;
7292}
7293#endif
7294
7295#ifdef __LITTLE_ENDIAN__
7296__ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1) {
7297  int32x2_t __ret;
7298  __ret = (int32x2_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
7299  return __ret;
7300}
7301#else
7302__ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1) {
7303  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7304  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7305  int32x2_t __ret;
7306  __ret = (int32x2_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
7307  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7308  return __ret;
7309}
7310#endif
7311
7312#ifdef __LITTLE_ENDIAN__
7313__ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1) {
7314  int16x4_t __ret;
7315  __ret = (int16x4_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
7316  return __ret;
7317}
7318#else
7319__ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1) {
7320  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7321  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7322  int16x4_t __ret;
7323  __ret = (int16x4_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
7324  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7325  return __ret;
7326}
7327#endif
7328
7329#ifdef __LITTLE_ENDIAN__
7330__ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7331  uint8x16_t __ret;
7332  __ret = (uint8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
7333  return __ret;
7334}
7335#else
7336__ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7337  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7338  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7339  uint8x16_t __ret;
7340  __ret = (uint8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
7341  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7342  return __ret;
7343}
7344#endif
7345
7346#ifdef __LITTLE_ENDIAN__
7347__ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7348  uint32x4_t __ret;
7349  __ret = (uint32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
7350  return __ret;
7351}
7352#else
7353__ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7354  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7355  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7356  uint32x4_t __ret;
7357  __ret = (uint32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
7358  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7359  return __ret;
7360}
7361#endif
7362
7363#ifdef __LITTLE_ENDIAN__
7364__ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7365  uint16x8_t __ret;
7366  __ret = (uint16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
7367  return __ret;
7368}
7369#else
7370__ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7371  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7372  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7373  uint16x8_t __ret;
7374  __ret = (uint16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
7375  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7376  return __ret;
7377}
7378#endif
7379
7380#ifdef __LITTLE_ENDIAN__
7381__ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1) {
7382  int8x16_t __ret;
7383  __ret = (int8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
7384  return __ret;
7385}
7386#else
7387__ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1) {
7388  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7389  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7390  int8x16_t __ret;
7391  __ret = (int8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
7392  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7393  return __ret;
7394}
7395#endif
7396
7397#ifdef __LITTLE_ENDIAN__
7398__ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1) {
7399  int32x4_t __ret;
7400  __ret = (int32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
7401  return __ret;
7402}
7403#else
7404__ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1) {
7405  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7406  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7407  int32x4_t __ret;
7408  __ret = (int32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
7409  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7410  return __ret;
7411}
7412#endif
7413
7414#ifdef __LITTLE_ENDIAN__
7415__ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1) {
7416  int16x8_t __ret;
7417  __ret = (int16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
7418  return __ret;
7419}
7420#else
7421__ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1) {
7422  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7423  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7424  int16x8_t __ret;
7425  __ret = (int16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
7426  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7427  return __ret;
7428}
7429#endif
7430
7431#ifdef __LITTLE_ENDIAN__
7432__ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
7433  uint8x8_t __ret;
7434  __ret = (uint8x8_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
7435  return __ret;
7436}
7437#else
7438__ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
7439  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7440  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7441  uint8x8_t __ret;
7442  __ret = (uint8x8_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
7443  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7444  return __ret;
7445}
7446#endif
7447
7448#ifdef __LITTLE_ENDIAN__
7449__ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
7450  uint32x2_t __ret;
7451  __ret = (uint32x2_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
7452  return __ret;
7453}
7454#else
7455__ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
7456  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7457  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7458  uint32x2_t __ret;
7459  __ret = (uint32x2_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
7460  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7461  return __ret;
7462}
7463#endif
7464
7465#ifdef __LITTLE_ENDIAN__
7466__ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
7467  uint16x4_t __ret;
7468  __ret = (uint16x4_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
7469  return __ret;
7470}
7471#else
7472__ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
7473  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7474  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7475  uint16x4_t __ret;
7476  __ret = (uint16x4_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
7477  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7478  return __ret;
7479}
7480#endif
7481
7482#ifdef __LITTLE_ENDIAN__
7483__ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1) {
7484  int8x8_t __ret;
7485  __ret = (int8x8_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
7486  return __ret;
7487}
7488#else
7489__ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1) {
7490  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7491  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7492  int8x8_t __ret;
7493  __ret = (int8x8_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
7494  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7495  return __ret;
7496}
7497#endif
7498
7499#ifdef __LITTLE_ENDIAN__
7500__ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1) {
7501  int32x2_t __ret;
7502  __ret = (int32x2_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
7503  return __ret;
7504}
7505#else
7506__ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1) {
7507  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7508  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7509  int32x2_t __ret;
7510  __ret = (int32x2_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
7511  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7512  return __ret;
7513}
7514#endif
7515
7516#ifdef __LITTLE_ENDIAN__
7517__ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1) {
7518  int16x4_t __ret;
7519  __ret = (int16x4_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
7520  return __ret;
7521}
7522#else
7523__ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1) {
7524  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7525  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7526  int16x4_t __ret;
7527  __ret = (int16x4_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
7528  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7529  return __ret;
7530}
7531#endif
7532
7533#ifdef __LITTLE_ENDIAN__
7534#define vld1_p8(__p0) __extension__ ({ \
7535  poly8x8_t __ret; \
7536  __ret = (poly8x8_t) __builtin_neon_vld1_v(__p0, 4); \
7537  __ret; \
7538})
7539#else
7540#define vld1_p8(__p0) __extension__ ({ \
7541  poly8x8_t __ret; \
7542  __ret = (poly8x8_t) __builtin_neon_vld1_v(__p0, 4); \
7543  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7544  __ret; \
7545})
7546#endif
7547
7548#ifdef __LITTLE_ENDIAN__
7549#define vld1_p16(__p0) __extension__ ({ \
7550  poly16x4_t __ret; \
7551  __ret = (poly16x4_t) __builtin_neon_vld1_v(__p0, 5); \
7552  __ret; \
7553})
7554#else
7555#define vld1_p16(__p0) __extension__ ({ \
7556  poly16x4_t __ret; \
7557  __ret = (poly16x4_t) __builtin_neon_vld1_v(__p0, 5); \
7558  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7559  __ret; \
7560})
7561#endif
7562
7563#ifdef __LITTLE_ENDIAN__
7564#define vld1q_p8(__p0) __extension__ ({ \
7565  poly8x16_t __ret; \
7566  __ret = (poly8x16_t) __builtin_neon_vld1q_v(__p0, 36); \
7567  __ret; \
7568})
7569#else
7570#define vld1q_p8(__p0) __extension__ ({ \
7571  poly8x16_t __ret; \
7572  __ret = (poly8x16_t) __builtin_neon_vld1q_v(__p0, 36); \
7573  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7574  __ret; \
7575})
7576#endif
7577
7578#ifdef __LITTLE_ENDIAN__
7579#define vld1q_p16(__p0) __extension__ ({ \
7580  poly16x8_t __ret; \
7581  __ret = (poly16x8_t) __builtin_neon_vld1q_v(__p0, 37); \
7582  __ret; \
7583})
7584#else
7585#define vld1q_p16(__p0) __extension__ ({ \
7586  poly16x8_t __ret; \
7587  __ret = (poly16x8_t) __builtin_neon_vld1q_v(__p0, 37); \
7588  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7589  __ret; \
7590})
7591#endif
7592
7593#ifdef __LITTLE_ENDIAN__
7594#define vld1q_u8(__p0) __extension__ ({ \
7595  uint8x16_t __ret; \
7596  __ret = (uint8x16_t) __builtin_neon_vld1q_v(__p0, 48); \
7597  __ret; \
7598})
7599#else
7600#define vld1q_u8(__p0) __extension__ ({ \
7601  uint8x16_t __ret; \
7602  __ret = (uint8x16_t) __builtin_neon_vld1q_v(__p0, 48); \
7603  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7604  __ret; \
7605})
7606#endif
7607
7608#ifdef __LITTLE_ENDIAN__
7609#define vld1q_u32(__p0) __extension__ ({ \
7610  uint32x4_t __ret; \
7611  __ret = (uint32x4_t) __builtin_neon_vld1q_v(__p0, 50); \
7612  __ret; \
7613})
7614#else
7615#define vld1q_u32(__p0) __extension__ ({ \
7616  uint32x4_t __ret; \
7617  __ret = (uint32x4_t) __builtin_neon_vld1q_v(__p0, 50); \
7618  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7619  __ret; \
7620})
7621#endif
7622
7623#ifdef __LITTLE_ENDIAN__
7624#define vld1q_u64(__p0) __extension__ ({ \
7625  uint64x2_t __ret; \
7626  __ret = (uint64x2_t) __builtin_neon_vld1q_v(__p0, 51); \
7627  __ret; \
7628})
7629#else
7630#define vld1q_u64(__p0) __extension__ ({ \
7631  uint64x2_t __ret; \
7632  __ret = (uint64x2_t) __builtin_neon_vld1q_v(__p0, 51); \
7633  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7634  __ret; \
7635})
7636#endif
7637
7638#ifdef __LITTLE_ENDIAN__
7639#define vld1q_u16(__p0) __extension__ ({ \
7640  uint16x8_t __ret; \
7641  __ret = (uint16x8_t) __builtin_neon_vld1q_v(__p0, 49); \
7642  __ret; \
7643})
7644#else
7645#define vld1q_u16(__p0) __extension__ ({ \
7646  uint16x8_t __ret; \
7647  __ret = (uint16x8_t) __builtin_neon_vld1q_v(__p0, 49); \
7648  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7649  __ret; \
7650})
7651#endif
7652
7653#ifdef __LITTLE_ENDIAN__
7654#define vld1q_s8(__p0) __extension__ ({ \
7655  int8x16_t __ret; \
7656  __ret = (int8x16_t) __builtin_neon_vld1q_v(__p0, 32); \
7657  __ret; \
7658})
7659#else
7660#define vld1q_s8(__p0) __extension__ ({ \
7661  int8x16_t __ret; \
7662  __ret = (int8x16_t) __builtin_neon_vld1q_v(__p0, 32); \
7663  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7664  __ret; \
7665})
7666#endif
7667
7668#ifdef __LITTLE_ENDIAN__
7669#define vld1q_f32(__p0) __extension__ ({ \
7670  float32x4_t __ret; \
7671  __ret = (float32x4_t) __builtin_neon_vld1q_v(__p0, 41); \
7672  __ret; \
7673})
7674#else
7675#define vld1q_f32(__p0) __extension__ ({ \
7676  float32x4_t __ret; \
7677  __ret = (float32x4_t) __builtin_neon_vld1q_v(__p0, 41); \
7678  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7679  __ret; \
7680})
7681#endif
7682
7683#ifdef __LITTLE_ENDIAN__
7684#define vld1q_f16(__p0) __extension__ ({ \
7685  float16x8_t __ret; \
7686  __ret = (float16x8_t) __builtin_neon_vld1q_v(__p0, 40); \
7687  __ret; \
7688})
7689#else
7690#define vld1q_f16(__p0) __extension__ ({ \
7691  float16x8_t __ret; \
7692  __ret = (float16x8_t) __builtin_neon_vld1q_v(__p0, 40); \
7693  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7694  __ret; \
7695})
7696#endif
7697
7698#ifdef __LITTLE_ENDIAN__
7699#define vld1q_s32(__p0) __extension__ ({ \
7700  int32x4_t __ret; \
7701  __ret = (int32x4_t) __builtin_neon_vld1q_v(__p0, 34); \
7702  __ret; \
7703})
7704#else
7705#define vld1q_s32(__p0) __extension__ ({ \
7706  int32x4_t __ret; \
7707  __ret = (int32x4_t) __builtin_neon_vld1q_v(__p0, 34); \
7708  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7709  __ret; \
7710})
7711#endif
7712
7713#ifdef __LITTLE_ENDIAN__
7714#define vld1q_s64(__p0) __extension__ ({ \
7715  int64x2_t __ret; \
7716  __ret = (int64x2_t) __builtin_neon_vld1q_v(__p0, 35); \
7717  __ret; \
7718})
7719#else
7720#define vld1q_s64(__p0) __extension__ ({ \
7721  int64x2_t __ret; \
7722  __ret = (int64x2_t) __builtin_neon_vld1q_v(__p0, 35); \
7723  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7724  __ret; \
7725})
7726#endif
7727
7728#ifdef __LITTLE_ENDIAN__
7729#define vld1q_s16(__p0) __extension__ ({ \
7730  int16x8_t __ret; \
7731  __ret = (int16x8_t) __builtin_neon_vld1q_v(__p0, 33); \
7732  __ret; \
7733})
7734#else
7735#define vld1q_s16(__p0) __extension__ ({ \
7736  int16x8_t __ret; \
7737  __ret = (int16x8_t) __builtin_neon_vld1q_v(__p0, 33); \
7738  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7739  __ret; \
7740})
7741#endif
7742
7743#ifdef __LITTLE_ENDIAN__
7744#define vld1_u8(__p0) __extension__ ({ \
7745  uint8x8_t __ret; \
7746  __ret = (uint8x8_t) __builtin_neon_vld1_v(__p0, 16); \
7747  __ret; \
7748})
7749#else
7750#define vld1_u8(__p0) __extension__ ({ \
7751  uint8x8_t __ret; \
7752  __ret = (uint8x8_t) __builtin_neon_vld1_v(__p0, 16); \
7753  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7754  __ret; \
7755})
7756#endif
7757
7758#ifdef __LITTLE_ENDIAN__
7759#define vld1_u32(__p0) __extension__ ({ \
7760  uint32x2_t __ret; \
7761  __ret = (uint32x2_t) __builtin_neon_vld1_v(__p0, 18); \
7762  __ret; \
7763})
7764#else
7765#define vld1_u32(__p0) __extension__ ({ \
7766  uint32x2_t __ret; \
7767  __ret = (uint32x2_t) __builtin_neon_vld1_v(__p0, 18); \
7768  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7769  __ret; \
7770})
7771#endif
7772
7773#ifdef __LITTLE_ENDIAN__
7774#define vld1_u64(__p0) __extension__ ({ \
7775  uint64x1_t __ret; \
7776  __ret = (uint64x1_t) __builtin_neon_vld1_v(__p0, 19); \
7777  __ret; \
7778})
7779#else
7780#define vld1_u64(__p0) __extension__ ({ \
7781  uint64x1_t __ret; \
7782  __ret = (uint64x1_t) __builtin_neon_vld1_v(__p0, 19); \
7783  __ret; \
7784})
7785#endif
7786
7787#ifdef __LITTLE_ENDIAN__
7788#define vld1_u16(__p0) __extension__ ({ \
7789  uint16x4_t __ret; \
7790  __ret = (uint16x4_t) __builtin_neon_vld1_v(__p0, 17); \
7791  __ret; \
7792})
7793#else
7794#define vld1_u16(__p0) __extension__ ({ \
7795  uint16x4_t __ret; \
7796  __ret = (uint16x4_t) __builtin_neon_vld1_v(__p0, 17); \
7797  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7798  __ret; \
7799})
7800#endif
7801
7802#ifdef __LITTLE_ENDIAN__
7803#define vld1_s8(__p0) __extension__ ({ \
7804  int8x8_t __ret; \
7805  __ret = (int8x8_t) __builtin_neon_vld1_v(__p0, 0); \
7806  __ret; \
7807})
7808#else
7809#define vld1_s8(__p0) __extension__ ({ \
7810  int8x8_t __ret; \
7811  __ret = (int8x8_t) __builtin_neon_vld1_v(__p0, 0); \
7812  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7813  __ret; \
7814})
7815#endif
7816
7817#ifdef __LITTLE_ENDIAN__
7818#define vld1_f32(__p0) __extension__ ({ \
7819  float32x2_t __ret; \
7820  __ret = (float32x2_t) __builtin_neon_vld1_v(__p0, 9); \
7821  __ret; \
7822})
7823#else
7824#define vld1_f32(__p0) __extension__ ({ \
7825  float32x2_t __ret; \
7826  __ret = (float32x2_t) __builtin_neon_vld1_v(__p0, 9); \
7827  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7828  __ret; \
7829})
7830#endif
7831
7832#ifdef __LITTLE_ENDIAN__
7833#define vld1_f16(__p0) __extension__ ({ \
7834  float16x4_t __ret; \
7835  __ret = (float16x4_t) __builtin_neon_vld1_v(__p0, 8); \
7836  __ret; \
7837})
7838#else
7839#define vld1_f16(__p0) __extension__ ({ \
7840  float16x4_t __ret; \
7841  __ret = (float16x4_t) __builtin_neon_vld1_v(__p0, 8); \
7842  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7843  __ret; \
7844})
7845#endif
7846
7847#ifdef __LITTLE_ENDIAN__
7848#define vld1_s32(__p0) __extension__ ({ \
7849  int32x2_t __ret; \
7850  __ret = (int32x2_t) __builtin_neon_vld1_v(__p0, 2); \
7851  __ret; \
7852})
7853#else
7854#define vld1_s32(__p0) __extension__ ({ \
7855  int32x2_t __ret; \
7856  __ret = (int32x2_t) __builtin_neon_vld1_v(__p0, 2); \
7857  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7858  __ret; \
7859})
7860#endif
7861
7862#ifdef __LITTLE_ENDIAN__
7863#define vld1_s64(__p0) __extension__ ({ \
7864  int64x1_t __ret; \
7865  __ret = (int64x1_t) __builtin_neon_vld1_v(__p0, 3); \
7866  __ret; \
7867})
7868#else
7869#define vld1_s64(__p0) __extension__ ({ \
7870  int64x1_t __ret; \
7871  __ret = (int64x1_t) __builtin_neon_vld1_v(__p0, 3); \
7872  __ret; \
7873})
7874#endif
7875
7876#ifdef __LITTLE_ENDIAN__
7877#define vld1_s16(__p0) __extension__ ({ \
7878  int16x4_t __ret; \
7879  __ret = (int16x4_t) __builtin_neon_vld1_v(__p0, 1); \
7880  __ret; \
7881})
7882#else
7883#define vld1_s16(__p0) __extension__ ({ \
7884  int16x4_t __ret; \
7885  __ret = (int16x4_t) __builtin_neon_vld1_v(__p0, 1); \
7886  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7887  __ret; \
7888})
7889#endif
7890
7891#ifdef __LITTLE_ENDIAN__
7892#define vld1_dup_p8(__p0) __extension__ ({ \
7893  poly8x8_t __ret; \
7894  __ret = (poly8x8_t) __builtin_neon_vld1_dup_v(__p0, 4); \
7895  __ret; \
7896})
7897#else
7898#define vld1_dup_p8(__p0) __extension__ ({ \
7899  poly8x8_t __ret; \
7900  __ret = (poly8x8_t) __builtin_neon_vld1_dup_v(__p0, 4); \
7901  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7902  __ret; \
7903})
7904#endif
7905
7906#ifdef __LITTLE_ENDIAN__
7907#define vld1_dup_p16(__p0) __extension__ ({ \
7908  poly16x4_t __ret; \
7909  __ret = (poly16x4_t) __builtin_neon_vld1_dup_v(__p0, 5); \
7910  __ret; \
7911})
7912#else
7913#define vld1_dup_p16(__p0) __extension__ ({ \
7914  poly16x4_t __ret; \
7915  __ret = (poly16x4_t) __builtin_neon_vld1_dup_v(__p0, 5); \
7916  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7917  __ret; \
7918})
7919#endif
7920
7921#ifdef __LITTLE_ENDIAN__
7922#define vld1q_dup_p8(__p0) __extension__ ({ \
7923  poly8x16_t __ret; \
7924  __ret = (poly8x16_t) __builtin_neon_vld1q_dup_v(__p0, 36); \
7925  __ret; \
7926})
7927#else
7928#define vld1q_dup_p8(__p0) __extension__ ({ \
7929  poly8x16_t __ret; \
7930  __ret = (poly8x16_t) __builtin_neon_vld1q_dup_v(__p0, 36); \
7931  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7932  __ret; \
7933})
7934#endif
7935
7936#ifdef __LITTLE_ENDIAN__
7937#define vld1q_dup_p16(__p0) __extension__ ({ \
7938  poly16x8_t __ret; \
7939  __ret = (poly16x8_t) __builtin_neon_vld1q_dup_v(__p0, 37); \
7940  __ret; \
7941})
7942#else
7943#define vld1q_dup_p16(__p0) __extension__ ({ \
7944  poly16x8_t __ret; \
7945  __ret = (poly16x8_t) __builtin_neon_vld1q_dup_v(__p0, 37); \
7946  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7947  __ret; \
7948})
7949#endif
7950
7951#ifdef __LITTLE_ENDIAN__
7952#define vld1q_dup_u8(__p0) __extension__ ({ \
7953  uint8x16_t __ret; \
7954  __ret = (uint8x16_t) __builtin_neon_vld1q_dup_v(__p0, 48); \
7955  __ret; \
7956})
7957#else
7958#define vld1q_dup_u8(__p0) __extension__ ({ \
7959  uint8x16_t __ret; \
7960  __ret = (uint8x16_t) __builtin_neon_vld1q_dup_v(__p0, 48); \
7961  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7962  __ret; \
7963})
7964#endif
7965
7966#ifdef __LITTLE_ENDIAN__
7967#define vld1q_dup_u32(__p0) __extension__ ({ \
7968  uint32x4_t __ret; \
7969  __ret = (uint32x4_t) __builtin_neon_vld1q_dup_v(__p0, 50); \
7970  __ret; \
7971})
7972#else
7973#define vld1q_dup_u32(__p0) __extension__ ({ \
7974  uint32x4_t __ret; \
7975  __ret = (uint32x4_t) __builtin_neon_vld1q_dup_v(__p0, 50); \
7976  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7977  __ret; \
7978})
7979#endif
7980
7981#ifdef __LITTLE_ENDIAN__
7982#define vld1q_dup_u64(__p0) __extension__ ({ \
7983  uint64x2_t __ret; \
7984  __ret = (uint64x2_t) __builtin_neon_vld1q_dup_v(__p0, 51); \
7985  __ret; \
7986})
7987#else
7988#define vld1q_dup_u64(__p0) __extension__ ({ \
7989  uint64x2_t __ret; \
7990  __ret = (uint64x2_t) __builtin_neon_vld1q_dup_v(__p0, 51); \
7991  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7992  __ret; \
7993})
7994#endif
7995
7996#ifdef __LITTLE_ENDIAN__
7997#define vld1q_dup_u16(__p0) __extension__ ({ \
7998  uint16x8_t __ret; \
7999  __ret = (uint16x8_t) __builtin_neon_vld1q_dup_v(__p0, 49); \
8000  __ret; \
8001})
8002#else
8003#define vld1q_dup_u16(__p0) __extension__ ({ \
8004  uint16x8_t __ret; \
8005  __ret = (uint16x8_t) __builtin_neon_vld1q_dup_v(__p0, 49); \
8006  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8007  __ret; \
8008})
8009#endif
8010
8011#ifdef __LITTLE_ENDIAN__
8012#define vld1q_dup_s8(__p0) __extension__ ({ \
8013  int8x16_t __ret; \
8014  __ret = (int8x16_t) __builtin_neon_vld1q_dup_v(__p0, 32); \
8015  __ret; \
8016})
8017#else
8018#define vld1q_dup_s8(__p0) __extension__ ({ \
8019  int8x16_t __ret; \
8020  __ret = (int8x16_t) __builtin_neon_vld1q_dup_v(__p0, 32); \
8021  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8022  __ret; \
8023})
8024#endif
8025
8026#ifdef __LITTLE_ENDIAN__
8027#define vld1q_dup_f32(__p0) __extension__ ({ \
8028  float32x4_t __ret; \
8029  __ret = (float32x4_t) __builtin_neon_vld1q_dup_v(__p0, 41); \
8030  __ret; \
8031})
8032#else
8033#define vld1q_dup_f32(__p0) __extension__ ({ \
8034  float32x4_t __ret; \
8035  __ret = (float32x4_t) __builtin_neon_vld1q_dup_v(__p0, 41); \
8036  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8037  __ret; \
8038})
8039#endif
8040
8041#ifdef __LITTLE_ENDIAN__
8042#define vld1q_dup_f16(__p0) __extension__ ({ \
8043  float16x8_t __ret; \
8044  __ret = (float16x8_t) __builtin_neon_vld1q_dup_v(__p0, 40); \
8045  __ret; \
8046})
8047#else
8048#define vld1q_dup_f16(__p0) __extension__ ({ \
8049  float16x8_t __ret; \
8050  __ret = (float16x8_t) __builtin_neon_vld1q_dup_v(__p0, 40); \
8051  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8052  __ret; \
8053})
8054#endif
8055
8056#ifdef __LITTLE_ENDIAN__
8057#define vld1q_dup_s32(__p0) __extension__ ({ \
8058  int32x4_t __ret; \
8059  __ret = (int32x4_t) __builtin_neon_vld1q_dup_v(__p0, 34); \
8060  __ret; \
8061})
8062#else
8063#define vld1q_dup_s32(__p0) __extension__ ({ \
8064  int32x4_t __ret; \
8065  __ret = (int32x4_t) __builtin_neon_vld1q_dup_v(__p0, 34); \
8066  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8067  __ret; \
8068})
8069#endif
8070
8071#ifdef __LITTLE_ENDIAN__
8072#define vld1q_dup_s64(__p0) __extension__ ({ \
8073  int64x2_t __ret; \
8074  __ret = (int64x2_t) __builtin_neon_vld1q_dup_v(__p0, 35); \
8075  __ret; \
8076})
8077#else
8078#define vld1q_dup_s64(__p0) __extension__ ({ \
8079  int64x2_t __ret; \
8080  __ret = (int64x2_t) __builtin_neon_vld1q_dup_v(__p0, 35); \
8081  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8082  __ret; \
8083})
8084#endif
8085
8086#ifdef __LITTLE_ENDIAN__
8087#define vld1q_dup_s16(__p0) __extension__ ({ \
8088  int16x8_t __ret; \
8089  __ret = (int16x8_t) __builtin_neon_vld1q_dup_v(__p0, 33); \
8090  __ret; \
8091})
8092#else
8093#define vld1q_dup_s16(__p0) __extension__ ({ \
8094  int16x8_t __ret; \
8095  __ret = (int16x8_t) __builtin_neon_vld1q_dup_v(__p0, 33); \
8096  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8097  __ret; \
8098})
8099#endif
8100
8101#ifdef __LITTLE_ENDIAN__
8102#define vld1_dup_u8(__p0) __extension__ ({ \
8103  uint8x8_t __ret; \
8104  __ret = (uint8x8_t) __builtin_neon_vld1_dup_v(__p0, 16); \
8105  __ret; \
8106})
8107#else
8108#define vld1_dup_u8(__p0) __extension__ ({ \
8109  uint8x8_t __ret; \
8110  __ret = (uint8x8_t) __builtin_neon_vld1_dup_v(__p0, 16); \
8111  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8112  __ret; \
8113})
8114#endif
8115
8116#ifdef __LITTLE_ENDIAN__
8117#define vld1_dup_u32(__p0) __extension__ ({ \
8118  uint32x2_t __ret; \
8119  __ret = (uint32x2_t) __builtin_neon_vld1_dup_v(__p0, 18); \
8120  __ret; \
8121})
8122#else
8123#define vld1_dup_u32(__p0) __extension__ ({ \
8124  uint32x2_t __ret; \
8125  __ret = (uint32x2_t) __builtin_neon_vld1_dup_v(__p0, 18); \
8126  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8127  __ret; \
8128})
8129#endif
8130
8131#ifdef __LITTLE_ENDIAN__
8132#define vld1_dup_u64(__p0) __extension__ ({ \
8133  uint64x1_t __ret; \
8134  __ret = (uint64x1_t) __builtin_neon_vld1_dup_v(__p0, 19); \
8135  __ret; \
8136})
8137#else
8138#define vld1_dup_u64(__p0) __extension__ ({ \
8139  uint64x1_t __ret; \
8140  __ret = (uint64x1_t) __builtin_neon_vld1_dup_v(__p0, 19); \
8141  __ret; \
8142})
8143#endif
8144
8145#ifdef __LITTLE_ENDIAN__
8146#define vld1_dup_u16(__p0) __extension__ ({ \
8147  uint16x4_t __ret; \
8148  __ret = (uint16x4_t) __builtin_neon_vld1_dup_v(__p0, 17); \
8149  __ret; \
8150})
8151#else
8152#define vld1_dup_u16(__p0) __extension__ ({ \
8153  uint16x4_t __ret; \
8154  __ret = (uint16x4_t) __builtin_neon_vld1_dup_v(__p0, 17); \
8155  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8156  __ret; \
8157})
8158#endif
8159
8160#ifdef __LITTLE_ENDIAN__
8161#define vld1_dup_s8(__p0) __extension__ ({ \
8162  int8x8_t __ret; \
8163  __ret = (int8x8_t) __builtin_neon_vld1_dup_v(__p0, 0); \
8164  __ret; \
8165})
8166#else
8167#define vld1_dup_s8(__p0) __extension__ ({ \
8168  int8x8_t __ret; \
8169  __ret = (int8x8_t) __builtin_neon_vld1_dup_v(__p0, 0); \
8170  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8171  __ret; \
8172})
8173#endif
8174
8175#ifdef __LITTLE_ENDIAN__
8176#define vld1_dup_f32(__p0) __extension__ ({ \
8177  float32x2_t __ret; \
8178  __ret = (float32x2_t) __builtin_neon_vld1_dup_v(__p0, 9); \
8179  __ret; \
8180})
8181#else
8182#define vld1_dup_f32(__p0) __extension__ ({ \
8183  float32x2_t __ret; \
8184  __ret = (float32x2_t) __builtin_neon_vld1_dup_v(__p0, 9); \
8185  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8186  __ret; \
8187})
8188#endif
8189
8190#ifdef __LITTLE_ENDIAN__
8191#define vld1_dup_f16(__p0) __extension__ ({ \
8192  float16x4_t __ret; \
8193  __ret = (float16x4_t) __builtin_neon_vld1_dup_v(__p0, 8); \
8194  __ret; \
8195})
8196#else
8197#define vld1_dup_f16(__p0) __extension__ ({ \
8198  float16x4_t __ret; \
8199  __ret = (float16x4_t) __builtin_neon_vld1_dup_v(__p0, 8); \
8200  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8201  __ret; \
8202})
8203#endif
8204
8205#ifdef __LITTLE_ENDIAN__
8206#define vld1_dup_s32(__p0) __extension__ ({ \
8207  int32x2_t __ret; \
8208  __ret = (int32x2_t) __builtin_neon_vld1_dup_v(__p0, 2); \
8209  __ret; \
8210})
8211#else
8212#define vld1_dup_s32(__p0) __extension__ ({ \
8213  int32x2_t __ret; \
8214  __ret = (int32x2_t) __builtin_neon_vld1_dup_v(__p0, 2); \
8215  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8216  __ret; \
8217})
8218#endif
8219
8220#ifdef __LITTLE_ENDIAN__
8221#define vld1_dup_s64(__p0) __extension__ ({ \
8222  int64x1_t __ret; \
8223  __ret = (int64x1_t) __builtin_neon_vld1_dup_v(__p0, 3); \
8224  __ret; \
8225})
8226#else
8227#define vld1_dup_s64(__p0) __extension__ ({ \
8228  int64x1_t __ret; \
8229  __ret = (int64x1_t) __builtin_neon_vld1_dup_v(__p0, 3); \
8230  __ret; \
8231})
8232#endif
8233
8234#ifdef __LITTLE_ENDIAN__
8235#define vld1_dup_s16(__p0) __extension__ ({ \
8236  int16x4_t __ret; \
8237  __ret = (int16x4_t) __builtin_neon_vld1_dup_v(__p0, 1); \
8238  __ret; \
8239})
8240#else
8241#define vld1_dup_s16(__p0) __extension__ ({ \
8242  int16x4_t __ret; \
8243  __ret = (int16x4_t) __builtin_neon_vld1_dup_v(__p0, 1); \
8244  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8245  __ret; \
8246})
8247#endif
8248
8249#ifdef __LITTLE_ENDIAN__
8250#define vld1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8251  poly8x8_t __s1 = __p1; \
8252  poly8x8_t __ret; \
8253  __ret = (poly8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 4); \
8254  __ret; \
8255})
8256#else
8257#define vld1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8258  poly8x8_t __s1 = __p1; \
8259  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8260  poly8x8_t __ret; \
8261  __ret = (poly8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 4); \
8262  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8263  __ret; \
8264})
8265#endif
8266
8267#ifdef __LITTLE_ENDIAN__
8268#define vld1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8269  poly16x4_t __s1 = __p1; \
8270  poly16x4_t __ret; \
8271  __ret = (poly16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 5); \
8272  __ret; \
8273})
8274#else
8275#define vld1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8276  poly16x4_t __s1 = __p1; \
8277  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8278  poly16x4_t __ret; \
8279  __ret = (poly16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 5); \
8280  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8281  __ret; \
8282})
8283#endif
8284
8285#ifdef __LITTLE_ENDIAN__
8286#define vld1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8287  poly8x16_t __s1 = __p1; \
8288  poly8x16_t __ret; \
8289  __ret = (poly8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 36); \
8290  __ret; \
8291})
8292#else
8293#define vld1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8294  poly8x16_t __s1 = __p1; \
8295  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8296  poly8x16_t __ret; \
8297  __ret = (poly8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 36); \
8298  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8299  __ret; \
8300})
8301#endif
8302
8303#ifdef __LITTLE_ENDIAN__
8304#define vld1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8305  poly16x8_t __s1 = __p1; \
8306  poly16x8_t __ret; \
8307  __ret = (poly16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 37); \
8308  __ret; \
8309})
8310#else
8311#define vld1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8312  poly16x8_t __s1 = __p1; \
8313  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8314  poly16x8_t __ret; \
8315  __ret = (poly16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 37); \
8316  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8317  __ret; \
8318})
8319#endif
8320
8321#ifdef __LITTLE_ENDIAN__
8322#define vld1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8323  uint8x16_t __s1 = __p1; \
8324  uint8x16_t __ret; \
8325  __ret = (uint8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 48); \
8326  __ret; \
8327})
8328#else
8329#define vld1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8330  uint8x16_t __s1 = __p1; \
8331  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8332  uint8x16_t __ret; \
8333  __ret = (uint8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 48); \
8334  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8335  __ret; \
8336})
8337#endif
8338
8339#ifdef __LITTLE_ENDIAN__
8340#define vld1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8341  uint32x4_t __s1 = __p1; \
8342  uint32x4_t __ret; \
8343  __ret = (uint32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 50); \
8344  __ret; \
8345})
8346#else
8347#define vld1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8348  uint32x4_t __s1 = __p1; \
8349  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8350  uint32x4_t __ret; \
8351  __ret = (uint32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 50); \
8352  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8353  __ret; \
8354})
8355#endif
8356
8357#ifdef __LITTLE_ENDIAN__
8358#define vld1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8359  uint64x2_t __s1 = __p1; \
8360  uint64x2_t __ret; \
8361  __ret = (uint64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 51); \
8362  __ret; \
8363})
8364#else
8365#define vld1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8366  uint64x2_t __s1 = __p1; \
8367  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8368  uint64x2_t __ret; \
8369  __ret = (uint64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 51); \
8370  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8371  __ret; \
8372})
8373#endif
8374
8375#ifdef __LITTLE_ENDIAN__
8376#define vld1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8377  uint16x8_t __s1 = __p1; \
8378  uint16x8_t __ret; \
8379  __ret = (uint16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 49); \
8380  __ret; \
8381})
8382#else
8383#define vld1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8384  uint16x8_t __s1 = __p1; \
8385  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8386  uint16x8_t __ret; \
8387  __ret = (uint16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 49); \
8388  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8389  __ret; \
8390})
8391#endif
8392
8393#ifdef __LITTLE_ENDIAN__
8394#define vld1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8395  int8x16_t __s1 = __p1; \
8396  int8x16_t __ret; \
8397  __ret = (int8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 32); \
8398  __ret; \
8399})
8400#else
8401#define vld1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8402  int8x16_t __s1 = __p1; \
8403  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8404  int8x16_t __ret; \
8405  __ret = (int8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 32); \
8406  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8407  __ret; \
8408})
8409#endif
8410
8411#ifdef __LITTLE_ENDIAN__
8412#define vld1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8413  float32x4_t __s1 = __p1; \
8414  float32x4_t __ret; \
8415  __ret = (float32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 41); \
8416  __ret; \
8417})
8418#else
8419#define vld1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8420  float32x4_t __s1 = __p1; \
8421  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8422  float32x4_t __ret; \
8423  __ret = (float32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 41); \
8424  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8425  __ret; \
8426})
8427#endif
8428
8429#ifdef __LITTLE_ENDIAN__
8430#define vld1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8431  float16x8_t __s1 = __p1; \
8432  float16x8_t __ret; \
8433  __ret = (float16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 40); \
8434  __ret; \
8435})
8436#else
8437#define vld1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8438  float16x8_t __s1 = __p1; \
8439  float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8440  float16x8_t __ret; \
8441  __ret = (float16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 40); \
8442  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8443  __ret; \
8444})
8445#endif
8446
8447#ifdef __LITTLE_ENDIAN__
8448#define vld1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8449  int32x4_t __s1 = __p1; \
8450  int32x4_t __ret; \
8451  __ret = (int32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 34); \
8452  __ret; \
8453})
8454#else
8455#define vld1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8456  int32x4_t __s1 = __p1; \
8457  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8458  int32x4_t __ret; \
8459  __ret = (int32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 34); \
8460  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8461  __ret; \
8462})
8463#endif
8464
8465#ifdef __LITTLE_ENDIAN__
8466#define vld1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8467  int64x2_t __s1 = __p1; \
8468  int64x2_t __ret; \
8469  __ret = (int64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 35); \
8470  __ret; \
8471})
8472#else
8473#define vld1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8474  int64x2_t __s1 = __p1; \
8475  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8476  int64x2_t __ret; \
8477  __ret = (int64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 35); \
8478  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8479  __ret; \
8480})
8481#endif
8482
8483#ifdef __LITTLE_ENDIAN__
8484#define vld1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8485  int16x8_t __s1 = __p1; \
8486  int16x8_t __ret; \
8487  __ret = (int16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 33); \
8488  __ret; \
8489})
8490#else
8491#define vld1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8492  int16x8_t __s1 = __p1; \
8493  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8494  int16x8_t __ret; \
8495  __ret = (int16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 33); \
8496  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8497  __ret; \
8498})
8499#endif
8500
8501#ifdef __LITTLE_ENDIAN__
8502#define vld1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8503  uint8x8_t __s1 = __p1; \
8504  uint8x8_t __ret; \
8505  __ret = (uint8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 16); \
8506  __ret; \
8507})
8508#else
8509#define vld1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8510  uint8x8_t __s1 = __p1; \
8511  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8512  uint8x8_t __ret; \
8513  __ret = (uint8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 16); \
8514  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8515  __ret; \
8516})
8517#endif
8518
8519#ifdef __LITTLE_ENDIAN__
8520#define vld1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8521  uint32x2_t __s1 = __p1; \
8522  uint32x2_t __ret; \
8523  __ret = (uint32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 18); \
8524  __ret; \
8525})
8526#else
8527#define vld1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8528  uint32x2_t __s1 = __p1; \
8529  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8530  uint32x2_t __ret; \
8531  __ret = (uint32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 18); \
8532  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8533  __ret; \
8534})
8535#endif
8536
8537#ifdef __LITTLE_ENDIAN__
8538#define vld1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8539  uint64x1_t __s1 = __p1; \
8540  uint64x1_t __ret; \
8541  __ret = (uint64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
8542  __ret; \
8543})
8544#else
8545#define vld1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8546  uint64x1_t __s1 = __p1; \
8547  uint64x1_t __ret; \
8548  __ret = (uint64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
8549  __ret; \
8550})
8551#endif
8552
8553#ifdef __LITTLE_ENDIAN__
8554#define vld1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8555  uint16x4_t __s1 = __p1; \
8556  uint16x4_t __ret; \
8557  __ret = (uint16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 17); \
8558  __ret; \
8559})
8560#else
8561#define vld1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8562  uint16x4_t __s1 = __p1; \
8563  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8564  uint16x4_t __ret; \
8565  __ret = (uint16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 17); \
8566  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8567  __ret; \
8568})
8569#endif
8570
8571#ifdef __LITTLE_ENDIAN__
8572#define vld1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8573  int8x8_t __s1 = __p1; \
8574  int8x8_t __ret; \
8575  __ret = (int8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 0); \
8576  __ret; \
8577})
8578#else
8579#define vld1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8580  int8x8_t __s1 = __p1; \
8581  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8582  int8x8_t __ret; \
8583  __ret = (int8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 0); \
8584  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8585  __ret; \
8586})
8587#endif
8588
8589#ifdef __LITTLE_ENDIAN__
8590#define vld1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8591  float32x2_t __s1 = __p1; \
8592  float32x2_t __ret; \
8593  __ret = (float32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 9); \
8594  __ret; \
8595})
8596#else
8597#define vld1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8598  float32x2_t __s1 = __p1; \
8599  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8600  float32x2_t __ret; \
8601  __ret = (float32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 9); \
8602  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8603  __ret; \
8604})
8605#endif
8606
8607#ifdef __LITTLE_ENDIAN__
8608#define vld1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8609  float16x4_t __s1 = __p1; \
8610  float16x4_t __ret; \
8611  __ret = (float16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 8); \
8612  __ret; \
8613})
8614#else
8615#define vld1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8616  float16x4_t __s1 = __p1; \
8617  float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8618  float16x4_t __ret; \
8619  __ret = (float16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 8); \
8620  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8621  __ret; \
8622})
8623#endif
8624
8625#ifdef __LITTLE_ENDIAN__
8626#define vld1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8627  int32x2_t __s1 = __p1; \
8628  int32x2_t __ret; \
8629  __ret = (int32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 2); \
8630  __ret; \
8631})
8632#else
8633#define vld1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8634  int32x2_t __s1 = __p1; \
8635  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8636  int32x2_t __ret; \
8637  __ret = (int32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 2); \
8638  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8639  __ret; \
8640})
8641#endif
8642
8643#ifdef __LITTLE_ENDIAN__
8644#define vld1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8645  int64x1_t __s1 = __p1; \
8646  int64x1_t __ret; \
8647  __ret = (int64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
8648  __ret; \
8649})
8650#else
8651#define vld1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8652  int64x1_t __s1 = __p1; \
8653  int64x1_t __ret; \
8654  __ret = (int64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
8655  __ret; \
8656})
8657#endif
8658
8659#ifdef __LITTLE_ENDIAN__
8660#define vld1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8661  int16x4_t __s1 = __p1; \
8662  int16x4_t __ret; \
8663  __ret = (int16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 1); \
8664  __ret; \
8665})
8666#else
8667#define vld1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8668  int16x4_t __s1 = __p1; \
8669  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8670  int16x4_t __ret; \
8671  __ret = (int16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 1); \
8672  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8673  __ret; \
8674})
8675#endif
8676
8677#ifdef __LITTLE_ENDIAN__
8678#define vld2_p8(__p0) __extension__ ({ \
8679  poly8x8x2_t __ret; \
8680  __builtin_neon_vld2_v(&__ret, __p0, 4); \
8681  __ret; \
8682})
8683#else
8684#define vld2_p8(__p0) __extension__ ({ \
8685  poly8x8x2_t __ret; \
8686  __builtin_neon_vld2_v(&__ret, __p0, 4); \
8687 \
8688  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8689  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8690  __ret; \
8691})
8692#endif
8693
8694#ifdef __LITTLE_ENDIAN__
8695#define vld2_p16(__p0) __extension__ ({ \
8696  poly16x4x2_t __ret; \
8697  __builtin_neon_vld2_v(&__ret, __p0, 5); \
8698  __ret; \
8699})
8700#else
8701#define vld2_p16(__p0) __extension__ ({ \
8702  poly16x4x2_t __ret; \
8703  __builtin_neon_vld2_v(&__ret, __p0, 5); \
8704 \
8705  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8706  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8707  __ret; \
8708})
8709#endif
8710
8711#ifdef __LITTLE_ENDIAN__
8712#define vld2q_p8(__p0) __extension__ ({ \
8713  poly8x16x2_t __ret; \
8714  __builtin_neon_vld2q_v(&__ret, __p0, 36); \
8715  __ret; \
8716})
8717#else
8718#define vld2q_p8(__p0) __extension__ ({ \
8719  poly8x16x2_t __ret; \
8720  __builtin_neon_vld2q_v(&__ret, __p0, 36); \
8721 \
8722  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8723  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8724  __ret; \
8725})
8726#endif
8727
8728#ifdef __LITTLE_ENDIAN__
8729#define vld2q_p16(__p0) __extension__ ({ \
8730  poly16x8x2_t __ret; \
8731  __builtin_neon_vld2q_v(&__ret, __p0, 37); \
8732  __ret; \
8733})
8734#else
8735#define vld2q_p16(__p0) __extension__ ({ \
8736  poly16x8x2_t __ret; \
8737  __builtin_neon_vld2q_v(&__ret, __p0, 37); \
8738 \
8739  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8740  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8741  __ret; \
8742})
8743#endif
8744
8745#ifdef __LITTLE_ENDIAN__
8746#define vld2q_u8(__p0) __extension__ ({ \
8747  uint8x16x2_t __ret; \
8748  __builtin_neon_vld2q_v(&__ret, __p0, 48); \
8749  __ret; \
8750})
8751#else
8752#define vld2q_u8(__p0) __extension__ ({ \
8753  uint8x16x2_t __ret; \
8754  __builtin_neon_vld2q_v(&__ret, __p0, 48); \
8755 \
8756  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8757  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8758  __ret; \
8759})
8760#endif
8761
8762#ifdef __LITTLE_ENDIAN__
8763#define vld2q_u32(__p0) __extension__ ({ \
8764  uint32x4x2_t __ret; \
8765  __builtin_neon_vld2q_v(&__ret, __p0, 50); \
8766  __ret; \
8767})
8768#else
8769#define vld2q_u32(__p0) __extension__ ({ \
8770  uint32x4x2_t __ret; \
8771  __builtin_neon_vld2q_v(&__ret, __p0, 50); \
8772 \
8773  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8774  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8775  __ret; \
8776})
8777#endif
8778
8779#ifdef __LITTLE_ENDIAN__
8780#define vld2q_u16(__p0) __extension__ ({ \
8781  uint16x8x2_t __ret; \
8782  __builtin_neon_vld2q_v(&__ret, __p0, 49); \
8783  __ret; \
8784})
8785#else
8786#define vld2q_u16(__p0) __extension__ ({ \
8787  uint16x8x2_t __ret; \
8788  __builtin_neon_vld2q_v(&__ret, __p0, 49); \
8789 \
8790  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8791  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8792  __ret; \
8793})
8794#endif
8795
8796#ifdef __LITTLE_ENDIAN__
8797#define vld2q_s8(__p0) __extension__ ({ \
8798  int8x16x2_t __ret; \
8799  __builtin_neon_vld2q_v(&__ret, __p0, 32); \
8800  __ret; \
8801})
8802#else
8803#define vld2q_s8(__p0) __extension__ ({ \
8804  int8x16x2_t __ret; \
8805  __builtin_neon_vld2q_v(&__ret, __p0, 32); \
8806 \
8807  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8808  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8809  __ret; \
8810})
8811#endif
8812
8813#ifdef __LITTLE_ENDIAN__
8814#define vld2q_f32(__p0) __extension__ ({ \
8815  float32x4x2_t __ret; \
8816  __builtin_neon_vld2q_v(&__ret, __p0, 41); \
8817  __ret; \
8818})
8819#else
8820#define vld2q_f32(__p0) __extension__ ({ \
8821  float32x4x2_t __ret; \
8822  __builtin_neon_vld2q_v(&__ret, __p0, 41); \
8823 \
8824  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8825  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8826  __ret; \
8827})
8828#endif
8829
8830#ifdef __LITTLE_ENDIAN__
8831#define vld2q_f16(__p0) __extension__ ({ \
8832  float16x8x2_t __ret; \
8833  __builtin_neon_vld2q_v(&__ret, __p0, 40); \
8834  __ret; \
8835})
8836#else
8837#define vld2q_f16(__p0) __extension__ ({ \
8838  float16x8x2_t __ret; \
8839  __builtin_neon_vld2q_v(&__ret, __p0, 40); \
8840 \
8841  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8842  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8843  __ret; \
8844})
8845#endif
8846
8847#ifdef __LITTLE_ENDIAN__
8848#define vld2q_s32(__p0) __extension__ ({ \
8849  int32x4x2_t __ret; \
8850  __builtin_neon_vld2q_v(&__ret, __p0, 34); \
8851  __ret; \
8852})
8853#else
8854#define vld2q_s32(__p0) __extension__ ({ \
8855  int32x4x2_t __ret; \
8856  __builtin_neon_vld2q_v(&__ret, __p0, 34); \
8857 \
8858  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8859  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8860  __ret; \
8861})
8862#endif
8863
8864#ifdef __LITTLE_ENDIAN__
8865#define vld2q_s16(__p0) __extension__ ({ \
8866  int16x8x2_t __ret; \
8867  __builtin_neon_vld2q_v(&__ret, __p0, 33); \
8868  __ret; \
8869})
8870#else
8871#define vld2q_s16(__p0) __extension__ ({ \
8872  int16x8x2_t __ret; \
8873  __builtin_neon_vld2q_v(&__ret, __p0, 33); \
8874 \
8875  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8876  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8877  __ret; \
8878})
8879#endif
8880
8881#ifdef __LITTLE_ENDIAN__
8882#define vld2_u8(__p0) __extension__ ({ \
8883  uint8x8x2_t __ret; \
8884  __builtin_neon_vld2_v(&__ret, __p0, 16); \
8885  __ret; \
8886})
8887#else
8888#define vld2_u8(__p0) __extension__ ({ \
8889  uint8x8x2_t __ret; \
8890  __builtin_neon_vld2_v(&__ret, __p0, 16); \
8891 \
8892  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8893  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8894  __ret; \
8895})
8896#endif
8897
8898#ifdef __LITTLE_ENDIAN__
8899#define vld2_u32(__p0) __extension__ ({ \
8900  uint32x2x2_t __ret; \
8901  __builtin_neon_vld2_v(&__ret, __p0, 18); \
8902  __ret; \
8903})
8904#else
8905#define vld2_u32(__p0) __extension__ ({ \
8906  uint32x2x2_t __ret; \
8907  __builtin_neon_vld2_v(&__ret, __p0, 18); \
8908 \
8909  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
8910  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
8911  __ret; \
8912})
8913#endif
8914
8915#ifdef __LITTLE_ENDIAN__
8916#define vld2_u64(__p0) __extension__ ({ \
8917  uint64x1x2_t __ret; \
8918  __builtin_neon_vld2_v(&__ret, __p0, 19); \
8919  __ret; \
8920})
8921#else
8922#define vld2_u64(__p0) __extension__ ({ \
8923  uint64x1x2_t __ret; \
8924  __builtin_neon_vld2_v(&__ret, __p0, 19); \
8925  __ret; \
8926})
8927#endif
8928
8929#ifdef __LITTLE_ENDIAN__
8930#define vld2_u16(__p0) __extension__ ({ \
8931  uint16x4x2_t __ret; \
8932  __builtin_neon_vld2_v(&__ret, __p0, 17); \
8933  __ret; \
8934})
8935#else
8936#define vld2_u16(__p0) __extension__ ({ \
8937  uint16x4x2_t __ret; \
8938  __builtin_neon_vld2_v(&__ret, __p0, 17); \
8939 \
8940  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8941  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8942  __ret; \
8943})
8944#endif
8945
8946#ifdef __LITTLE_ENDIAN__
8947#define vld2_s8(__p0) __extension__ ({ \
8948  int8x8x2_t __ret; \
8949  __builtin_neon_vld2_v(&__ret, __p0, 0); \
8950  __ret; \
8951})
8952#else
8953#define vld2_s8(__p0) __extension__ ({ \
8954  int8x8x2_t __ret; \
8955  __builtin_neon_vld2_v(&__ret, __p0, 0); \
8956 \
8957  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8958  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8959  __ret; \
8960})
8961#endif
8962
8963#ifdef __LITTLE_ENDIAN__
8964#define vld2_f32(__p0) __extension__ ({ \
8965  float32x2x2_t __ret; \
8966  __builtin_neon_vld2_v(&__ret, __p0, 9); \
8967  __ret; \
8968})
8969#else
8970#define vld2_f32(__p0) __extension__ ({ \
8971  float32x2x2_t __ret; \
8972  __builtin_neon_vld2_v(&__ret, __p0, 9); \
8973 \
8974  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
8975  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
8976  __ret; \
8977})
8978#endif
8979
8980#ifdef __LITTLE_ENDIAN__
8981#define vld2_f16(__p0) __extension__ ({ \
8982  float16x4x2_t __ret; \
8983  __builtin_neon_vld2_v(&__ret, __p0, 8); \
8984  __ret; \
8985})
8986#else
8987#define vld2_f16(__p0) __extension__ ({ \
8988  float16x4x2_t __ret; \
8989  __builtin_neon_vld2_v(&__ret, __p0, 8); \
8990 \
8991  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8992  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8993  __ret; \
8994})
8995#endif
8996
8997#ifdef __LITTLE_ENDIAN__
8998#define vld2_s32(__p0) __extension__ ({ \
8999  int32x2x2_t __ret; \
9000  __builtin_neon_vld2_v(&__ret, __p0, 2); \
9001  __ret; \
9002})
9003#else
9004#define vld2_s32(__p0) __extension__ ({ \
9005  int32x2x2_t __ret; \
9006  __builtin_neon_vld2_v(&__ret, __p0, 2); \
9007 \
9008  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9009  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9010  __ret; \
9011})
9012#endif
9013
9014#ifdef __LITTLE_ENDIAN__
9015#define vld2_s64(__p0) __extension__ ({ \
9016  int64x1x2_t __ret; \
9017  __builtin_neon_vld2_v(&__ret, __p0, 3); \
9018  __ret; \
9019})
9020#else
9021#define vld2_s64(__p0) __extension__ ({ \
9022  int64x1x2_t __ret; \
9023  __builtin_neon_vld2_v(&__ret, __p0, 3); \
9024  __ret; \
9025})
9026#endif
9027
9028#ifdef __LITTLE_ENDIAN__
9029#define vld2_s16(__p0) __extension__ ({ \
9030  int16x4x2_t __ret; \
9031  __builtin_neon_vld2_v(&__ret, __p0, 1); \
9032  __ret; \
9033})
9034#else
9035#define vld2_s16(__p0) __extension__ ({ \
9036  int16x4x2_t __ret; \
9037  __builtin_neon_vld2_v(&__ret, __p0, 1); \
9038 \
9039  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9040  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9041  __ret; \
9042})
9043#endif
9044
9045#ifdef __LITTLE_ENDIAN__
9046#define vld2_dup_p8(__p0) __extension__ ({ \
9047  poly8x8x2_t __ret; \
9048  __builtin_neon_vld2_dup_v(&__ret, __p0, 4); \
9049  __ret; \
9050})
9051#else
9052#define vld2_dup_p8(__p0) __extension__ ({ \
9053  poly8x8x2_t __ret; \
9054  __builtin_neon_vld2_dup_v(&__ret, __p0, 4); \
9055 \
9056  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9057  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9058  __ret; \
9059})
9060#endif
9061
9062#ifdef __LITTLE_ENDIAN__
9063#define vld2_dup_p16(__p0) __extension__ ({ \
9064  poly16x4x2_t __ret; \
9065  __builtin_neon_vld2_dup_v(&__ret, __p0, 5); \
9066  __ret; \
9067})
9068#else
9069#define vld2_dup_p16(__p0) __extension__ ({ \
9070  poly16x4x2_t __ret; \
9071  __builtin_neon_vld2_dup_v(&__ret, __p0, 5); \
9072 \
9073  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9074  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9075  __ret; \
9076})
9077#endif
9078
9079#ifdef __LITTLE_ENDIAN__
9080#define vld2_dup_u8(__p0) __extension__ ({ \
9081  uint8x8x2_t __ret; \
9082  __builtin_neon_vld2_dup_v(&__ret, __p0, 16); \
9083  __ret; \
9084})
9085#else
9086#define vld2_dup_u8(__p0) __extension__ ({ \
9087  uint8x8x2_t __ret; \
9088  __builtin_neon_vld2_dup_v(&__ret, __p0, 16); \
9089 \
9090  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9091  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9092  __ret; \
9093})
9094#endif
9095
9096#ifdef __LITTLE_ENDIAN__
9097#define vld2_dup_u32(__p0) __extension__ ({ \
9098  uint32x2x2_t __ret; \
9099  __builtin_neon_vld2_dup_v(&__ret, __p0, 18); \
9100  __ret; \
9101})
9102#else
9103#define vld2_dup_u32(__p0) __extension__ ({ \
9104  uint32x2x2_t __ret; \
9105  __builtin_neon_vld2_dup_v(&__ret, __p0, 18); \
9106 \
9107  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9108  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9109  __ret; \
9110})
9111#endif
9112
9113#ifdef __LITTLE_ENDIAN__
9114#define vld2_dup_u64(__p0) __extension__ ({ \
9115  uint64x1x2_t __ret; \
9116  __builtin_neon_vld2_dup_v(&__ret, __p0, 19); \
9117  __ret; \
9118})
9119#else
9120#define vld2_dup_u64(__p0) __extension__ ({ \
9121  uint64x1x2_t __ret; \
9122  __builtin_neon_vld2_dup_v(&__ret, __p0, 19); \
9123  __ret; \
9124})
9125#endif
9126
9127#ifdef __LITTLE_ENDIAN__
9128#define vld2_dup_u16(__p0) __extension__ ({ \
9129  uint16x4x2_t __ret; \
9130  __builtin_neon_vld2_dup_v(&__ret, __p0, 17); \
9131  __ret; \
9132})
9133#else
9134#define vld2_dup_u16(__p0) __extension__ ({ \
9135  uint16x4x2_t __ret; \
9136  __builtin_neon_vld2_dup_v(&__ret, __p0, 17); \
9137 \
9138  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9139  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9140  __ret; \
9141})
9142#endif
9143
9144#ifdef __LITTLE_ENDIAN__
9145#define vld2_dup_s8(__p0) __extension__ ({ \
9146  int8x8x2_t __ret; \
9147  __builtin_neon_vld2_dup_v(&__ret, __p0, 0); \
9148  __ret; \
9149})
9150#else
9151#define vld2_dup_s8(__p0) __extension__ ({ \
9152  int8x8x2_t __ret; \
9153  __builtin_neon_vld2_dup_v(&__ret, __p0, 0); \
9154 \
9155  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9156  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9157  __ret; \
9158})
9159#endif
9160
9161#ifdef __LITTLE_ENDIAN__
9162#define vld2_dup_f32(__p0) __extension__ ({ \
9163  float32x2x2_t __ret; \
9164  __builtin_neon_vld2_dup_v(&__ret, __p0, 9); \
9165  __ret; \
9166})
9167#else
9168#define vld2_dup_f32(__p0) __extension__ ({ \
9169  float32x2x2_t __ret; \
9170  __builtin_neon_vld2_dup_v(&__ret, __p0, 9); \
9171 \
9172  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9173  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9174  __ret; \
9175})
9176#endif
9177
9178#ifdef __LITTLE_ENDIAN__
9179#define vld2_dup_f16(__p0) __extension__ ({ \
9180  float16x4x2_t __ret; \
9181  __builtin_neon_vld2_dup_v(&__ret, __p0, 8); \
9182  __ret; \
9183})
9184#else
9185#define vld2_dup_f16(__p0) __extension__ ({ \
9186  float16x4x2_t __ret; \
9187  __builtin_neon_vld2_dup_v(&__ret, __p0, 8); \
9188 \
9189  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9190  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9191  __ret; \
9192})
9193#endif
9194
9195#ifdef __LITTLE_ENDIAN__
9196#define vld2_dup_s32(__p0) __extension__ ({ \
9197  int32x2x2_t __ret; \
9198  __builtin_neon_vld2_dup_v(&__ret, __p0, 2); \
9199  __ret; \
9200})
9201#else
9202#define vld2_dup_s32(__p0) __extension__ ({ \
9203  int32x2x2_t __ret; \
9204  __builtin_neon_vld2_dup_v(&__ret, __p0, 2); \
9205 \
9206  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9207  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9208  __ret; \
9209})
9210#endif
9211
9212#ifdef __LITTLE_ENDIAN__
9213#define vld2_dup_s64(__p0) __extension__ ({ \
9214  int64x1x2_t __ret; \
9215  __builtin_neon_vld2_dup_v(&__ret, __p0, 3); \
9216  __ret; \
9217})
9218#else
9219#define vld2_dup_s64(__p0) __extension__ ({ \
9220  int64x1x2_t __ret; \
9221  __builtin_neon_vld2_dup_v(&__ret, __p0, 3); \
9222  __ret; \
9223})
9224#endif
9225
9226#ifdef __LITTLE_ENDIAN__
9227#define vld2_dup_s16(__p0) __extension__ ({ \
9228  int16x4x2_t __ret; \
9229  __builtin_neon_vld2_dup_v(&__ret, __p0, 1); \
9230  __ret; \
9231})
9232#else
9233#define vld2_dup_s16(__p0) __extension__ ({ \
9234  int16x4x2_t __ret; \
9235  __builtin_neon_vld2_dup_v(&__ret, __p0, 1); \
9236 \
9237  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9238  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9239  __ret; \
9240})
9241#endif
9242
9243#ifdef __LITTLE_ENDIAN__
9244#define vld2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
9245  poly8x8x2_t __s1 = __p1; \
9246  poly8x8x2_t __ret; \
9247  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 4); \
9248  __ret; \
9249})
9250#else
9251#define vld2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
9252  poly8x8x2_t __s1 = __p1; \
9253  poly8x8x2_t __rev1; \
9254  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9255  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9256  poly8x8x2_t __ret; \
9257  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 4); \
9258 \
9259  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9260  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9261  __ret; \
9262})
9263#endif
9264
9265#ifdef __LITTLE_ENDIAN__
9266#define vld2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9267  poly16x4x2_t __s1 = __p1; \
9268  poly16x4x2_t __ret; \
9269  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 5); \
9270  __ret; \
9271})
9272#else
9273#define vld2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9274  poly16x4x2_t __s1 = __p1; \
9275  poly16x4x2_t __rev1; \
9276  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9277  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9278  poly16x4x2_t __ret; \
9279  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 5); \
9280 \
9281  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9282  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9283  __ret; \
9284})
9285#endif
9286
9287#ifdef __LITTLE_ENDIAN__
9288#define vld2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9289  poly16x8x2_t __s1 = __p1; \
9290  poly16x8x2_t __ret; \
9291  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 37); \
9292  __ret; \
9293})
9294#else
9295#define vld2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9296  poly16x8x2_t __s1 = __p1; \
9297  poly16x8x2_t __rev1; \
9298  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9299  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9300  poly16x8x2_t __ret; \
9301  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 37); \
9302 \
9303  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9304  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9305  __ret; \
9306})
9307#endif
9308
9309#ifdef __LITTLE_ENDIAN__
9310#define vld2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9311  uint32x4x2_t __s1 = __p1; \
9312  uint32x4x2_t __ret; \
9313  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 50); \
9314  __ret; \
9315})
9316#else
9317#define vld2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9318  uint32x4x2_t __s1 = __p1; \
9319  uint32x4x2_t __rev1; \
9320  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9321  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9322  uint32x4x2_t __ret; \
9323  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 50); \
9324 \
9325  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9326  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9327  __ret; \
9328})
9329#endif
9330
9331#ifdef __LITTLE_ENDIAN__
9332#define vld2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9333  uint16x8x2_t __s1 = __p1; \
9334  uint16x8x2_t __ret; \
9335  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 49); \
9336  __ret; \
9337})
9338#else
9339#define vld2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9340  uint16x8x2_t __s1 = __p1; \
9341  uint16x8x2_t __rev1; \
9342  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9343  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9344  uint16x8x2_t __ret; \
9345  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 49); \
9346 \
9347  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9348  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9349  __ret; \
9350})
9351#endif
9352
9353#ifdef __LITTLE_ENDIAN__
9354#define vld2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9355  float32x4x2_t __s1 = __p1; \
9356  float32x4x2_t __ret; \
9357  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 41); \
9358  __ret; \
9359})
9360#else
9361#define vld2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9362  float32x4x2_t __s1 = __p1; \
9363  float32x4x2_t __rev1; \
9364  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9365  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9366  float32x4x2_t __ret; \
9367  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 41); \
9368 \
9369  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9370  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9371  __ret; \
9372})
9373#endif
9374
9375#ifdef __LITTLE_ENDIAN__
9376#define vld2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9377  float16x8x2_t __s1 = __p1; \
9378  float16x8x2_t __ret; \
9379  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 40); \
9380  __ret; \
9381})
9382#else
9383#define vld2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9384  float16x8x2_t __s1 = __p1; \
9385  float16x8x2_t __rev1; \
9386  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9387  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9388  float16x8x2_t __ret; \
9389  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 40); \
9390 \
9391  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9392  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9393  __ret; \
9394})
9395#endif
9396
9397#ifdef __LITTLE_ENDIAN__
9398#define vld2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9399  int32x4x2_t __s1 = __p1; \
9400  int32x4x2_t __ret; \
9401  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 34); \
9402  __ret; \
9403})
9404#else
9405#define vld2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9406  int32x4x2_t __s1 = __p1; \
9407  int32x4x2_t __rev1; \
9408  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9409  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9410  int32x4x2_t __ret; \
9411  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 34); \
9412 \
9413  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9414  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9415  __ret; \
9416})
9417#endif
9418
9419#ifdef __LITTLE_ENDIAN__
9420#define vld2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9421  int16x8x2_t __s1 = __p1; \
9422  int16x8x2_t __ret; \
9423  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 33); \
9424  __ret; \
9425})
9426#else
9427#define vld2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9428  int16x8x2_t __s1 = __p1; \
9429  int16x8x2_t __rev1; \
9430  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9431  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9432  int16x8x2_t __ret; \
9433  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 33); \
9434 \
9435  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9436  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9437  __ret; \
9438})
9439#endif
9440
9441#ifdef __LITTLE_ENDIAN__
9442#define vld2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9443  uint8x8x2_t __s1 = __p1; \
9444  uint8x8x2_t __ret; \
9445  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 16); \
9446  __ret; \
9447})
9448#else
9449#define vld2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9450  uint8x8x2_t __s1 = __p1; \
9451  uint8x8x2_t __rev1; \
9452  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9453  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9454  uint8x8x2_t __ret; \
9455  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 16); \
9456 \
9457  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9458  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9459  __ret; \
9460})
9461#endif
9462
9463#ifdef __LITTLE_ENDIAN__
9464#define vld2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9465  uint32x2x2_t __s1 = __p1; \
9466  uint32x2x2_t __ret; \
9467  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 18); \
9468  __ret; \
9469})
9470#else
9471#define vld2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9472  uint32x2x2_t __s1 = __p1; \
9473  uint32x2x2_t __rev1; \
9474  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
9475  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
9476  uint32x2x2_t __ret; \
9477  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 18); \
9478 \
9479  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9480  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9481  __ret; \
9482})
9483#endif
9484
9485#ifdef __LITTLE_ENDIAN__
9486#define vld2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9487  uint16x4x2_t __s1 = __p1; \
9488  uint16x4x2_t __ret; \
9489  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 17); \
9490  __ret; \
9491})
9492#else
9493#define vld2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9494  uint16x4x2_t __s1 = __p1; \
9495  uint16x4x2_t __rev1; \
9496  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9497  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9498  uint16x4x2_t __ret; \
9499  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 17); \
9500 \
9501  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9502  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9503  __ret; \
9504})
9505#endif
9506
9507#ifdef __LITTLE_ENDIAN__
9508#define vld2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9509  int8x8x2_t __s1 = __p1; \
9510  int8x8x2_t __ret; \
9511  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 0); \
9512  __ret; \
9513})
9514#else
9515#define vld2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9516  int8x8x2_t __s1 = __p1; \
9517  int8x8x2_t __rev1; \
9518  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9519  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9520  int8x8x2_t __ret; \
9521  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 0); \
9522 \
9523  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9524  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9525  __ret; \
9526})
9527#endif
9528
9529#ifdef __LITTLE_ENDIAN__
9530#define vld2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9531  float32x2x2_t __s1 = __p1; \
9532  float32x2x2_t __ret; \
9533  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 9); \
9534  __ret; \
9535})
9536#else
9537#define vld2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9538  float32x2x2_t __s1 = __p1; \
9539  float32x2x2_t __rev1; \
9540  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
9541  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
9542  float32x2x2_t __ret; \
9543  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 9); \
9544 \
9545  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9546  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9547  __ret; \
9548})
9549#endif
9550
9551#ifdef __LITTLE_ENDIAN__
9552#define vld2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9553  float16x4x2_t __s1 = __p1; \
9554  float16x4x2_t __ret; \
9555  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 8); \
9556  __ret; \
9557})
9558#else
9559#define vld2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9560  float16x4x2_t __s1 = __p1; \
9561  float16x4x2_t __rev1; \
9562  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9563  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9564  float16x4x2_t __ret; \
9565  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 8); \
9566 \
9567  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9568  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9569  __ret; \
9570})
9571#endif
9572
9573#ifdef __LITTLE_ENDIAN__
9574#define vld2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9575  int32x2x2_t __s1 = __p1; \
9576  int32x2x2_t __ret; \
9577  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 2); \
9578  __ret; \
9579})
9580#else
9581#define vld2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9582  int32x2x2_t __s1 = __p1; \
9583  int32x2x2_t __rev1; \
9584  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
9585  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
9586  int32x2x2_t __ret; \
9587  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 2); \
9588 \
9589  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9590  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9591  __ret; \
9592})
9593#endif
9594
9595#ifdef __LITTLE_ENDIAN__
9596#define vld2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9597  int16x4x2_t __s1 = __p1; \
9598  int16x4x2_t __ret; \
9599  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 1); \
9600  __ret; \
9601})
9602#else
9603#define vld2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9604  int16x4x2_t __s1 = __p1; \
9605  int16x4x2_t __rev1; \
9606  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9607  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9608  int16x4x2_t __ret; \
9609  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 1); \
9610 \
9611  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9612  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9613  __ret; \
9614})
9615#endif
9616
9617#ifdef __LITTLE_ENDIAN__
9618#define vld3_p8(__p0) __extension__ ({ \
9619  poly8x8x3_t __ret; \
9620  __builtin_neon_vld3_v(&__ret, __p0, 4); \
9621  __ret; \
9622})
9623#else
9624#define vld3_p8(__p0) __extension__ ({ \
9625  poly8x8x3_t __ret; \
9626  __builtin_neon_vld3_v(&__ret, __p0, 4); \
9627 \
9628  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9629  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9630  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9631  __ret; \
9632})
9633#endif
9634
9635#ifdef __LITTLE_ENDIAN__
9636#define vld3_p16(__p0) __extension__ ({ \
9637  poly16x4x3_t __ret; \
9638  __builtin_neon_vld3_v(&__ret, __p0, 5); \
9639  __ret; \
9640})
9641#else
9642#define vld3_p16(__p0) __extension__ ({ \
9643  poly16x4x3_t __ret; \
9644  __builtin_neon_vld3_v(&__ret, __p0, 5); \
9645 \
9646  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9647  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9648  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9649  __ret; \
9650})
9651#endif
9652
9653#ifdef __LITTLE_ENDIAN__
9654#define vld3q_p8(__p0) __extension__ ({ \
9655  poly8x16x3_t __ret; \
9656  __builtin_neon_vld3q_v(&__ret, __p0, 36); \
9657  __ret; \
9658})
9659#else
9660#define vld3q_p8(__p0) __extension__ ({ \
9661  poly8x16x3_t __ret; \
9662  __builtin_neon_vld3q_v(&__ret, __p0, 36); \
9663 \
9664  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9665  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9666  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9667  __ret; \
9668})
9669#endif
9670
9671#ifdef __LITTLE_ENDIAN__
9672#define vld3q_p16(__p0) __extension__ ({ \
9673  poly16x8x3_t __ret; \
9674  __builtin_neon_vld3q_v(&__ret, __p0, 37); \
9675  __ret; \
9676})
9677#else
9678#define vld3q_p16(__p0) __extension__ ({ \
9679  poly16x8x3_t __ret; \
9680  __builtin_neon_vld3q_v(&__ret, __p0, 37); \
9681 \
9682  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9683  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9684  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9685  __ret; \
9686})
9687#endif
9688
9689#ifdef __LITTLE_ENDIAN__
9690#define vld3q_u8(__p0) __extension__ ({ \
9691  uint8x16x3_t __ret; \
9692  __builtin_neon_vld3q_v(&__ret, __p0, 48); \
9693  __ret; \
9694})
9695#else
9696#define vld3q_u8(__p0) __extension__ ({ \
9697  uint8x16x3_t __ret; \
9698  __builtin_neon_vld3q_v(&__ret, __p0, 48); \
9699 \
9700  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9701  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9702  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9703  __ret; \
9704})
9705#endif
9706
9707#ifdef __LITTLE_ENDIAN__
9708#define vld3q_u32(__p0) __extension__ ({ \
9709  uint32x4x3_t __ret; \
9710  __builtin_neon_vld3q_v(&__ret, __p0, 50); \
9711  __ret; \
9712})
9713#else
9714#define vld3q_u32(__p0) __extension__ ({ \
9715  uint32x4x3_t __ret; \
9716  __builtin_neon_vld3q_v(&__ret, __p0, 50); \
9717 \
9718  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9719  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9720  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9721  __ret; \
9722})
9723#endif
9724
9725#ifdef __LITTLE_ENDIAN__
9726#define vld3q_u16(__p0) __extension__ ({ \
9727  uint16x8x3_t __ret; \
9728  __builtin_neon_vld3q_v(&__ret, __p0, 49); \
9729  __ret; \
9730})
9731#else
9732#define vld3q_u16(__p0) __extension__ ({ \
9733  uint16x8x3_t __ret; \
9734  __builtin_neon_vld3q_v(&__ret, __p0, 49); \
9735 \
9736  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9737  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9738  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9739  __ret; \
9740})
9741#endif
9742
9743#ifdef __LITTLE_ENDIAN__
9744#define vld3q_s8(__p0) __extension__ ({ \
9745  int8x16x3_t __ret; \
9746  __builtin_neon_vld3q_v(&__ret, __p0, 32); \
9747  __ret; \
9748})
9749#else
9750#define vld3q_s8(__p0) __extension__ ({ \
9751  int8x16x3_t __ret; \
9752  __builtin_neon_vld3q_v(&__ret, __p0, 32); \
9753 \
9754  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9755  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9756  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9757  __ret; \
9758})
9759#endif
9760
9761#ifdef __LITTLE_ENDIAN__
9762#define vld3q_f32(__p0) __extension__ ({ \
9763  float32x4x3_t __ret; \
9764  __builtin_neon_vld3q_v(&__ret, __p0, 41); \
9765  __ret; \
9766})
9767#else
9768#define vld3q_f32(__p0) __extension__ ({ \
9769  float32x4x3_t __ret; \
9770  __builtin_neon_vld3q_v(&__ret, __p0, 41); \
9771 \
9772  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9773  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9774  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9775  __ret; \
9776})
9777#endif
9778
9779#ifdef __LITTLE_ENDIAN__
9780#define vld3q_f16(__p0) __extension__ ({ \
9781  float16x8x3_t __ret; \
9782  __builtin_neon_vld3q_v(&__ret, __p0, 40); \
9783  __ret; \
9784})
9785#else
9786#define vld3q_f16(__p0) __extension__ ({ \
9787  float16x8x3_t __ret; \
9788  __builtin_neon_vld3q_v(&__ret, __p0, 40); \
9789 \
9790  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9791  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9792  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9793  __ret; \
9794})
9795#endif
9796
9797#ifdef __LITTLE_ENDIAN__
9798#define vld3q_s32(__p0) __extension__ ({ \
9799  int32x4x3_t __ret; \
9800  __builtin_neon_vld3q_v(&__ret, __p0, 34); \
9801  __ret; \
9802})
9803#else
9804#define vld3q_s32(__p0) __extension__ ({ \
9805  int32x4x3_t __ret; \
9806  __builtin_neon_vld3q_v(&__ret, __p0, 34); \
9807 \
9808  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9809  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9810  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9811  __ret; \
9812})
9813#endif
9814
9815#ifdef __LITTLE_ENDIAN__
9816#define vld3q_s16(__p0) __extension__ ({ \
9817  int16x8x3_t __ret; \
9818  __builtin_neon_vld3q_v(&__ret, __p0, 33); \
9819  __ret; \
9820})
9821#else
9822#define vld3q_s16(__p0) __extension__ ({ \
9823  int16x8x3_t __ret; \
9824  __builtin_neon_vld3q_v(&__ret, __p0, 33); \
9825 \
9826  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9827  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9828  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9829  __ret; \
9830})
9831#endif
9832
9833#ifdef __LITTLE_ENDIAN__
9834#define vld3_u8(__p0) __extension__ ({ \
9835  uint8x8x3_t __ret; \
9836  __builtin_neon_vld3_v(&__ret, __p0, 16); \
9837  __ret; \
9838})
9839#else
9840#define vld3_u8(__p0) __extension__ ({ \
9841  uint8x8x3_t __ret; \
9842  __builtin_neon_vld3_v(&__ret, __p0, 16); \
9843 \
9844  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9845  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9846  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9847  __ret; \
9848})
9849#endif
9850
9851#ifdef __LITTLE_ENDIAN__
9852#define vld3_u32(__p0) __extension__ ({ \
9853  uint32x2x3_t __ret; \
9854  __builtin_neon_vld3_v(&__ret, __p0, 18); \
9855  __ret; \
9856})
9857#else
9858#define vld3_u32(__p0) __extension__ ({ \
9859  uint32x2x3_t __ret; \
9860  __builtin_neon_vld3_v(&__ret, __p0, 18); \
9861 \
9862  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9863  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9864  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9865  __ret; \
9866})
9867#endif
9868
9869#ifdef __LITTLE_ENDIAN__
9870#define vld3_u64(__p0) __extension__ ({ \
9871  uint64x1x3_t __ret; \
9872  __builtin_neon_vld3_v(&__ret, __p0, 19); \
9873  __ret; \
9874})
9875#else
9876#define vld3_u64(__p0) __extension__ ({ \
9877  uint64x1x3_t __ret; \
9878  __builtin_neon_vld3_v(&__ret, __p0, 19); \
9879  __ret; \
9880})
9881#endif
9882
9883#ifdef __LITTLE_ENDIAN__
9884#define vld3_u16(__p0) __extension__ ({ \
9885  uint16x4x3_t __ret; \
9886  __builtin_neon_vld3_v(&__ret, __p0, 17); \
9887  __ret; \
9888})
9889#else
9890#define vld3_u16(__p0) __extension__ ({ \
9891  uint16x4x3_t __ret; \
9892  __builtin_neon_vld3_v(&__ret, __p0, 17); \
9893 \
9894  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9895  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9896  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9897  __ret; \
9898})
9899#endif
9900
9901#ifdef __LITTLE_ENDIAN__
9902#define vld3_s8(__p0) __extension__ ({ \
9903  int8x8x3_t __ret; \
9904  __builtin_neon_vld3_v(&__ret, __p0, 0); \
9905  __ret; \
9906})
9907#else
9908#define vld3_s8(__p0) __extension__ ({ \
9909  int8x8x3_t __ret; \
9910  __builtin_neon_vld3_v(&__ret, __p0, 0); \
9911 \
9912  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9913  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9914  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9915  __ret; \
9916})
9917#endif
9918
9919#ifdef __LITTLE_ENDIAN__
9920#define vld3_f32(__p0) __extension__ ({ \
9921  float32x2x3_t __ret; \
9922  __builtin_neon_vld3_v(&__ret, __p0, 9); \
9923  __ret; \
9924})
9925#else
9926#define vld3_f32(__p0) __extension__ ({ \
9927  float32x2x3_t __ret; \
9928  __builtin_neon_vld3_v(&__ret, __p0, 9); \
9929 \
9930  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9931  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9932  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9933  __ret; \
9934})
9935#endif
9936
9937#ifdef __LITTLE_ENDIAN__
9938#define vld3_f16(__p0) __extension__ ({ \
9939  float16x4x3_t __ret; \
9940  __builtin_neon_vld3_v(&__ret, __p0, 8); \
9941  __ret; \
9942})
9943#else
9944#define vld3_f16(__p0) __extension__ ({ \
9945  float16x4x3_t __ret; \
9946  __builtin_neon_vld3_v(&__ret, __p0, 8); \
9947 \
9948  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9949  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9950  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9951  __ret; \
9952})
9953#endif
9954
9955#ifdef __LITTLE_ENDIAN__
9956#define vld3_s32(__p0) __extension__ ({ \
9957  int32x2x3_t __ret; \
9958  __builtin_neon_vld3_v(&__ret, __p0, 2); \
9959  __ret; \
9960})
9961#else
9962#define vld3_s32(__p0) __extension__ ({ \
9963  int32x2x3_t __ret; \
9964  __builtin_neon_vld3_v(&__ret, __p0, 2); \
9965 \
9966  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9967  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9968  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9969  __ret; \
9970})
9971#endif
9972
9973#ifdef __LITTLE_ENDIAN__
9974#define vld3_s64(__p0) __extension__ ({ \
9975  int64x1x3_t __ret; \
9976  __builtin_neon_vld3_v(&__ret, __p0, 3); \
9977  __ret; \
9978})
9979#else
9980#define vld3_s64(__p0) __extension__ ({ \
9981  int64x1x3_t __ret; \
9982  __builtin_neon_vld3_v(&__ret, __p0, 3); \
9983  __ret; \
9984})
9985#endif
9986
9987#ifdef __LITTLE_ENDIAN__
9988#define vld3_s16(__p0) __extension__ ({ \
9989  int16x4x3_t __ret; \
9990  __builtin_neon_vld3_v(&__ret, __p0, 1); \
9991  __ret; \
9992})
9993#else
9994#define vld3_s16(__p0) __extension__ ({ \
9995  int16x4x3_t __ret; \
9996  __builtin_neon_vld3_v(&__ret, __p0, 1); \
9997 \
9998  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9999  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10000  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10001  __ret; \
10002})
10003#endif
10004
10005#ifdef __LITTLE_ENDIAN__
10006#define vld3_dup_p8(__p0) __extension__ ({ \
10007  poly8x8x3_t __ret; \
10008  __builtin_neon_vld3_dup_v(&__ret, __p0, 4); \
10009  __ret; \
10010})
10011#else
10012#define vld3_dup_p8(__p0) __extension__ ({ \
10013  poly8x8x3_t __ret; \
10014  __builtin_neon_vld3_dup_v(&__ret, __p0, 4); \
10015 \
10016  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10017  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10018  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10019  __ret; \
10020})
10021#endif
10022
10023#ifdef __LITTLE_ENDIAN__
10024#define vld3_dup_p16(__p0) __extension__ ({ \
10025  poly16x4x3_t __ret; \
10026  __builtin_neon_vld3_dup_v(&__ret, __p0, 5); \
10027  __ret; \
10028})
10029#else
10030#define vld3_dup_p16(__p0) __extension__ ({ \
10031  poly16x4x3_t __ret; \
10032  __builtin_neon_vld3_dup_v(&__ret, __p0, 5); \
10033 \
10034  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10035  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10036  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10037  __ret; \
10038})
10039#endif
10040
10041#ifdef __LITTLE_ENDIAN__
10042#define vld3_dup_u8(__p0) __extension__ ({ \
10043  uint8x8x3_t __ret; \
10044  __builtin_neon_vld3_dup_v(&__ret, __p0, 16); \
10045  __ret; \
10046})
10047#else
10048#define vld3_dup_u8(__p0) __extension__ ({ \
10049  uint8x8x3_t __ret; \
10050  __builtin_neon_vld3_dup_v(&__ret, __p0, 16); \
10051 \
10052  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10053  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10054  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10055  __ret; \
10056})
10057#endif
10058
10059#ifdef __LITTLE_ENDIAN__
10060#define vld3_dup_u32(__p0) __extension__ ({ \
10061  uint32x2x3_t __ret; \
10062  __builtin_neon_vld3_dup_v(&__ret, __p0, 18); \
10063  __ret; \
10064})
10065#else
10066#define vld3_dup_u32(__p0) __extension__ ({ \
10067  uint32x2x3_t __ret; \
10068  __builtin_neon_vld3_dup_v(&__ret, __p0, 18); \
10069 \
10070  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10071  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10072  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10073  __ret; \
10074})
10075#endif
10076
10077#ifdef __LITTLE_ENDIAN__
10078#define vld3_dup_u64(__p0) __extension__ ({ \
10079  uint64x1x3_t __ret; \
10080  __builtin_neon_vld3_dup_v(&__ret, __p0, 19); \
10081  __ret; \
10082})
10083#else
10084#define vld3_dup_u64(__p0) __extension__ ({ \
10085  uint64x1x3_t __ret; \
10086  __builtin_neon_vld3_dup_v(&__ret, __p0, 19); \
10087  __ret; \
10088})
10089#endif
10090
10091#ifdef __LITTLE_ENDIAN__
10092#define vld3_dup_u16(__p0) __extension__ ({ \
10093  uint16x4x3_t __ret; \
10094  __builtin_neon_vld3_dup_v(&__ret, __p0, 17); \
10095  __ret; \
10096})
10097#else
10098#define vld3_dup_u16(__p0) __extension__ ({ \
10099  uint16x4x3_t __ret; \
10100  __builtin_neon_vld3_dup_v(&__ret, __p0, 17); \
10101 \
10102  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10103  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10104  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10105  __ret; \
10106})
10107#endif
10108
10109#ifdef __LITTLE_ENDIAN__
10110#define vld3_dup_s8(__p0) __extension__ ({ \
10111  int8x8x3_t __ret; \
10112  __builtin_neon_vld3_dup_v(&__ret, __p0, 0); \
10113  __ret; \
10114})
10115#else
10116#define vld3_dup_s8(__p0) __extension__ ({ \
10117  int8x8x3_t __ret; \
10118  __builtin_neon_vld3_dup_v(&__ret, __p0, 0); \
10119 \
10120  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10121  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10122  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10123  __ret; \
10124})
10125#endif
10126
10127#ifdef __LITTLE_ENDIAN__
10128#define vld3_dup_f32(__p0) __extension__ ({ \
10129  float32x2x3_t __ret; \
10130  __builtin_neon_vld3_dup_v(&__ret, __p0, 9); \
10131  __ret; \
10132})
10133#else
10134#define vld3_dup_f32(__p0) __extension__ ({ \
10135  float32x2x3_t __ret; \
10136  __builtin_neon_vld3_dup_v(&__ret, __p0, 9); \
10137 \
10138  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10139  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10140  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10141  __ret; \
10142})
10143#endif
10144
10145#ifdef __LITTLE_ENDIAN__
10146#define vld3_dup_f16(__p0) __extension__ ({ \
10147  float16x4x3_t __ret; \
10148  __builtin_neon_vld3_dup_v(&__ret, __p0, 8); \
10149  __ret; \
10150})
10151#else
10152#define vld3_dup_f16(__p0) __extension__ ({ \
10153  float16x4x3_t __ret; \
10154  __builtin_neon_vld3_dup_v(&__ret, __p0, 8); \
10155 \
10156  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10157  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10158  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10159  __ret; \
10160})
10161#endif
10162
10163#ifdef __LITTLE_ENDIAN__
10164#define vld3_dup_s32(__p0) __extension__ ({ \
10165  int32x2x3_t __ret; \
10166  __builtin_neon_vld3_dup_v(&__ret, __p0, 2); \
10167  __ret; \
10168})
10169#else
10170#define vld3_dup_s32(__p0) __extension__ ({ \
10171  int32x2x3_t __ret; \
10172  __builtin_neon_vld3_dup_v(&__ret, __p0, 2); \
10173 \
10174  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10175  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10176  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10177  __ret; \
10178})
10179#endif
10180
10181#ifdef __LITTLE_ENDIAN__
10182#define vld3_dup_s64(__p0) __extension__ ({ \
10183  int64x1x3_t __ret; \
10184  __builtin_neon_vld3_dup_v(&__ret, __p0, 3); \
10185  __ret; \
10186})
10187#else
10188#define vld3_dup_s64(__p0) __extension__ ({ \
10189  int64x1x3_t __ret; \
10190  __builtin_neon_vld3_dup_v(&__ret, __p0, 3); \
10191  __ret; \
10192})
10193#endif
10194
10195#ifdef __LITTLE_ENDIAN__
10196#define vld3_dup_s16(__p0) __extension__ ({ \
10197  int16x4x3_t __ret; \
10198  __builtin_neon_vld3_dup_v(&__ret, __p0, 1); \
10199  __ret; \
10200})
10201#else
10202#define vld3_dup_s16(__p0) __extension__ ({ \
10203  int16x4x3_t __ret; \
10204  __builtin_neon_vld3_dup_v(&__ret, __p0, 1); \
10205 \
10206  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10207  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10208  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10209  __ret; \
10210})
10211#endif
10212
10213#ifdef __LITTLE_ENDIAN__
10214#define vld3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
10215  poly8x8x3_t __s1 = __p1; \
10216  poly8x8x3_t __ret; \
10217  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 4); \
10218  __ret; \
10219})
10220#else
10221#define vld3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
10222  poly8x8x3_t __s1 = __p1; \
10223  poly8x8x3_t __rev1; \
10224  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10225  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10226  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10227  poly8x8x3_t __ret; \
10228  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 4); \
10229 \
10230  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10231  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10232  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10233  __ret; \
10234})
10235#endif
10236
10237#ifdef __LITTLE_ENDIAN__
10238#define vld3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10239  poly16x4x3_t __s1 = __p1; \
10240  poly16x4x3_t __ret; \
10241  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 5); \
10242  __ret; \
10243})
10244#else
10245#define vld3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10246  poly16x4x3_t __s1 = __p1; \
10247  poly16x4x3_t __rev1; \
10248  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10249  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10250  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10251  poly16x4x3_t __ret; \
10252  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 5); \
10253 \
10254  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10255  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10256  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10257  __ret; \
10258})
10259#endif
10260
10261#ifdef __LITTLE_ENDIAN__
10262#define vld3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10263  poly16x8x3_t __s1 = __p1; \
10264  poly16x8x3_t __ret; \
10265  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 37); \
10266  __ret; \
10267})
10268#else
10269#define vld3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10270  poly16x8x3_t __s1 = __p1; \
10271  poly16x8x3_t __rev1; \
10272  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10273  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10274  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10275  poly16x8x3_t __ret; \
10276  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 37); \
10277 \
10278  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10279  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10280  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10281  __ret; \
10282})
10283#endif
10284
10285#ifdef __LITTLE_ENDIAN__
10286#define vld3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10287  uint32x4x3_t __s1 = __p1; \
10288  uint32x4x3_t __ret; \
10289  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 50); \
10290  __ret; \
10291})
10292#else
10293#define vld3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10294  uint32x4x3_t __s1 = __p1; \
10295  uint32x4x3_t __rev1; \
10296  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10297  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10298  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10299  uint32x4x3_t __ret; \
10300  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 50); \
10301 \
10302  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10303  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10304  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10305  __ret; \
10306})
10307#endif
10308
10309#ifdef __LITTLE_ENDIAN__
10310#define vld3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10311  uint16x8x3_t __s1 = __p1; \
10312  uint16x8x3_t __ret; \
10313  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 49); \
10314  __ret; \
10315})
10316#else
10317#define vld3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10318  uint16x8x3_t __s1 = __p1; \
10319  uint16x8x3_t __rev1; \
10320  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10321  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10322  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10323  uint16x8x3_t __ret; \
10324  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 49); \
10325 \
10326  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10327  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10328  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10329  __ret; \
10330})
10331#endif
10332
10333#ifdef __LITTLE_ENDIAN__
10334#define vld3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10335  float32x4x3_t __s1 = __p1; \
10336  float32x4x3_t __ret; \
10337  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 41); \
10338  __ret; \
10339})
10340#else
10341#define vld3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10342  float32x4x3_t __s1 = __p1; \
10343  float32x4x3_t __rev1; \
10344  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10345  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10346  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10347  float32x4x3_t __ret; \
10348  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 41); \
10349 \
10350  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10351  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10352  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10353  __ret; \
10354})
10355#endif
10356
10357#ifdef __LITTLE_ENDIAN__
10358#define vld3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10359  float16x8x3_t __s1 = __p1; \
10360  float16x8x3_t __ret; \
10361  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 40); \
10362  __ret; \
10363})
10364#else
10365#define vld3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10366  float16x8x3_t __s1 = __p1; \
10367  float16x8x3_t __rev1; \
10368  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10369  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10370  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10371  float16x8x3_t __ret; \
10372  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 40); \
10373 \
10374  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10375  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10376  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10377  __ret; \
10378})
10379#endif
10380
10381#ifdef __LITTLE_ENDIAN__
10382#define vld3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10383  int32x4x3_t __s1 = __p1; \
10384  int32x4x3_t __ret; \
10385  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 34); \
10386  __ret; \
10387})
10388#else
10389#define vld3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10390  int32x4x3_t __s1 = __p1; \
10391  int32x4x3_t __rev1; \
10392  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10393  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10394  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10395  int32x4x3_t __ret; \
10396  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 34); \
10397 \
10398  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10399  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10400  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10401  __ret; \
10402})
10403#endif
10404
10405#ifdef __LITTLE_ENDIAN__
10406#define vld3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10407  int16x8x3_t __s1 = __p1; \
10408  int16x8x3_t __ret; \
10409  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 33); \
10410  __ret; \
10411})
10412#else
10413#define vld3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10414  int16x8x3_t __s1 = __p1; \
10415  int16x8x3_t __rev1; \
10416  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10417  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10418  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10419  int16x8x3_t __ret; \
10420  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 33); \
10421 \
10422  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10423  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10424  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10425  __ret; \
10426})
10427#endif
10428
10429#ifdef __LITTLE_ENDIAN__
10430#define vld3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
10431  uint8x8x3_t __s1 = __p1; \
10432  uint8x8x3_t __ret; \
10433  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 16); \
10434  __ret; \
10435})
10436#else
10437#define vld3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
10438  uint8x8x3_t __s1 = __p1; \
10439  uint8x8x3_t __rev1; \
10440  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10441  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10442  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10443  uint8x8x3_t __ret; \
10444  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 16); \
10445 \
10446  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10447  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10448  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10449  __ret; \
10450})
10451#endif
10452
10453#ifdef __LITTLE_ENDIAN__
10454#define vld3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10455  uint32x2x3_t __s1 = __p1; \
10456  uint32x2x3_t __ret; \
10457  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 18); \
10458  __ret; \
10459})
10460#else
10461#define vld3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10462  uint32x2x3_t __s1 = __p1; \
10463  uint32x2x3_t __rev1; \
10464  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
10465  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
10466  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
10467  uint32x2x3_t __ret; \
10468  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 18); \
10469 \
10470  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10471  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10472  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10473  __ret; \
10474})
10475#endif
10476
10477#ifdef __LITTLE_ENDIAN__
10478#define vld3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10479  uint16x4x3_t __s1 = __p1; \
10480  uint16x4x3_t __ret; \
10481  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 17); \
10482  __ret; \
10483})
10484#else
10485#define vld3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10486  uint16x4x3_t __s1 = __p1; \
10487  uint16x4x3_t __rev1; \
10488  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10489  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10490  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10491  uint16x4x3_t __ret; \
10492  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 17); \
10493 \
10494  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10495  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10496  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10497  __ret; \
10498})
10499#endif
10500
10501#ifdef __LITTLE_ENDIAN__
10502#define vld3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
10503  int8x8x3_t __s1 = __p1; \
10504  int8x8x3_t __ret; \
10505  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 0); \
10506  __ret; \
10507})
10508#else
10509#define vld3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
10510  int8x8x3_t __s1 = __p1; \
10511  int8x8x3_t __rev1; \
10512  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10513  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10514  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10515  int8x8x3_t __ret; \
10516  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 0); \
10517 \
10518  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10519  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10520  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10521  __ret; \
10522})
10523#endif
10524
10525#ifdef __LITTLE_ENDIAN__
10526#define vld3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10527  float32x2x3_t __s1 = __p1; \
10528  float32x2x3_t __ret; \
10529  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 9); \
10530  __ret; \
10531})
10532#else
10533#define vld3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10534  float32x2x3_t __s1 = __p1; \
10535  float32x2x3_t __rev1; \
10536  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
10537  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
10538  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
10539  float32x2x3_t __ret; \
10540  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 9); \
10541 \
10542  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10543  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10544  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10545  __ret; \
10546})
10547#endif
10548
10549#ifdef __LITTLE_ENDIAN__
10550#define vld3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10551  float16x4x3_t __s1 = __p1; \
10552  float16x4x3_t __ret; \
10553  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 8); \
10554  __ret; \
10555})
10556#else
10557#define vld3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10558  float16x4x3_t __s1 = __p1; \
10559  float16x4x3_t __rev1; \
10560  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10561  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10562  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10563  float16x4x3_t __ret; \
10564  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 8); \
10565 \
10566  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10567  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10568  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10569  __ret; \
10570})
10571#endif
10572
10573#ifdef __LITTLE_ENDIAN__
10574#define vld3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10575  int32x2x3_t __s1 = __p1; \
10576  int32x2x3_t __ret; \
10577  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 2); \
10578  __ret; \
10579})
10580#else
10581#define vld3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10582  int32x2x3_t __s1 = __p1; \
10583  int32x2x3_t __rev1; \
10584  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
10585  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
10586  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
10587  int32x2x3_t __ret; \
10588  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 2); \
10589 \
10590  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10591  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10592  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10593  __ret; \
10594})
10595#endif
10596
10597#ifdef __LITTLE_ENDIAN__
10598#define vld3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10599  int16x4x3_t __s1 = __p1; \
10600  int16x4x3_t __ret; \
10601  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 1); \
10602  __ret; \
10603})
10604#else
10605#define vld3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10606  int16x4x3_t __s1 = __p1; \
10607  int16x4x3_t __rev1; \
10608  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10609  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10610  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10611  int16x4x3_t __ret; \
10612  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 1); \
10613 \
10614  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10615  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10616  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10617  __ret; \
10618})
10619#endif
10620
10621#ifdef __LITTLE_ENDIAN__
10622#define vld4_p8(__p0) __extension__ ({ \
10623  poly8x8x4_t __ret; \
10624  __builtin_neon_vld4_v(&__ret, __p0, 4); \
10625  __ret; \
10626})
10627#else
10628#define vld4_p8(__p0) __extension__ ({ \
10629  poly8x8x4_t __ret; \
10630  __builtin_neon_vld4_v(&__ret, __p0, 4); \
10631 \
10632  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10633  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10634  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10635  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10636  __ret; \
10637})
10638#endif
10639
10640#ifdef __LITTLE_ENDIAN__
10641#define vld4_p16(__p0) __extension__ ({ \
10642  poly16x4x4_t __ret; \
10643  __builtin_neon_vld4_v(&__ret, __p0, 5); \
10644  __ret; \
10645})
10646#else
10647#define vld4_p16(__p0) __extension__ ({ \
10648  poly16x4x4_t __ret; \
10649  __builtin_neon_vld4_v(&__ret, __p0, 5); \
10650 \
10651  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10652  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10653  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10654  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10655  __ret; \
10656})
10657#endif
10658
10659#ifdef __LITTLE_ENDIAN__
10660#define vld4q_p8(__p0) __extension__ ({ \
10661  poly8x16x4_t __ret; \
10662  __builtin_neon_vld4q_v(&__ret, __p0, 36); \
10663  __ret; \
10664})
10665#else
10666#define vld4q_p8(__p0) __extension__ ({ \
10667  poly8x16x4_t __ret; \
10668  __builtin_neon_vld4q_v(&__ret, __p0, 36); \
10669 \
10670  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10671  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10672  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10673  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10674  __ret; \
10675})
10676#endif
10677
10678#ifdef __LITTLE_ENDIAN__
10679#define vld4q_p16(__p0) __extension__ ({ \
10680  poly16x8x4_t __ret; \
10681  __builtin_neon_vld4q_v(&__ret, __p0, 37); \
10682  __ret; \
10683})
10684#else
10685#define vld4q_p16(__p0) __extension__ ({ \
10686  poly16x8x4_t __ret; \
10687  __builtin_neon_vld4q_v(&__ret, __p0, 37); \
10688 \
10689  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10690  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10691  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10692  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10693  __ret; \
10694})
10695#endif
10696
10697#ifdef __LITTLE_ENDIAN__
10698#define vld4q_u8(__p0) __extension__ ({ \
10699  uint8x16x4_t __ret; \
10700  __builtin_neon_vld4q_v(&__ret, __p0, 48); \
10701  __ret; \
10702})
10703#else
10704#define vld4q_u8(__p0) __extension__ ({ \
10705  uint8x16x4_t __ret; \
10706  __builtin_neon_vld4q_v(&__ret, __p0, 48); \
10707 \
10708  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10709  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10710  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10711  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10712  __ret; \
10713})
10714#endif
10715
10716#ifdef __LITTLE_ENDIAN__
10717#define vld4q_u32(__p0) __extension__ ({ \
10718  uint32x4x4_t __ret; \
10719  __builtin_neon_vld4q_v(&__ret, __p0, 50); \
10720  __ret; \
10721})
10722#else
10723#define vld4q_u32(__p0) __extension__ ({ \
10724  uint32x4x4_t __ret; \
10725  __builtin_neon_vld4q_v(&__ret, __p0, 50); \
10726 \
10727  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10728  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10729  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10730  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10731  __ret; \
10732})
10733#endif
10734
10735#ifdef __LITTLE_ENDIAN__
10736#define vld4q_u16(__p0) __extension__ ({ \
10737  uint16x8x4_t __ret; \
10738  __builtin_neon_vld4q_v(&__ret, __p0, 49); \
10739  __ret; \
10740})
10741#else
10742#define vld4q_u16(__p0) __extension__ ({ \
10743  uint16x8x4_t __ret; \
10744  __builtin_neon_vld4q_v(&__ret, __p0, 49); \
10745 \
10746  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10747  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10748  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10749  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10750  __ret; \
10751})
10752#endif
10753
10754#ifdef __LITTLE_ENDIAN__
10755#define vld4q_s8(__p0) __extension__ ({ \
10756  int8x16x4_t __ret; \
10757  __builtin_neon_vld4q_v(&__ret, __p0, 32); \
10758  __ret; \
10759})
10760#else
10761#define vld4q_s8(__p0) __extension__ ({ \
10762  int8x16x4_t __ret; \
10763  __builtin_neon_vld4q_v(&__ret, __p0, 32); \
10764 \
10765  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10766  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10767  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10768  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10769  __ret; \
10770})
10771#endif
10772
10773#ifdef __LITTLE_ENDIAN__
10774#define vld4q_f32(__p0) __extension__ ({ \
10775  float32x4x4_t __ret; \
10776  __builtin_neon_vld4q_v(&__ret, __p0, 41); \
10777  __ret; \
10778})
10779#else
10780#define vld4q_f32(__p0) __extension__ ({ \
10781  float32x4x4_t __ret; \
10782  __builtin_neon_vld4q_v(&__ret, __p0, 41); \
10783 \
10784  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10785  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10786  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10787  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10788  __ret; \
10789})
10790#endif
10791
10792#ifdef __LITTLE_ENDIAN__
10793#define vld4q_f16(__p0) __extension__ ({ \
10794  float16x8x4_t __ret; \
10795  __builtin_neon_vld4q_v(&__ret, __p0, 40); \
10796  __ret; \
10797})
10798#else
10799#define vld4q_f16(__p0) __extension__ ({ \
10800  float16x8x4_t __ret; \
10801  __builtin_neon_vld4q_v(&__ret, __p0, 40); \
10802 \
10803  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10804  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10805  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10806  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10807  __ret; \
10808})
10809#endif
10810
10811#ifdef __LITTLE_ENDIAN__
10812#define vld4q_s32(__p0) __extension__ ({ \
10813  int32x4x4_t __ret; \
10814  __builtin_neon_vld4q_v(&__ret, __p0, 34); \
10815  __ret; \
10816})
10817#else
10818#define vld4q_s32(__p0) __extension__ ({ \
10819  int32x4x4_t __ret; \
10820  __builtin_neon_vld4q_v(&__ret, __p0, 34); \
10821 \
10822  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10823  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10824  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10825  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10826  __ret; \
10827})
10828#endif
10829
10830#ifdef __LITTLE_ENDIAN__
10831#define vld4q_s16(__p0) __extension__ ({ \
10832  int16x8x4_t __ret; \
10833  __builtin_neon_vld4q_v(&__ret, __p0, 33); \
10834  __ret; \
10835})
10836#else
10837#define vld4q_s16(__p0) __extension__ ({ \
10838  int16x8x4_t __ret; \
10839  __builtin_neon_vld4q_v(&__ret, __p0, 33); \
10840 \
10841  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10842  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10843  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10844  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10845  __ret; \
10846})
10847#endif
10848
10849#ifdef __LITTLE_ENDIAN__
10850#define vld4_u8(__p0) __extension__ ({ \
10851  uint8x8x4_t __ret; \
10852  __builtin_neon_vld4_v(&__ret, __p0, 16); \
10853  __ret; \
10854})
10855#else
10856#define vld4_u8(__p0) __extension__ ({ \
10857  uint8x8x4_t __ret; \
10858  __builtin_neon_vld4_v(&__ret, __p0, 16); \
10859 \
10860  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10861  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10862  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10863  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10864  __ret; \
10865})
10866#endif
10867
10868#ifdef __LITTLE_ENDIAN__
10869#define vld4_u32(__p0) __extension__ ({ \
10870  uint32x2x4_t __ret; \
10871  __builtin_neon_vld4_v(&__ret, __p0, 18); \
10872  __ret; \
10873})
10874#else
10875#define vld4_u32(__p0) __extension__ ({ \
10876  uint32x2x4_t __ret; \
10877  __builtin_neon_vld4_v(&__ret, __p0, 18); \
10878 \
10879  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10880  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10881  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10882  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10883  __ret; \
10884})
10885#endif
10886
10887#ifdef __LITTLE_ENDIAN__
10888#define vld4_u64(__p0) __extension__ ({ \
10889  uint64x1x4_t __ret; \
10890  __builtin_neon_vld4_v(&__ret, __p0, 19); \
10891  __ret; \
10892})
10893#else
10894#define vld4_u64(__p0) __extension__ ({ \
10895  uint64x1x4_t __ret; \
10896  __builtin_neon_vld4_v(&__ret, __p0, 19); \
10897  __ret; \
10898})
10899#endif
10900
10901#ifdef __LITTLE_ENDIAN__
10902#define vld4_u16(__p0) __extension__ ({ \
10903  uint16x4x4_t __ret; \
10904  __builtin_neon_vld4_v(&__ret, __p0, 17); \
10905  __ret; \
10906})
10907#else
10908#define vld4_u16(__p0) __extension__ ({ \
10909  uint16x4x4_t __ret; \
10910  __builtin_neon_vld4_v(&__ret, __p0, 17); \
10911 \
10912  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10913  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10914  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10915  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10916  __ret; \
10917})
10918#endif
10919
10920#ifdef __LITTLE_ENDIAN__
10921#define vld4_s8(__p0) __extension__ ({ \
10922  int8x8x4_t __ret; \
10923  __builtin_neon_vld4_v(&__ret, __p0, 0); \
10924  __ret; \
10925})
10926#else
10927#define vld4_s8(__p0) __extension__ ({ \
10928  int8x8x4_t __ret; \
10929  __builtin_neon_vld4_v(&__ret, __p0, 0); \
10930 \
10931  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10932  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10933  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10934  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10935  __ret; \
10936})
10937#endif
10938
10939#ifdef __LITTLE_ENDIAN__
10940#define vld4_f32(__p0) __extension__ ({ \
10941  float32x2x4_t __ret; \
10942  __builtin_neon_vld4_v(&__ret, __p0, 9); \
10943  __ret; \
10944})
10945#else
10946#define vld4_f32(__p0) __extension__ ({ \
10947  float32x2x4_t __ret; \
10948  __builtin_neon_vld4_v(&__ret, __p0, 9); \
10949 \
10950  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10951  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10952  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10953  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10954  __ret; \
10955})
10956#endif
10957
10958#ifdef __LITTLE_ENDIAN__
10959#define vld4_f16(__p0) __extension__ ({ \
10960  float16x4x4_t __ret; \
10961  __builtin_neon_vld4_v(&__ret, __p0, 8); \
10962  __ret; \
10963})
10964#else
10965#define vld4_f16(__p0) __extension__ ({ \
10966  float16x4x4_t __ret; \
10967  __builtin_neon_vld4_v(&__ret, __p0, 8); \
10968 \
10969  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10970  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10971  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10972  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10973  __ret; \
10974})
10975#endif
10976
10977#ifdef __LITTLE_ENDIAN__
10978#define vld4_s32(__p0) __extension__ ({ \
10979  int32x2x4_t __ret; \
10980  __builtin_neon_vld4_v(&__ret, __p0, 2); \
10981  __ret; \
10982})
10983#else
10984#define vld4_s32(__p0) __extension__ ({ \
10985  int32x2x4_t __ret; \
10986  __builtin_neon_vld4_v(&__ret, __p0, 2); \
10987 \
10988  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10989  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10990  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10991  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10992  __ret; \
10993})
10994#endif
10995
10996#ifdef __LITTLE_ENDIAN__
10997#define vld4_s64(__p0) __extension__ ({ \
10998  int64x1x4_t __ret; \
10999  __builtin_neon_vld4_v(&__ret, __p0, 3); \
11000  __ret; \
11001})
11002#else
11003#define vld4_s64(__p0) __extension__ ({ \
11004  int64x1x4_t __ret; \
11005  __builtin_neon_vld4_v(&__ret, __p0, 3); \
11006  __ret; \
11007})
11008#endif
11009
11010#ifdef __LITTLE_ENDIAN__
11011#define vld4_s16(__p0) __extension__ ({ \
11012  int16x4x4_t __ret; \
11013  __builtin_neon_vld4_v(&__ret, __p0, 1); \
11014  __ret; \
11015})
11016#else
11017#define vld4_s16(__p0) __extension__ ({ \
11018  int16x4x4_t __ret; \
11019  __builtin_neon_vld4_v(&__ret, __p0, 1); \
11020 \
11021  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11022  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11023  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11024  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11025  __ret; \
11026})
11027#endif
11028
11029#ifdef __LITTLE_ENDIAN__
11030#define vld4_dup_p8(__p0) __extension__ ({ \
11031  poly8x8x4_t __ret; \
11032  __builtin_neon_vld4_dup_v(&__ret, __p0, 4); \
11033  __ret; \
11034})
11035#else
11036#define vld4_dup_p8(__p0) __extension__ ({ \
11037  poly8x8x4_t __ret; \
11038  __builtin_neon_vld4_dup_v(&__ret, __p0, 4); \
11039 \
11040  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11041  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11042  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11043  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11044  __ret; \
11045})
11046#endif
11047
11048#ifdef __LITTLE_ENDIAN__
11049#define vld4_dup_p16(__p0) __extension__ ({ \
11050  poly16x4x4_t __ret; \
11051  __builtin_neon_vld4_dup_v(&__ret, __p0, 5); \
11052  __ret; \
11053})
11054#else
11055#define vld4_dup_p16(__p0) __extension__ ({ \
11056  poly16x4x4_t __ret; \
11057  __builtin_neon_vld4_dup_v(&__ret, __p0, 5); \
11058 \
11059  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11060  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11061  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11062  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11063  __ret; \
11064})
11065#endif
11066
11067#ifdef __LITTLE_ENDIAN__
11068#define vld4_dup_u8(__p0) __extension__ ({ \
11069  uint8x8x4_t __ret; \
11070  __builtin_neon_vld4_dup_v(&__ret, __p0, 16); \
11071  __ret; \
11072})
11073#else
11074#define vld4_dup_u8(__p0) __extension__ ({ \
11075  uint8x8x4_t __ret; \
11076  __builtin_neon_vld4_dup_v(&__ret, __p0, 16); \
11077 \
11078  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11079  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11080  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11081  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11082  __ret; \
11083})
11084#endif
11085
11086#ifdef __LITTLE_ENDIAN__
11087#define vld4_dup_u32(__p0) __extension__ ({ \
11088  uint32x2x4_t __ret; \
11089  __builtin_neon_vld4_dup_v(&__ret, __p0, 18); \
11090  __ret; \
11091})
11092#else
11093#define vld4_dup_u32(__p0) __extension__ ({ \
11094  uint32x2x4_t __ret; \
11095  __builtin_neon_vld4_dup_v(&__ret, __p0, 18); \
11096 \
11097  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11098  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11099  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11100  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11101  __ret; \
11102})
11103#endif
11104
11105#ifdef __LITTLE_ENDIAN__
11106#define vld4_dup_u64(__p0) __extension__ ({ \
11107  uint64x1x4_t __ret; \
11108  __builtin_neon_vld4_dup_v(&__ret, __p0, 19); \
11109  __ret; \
11110})
11111#else
11112#define vld4_dup_u64(__p0) __extension__ ({ \
11113  uint64x1x4_t __ret; \
11114  __builtin_neon_vld4_dup_v(&__ret, __p0, 19); \
11115  __ret; \
11116})
11117#endif
11118
11119#ifdef __LITTLE_ENDIAN__
11120#define vld4_dup_u16(__p0) __extension__ ({ \
11121  uint16x4x4_t __ret; \
11122  __builtin_neon_vld4_dup_v(&__ret, __p0, 17); \
11123  __ret; \
11124})
11125#else
11126#define vld4_dup_u16(__p0) __extension__ ({ \
11127  uint16x4x4_t __ret; \
11128  __builtin_neon_vld4_dup_v(&__ret, __p0, 17); \
11129 \
11130  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11131  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11132  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11133  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11134  __ret; \
11135})
11136#endif
11137
11138#ifdef __LITTLE_ENDIAN__
11139#define vld4_dup_s8(__p0) __extension__ ({ \
11140  int8x8x4_t __ret; \
11141  __builtin_neon_vld4_dup_v(&__ret, __p0, 0); \
11142  __ret; \
11143})
11144#else
11145#define vld4_dup_s8(__p0) __extension__ ({ \
11146  int8x8x4_t __ret; \
11147  __builtin_neon_vld4_dup_v(&__ret, __p0, 0); \
11148 \
11149  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11150  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11151  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11152  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11153  __ret; \
11154})
11155#endif
11156
11157#ifdef __LITTLE_ENDIAN__
11158#define vld4_dup_f32(__p0) __extension__ ({ \
11159  float32x2x4_t __ret; \
11160  __builtin_neon_vld4_dup_v(&__ret, __p0, 9); \
11161  __ret; \
11162})
11163#else
11164#define vld4_dup_f32(__p0) __extension__ ({ \
11165  float32x2x4_t __ret; \
11166  __builtin_neon_vld4_dup_v(&__ret, __p0, 9); \
11167 \
11168  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11169  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11170  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11171  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11172  __ret; \
11173})
11174#endif
11175
11176#ifdef __LITTLE_ENDIAN__
11177#define vld4_dup_f16(__p0) __extension__ ({ \
11178  float16x4x4_t __ret; \
11179  __builtin_neon_vld4_dup_v(&__ret, __p0, 8); \
11180  __ret; \
11181})
11182#else
11183#define vld4_dup_f16(__p0) __extension__ ({ \
11184  float16x4x4_t __ret; \
11185  __builtin_neon_vld4_dup_v(&__ret, __p0, 8); \
11186 \
11187  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11188  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11189  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11190  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11191  __ret; \
11192})
11193#endif
11194
11195#ifdef __LITTLE_ENDIAN__
11196#define vld4_dup_s32(__p0) __extension__ ({ \
11197  int32x2x4_t __ret; \
11198  __builtin_neon_vld4_dup_v(&__ret, __p0, 2); \
11199  __ret; \
11200})
11201#else
11202#define vld4_dup_s32(__p0) __extension__ ({ \
11203  int32x2x4_t __ret; \
11204  __builtin_neon_vld4_dup_v(&__ret, __p0, 2); \
11205 \
11206  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11207  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11208  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11209  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11210  __ret; \
11211})
11212#endif
11213
11214#ifdef __LITTLE_ENDIAN__
11215#define vld4_dup_s64(__p0) __extension__ ({ \
11216  int64x1x4_t __ret; \
11217  __builtin_neon_vld4_dup_v(&__ret, __p0, 3); \
11218  __ret; \
11219})
11220#else
11221#define vld4_dup_s64(__p0) __extension__ ({ \
11222  int64x1x4_t __ret; \
11223  __builtin_neon_vld4_dup_v(&__ret, __p0, 3); \
11224  __ret; \
11225})
11226#endif
11227
11228#ifdef __LITTLE_ENDIAN__
11229#define vld4_dup_s16(__p0) __extension__ ({ \
11230  int16x4x4_t __ret; \
11231  __builtin_neon_vld4_dup_v(&__ret, __p0, 1); \
11232  __ret; \
11233})
11234#else
11235#define vld4_dup_s16(__p0) __extension__ ({ \
11236  int16x4x4_t __ret; \
11237  __builtin_neon_vld4_dup_v(&__ret, __p0, 1); \
11238 \
11239  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11240  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11241  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11242  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11243  __ret; \
11244})
11245#endif
11246
11247#ifdef __LITTLE_ENDIAN__
11248#define vld4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
11249  poly8x8x4_t __s1 = __p1; \
11250  poly8x8x4_t __ret; \
11251  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 4); \
11252  __ret; \
11253})
11254#else
11255#define vld4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
11256  poly8x8x4_t __s1 = __p1; \
11257  poly8x8x4_t __rev1; \
11258  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11259  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11260  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11261  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11262  poly8x8x4_t __ret; \
11263  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 4); \
11264 \
11265  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11266  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11267  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11268  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11269  __ret; \
11270})
11271#endif
11272
11273#ifdef __LITTLE_ENDIAN__
11274#define vld4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11275  poly16x4x4_t __s1 = __p1; \
11276  poly16x4x4_t __ret; \
11277  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 5); \
11278  __ret; \
11279})
11280#else
11281#define vld4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11282  poly16x4x4_t __s1 = __p1; \
11283  poly16x4x4_t __rev1; \
11284  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11285  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11286  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11287  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11288  poly16x4x4_t __ret; \
11289  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 5); \
11290 \
11291  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11292  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11293  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11294  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11295  __ret; \
11296})
11297#endif
11298
11299#ifdef __LITTLE_ENDIAN__
11300#define vld4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11301  poly16x8x4_t __s1 = __p1; \
11302  poly16x8x4_t __ret; \
11303  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 37); \
11304  __ret; \
11305})
11306#else
11307#define vld4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11308  poly16x8x4_t __s1 = __p1; \
11309  poly16x8x4_t __rev1; \
11310  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11311  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11312  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11313  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11314  poly16x8x4_t __ret; \
11315  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 37); \
11316 \
11317  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11318  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11319  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11320  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11321  __ret; \
11322})
11323#endif
11324
11325#ifdef __LITTLE_ENDIAN__
11326#define vld4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11327  uint32x4x4_t __s1 = __p1; \
11328  uint32x4x4_t __ret; \
11329  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 50); \
11330  __ret; \
11331})
11332#else
11333#define vld4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11334  uint32x4x4_t __s1 = __p1; \
11335  uint32x4x4_t __rev1; \
11336  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11337  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11338  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11339  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11340  uint32x4x4_t __ret; \
11341  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 50); \
11342 \
11343  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11344  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11345  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11346  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11347  __ret; \
11348})
11349#endif
11350
11351#ifdef __LITTLE_ENDIAN__
11352#define vld4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11353  uint16x8x4_t __s1 = __p1; \
11354  uint16x8x4_t __ret; \
11355  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 49); \
11356  __ret; \
11357})
11358#else
11359#define vld4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11360  uint16x8x4_t __s1 = __p1; \
11361  uint16x8x4_t __rev1; \
11362  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11363  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11364  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11365  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11366  uint16x8x4_t __ret; \
11367  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 49); \
11368 \
11369  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11370  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11371  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11372  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11373  __ret; \
11374})
11375#endif
11376
11377#ifdef __LITTLE_ENDIAN__
11378#define vld4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11379  float32x4x4_t __s1 = __p1; \
11380  float32x4x4_t __ret; \
11381  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 41); \
11382  __ret; \
11383})
11384#else
11385#define vld4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11386  float32x4x4_t __s1 = __p1; \
11387  float32x4x4_t __rev1; \
11388  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11389  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11390  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11391  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11392  float32x4x4_t __ret; \
11393  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 41); \
11394 \
11395  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11396  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11397  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11398  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11399  __ret; \
11400})
11401#endif
11402
11403#ifdef __LITTLE_ENDIAN__
11404#define vld4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11405  float16x8x4_t __s1 = __p1; \
11406  float16x8x4_t __ret; \
11407  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 40); \
11408  __ret; \
11409})
11410#else
11411#define vld4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11412  float16x8x4_t __s1 = __p1; \
11413  float16x8x4_t __rev1; \
11414  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11415  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11416  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11417  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11418  float16x8x4_t __ret; \
11419  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 40); \
11420 \
11421  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11422  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11423  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11424  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11425  __ret; \
11426})
11427#endif
11428
11429#ifdef __LITTLE_ENDIAN__
11430#define vld4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11431  int32x4x4_t __s1 = __p1; \
11432  int32x4x4_t __ret; \
11433  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 34); \
11434  __ret; \
11435})
11436#else
11437#define vld4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11438  int32x4x4_t __s1 = __p1; \
11439  int32x4x4_t __rev1; \
11440  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11441  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11442  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11443  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11444  int32x4x4_t __ret; \
11445  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 34); \
11446 \
11447  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11448  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11449  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11450  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11451  __ret; \
11452})
11453#endif
11454
11455#ifdef __LITTLE_ENDIAN__
11456#define vld4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11457  int16x8x4_t __s1 = __p1; \
11458  int16x8x4_t __ret; \
11459  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 33); \
11460  __ret; \
11461})
11462#else
11463#define vld4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11464  int16x8x4_t __s1 = __p1; \
11465  int16x8x4_t __rev1; \
11466  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11467  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11468  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11469  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11470  int16x8x4_t __ret; \
11471  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 33); \
11472 \
11473  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11474  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11475  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11476  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11477  __ret; \
11478})
11479#endif
11480
11481#ifdef __LITTLE_ENDIAN__
11482#define vld4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
11483  uint8x8x4_t __s1 = __p1; \
11484  uint8x8x4_t __ret; \
11485  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 16); \
11486  __ret; \
11487})
11488#else
11489#define vld4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
11490  uint8x8x4_t __s1 = __p1; \
11491  uint8x8x4_t __rev1; \
11492  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11493  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11494  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11495  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11496  uint8x8x4_t __ret; \
11497  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 16); \
11498 \
11499  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11500  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11501  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11502  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11503  __ret; \
11504})
11505#endif
11506
11507#ifdef __LITTLE_ENDIAN__
11508#define vld4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11509  uint32x2x4_t __s1 = __p1; \
11510  uint32x2x4_t __ret; \
11511  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 18); \
11512  __ret; \
11513})
11514#else
11515#define vld4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11516  uint32x2x4_t __s1 = __p1; \
11517  uint32x2x4_t __rev1; \
11518  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11519  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11520  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
11521  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
11522  uint32x2x4_t __ret; \
11523  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 18); \
11524 \
11525  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11526  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11527  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11528  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11529  __ret; \
11530})
11531#endif
11532
11533#ifdef __LITTLE_ENDIAN__
11534#define vld4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11535  uint16x4x4_t __s1 = __p1; \
11536  uint16x4x4_t __ret; \
11537  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 17); \
11538  __ret; \
11539})
11540#else
11541#define vld4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11542  uint16x4x4_t __s1 = __p1; \
11543  uint16x4x4_t __rev1; \
11544  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11545  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11546  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11547  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11548  uint16x4x4_t __ret; \
11549  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 17); \
11550 \
11551  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11552  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11553  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11554  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11555  __ret; \
11556})
11557#endif
11558
11559#ifdef __LITTLE_ENDIAN__
11560#define vld4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
11561  int8x8x4_t __s1 = __p1; \
11562  int8x8x4_t __ret; \
11563  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 0); \
11564  __ret; \
11565})
11566#else
11567#define vld4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
11568  int8x8x4_t __s1 = __p1; \
11569  int8x8x4_t __rev1; \
11570  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11571  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11572  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11573  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11574  int8x8x4_t __ret; \
11575  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 0); \
11576 \
11577  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11578  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11579  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11580  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11581  __ret; \
11582})
11583#endif
11584
11585#ifdef __LITTLE_ENDIAN__
11586#define vld4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11587  float32x2x4_t __s1 = __p1; \
11588  float32x2x4_t __ret; \
11589  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 9); \
11590  __ret; \
11591})
11592#else
11593#define vld4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11594  float32x2x4_t __s1 = __p1; \
11595  float32x2x4_t __rev1; \
11596  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11597  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11598  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
11599  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
11600  float32x2x4_t __ret; \
11601  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 9); \
11602 \
11603  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11604  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11605  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11606  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11607  __ret; \
11608})
11609#endif
11610
11611#ifdef __LITTLE_ENDIAN__
11612#define vld4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11613  float16x4x4_t __s1 = __p1; \
11614  float16x4x4_t __ret; \
11615  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 8); \
11616  __ret; \
11617})
11618#else
11619#define vld4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11620  float16x4x4_t __s1 = __p1; \
11621  float16x4x4_t __rev1; \
11622  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11623  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11624  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11625  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11626  float16x4x4_t __ret; \
11627  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 8); \
11628 \
11629  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11630  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11631  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11632  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11633  __ret; \
11634})
11635#endif
11636
11637#ifdef __LITTLE_ENDIAN__
11638#define vld4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11639  int32x2x4_t __s1 = __p1; \
11640  int32x2x4_t __ret; \
11641  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 2); \
11642  __ret; \
11643})
11644#else
11645#define vld4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11646  int32x2x4_t __s1 = __p1; \
11647  int32x2x4_t __rev1; \
11648  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11649  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11650  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
11651  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
11652  int32x2x4_t __ret; \
11653  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 2); \
11654 \
11655  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11656  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11657  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11658  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11659  __ret; \
11660})
11661#endif
11662
11663#ifdef __LITTLE_ENDIAN__
11664#define vld4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11665  int16x4x4_t __s1 = __p1; \
11666  int16x4x4_t __ret; \
11667  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 1); \
11668  __ret; \
11669})
11670#else
11671#define vld4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11672  int16x4x4_t __s1 = __p1; \
11673  int16x4x4_t __rev1; \
11674  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11675  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11676  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11677  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11678  int16x4x4_t __ret; \
11679  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 1); \
11680 \
11681  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11682  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11683  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11684  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11685  __ret; \
11686})
11687#endif
11688
11689#ifdef __LITTLE_ENDIAN__
11690__ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11691  uint8x16_t __ret;
11692  __ret = (uint8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
11693  return __ret;
11694}
11695#else
11696__ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11697  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11698  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11699  uint8x16_t __ret;
11700  __ret = (uint8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
11701  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11702  return __ret;
11703}
11704#endif
11705
11706#ifdef __LITTLE_ENDIAN__
11707__ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11708  uint32x4_t __ret;
11709  __ret = (uint32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
11710  return __ret;
11711}
11712#else
11713__ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11714  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11715  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11716  uint32x4_t __ret;
11717  __ret = (uint32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
11718  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11719  return __ret;
11720}
11721#endif
11722
11723#ifdef __LITTLE_ENDIAN__
11724__ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11725  uint16x8_t __ret;
11726  __ret = (uint16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
11727  return __ret;
11728}
11729#else
11730__ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11731  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11732  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11733  uint16x8_t __ret;
11734  __ret = (uint16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
11735  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11736  return __ret;
11737}
11738#endif
11739
11740#ifdef __LITTLE_ENDIAN__
11741__ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
11742  int8x16_t __ret;
11743  __ret = (int8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
11744  return __ret;
11745}
11746#else
11747__ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
11748  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11749  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11750  int8x16_t __ret;
11751  __ret = (int8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
11752  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11753  return __ret;
11754}
11755#endif
11756
11757#ifdef __LITTLE_ENDIAN__
11758__ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
11759  float32x4_t __ret;
11760  __ret = (float32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
11761  return __ret;
11762}
11763#else
11764__ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
11765  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11766  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11767  float32x4_t __ret;
11768  __ret = (float32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
11769  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11770  return __ret;
11771}
11772#endif
11773
11774#ifdef __LITTLE_ENDIAN__
11775__ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
11776  int32x4_t __ret;
11777  __ret = (int32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
11778  return __ret;
11779}
11780#else
11781__ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
11782  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11783  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11784  int32x4_t __ret;
11785  __ret = (int32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
11786  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11787  return __ret;
11788}
11789#endif
11790
11791#ifdef __LITTLE_ENDIAN__
11792__ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
11793  int16x8_t __ret;
11794  __ret = (int16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
11795  return __ret;
11796}
11797#else
11798__ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
11799  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11800  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11801  int16x8_t __ret;
11802  __ret = (int16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
11803  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11804  return __ret;
11805}
11806#endif
11807
11808#ifdef __LITTLE_ENDIAN__
11809__ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
11810  uint8x8_t __ret;
11811  __ret = (uint8x8_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
11812  return __ret;
11813}
11814#else
11815__ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
11816  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11817  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11818  uint8x8_t __ret;
11819  __ret = (uint8x8_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
11820  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11821  return __ret;
11822}
11823#endif
11824
11825#ifdef __LITTLE_ENDIAN__
11826__ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
11827  uint32x2_t __ret;
11828  __ret = (uint32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
11829  return __ret;
11830}
11831#else
11832__ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
11833  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
11834  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
11835  uint32x2_t __ret;
11836  __ret = (uint32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
11837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
11838  return __ret;
11839}
11840#endif
11841
11842#ifdef __LITTLE_ENDIAN__
11843__ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
11844  uint16x4_t __ret;
11845  __ret = (uint16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
11846  return __ret;
11847}
11848#else
11849__ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
11850  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11851  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11852  uint16x4_t __ret;
11853  __ret = (uint16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
11854  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11855  return __ret;
11856}
11857#endif
11858
11859#ifdef __LITTLE_ENDIAN__
11860__ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1) {
11861  int8x8_t __ret;
11862  __ret = (int8x8_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
11863  return __ret;
11864}
11865#else
11866__ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1) {
11867  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11868  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11869  int8x8_t __ret;
11870  __ret = (int8x8_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
11871  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11872  return __ret;
11873}
11874#endif
11875
11876#ifdef __LITTLE_ENDIAN__
11877__ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1) {
11878  float32x2_t __ret;
11879  __ret = (float32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
11880  return __ret;
11881}
11882#else
11883__ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1) {
11884  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
11885  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
11886  float32x2_t __ret;
11887  __ret = (float32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
11888  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
11889  return __ret;
11890}
11891#endif
11892
11893#ifdef __LITTLE_ENDIAN__
11894__ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1) {
11895  int32x2_t __ret;
11896  __ret = (int32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
11897  return __ret;
11898}
11899#else
11900__ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1) {
11901  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
11902  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
11903  int32x2_t __ret;
11904  __ret = (int32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
11905  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
11906  return __ret;
11907}
11908#endif
11909
11910#ifdef __LITTLE_ENDIAN__
11911__ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1) {
11912  int16x4_t __ret;
11913  __ret = (int16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
11914  return __ret;
11915}
11916#else
11917__ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1) {
11918  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11919  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11920  int16x4_t __ret;
11921  __ret = (int16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
11922  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11923  return __ret;
11924}
11925#endif
11926
11927#ifdef __LITTLE_ENDIAN__
11928__ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11929  uint8x16_t __ret;
11930  __ret = (uint8x16_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
11931  return __ret;
11932}
11933#else
11934__ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11935  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11936  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11937  uint8x16_t __ret;
11938  __ret = (uint8x16_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
11939  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11940  return __ret;
11941}
11942#endif
11943
11944#ifdef __LITTLE_ENDIAN__
11945__ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11946  uint32x4_t __ret;
11947  __ret = (uint32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
11948  return __ret;
11949}
11950#else
11951__ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11952  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11953  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11954  uint32x4_t __ret;
11955  __ret = (uint32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
11956  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11957  return __ret;
11958}
11959#endif
11960
11961#ifdef __LITTLE_ENDIAN__
11962__ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11963  uint16x8_t __ret;
11964  __ret = (uint16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
11965  return __ret;
11966}
11967#else
11968__ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11969  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11970  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11971  uint16x8_t __ret;
11972  __ret = (uint16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
11973  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11974  return __ret;
11975}
11976#endif
11977
11978#ifdef __LITTLE_ENDIAN__
11979__ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1) {
11980  int8x16_t __ret;
11981  __ret = (int8x16_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
11982  return __ret;
11983}
11984#else
11985__ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1) {
11986  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11987  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11988  int8x16_t __ret;
11989  __ret = (int8x16_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
11990  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11991  return __ret;
11992}
11993#endif
11994
11995#ifdef __LITTLE_ENDIAN__
11996__ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1) {
11997  float32x4_t __ret;
11998  __ret = (float32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
11999  return __ret;
12000}
12001#else
12002__ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1) {
12003  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12004  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12005  float32x4_t __ret;
12006  __ret = (float32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
12007  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12008  return __ret;
12009}
12010#endif
12011
12012#ifdef __LITTLE_ENDIAN__
12013__ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1) {
12014  int32x4_t __ret;
12015  __ret = (int32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
12016  return __ret;
12017}
12018#else
12019__ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1) {
12020  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12021  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12022  int32x4_t __ret;
12023  __ret = (int32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
12024  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12025  return __ret;
12026}
12027#endif
12028
12029#ifdef __LITTLE_ENDIAN__
12030__ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1) {
12031  int16x8_t __ret;
12032  __ret = (int16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
12033  return __ret;
12034}
12035#else
12036__ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1) {
12037  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12038  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12039  int16x8_t __ret;
12040  __ret = (int16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
12041  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12042  return __ret;
12043}
12044#endif
12045
12046#ifdef __LITTLE_ENDIAN__
12047__ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
12048  uint8x8_t __ret;
12049  __ret = (uint8x8_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
12050  return __ret;
12051}
12052#else
12053__ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
12054  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12055  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12056  uint8x8_t __ret;
12057  __ret = (uint8x8_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
12058  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12059  return __ret;
12060}
12061#endif
12062
12063#ifdef __LITTLE_ENDIAN__
12064__ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
12065  uint32x2_t __ret;
12066  __ret = (uint32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
12067  return __ret;
12068}
12069#else
12070__ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
12071  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12072  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12073  uint32x2_t __ret;
12074  __ret = (uint32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
12075  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12076  return __ret;
12077}
12078#endif
12079
12080#ifdef __LITTLE_ENDIAN__
12081__ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
12082  uint16x4_t __ret;
12083  __ret = (uint16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
12084  return __ret;
12085}
12086#else
12087__ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
12088  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12089  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12090  uint16x4_t __ret;
12091  __ret = (uint16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
12092  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12093  return __ret;
12094}
12095#endif
12096
12097#ifdef __LITTLE_ENDIAN__
12098__ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1) {
12099  int8x8_t __ret;
12100  __ret = (int8x8_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
12101  return __ret;
12102}
12103#else
12104__ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1) {
12105  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12106  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12107  int8x8_t __ret;
12108  __ret = (int8x8_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
12109  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12110  return __ret;
12111}
12112#endif
12113
12114#ifdef __LITTLE_ENDIAN__
12115__ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1) {
12116  float32x2_t __ret;
12117  __ret = (float32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
12118  return __ret;
12119}
12120#else
12121__ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1) {
12122  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12123  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12124  float32x2_t __ret;
12125  __ret = (float32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
12126  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12127  return __ret;
12128}
12129#endif
12130
12131#ifdef __LITTLE_ENDIAN__
12132__ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1) {
12133  int32x2_t __ret;
12134  __ret = (int32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
12135  return __ret;
12136}
12137#else
12138__ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1) {
12139  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12140  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12141  int32x2_t __ret;
12142  __ret = (int32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
12143  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12144  return __ret;
12145}
12146#endif
12147
12148#ifdef __LITTLE_ENDIAN__
12149__ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1) {
12150  int16x4_t __ret;
12151  __ret = (int16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
12152  return __ret;
12153}
12154#else
12155__ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1) {
12156  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12157  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12158  int16x4_t __ret;
12159  __ret = (int16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
12160  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12161  return __ret;
12162}
12163#endif
12164
12165#ifdef __LITTLE_ENDIAN__
12166__ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12167  uint8x16_t __ret;
12168  __ret = __p0 + __p1 * __p2;
12169  return __ret;
12170}
12171#else
12172__ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12173  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12174  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12175  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12176  uint8x16_t __ret;
12177  __ret = __rev0 + __rev1 * __rev2;
12178  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12179  return __ret;
12180}
12181#endif
12182
12183#ifdef __LITTLE_ENDIAN__
12184__ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12185  uint32x4_t __ret;
12186  __ret = __p0 + __p1 * __p2;
12187  return __ret;
12188}
12189#else
12190__ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12191  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12192  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12193  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12194  uint32x4_t __ret;
12195  __ret = __rev0 + __rev1 * __rev2;
12196  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12197  return __ret;
12198}
12199#endif
12200
12201#ifdef __LITTLE_ENDIAN__
12202__ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12203  uint16x8_t __ret;
12204  __ret = __p0 + __p1 * __p2;
12205  return __ret;
12206}
12207#else
12208__ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12209  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12210  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12211  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12212  uint16x8_t __ret;
12213  __ret = __rev0 + __rev1 * __rev2;
12214  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12215  return __ret;
12216}
12217#endif
12218
12219#ifdef __LITTLE_ENDIAN__
12220__ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12221  int8x16_t __ret;
12222  __ret = __p0 + __p1 * __p2;
12223  return __ret;
12224}
12225#else
12226__ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12227  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12228  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12229  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12230  int8x16_t __ret;
12231  __ret = __rev0 + __rev1 * __rev2;
12232  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12233  return __ret;
12234}
12235#endif
12236
12237#ifdef __LITTLE_ENDIAN__
12238__ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12239  float32x4_t __ret;
12240  __ret = __p0 + __p1 * __p2;
12241  return __ret;
12242}
12243#else
12244__ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12245  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12246  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12247  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12248  float32x4_t __ret;
12249  __ret = __rev0 + __rev1 * __rev2;
12250  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12251  return __ret;
12252}
12253#endif
12254
12255#ifdef __LITTLE_ENDIAN__
12256__ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12257  int32x4_t __ret;
12258  __ret = __p0 + __p1 * __p2;
12259  return __ret;
12260}
12261#else
12262__ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12263  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12264  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12265  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12266  int32x4_t __ret;
12267  __ret = __rev0 + __rev1 * __rev2;
12268  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12269  return __ret;
12270}
12271#endif
12272
12273#ifdef __LITTLE_ENDIAN__
12274__ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12275  int16x8_t __ret;
12276  __ret = __p0 + __p1 * __p2;
12277  return __ret;
12278}
12279#else
12280__ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12281  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12282  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12283  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12284  int16x8_t __ret;
12285  __ret = __rev0 + __rev1 * __rev2;
12286  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12287  return __ret;
12288}
12289#endif
12290
12291#ifdef __LITTLE_ENDIAN__
12292__ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12293  uint8x8_t __ret;
12294  __ret = __p0 + __p1 * __p2;
12295  return __ret;
12296}
12297#else
12298__ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12299  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12300  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12301  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12302  uint8x8_t __ret;
12303  __ret = __rev0 + __rev1 * __rev2;
12304  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12305  return __ret;
12306}
12307#endif
12308
12309#ifdef __LITTLE_ENDIAN__
12310__ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12311  uint32x2_t __ret;
12312  __ret = __p0 + __p1 * __p2;
12313  return __ret;
12314}
12315#else
12316__ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12317  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12318  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12319  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12320  uint32x2_t __ret;
12321  __ret = __rev0 + __rev1 * __rev2;
12322  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12323  return __ret;
12324}
12325#endif
12326
12327#ifdef __LITTLE_ENDIAN__
12328__ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12329  uint16x4_t __ret;
12330  __ret = __p0 + __p1 * __p2;
12331  return __ret;
12332}
12333#else
12334__ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12335  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12336  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12337  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12338  uint16x4_t __ret;
12339  __ret = __rev0 + __rev1 * __rev2;
12340  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12341  return __ret;
12342}
12343#endif
12344
12345#ifdef __LITTLE_ENDIAN__
12346__ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
12347  int8x8_t __ret;
12348  __ret = __p0 + __p1 * __p2;
12349  return __ret;
12350}
12351#else
12352__ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
12353  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12354  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12355  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12356  int8x8_t __ret;
12357  __ret = __rev0 + __rev1 * __rev2;
12358  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12359  return __ret;
12360}
12361#endif
12362
12363#ifdef __LITTLE_ENDIAN__
12364__ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
12365  float32x2_t __ret;
12366  __ret = __p0 + __p1 * __p2;
12367  return __ret;
12368}
12369#else
12370__ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
12371  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12372  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12373  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12374  float32x2_t __ret;
12375  __ret = __rev0 + __rev1 * __rev2;
12376  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12377  return __ret;
12378}
12379#endif
12380
12381#ifdef __LITTLE_ENDIAN__
12382__ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
12383  int32x2_t __ret;
12384  __ret = __p0 + __p1 * __p2;
12385  return __ret;
12386}
12387#else
12388__ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
12389  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12390  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12391  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12392  int32x2_t __ret;
12393  __ret = __rev0 + __rev1 * __rev2;
12394  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12395  return __ret;
12396}
12397#endif
12398
12399#ifdef __LITTLE_ENDIAN__
12400__ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
12401  int16x4_t __ret;
12402  __ret = __p0 + __p1 * __p2;
12403  return __ret;
12404}
12405#else
12406__ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
12407  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12408  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12409  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12410  int16x4_t __ret;
12411  __ret = __rev0 + __rev1 * __rev2;
12412  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12413  return __ret;
12414}
12415#endif
12416
12417#ifdef __LITTLE_ENDIAN__
12418#define vmlaq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12419  uint32x4_t __s0 = __p0; \
12420  uint32x4_t __s1 = __p1; \
12421  uint32x2_t __s2 = __p2; \
12422  uint32x4_t __ret; \
12423  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12424  __ret; \
12425})
12426#else
12427#define vmlaq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12428  uint32x4_t __s0 = __p0; \
12429  uint32x4_t __s1 = __p1; \
12430  uint32x2_t __s2 = __p2; \
12431  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12432  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12433  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12434  uint32x4_t __ret; \
12435  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12436  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12437  __ret; \
12438})
12439#endif
12440
12441#ifdef __LITTLE_ENDIAN__
12442#define vmlaq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12443  uint16x8_t __s0 = __p0; \
12444  uint16x8_t __s1 = __p1; \
12445  uint16x4_t __s2 = __p2; \
12446  uint16x8_t __ret; \
12447  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12448  __ret; \
12449})
12450#else
12451#define vmlaq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12452  uint16x8_t __s0 = __p0; \
12453  uint16x8_t __s1 = __p1; \
12454  uint16x4_t __s2 = __p2; \
12455  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
12456  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
12457  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12458  uint16x8_t __ret; \
12459  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12460  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
12461  __ret; \
12462})
12463#endif
12464
12465#ifdef __LITTLE_ENDIAN__
12466#define vmlaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12467  float32x4_t __s0 = __p0; \
12468  float32x4_t __s1 = __p1; \
12469  float32x2_t __s2 = __p2; \
12470  float32x4_t __ret; \
12471  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12472  __ret; \
12473})
12474#else
12475#define vmlaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12476  float32x4_t __s0 = __p0; \
12477  float32x4_t __s1 = __p1; \
12478  float32x2_t __s2 = __p2; \
12479  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12480  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12481  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12482  float32x4_t __ret; \
12483  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12484  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12485  __ret; \
12486})
12487#endif
12488
12489#ifdef __LITTLE_ENDIAN__
12490#define vmlaq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12491  int32x4_t __s0 = __p0; \
12492  int32x4_t __s1 = __p1; \
12493  int32x2_t __s2 = __p2; \
12494  int32x4_t __ret; \
12495  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12496  __ret; \
12497})
12498#else
12499#define vmlaq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12500  int32x4_t __s0 = __p0; \
12501  int32x4_t __s1 = __p1; \
12502  int32x2_t __s2 = __p2; \
12503  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12504  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12505  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12506  int32x4_t __ret; \
12507  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12508  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12509  __ret; \
12510})
12511#endif
12512
12513#ifdef __LITTLE_ENDIAN__
12514#define vmlaq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12515  int16x8_t __s0 = __p0; \
12516  int16x8_t __s1 = __p1; \
12517  int16x4_t __s2 = __p2; \
12518  int16x8_t __ret; \
12519  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12520  __ret; \
12521})
12522#else
12523#define vmlaq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12524  int16x8_t __s0 = __p0; \
12525  int16x8_t __s1 = __p1; \
12526  int16x4_t __s2 = __p2; \
12527  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
12528  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
12529  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12530  int16x8_t __ret; \
12531  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12532  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
12533  __ret; \
12534})
12535#endif
12536
12537#ifdef __LITTLE_ENDIAN__
12538#define vmla_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12539  uint32x2_t __s0 = __p0; \
12540  uint32x2_t __s1 = __p1; \
12541  uint32x2_t __s2 = __p2; \
12542  uint32x2_t __ret; \
12543  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
12544  __ret; \
12545})
12546#else
12547#define vmla_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12548  uint32x2_t __s0 = __p0; \
12549  uint32x2_t __s1 = __p1; \
12550  uint32x2_t __s2 = __p2; \
12551  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
12552  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
12553  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12554  uint32x2_t __ret; \
12555  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
12556  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
12557  __ret; \
12558})
12559#endif
12560
12561#ifdef __LITTLE_ENDIAN__
12562#define vmla_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12563  uint16x4_t __s0 = __p0; \
12564  uint16x4_t __s1 = __p1; \
12565  uint16x4_t __s2 = __p2; \
12566  uint16x4_t __ret; \
12567  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12568  __ret; \
12569})
12570#else
12571#define vmla_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12572  uint16x4_t __s0 = __p0; \
12573  uint16x4_t __s1 = __p1; \
12574  uint16x4_t __s2 = __p2; \
12575  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12576  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12577  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12578  uint16x4_t __ret; \
12579  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12580  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12581  __ret; \
12582})
12583#endif
12584
12585#ifdef __LITTLE_ENDIAN__
12586#define vmla_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12587  float32x2_t __s0 = __p0; \
12588  float32x2_t __s1 = __p1; \
12589  float32x2_t __s2 = __p2; \
12590  float32x2_t __ret; \
12591  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
12592  __ret; \
12593})
12594#else
12595#define vmla_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12596  float32x2_t __s0 = __p0; \
12597  float32x2_t __s1 = __p1; \
12598  float32x2_t __s2 = __p2; \
12599  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
12600  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
12601  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12602  float32x2_t __ret; \
12603  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
12604  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
12605  __ret; \
12606})
12607#endif
12608
12609#ifdef __LITTLE_ENDIAN__
12610#define vmla_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12611  int32x2_t __s0 = __p0; \
12612  int32x2_t __s1 = __p1; \
12613  int32x2_t __s2 = __p2; \
12614  int32x2_t __ret; \
12615  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
12616  __ret; \
12617})
12618#else
12619#define vmla_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12620  int32x2_t __s0 = __p0; \
12621  int32x2_t __s1 = __p1; \
12622  int32x2_t __s2 = __p2; \
12623  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
12624  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
12625  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12626  int32x2_t __ret; \
12627  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
12628  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
12629  __ret; \
12630})
12631#endif
12632
12633#ifdef __LITTLE_ENDIAN__
12634#define vmla_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12635  int16x4_t __s0 = __p0; \
12636  int16x4_t __s1 = __p1; \
12637  int16x4_t __s2 = __p2; \
12638  int16x4_t __ret; \
12639  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12640  __ret; \
12641})
12642#else
12643#define vmla_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12644  int16x4_t __s0 = __p0; \
12645  int16x4_t __s1 = __p1; \
12646  int16x4_t __s2 = __p2; \
12647  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12648  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12649  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12650  int16x4_t __ret; \
12651  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12652  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12653  __ret; \
12654})
12655#endif
12656
12657#ifdef __LITTLE_ENDIAN__
12658__ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
12659  uint32x4_t __ret;
12660  __ret = __p0 + __p1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
12661  return __ret;
12662}
12663#else
12664__ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
12665  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12666  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12667  uint32x4_t __ret;
12668  __ret = __rev0 + __rev1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
12669  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12670  return __ret;
12671}
12672#endif
12673
12674#ifdef __LITTLE_ENDIAN__
12675__ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
12676  uint16x8_t __ret;
12677  __ret = __p0 + __p1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12678  return __ret;
12679}
12680#else
12681__ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
12682  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12683  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12684  uint16x8_t __ret;
12685  __ret = __rev0 + __rev1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12686  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12687  return __ret;
12688}
12689#endif
12690
12691#ifdef __LITTLE_ENDIAN__
12692__ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
12693  float32x4_t __ret;
12694  __ret = __p0 + __p1 * (float32x4_t) {__p2, __p2, __p2, __p2};
12695  return __ret;
12696}
12697#else
12698__ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
12699  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12700  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12701  float32x4_t __ret;
12702  __ret = __rev0 + __rev1 * (float32x4_t) {__p2, __p2, __p2, __p2};
12703  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12704  return __ret;
12705}
12706#endif
12707
12708#ifdef __LITTLE_ENDIAN__
12709__ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
12710  int32x4_t __ret;
12711  __ret = __p0 + __p1 * (int32x4_t) {__p2, __p2, __p2, __p2};
12712  return __ret;
12713}
12714#else
12715__ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
12716  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12717  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12718  int32x4_t __ret;
12719  __ret = __rev0 + __rev1 * (int32x4_t) {__p2, __p2, __p2, __p2};
12720  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12721  return __ret;
12722}
12723#endif
12724
12725#ifdef __LITTLE_ENDIAN__
12726__ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
12727  int16x8_t __ret;
12728  __ret = __p0 + __p1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12729  return __ret;
12730}
12731#else
12732__ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
12733  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12734  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12735  int16x8_t __ret;
12736  __ret = __rev0 + __rev1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12737  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12738  return __ret;
12739}
12740#endif
12741
12742#ifdef __LITTLE_ENDIAN__
12743__ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
12744  uint32x2_t __ret;
12745  __ret = __p0 + __p1 * (uint32x2_t) {__p2, __p2};
12746  return __ret;
12747}
12748#else
12749__ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
12750  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12751  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12752  uint32x2_t __ret;
12753  __ret = __rev0 + __rev1 * (uint32x2_t) {__p2, __p2};
12754  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12755  return __ret;
12756}
12757#endif
12758
12759#ifdef __LITTLE_ENDIAN__
12760__ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
12761  uint16x4_t __ret;
12762  __ret = __p0 + __p1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
12763  return __ret;
12764}
12765#else
12766__ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
12767  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12768  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12769  uint16x4_t __ret;
12770  __ret = __rev0 + __rev1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
12771  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12772  return __ret;
12773}
12774#endif
12775
12776#ifdef __LITTLE_ENDIAN__
12777__ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
12778  float32x2_t __ret;
12779  __ret = __p0 + __p1 * (float32x2_t) {__p2, __p2};
12780  return __ret;
12781}
12782#else
12783__ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
12784  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12785  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12786  float32x2_t __ret;
12787  __ret = __rev0 + __rev1 * (float32x2_t) {__p2, __p2};
12788  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12789  return __ret;
12790}
12791#endif
12792
12793#ifdef __LITTLE_ENDIAN__
12794__ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
12795  int32x2_t __ret;
12796  __ret = __p0 + __p1 * (int32x2_t) {__p2, __p2};
12797  return __ret;
12798}
12799#else
12800__ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
12801  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12802  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12803  int32x2_t __ret;
12804  __ret = __rev0 + __rev1 * (int32x2_t) {__p2, __p2};
12805  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12806  return __ret;
12807}
12808#endif
12809
12810#ifdef __LITTLE_ENDIAN__
12811__ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
12812  int16x4_t __ret;
12813  __ret = __p0 + __p1 * (int16x4_t) {__p2, __p2, __p2, __p2};
12814  return __ret;
12815}
12816#else
12817__ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
12818  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12819  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12820  int16x4_t __ret;
12821  __ret = __rev0 + __rev1 * (int16x4_t) {__p2, __p2, __p2, __p2};
12822  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12823  return __ret;
12824}
12825#endif
12826
12827#ifdef __LITTLE_ENDIAN__
12828__ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12829  uint8x16_t __ret;
12830  __ret = __p0 - __p1 * __p2;
12831  return __ret;
12832}
12833#else
12834__ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12835  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12836  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12837  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12838  uint8x16_t __ret;
12839  __ret = __rev0 - __rev1 * __rev2;
12840  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12841  return __ret;
12842}
12843#endif
12844
12845#ifdef __LITTLE_ENDIAN__
12846__ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12847  uint32x4_t __ret;
12848  __ret = __p0 - __p1 * __p2;
12849  return __ret;
12850}
12851#else
12852__ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12853  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12854  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12855  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12856  uint32x4_t __ret;
12857  __ret = __rev0 - __rev1 * __rev2;
12858  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12859  return __ret;
12860}
12861#endif
12862
12863#ifdef __LITTLE_ENDIAN__
12864__ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12865  uint16x8_t __ret;
12866  __ret = __p0 - __p1 * __p2;
12867  return __ret;
12868}
12869#else
12870__ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12871  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12872  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12873  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12874  uint16x8_t __ret;
12875  __ret = __rev0 - __rev1 * __rev2;
12876  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12877  return __ret;
12878}
12879#endif
12880
12881#ifdef __LITTLE_ENDIAN__
12882__ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12883  int8x16_t __ret;
12884  __ret = __p0 - __p1 * __p2;
12885  return __ret;
12886}
12887#else
12888__ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12889  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12890  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12891  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12892  int8x16_t __ret;
12893  __ret = __rev0 - __rev1 * __rev2;
12894  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12895  return __ret;
12896}
12897#endif
12898
12899#ifdef __LITTLE_ENDIAN__
12900__ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12901  float32x4_t __ret;
12902  __ret = __p0 - __p1 * __p2;
12903  return __ret;
12904}
12905#else
12906__ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12907  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12908  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12909  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12910  float32x4_t __ret;
12911  __ret = __rev0 - __rev1 * __rev2;
12912  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12913  return __ret;
12914}
12915#endif
12916
12917#ifdef __LITTLE_ENDIAN__
12918__ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12919  int32x4_t __ret;
12920  __ret = __p0 - __p1 * __p2;
12921  return __ret;
12922}
12923#else
12924__ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12925  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12926  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12927  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12928  int32x4_t __ret;
12929  __ret = __rev0 - __rev1 * __rev2;
12930  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12931  return __ret;
12932}
12933#endif
12934
12935#ifdef __LITTLE_ENDIAN__
12936__ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12937  int16x8_t __ret;
12938  __ret = __p0 - __p1 * __p2;
12939  return __ret;
12940}
12941#else
12942__ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12943  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12944  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12945  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12946  int16x8_t __ret;
12947  __ret = __rev0 - __rev1 * __rev2;
12948  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12949  return __ret;
12950}
12951#endif
12952
12953#ifdef __LITTLE_ENDIAN__
12954__ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12955  uint8x8_t __ret;
12956  __ret = __p0 - __p1 * __p2;
12957  return __ret;
12958}
12959#else
12960__ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12961  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12962  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12963  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12964  uint8x8_t __ret;
12965  __ret = __rev0 - __rev1 * __rev2;
12966  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12967  return __ret;
12968}
12969#endif
12970
12971#ifdef __LITTLE_ENDIAN__
12972__ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12973  uint32x2_t __ret;
12974  __ret = __p0 - __p1 * __p2;
12975  return __ret;
12976}
12977#else
12978__ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12979  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12980  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12981  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12982  uint32x2_t __ret;
12983  __ret = __rev0 - __rev1 * __rev2;
12984  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12985  return __ret;
12986}
12987#endif
12988
12989#ifdef __LITTLE_ENDIAN__
12990__ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12991  uint16x4_t __ret;
12992  __ret = __p0 - __p1 * __p2;
12993  return __ret;
12994}
12995#else
12996__ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12997  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12998  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12999  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
13000  uint16x4_t __ret;
13001  __ret = __rev0 - __rev1 * __rev2;
13002  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13003  return __ret;
13004}
13005#endif
13006
13007#ifdef __LITTLE_ENDIAN__
13008__ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
13009  int8x8_t __ret;
13010  __ret = __p0 - __p1 * __p2;
13011  return __ret;
13012}
13013#else
13014__ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
13015  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13016  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13017  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
13018  int8x8_t __ret;
13019  __ret = __rev0 - __rev1 * __rev2;
13020  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13021  return __ret;
13022}
13023#endif
13024
13025#ifdef __LITTLE_ENDIAN__
13026__ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
13027  float32x2_t __ret;
13028  __ret = __p0 - __p1 * __p2;
13029  return __ret;
13030}
13031#else
13032__ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
13033  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13034  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13035  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
13036  float32x2_t __ret;
13037  __ret = __rev0 - __rev1 * __rev2;
13038  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13039  return __ret;
13040}
13041#endif
13042
13043#ifdef __LITTLE_ENDIAN__
13044__ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
13045  int32x2_t __ret;
13046  __ret = __p0 - __p1 * __p2;
13047  return __ret;
13048}
13049#else
13050__ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
13051  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13052  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13053  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
13054  int32x2_t __ret;
13055  __ret = __rev0 - __rev1 * __rev2;
13056  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13057  return __ret;
13058}
13059#endif
13060
13061#ifdef __LITTLE_ENDIAN__
13062__ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
13063  int16x4_t __ret;
13064  __ret = __p0 - __p1 * __p2;
13065  return __ret;
13066}
13067#else
13068__ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
13069  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13070  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13071  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
13072  int16x4_t __ret;
13073  __ret = __rev0 - __rev1 * __rev2;
13074  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13075  return __ret;
13076}
13077#endif
13078
13079#ifdef __LITTLE_ENDIAN__
13080#define vmlsq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13081  uint32x4_t __s0 = __p0; \
13082  uint32x4_t __s1 = __p1; \
13083  uint32x2_t __s2 = __p2; \
13084  uint32x4_t __ret; \
13085  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13086  __ret; \
13087})
13088#else
13089#define vmlsq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13090  uint32x4_t __s0 = __p0; \
13091  uint32x4_t __s1 = __p1; \
13092  uint32x2_t __s2 = __p2; \
13093  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13094  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13095  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13096  uint32x4_t __ret; \
13097  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13098  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13099  __ret; \
13100})
13101#endif
13102
13103#ifdef __LITTLE_ENDIAN__
13104#define vmlsq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13105  uint16x8_t __s0 = __p0; \
13106  uint16x8_t __s1 = __p1; \
13107  uint16x4_t __s2 = __p2; \
13108  uint16x8_t __ret; \
13109  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13110  __ret; \
13111})
13112#else
13113#define vmlsq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13114  uint16x8_t __s0 = __p0; \
13115  uint16x8_t __s1 = __p1; \
13116  uint16x4_t __s2 = __p2; \
13117  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
13118  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
13119  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13120  uint16x8_t __ret; \
13121  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13122  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
13123  __ret; \
13124})
13125#endif
13126
13127#ifdef __LITTLE_ENDIAN__
13128#define vmlsq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13129  float32x4_t __s0 = __p0; \
13130  float32x4_t __s1 = __p1; \
13131  float32x2_t __s2 = __p2; \
13132  float32x4_t __ret; \
13133  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13134  __ret; \
13135})
13136#else
13137#define vmlsq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13138  float32x4_t __s0 = __p0; \
13139  float32x4_t __s1 = __p1; \
13140  float32x2_t __s2 = __p2; \
13141  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13142  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13143  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13144  float32x4_t __ret; \
13145  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13146  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13147  __ret; \
13148})
13149#endif
13150
13151#ifdef __LITTLE_ENDIAN__
13152#define vmlsq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13153  int32x4_t __s0 = __p0; \
13154  int32x4_t __s1 = __p1; \
13155  int32x2_t __s2 = __p2; \
13156  int32x4_t __ret; \
13157  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13158  __ret; \
13159})
13160#else
13161#define vmlsq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13162  int32x4_t __s0 = __p0; \
13163  int32x4_t __s1 = __p1; \
13164  int32x2_t __s2 = __p2; \
13165  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13166  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13167  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13168  int32x4_t __ret; \
13169  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13170  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13171  __ret; \
13172})
13173#endif
13174
13175#ifdef __LITTLE_ENDIAN__
13176#define vmlsq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13177  int16x8_t __s0 = __p0; \
13178  int16x8_t __s1 = __p1; \
13179  int16x4_t __s2 = __p2; \
13180  int16x8_t __ret; \
13181  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13182  __ret; \
13183})
13184#else
13185#define vmlsq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13186  int16x8_t __s0 = __p0; \
13187  int16x8_t __s1 = __p1; \
13188  int16x4_t __s2 = __p2; \
13189  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
13190  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
13191  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13192  int16x8_t __ret; \
13193  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13194  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
13195  __ret; \
13196})
13197#endif
13198
13199#ifdef __LITTLE_ENDIAN__
13200#define vmls_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13201  uint32x2_t __s0 = __p0; \
13202  uint32x2_t __s1 = __p1; \
13203  uint32x2_t __s2 = __p2; \
13204  uint32x2_t __ret; \
13205  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
13206  __ret; \
13207})
13208#else
13209#define vmls_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13210  uint32x2_t __s0 = __p0; \
13211  uint32x2_t __s1 = __p1; \
13212  uint32x2_t __s2 = __p2; \
13213  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
13214  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
13215  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13216  uint32x2_t __ret; \
13217  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
13218  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
13219  __ret; \
13220})
13221#endif
13222
13223#ifdef __LITTLE_ENDIAN__
13224#define vmls_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13225  uint16x4_t __s0 = __p0; \
13226  uint16x4_t __s1 = __p1; \
13227  uint16x4_t __s2 = __p2; \
13228  uint16x4_t __ret; \
13229  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13230  __ret; \
13231})
13232#else
13233#define vmls_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13234  uint16x4_t __s0 = __p0; \
13235  uint16x4_t __s1 = __p1; \
13236  uint16x4_t __s2 = __p2; \
13237  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13238  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13239  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13240  uint16x4_t __ret; \
13241  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13242  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13243  __ret; \
13244})
13245#endif
13246
13247#ifdef __LITTLE_ENDIAN__
13248#define vmls_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13249  float32x2_t __s0 = __p0; \
13250  float32x2_t __s1 = __p1; \
13251  float32x2_t __s2 = __p2; \
13252  float32x2_t __ret; \
13253  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
13254  __ret; \
13255})
13256#else
13257#define vmls_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13258  float32x2_t __s0 = __p0; \
13259  float32x2_t __s1 = __p1; \
13260  float32x2_t __s2 = __p2; \
13261  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
13262  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
13263  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13264  float32x2_t __ret; \
13265  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
13266  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
13267  __ret; \
13268})
13269#endif
13270
13271#ifdef __LITTLE_ENDIAN__
13272#define vmls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13273  int32x2_t __s0 = __p0; \
13274  int32x2_t __s1 = __p1; \
13275  int32x2_t __s2 = __p2; \
13276  int32x2_t __ret; \
13277  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
13278  __ret; \
13279})
13280#else
13281#define vmls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13282  int32x2_t __s0 = __p0; \
13283  int32x2_t __s1 = __p1; \
13284  int32x2_t __s2 = __p2; \
13285  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
13286  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
13287  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13288  int32x2_t __ret; \
13289  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
13290  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
13291  __ret; \
13292})
13293#endif
13294
13295#ifdef __LITTLE_ENDIAN__
13296#define vmls_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13297  int16x4_t __s0 = __p0; \
13298  int16x4_t __s1 = __p1; \
13299  int16x4_t __s2 = __p2; \
13300  int16x4_t __ret; \
13301  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13302  __ret; \
13303})
13304#else
13305#define vmls_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13306  int16x4_t __s0 = __p0; \
13307  int16x4_t __s1 = __p1; \
13308  int16x4_t __s2 = __p2; \
13309  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13310  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13311  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13312  int16x4_t __ret; \
13313  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13314  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13315  __ret; \
13316})
13317#endif
13318
13319#ifdef __LITTLE_ENDIAN__
13320__ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
13321  uint32x4_t __ret;
13322  __ret = __p0 - __p1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
13323  return __ret;
13324}
13325#else
13326__ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
13327  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13328  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13329  uint32x4_t __ret;
13330  __ret = __rev0 - __rev1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
13331  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13332  return __ret;
13333}
13334#endif
13335
13336#ifdef __LITTLE_ENDIAN__
13337__ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
13338  uint16x8_t __ret;
13339  __ret = __p0 - __p1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13340  return __ret;
13341}
13342#else
13343__ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
13344  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13345  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13346  uint16x8_t __ret;
13347  __ret = __rev0 - __rev1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13348  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13349  return __ret;
13350}
13351#endif
13352
13353#ifdef __LITTLE_ENDIAN__
13354__ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
13355  float32x4_t __ret;
13356  __ret = __p0 - __p1 * (float32x4_t) {__p2, __p2, __p2, __p2};
13357  return __ret;
13358}
13359#else
13360__ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
13361  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13362  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13363  float32x4_t __ret;
13364  __ret = __rev0 - __rev1 * (float32x4_t) {__p2, __p2, __p2, __p2};
13365  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13366  return __ret;
13367}
13368#endif
13369
13370#ifdef __LITTLE_ENDIAN__
13371__ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
13372  int32x4_t __ret;
13373  __ret = __p0 - __p1 * (int32x4_t) {__p2, __p2, __p2, __p2};
13374  return __ret;
13375}
13376#else
13377__ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
13378  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13379  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13380  int32x4_t __ret;
13381  __ret = __rev0 - __rev1 * (int32x4_t) {__p2, __p2, __p2, __p2};
13382  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13383  return __ret;
13384}
13385#endif
13386
13387#ifdef __LITTLE_ENDIAN__
13388__ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
13389  int16x8_t __ret;
13390  __ret = __p0 - __p1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13391  return __ret;
13392}
13393#else
13394__ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
13395  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13396  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13397  int16x8_t __ret;
13398  __ret = __rev0 - __rev1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13399  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13400  return __ret;
13401}
13402#endif
13403
13404#ifdef __LITTLE_ENDIAN__
13405__ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
13406  uint32x2_t __ret;
13407  __ret = __p0 - __p1 * (uint32x2_t) {__p2, __p2};
13408  return __ret;
13409}
13410#else
13411__ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
13412  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13413  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13414  uint32x2_t __ret;
13415  __ret = __rev0 - __rev1 * (uint32x2_t) {__p2, __p2};
13416  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13417  return __ret;
13418}
13419#endif
13420
13421#ifdef __LITTLE_ENDIAN__
13422__ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
13423  uint16x4_t __ret;
13424  __ret = __p0 - __p1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
13425  return __ret;
13426}
13427#else
13428__ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
13429  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13430  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13431  uint16x4_t __ret;
13432  __ret = __rev0 - __rev1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
13433  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13434  return __ret;
13435}
13436#endif
13437
13438#ifdef __LITTLE_ENDIAN__
13439__ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
13440  float32x2_t __ret;
13441  __ret = __p0 - __p1 * (float32x2_t) {__p2, __p2};
13442  return __ret;
13443}
13444#else
13445__ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
13446  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13447  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13448  float32x2_t __ret;
13449  __ret = __rev0 - __rev1 * (float32x2_t) {__p2, __p2};
13450  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13451  return __ret;
13452}
13453#endif
13454
13455#ifdef __LITTLE_ENDIAN__
13456__ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
13457  int32x2_t __ret;
13458  __ret = __p0 - __p1 * (int32x2_t) {__p2, __p2};
13459  return __ret;
13460}
13461#else
13462__ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
13463  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13464  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13465  int32x2_t __ret;
13466  __ret = __rev0 - __rev1 * (int32x2_t) {__p2, __p2};
13467  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13468  return __ret;
13469}
13470#endif
13471
13472#ifdef __LITTLE_ENDIAN__
13473__ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
13474  int16x4_t __ret;
13475  __ret = __p0 - __p1 * (int16x4_t) {__p2, __p2, __p2, __p2};
13476  return __ret;
13477}
13478#else
13479__ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
13480  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13481  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13482  int16x4_t __ret;
13483  __ret = __rev0 - __rev1 * (int16x4_t) {__p2, __p2, __p2, __p2};
13484  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13485  return __ret;
13486}
13487#endif
13488
13489#ifdef __LITTLE_ENDIAN__
13490__ai poly8x8_t vmov_n_p8(poly8_t __p0) {
13491  poly8x8_t __ret;
13492  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13493  return __ret;
13494}
13495#else
13496__ai poly8x8_t vmov_n_p8(poly8_t __p0) {
13497  poly8x8_t __ret;
13498  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13499  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13500  return __ret;
13501}
13502#endif
13503
13504#ifdef __LITTLE_ENDIAN__
13505__ai poly16x4_t vmov_n_p16(poly16_t __p0) {
13506  poly16x4_t __ret;
13507  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
13508  return __ret;
13509}
13510#else
13511__ai poly16x4_t vmov_n_p16(poly16_t __p0) {
13512  poly16x4_t __ret;
13513  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
13514  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13515  return __ret;
13516}
13517#endif
13518
13519#ifdef __LITTLE_ENDIAN__
13520__ai poly8x16_t vmovq_n_p8(poly8_t __p0) {
13521  poly8x16_t __ret;
13522  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13523  return __ret;
13524}
13525#else
13526__ai poly8x16_t vmovq_n_p8(poly8_t __p0) {
13527  poly8x16_t __ret;
13528  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13529  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13530  return __ret;
13531}
13532#endif
13533
13534#ifdef __LITTLE_ENDIAN__
13535__ai poly16x8_t vmovq_n_p16(poly16_t __p0) {
13536  poly16x8_t __ret;
13537  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13538  return __ret;
13539}
13540#else
13541__ai poly16x8_t vmovq_n_p16(poly16_t __p0) {
13542  poly16x8_t __ret;
13543  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13544  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13545  return __ret;
13546}
13547#endif
13548
13549#ifdef __LITTLE_ENDIAN__
13550__ai uint8x16_t vmovq_n_u8(uint8_t __p0) {
13551  uint8x16_t __ret;
13552  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13553  return __ret;
13554}
13555#else
13556__ai uint8x16_t vmovq_n_u8(uint8_t __p0) {
13557  uint8x16_t __ret;
13558  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13559  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13560  return __ret;
13561}
13562#endif
13563
13564#ifdef __LITTLE_ENDIAN__
13565__ai uint32x4_t vmovq_n_u32(uint32_t __p0) {
13566  uint32x4_t __ret;
13567  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
13568  return __ret;
13569}
13570#else
13571__ai uint32x4_t vmovq_n_u32(uint32_t __p0) {
13572  uint32x4_t __ret;
13573  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
13574  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13575  return __ret;
13576}
13577#endif
13578
13579#ifdef __LITTLE_ENDIAN__
13580__ai uint64x2_t vmovq_n_u64(uint64_t __p0) {
13581  uint64x2_t __ret;
13582  __ret = (uint64x2_t) {__p0, __p0};
13583  return __ret;
13584}
13585#else
13586__ai uint64x2_t vmovq_n_u64(uint64_t __p0) {
13587  uint64x2_t __ret;
13588  __ret = (uint64x2_t) {__p0, __p0};
13589  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13590  return __ret;
13591}
13592#endif
13593
13594#ifdef __LITTLE_ENDIAN__
13595__ai uint16x8_t vmovq_n_u16(uint16_t __p0) {
13596  uint16x8_t __ret;
13597  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13598  return __ret;
13599}
13600#else
13601__ai uint16x8_t vmovq_n_u16(uint16_t __p0) {
13602  uint16x8_t __ret;
13603  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13604  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13605  return __ret;
13606}
13607#endif
13608
13609#ifdef __LITTLE_ENDIAN__
13610__ai int8x16_t vmovq_n_s8(int8_t __p0) {
13611  int8x16_t __ret;
13612  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13613  return __ret;
13614}
13615#else
13616__ai int8x16_t vmovq_n_s8(int8_t __p0) {
13617  int8x16_t __ret;
13618  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13619  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13620  return __ret;
13621}
13622#endif
13623
13624#ifdef __LITTLE_ENDIAN__
13625__ai float32x4_t vmovq_n_f32(float32_t __p0) {
13626  float32x4_t __ret;
13627  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
13628  return __ret;
13629}
13630#else
13631__ai float32x4_t vmovq_n_f32(float32_t __p0) {
13632  float32x4_t __ret;
13633  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
13634  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13635  return __ret;
13636}
13637#endif
13638
13639#ifdef __LITTLE_ENDIAN__
13640#define vmovq_n_f16(__p0) __extension__ ({ \
13641  float16_t __s0 = __p0; \
13642  float16x8_t __ret; \
13643  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
13644  __ret; \
13645})
13646#else
13647#define vmovq_n_f16(__p0) __extension__ ({ \
13648  float16_t __s0 = __p0; \
13649  float16x8_t __ret; \
13650  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
13651  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
13652  __ret; \
13653})
13654#endif
13655
13656#ifdef __LITTLE_ENDIAN__
13657__ai int32x4_t vmovq_n_s32(int32_t __p0) {
13658  int32x4_t __ret;
13659  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
13660  return __ret;
13661}
13662#else
13663__ai int32x4_t vmovq_n_s32(int32_t __p0) {
13664  int32x4_t __ret;
13665  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
13666  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13667  return __ret;
13668}
13669#endif
13670
13671#ifdef __LITTLE_ENDIAN__
13672__ai int64x2_t vmovq_n_s64(int64_t __p0) {
13673  int64x2_t __ret;
13674  __ret = (int64x2_t) {__p0, __p0};
13675  return __ret;
13676}
13677#else
13678__ai int64x2_t vmovq_n_s64(int64_t __p0) {
13679  int64x2_t __ret;
13680  __ret = (int64x2_t) {__p0, __p0};
13681  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13682  return __ret;
13683}
13684#endif
13685
13686#ifdef __LITTLE_ENDIAN__
13687__ai int16x8_t vmovq_n_s16(int16_t __p0) {
13688  int16x8_t __ret;
13689  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13690  return __ret;
13691}
13692#else
13693__ai int16x8_t vmovq_n_s16(int16_t __p0) {
13694  int16x8_t __ret;
13695  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13696  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13697  return __ret;
13698}
13699#endif
13700
13701#ifdef __LITTLE_ENDIAN__
13702__ai uint8x8_t vmov_n_u8(uint8_t __p0) {
13703  uint8x8_t __ret;
13704  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13705  return __ret;
13706}
13707#else
13708__ai uint8x8_t vmov_n_u8(uint8_t __p0) {
13709  uint8x8_t __ret;
13710  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13711  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13712  return __ret;
13713}
13714#endif
13715
13716#ifdef __LITTLE_ENDIAN__
13717__ai uint32x2_t vmov_n_u32(uint32_t __p0) {
13718  uint32x2_t __ret;
13719  __ret = (uint32x2_t) {__p0, __p0};
13720  return __ret;
13721}
13722#else
13723__ai uint32x2_t vmov_n_u32(uint32_t __p0) {
13724  uint32x2_t __ret;
13725  __ret = (uint32x2_t) {__p0, __p0};
13726  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13727  return __ret;
13728}
13729#endif
13730
13731#ifdef __LITTLE_ENDIAN__
13732__ai uint64x1_t vmov_n_u64(uint64_t __p0) {
13733  uint64x1_t __ret;
13734  __ret = (uint64x1_t) {__p0};
13735  return __ret;
13736}
13737#else
13738__ai uint64x1_t vmov_n_u64(uint64_t __p0) {
13739  uint64x1_t __ret;
13740  __ret = (uint64x1_t) {__p0};
13741  return __ret;
13742}
13743#endif
13744
13745#ifdef __LITTLE_ENDIAN__
13746__ai uint16x4_t vmov_n_u16(uint16_t __p0) {
13747  uint16x4_t __ret;
13748  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
13749  return __ret;
13750}
13751#else
13752__ai uint16x4_t vmov_n_u16(uint16_t __p0) {
13753  uint16x4_t __ret;
13754  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
13755  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13756  return __ret;
13757}
13758#endif
13759
13760#ifdef __LITTLE_ENDIAN__
13761__ai int8x8_t vmov_n_s8(int8_t __p0) {
13762  int8x8_t __ret;
13763  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13764  return __ret;
13765}
13766#else
13767__ai int8x8_t vmov_n_s8(int8_t __p0) {
13768  int8x8_t __ret;
13769  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13770  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13771  return __ret;
13772}
13773#endif
13774
13775#ifdef __LITTLE_ENDIAN__
13776__ai float32x2_t vmov_n_f32(float32_t __p0) {
13777  float32x2_t __ret;
13778  __ret = (float32x2_t) {__p0, __p0};
13779  return __ret;
13780}
13781#else
13782__ai float32x2_t vmov_n_f32(float32_t __p0) {
13783  float32x2_t __ret;
13784  __ret = (float32x2_t) {__p0, __p0};
13785  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13786  return __ret;
13787}
13788#endif
13789
13790#ifdef __LITTLE_ENDIAN__
13791#define vmov_n_f16(__p0) __extension__ ({ \
13792  float16_t __s0 = __p0; \
13793  float16x4_t __ret; \
13794  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
13795  __ret; \
13796})
13797#else
13798#define vmov_n_f16(__p0) __extension__ ({ \
13799  float16_t __s0 = __p0; \
13800  float16x4_t __ret; \
13801  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
13802  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13803  __ret; \
13804})
13805#endif
13806
13807#ifdef __LITTLE_ENDIAN__
13808__ai int32x2_t vmov_n_s32(int32_t __p0) {
13809  int32x2_t __ret;
13810  __ret = (int32x2_t) {__p0, __p0};
13811  return __ret;
13812}
13813#else
13814__ai int32x2_t vmov_n_s32(int32_t __p0) {
13815  int32x2_t __ret;
13816  __ret = (int32x2_t) {__p0, __p0};
13817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13818  return __ret;
13819}
13820#endif
13821
13822#ifdef __LITTLE_ENDIAN__
13823__ai int64x1_t vmov_n_s64(int64_t __p0) {
13824  int64x1_t __ret;
13825  __ret = (int64x1_t) {__p0};
13826  return __ret;
13827}
13828#else
13829__ai int64x1_t vmov_n_s64(int64_t __p0) {
13830  int64x1_t __ret;
13831  __ret = (int64x1_t) {__p0};
13832  return __ret;
13833}
13834#endif
13835
13836#ifdef __LITTLE_ENDIAN__
13837__ai int16x4_t vmov_n_s16(int16_t __p0) {
13838  int16x4_t __ret;
13839  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
13840  return __ret;
13841}
13842#else
13843__ai int16x4_t vmov_n_s16(int16_t __p0) {
13844  int16x4_t __ret;
13845  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
13846  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13847  return __ret;
13848}
13849#endif
13850
13851#ifdef __LITTLE_ENDIAN__
13852__ai uint16x8_t vmovl_u8(uint8x8_t __p0) {
13853  uint16x8_t __ret;
13854  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 49);
13855  return __ret;
13856}
13857#else
13858__ai uint16x8_t vmovl_u8(uint8x8_t __p0) {
13859  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13860  uint16x8_t __ret;
13861  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 49);
13862  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13863  return __ret;
13864}
13865__ai uint16x8_t __noswap_vmovl_u8(uint8x8_t __p0) {
13866  uint16x8_t __ret;
13867  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 49);
13868  return __ret;
13869}
13870#endif
13871
13872#ifdef __LITTLE_ENDIAN__
13873__ai uint64x2_t vmovl_u32(uint32x2_t __p0) {
13874  uint64x2_t __ret;
13875  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 51);
13876  return __ret;
13877}
13878#else
13879__ai uint64x2_t vmovl_u32(uint32x2_t __p0) {
13880  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13881  uint64x2_t __ret;
13882  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 51);
13883  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13884  return __ret;
13885}
13886__ai uint64x2_t __noswap_vmovl_u32(uint32x2_t __p0) {
13887  uint64x2_t __ret;
13888  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 51);
13889  return __ret;
13890}
13891#endif
13892
13893#ifdef __LITTLE_ENDIAN__
13894__ai uint32x4_t vmovl_u16(uint16x4_t __p0) {
13895  uint32x4_t __ret;
13896  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 50);
13897  return __ret;
13898}
13899#else
13900__ai uint32x4_t vmovl_u16(uint16x4_t __p0) {
13901  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13902  uint32x4_t __ret;
13903  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 50);
13904  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13905  return __ret;
13906}
13907__ai uint32x4_t __noswap_vmovl_u16(uint16x4_t __p0) {
13908  uint32x4_t __ret;
13909  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 50);
13910  return __ret;
13911}
13912#endif
13913
13914#ifdef __LITTLE_ENDIAN__
13915__ai int16x8_t vmovl_s8(int8x8_t __p0) {
13916  int16x8_t __ret;
13917  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 33);
13918  return __ret;
13919}
13920#else
13921__ai int16x8_t vmovl_s8(int8x8_t __p0) {
13922  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13923  int16x8_t __ret;
13924  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 33);
13925  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13926  return __ret;
13927}
13928__ai int16x8_t __noswap_vmovl_s8(int8x8_t __p0) {
13929  int16x8_t __ret;
13930  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 33);
13931  return __ret;
13932}
13933#endif
13934
13935#ifdef __LITTLE_ENDIAN__
13936__ai int64x2_t vmovl_s32(int32x2_t __p0) {
13937  int64x2_t __ret;
13938  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 35);
13939  return __ret;
13940}
13941#else
13942__ai int64x2_t vmovl_s32(int32x2_t __p0) {
13943  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13944  int64x2_t __ret;
13945  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 35);
13946  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13947  return __ret;
13948}
13949__ai int64x2_t __noswap_vmovl_s32(int32x2_t __p0) {
13950  int64x2_t __ret;
13951  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 35);
13952  return __ret;
13953}
13954#endif
13955
13956#ifdef __LITTLE_ENDIAN__
13957__ai int32x4_t vmovl_s16(int16x4_t __p0) {
13958  int32x4_t __ret;
13959  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 34);
13960  return __ret;
13961}
13962#else
13963__ai int32x4_t vmovl_s16(int16x4_t __p0) {
13964  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13965  int32x4_t __ret;
13966  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 34);
13967  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13968  return __ret;
13969}
13970__ai int32x4_t __noswap_vmovl_s16(int16x4_t __p0) {
13971  int32x4_t __ret;
13972  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 34);
13973  return __ret;
13974}
13975#endif
13976
13977#ifdef __LITTLE_ENDIAN__
13978__ai uint16x4_t vmovn_u32(uint32x4_t __p0) {
13979  uint16x4_t __ret;
13980  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 17);
13981  return __ret;
13982}
13983#else
13984__ai uint16x4_t vmovn_u32(uint32x4_t __p0) {
13985  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13986  uint16x4_t __ret;
13987  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 17);
13988  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13989  return __ret;
13990}
13991__ai uint16x4_t __noswap_vmovn_u32(uint32x4_t __p0) {
13992  uint16x4_t __ret;
13993  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 17);
13994  return __ret;
13995}
13996#endif
13997
13998#ifdef __LITTLE_ENDIAN__
13999__ai uint32x2_t vmovn_u64(uint64x2_t __p0) {
14000  uint32x2_t __ret;
14001  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 18);
14002  return __ret;
14003}
14004#else
14005__ai uint32x2_t vmovn_u64(uint64x2_t __p0) {
14006  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14007  uint32x2_t __ret;
14008  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 18);
14009  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14010  return __ret;
14011}
14012__ai uint32x2_t __noswap_vmovn_u64(uint64x2_t __p0) {
14013  uint32x2_t __ret;
14014  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 18);
14015  return __ret;
14016}
14017#endif
14018
14019#ifdef __LITTLE_ENDIAN__
14020__ai uint8x8_t vmovn_u16(uint16x8_t __p0) {
14021  uint8x8_t __ret;
14022  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 16);
14023  return __ret;
14024}
14025#else
14026__ai uint8x8_t vmovn_u16(uint16x8_t __p0) {
14027  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14028  uint8x8_t __ret;
14029  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 16);
14030  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14031  return __ret;
14032}
14033__ai uint8x8_t __noswap_vmovn_u16(uint16x8_t __p0) {
14034  uint8x8_t __ret;
14035  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 16);
14036  return __ret;
14037}
14038#endif
14039
14040#ifdef __LITTLE_ENDIAN__
14041__ai int16x4_t vmovn_s32(int32x4_t __p0) {
14042  int16x4_t __ret;
14043  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 1);
14044  return __ret;
14045}
14046#else
14047__ai int16x4_t vmovn_s32(int32x4_t __p0) {
14048  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14049  int16x4_t __ret;
14050  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 1);
14051  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14052  return __ret;
14053}
14054__ai int16x4_t __noswap_vmovn_s32(int32x4_t __p0) {
14055  int16x4_t __ret;
14056  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 1);
14057  return __ret;
14058}
14059#endif
14060
14061#ifdef __LITTLE_ENDIAN__
14062__ai int32x2_t vmovn_s64(int64x2_t __p0) {
14063  int32x2_t __ret;
14064  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 2);
14065  return __ret;
14066}
14067#else
14068__ai int32x2_t vmovn_s64(int64x2_t __p0) {
14069  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14070  int32x2_t __ret;
14071  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 2);
14072  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14073  return __ret;
14074}
14075__ai int32x2_t __noswap_vmovn_s64(int64x2_t __p0) {
14076  int32x2_t __ret;
14077  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 2);
14078  return __ret;
14079}
14080#endif
14081
14082#ifdef __LITTLE_ENDIAN__
14083__ai int8x8_t vmovn_s16(int16x8_t __p0) {
14084  int8x8_t __ret;
14085  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 0);
14086  return __ret;
14087}
14088#else
14089__ai int8x8_t vmovn_s16(int16x8_t __p0) {
14090  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14091  int8x8_t __ret;
14092  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 0);
14093  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14094  return __ret;
14095}
14096__ai int8x8_t __noswap_vmovn_s16(int16x8_t __p0) {
14097  int8x8_t __ret;
14098  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 0);
14099  return __ret;
14100}
14101#endif
14102
14103#ifdef __LITTLE_ENDIAN__
14104__ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1) {
14105  uint8x16_t __ret;
14106  __ret = __p0 * __p1;
14107  return __ret;
14108}
14109#else
14110__ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1) {
14111  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14112  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14113  uint8x16_t __ret;
14114  __ret = __rev0 * __rev1;
14115  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14116  return __ret;
14117}
14118#endif
14119
14120#ifdef __LITTLE_ENDIAN__
14121__ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1) {
14122  uint32x4_t __ret;
14123  __ret = __p0 * __p1;
14124  return __ret;
14125}
14126#else
14127__ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1) {
14128  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14129  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14130  uint32x4_t __ret;
14131  __ret = __rev0 * __rev1;
14132  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14133  return __ret;
14134}
14135#endif
14136
14137#ifdef __LITTLE_ENDIAN__
14138__ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1) {
14139  uint16x8_t __ret;
14140  __ret = __p0 * __p1;
14141  return __ret;
14142}
14143#else
14144__ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1) {
14145  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14146  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14147  uint16x8_t __ret;
14148  __ret = __rev0 * __rev1;
14149  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14150  return __ret;
14151}
14152#endif
14153
14154#ifdef __LITTLE_ENDIAN__
14155__ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1) {
14156  int8x16_t __ret;
14157  __ret = __p0 * __p1;
14158  return __ret;
14159}
14160#else
14161__ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1) {
14162  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14163  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14164  int8x16_t __ret;
14165  __ret = __rev0 * __rev1;
14166  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14167  return __ret;
14168}
14169#endif
14170
14171#ifdef __LITTLE_ENDIAN__
14172__ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1) {
14173  float32x4_t __ret;
14174  __ret = __p0 * __p1;
14175  return __ret;
14176}
14177#else
14178__ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1) {
14179  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14180  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14181  float32x4_t __ret;
14182  __ret = __rev0 * __rev1;
14183  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14184  return __ret;
14185}
14186#endif
14187
14188#ifdef __LITTLE_ENDIAN__
14189__ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1) {
14190  int32x4_t __ret;
14191  __ret = __p0 * __p1;
14192  return __ret;
14193}
14194#else
14195__ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1) {
14196  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14197  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14198  int32x4_t __ret;
14199  __ret = __rev0 * __rev1;
14200  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14201  return __ret;
14202}
14203#endif
14204
14205#ifdef __LITTLE_ENDIAN__
14206__ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1) {
14207  int16x8_t __ret;
14208  __ret = __p0 * __p1;
14209  return __ret;
14210}
14211#else
14212__ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1) {
14213  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14214  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14215  int16x8_t __ret;
14216  __ret = __rev0 * __rev1;
14217  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14218  return __ret;
14219}
14220#endif
14221
14222#ifdef __LITTLE_ENDIAN__
14223__ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1) {
14224  uint8x8_t __ret;
14225  __ret = __p0 * __p1;
14226  return __ret;
14227}
14228#else
14229__ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1) {
14230  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14231  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14232  uint8x8_t __ret;
14233  __ret = __rev0 * __rev1;
14234  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14235  return __ret;
14236}
14237#endif
14238
14239#ifdef __LITTLE_ENDIAN__
14240__ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1) {
14241  uint32x2_t __ret;
14242  __ret = __p0 * __p1;
14243  return __ret;
14244}
14245#else
14246__ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1) {
14247  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14248  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14249  uint32x2_t __ret;
14250  __ret = __rev0 * __rev1;
14251  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14252  return __ret;
14253}
14254#endif
14255
14256#ifdef __LITTLE_ENDIAN__
14257__ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1) {
14258  uint16x4_t __ret;
14259  __ret = __p0 * __p1;
14260  return __ret;
14261}
14262#else
14263__ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1) {
14264  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14265  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14266  uint16x4_t __ret;
14267  __ret = __rev0 * __rev1;
14268  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14269  return __ret;
14270}
14271#endif
14272
14273#ifdef __LITTLE_ENDIAN__
14274__ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1) {
14275  int8x8_t __ret;
14276  __ret = __p0 * __p1;
14277  return __ret;
14278}
14279#else
14280__ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1) {
14281  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14282  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14283  int8x8_t __ret;
14284  __ret = __rev0 * __rev1;
14285  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14286  return __ret;
14287}
14288#endif
14289
14290#ifdef __LITTLE_ENDIAN__
14291__ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1) {
14292  float32x2_t __ret;
14293  __ret = __p0 * __p1;
14294  return __ret;
14295}
14296#else
14297__ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1) {
14298  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14299  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14300  float32x2_t __ret;
14301  __ret = __rev0 * __rev1;
14302  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14303  return __ret;
14304}
14305#endif
14306
14307#ifdef __LITTLE_ENDIAN__
14308__ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1) {
14309  int32x2_t __ret;
14310  __ret = __p0 * __p1;
14311  return __ret;
14312}
14313#else
14314__ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1) {
14315  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14316  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14317  int32x2_t __ret;
14318  __ret = __rev0 * __rev1;
14319  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14320  return __ret;
14321}
14322#endif
14323
14324#ifdef __LITTLE_ENDIAN__
14325__ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1) {
14326  int16x4_t __ret;
14327  __ret = __p0 * __p1;
14328  return __ret;
14329}
14330#else
14331__ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1) {
14332  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14333  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14334  int16x4_t __ret;
14335  __ret = __rev0 * __rev1;
14336  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14337  return __ret;
14338}
14339#endif
14340
14341#ifdef __LITTLE_ENDIAN__
14342__ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1) {
14343  poly8x8_t __ret;
14344  __ret = (poly8x8_t) __builtin_neon_vmul_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
14345  return __ret;
14346}
14347#else
14348__ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1) {
14349  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14350  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14351  poly8x8_t __ret;
14352  __ret = (poly8x8_t) __builtin_neon_vmul_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
14353  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14354  return __ret;
14355}
14356#endif
14357
14358#ifdef __LITTLE_ENDIAN__
14359__ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1) {
14360  poly8x16_t __ret;
14361  __ret = (poly8x16_t) __builtin_neon_vmulq_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
14362  return __ret;
14363}
14364#else
14365__ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1) {
14366  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14367  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14368  poly8x16_t __ret;
14369  __ret = (poly8x16_t) __builtin_neon_vmulq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
14370  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14371  return __ret;
14372}
14373#endif
14374
14375#ifdef __LITTLE_ENDIAN__
14376#define vmulq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14377  uint32x4_t __s0 = __p0; \
14378  uint32x2_t __s1 = __p1; \
14379  uint32x4_t __ret; \
14380  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14381  __ret; \
14382})
14383#else
14384#define vmulq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14385  uint32x4_t __s0 = __p0; \
14386  uint32x2_t __s1 = __p1; \
14387  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14388  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14389  uint32x4_t __ret; \
14390  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14391  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14392  __ret; \
14393})
14394#endif
14395
14396#ifdef __LITTLE_ENDIAN__
14397#define vmulq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14398  uint16x8_t __s0 = __p0; \
14399  uint16x4_t __s1 = __p1; \
14400  uint16x8_t __ret; \
14401  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14402  __ret; \
14403})
14404#else
14405#define vmulq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14406  uint16x8_t __s0 = __p0; \
14407  uint16x4_t __s1 = __p1; \
14408  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
14409  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14410  uint16x8_t __ret; \
14411  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14412  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
14413  __ret; \
14414})
14415#endif
14416
14417#ifdef __LITTLE_ENDIAN__
14418#define vmulq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14419  float32x4_t __s0 = __p0; \
14420  float32x2_t __s1 = __p1; \
14421  float32x4_t __ret; \
14422  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14423  __ret; \
14424})
14425#else
14426#define vmulq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14427  float32x4_t __s0 = __p0; \
14428  float32x2_t __s1 = __p1; \
14429  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14430  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14431  float32x4_t __ret; \
14432  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14433  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14434  __ret; \
14435})
14436#endif
14437
14438#ifdef __LITTLE_ENDIAN__
14439#define vmulq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14440  int32x4_t __s0 = __p0; \
14441  int32x2_t __s1 = __p1; \
14442  int32x4_t __ret; \
14443  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14444  __ret; \
14445})
14446#else
14447#define vmulq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14448  int32x4_t __s0 = __p0; \
14449  int32x2_t __s1 = __p1; \
14450  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14451  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14452  int32x4_t __ret; \
14453  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14454  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14455  __ret; \
14456})
14457#endif
14458
14459#ifdef __LITTLE_ENDIAN__
14460#define vmulq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14461  int16x8_t __s0 = __p0; \
14462  int16x4_t __s1 = __p1; \
14463  int16x8_t __ret; \
14464  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14465  __ret; \
14466})
14467#else
14468#define vmulq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14469  int16x8_t __s0 = __p0; \
14470  int16x4_t __s1 = __p1; \
14471  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
14472  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14473  int16x8_t __ret; \
14474  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14475  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
14476  __ret; \
14477})
14478#endif
14479
14480#ifdef __LITTLE_ENDIAN__
14481#define vmul_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14482  uint32x2_t __s0 = __p0; \
14483  uint32x2_t __s1 = __p1; \
14484  uint32x2_t __ret; \
14485  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
14486  __ret; \
14487})
14488#else
14489#define vmul_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14490  uint32x2_t __s0 = __p0; \
14491  uint32x2_t __s1 = __p1; \
14492  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14493  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14494  uint32x2_t __ret; \
14495  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
14496  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14497  __ret; \
14498})
14499#endif
14500
14501#ifdef __LITTLE_ENDIAN__
14502#define vmul_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14503  uint16x4_t __s0 = __p0; \
14504  uint16x4_t __s1 = __p1; \
14505  uint16x4_t __ret; \
14506  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14507  __ret; \
14508})
14509#else
14510#define vmul_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14511  uint16x4_t __s0 = __p0; \
14512  uint16x4_t __s1 = __p1; \
14513  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14514  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14515  uint16x4_t __ret; \
14516  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14517  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14518  __ret; \
14519})
14520#endif
14521
14522#ifdef __LITTLE_ENDIAN__
14523#define vmul_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14524  float32x2_t __s0 = __p0; \
14525  float32x2_t __s1 = __p1; \
14526  float32x2_t __ret; \
14527  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
14528  __ret; \
14529})
14530#else
14531#define vmul_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14532  float32x2_t __s0 = __p0; \
14533  float32x2_t __s1 = __p1; \
14534  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14535  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14536  float32x2_t __ret; \
14537  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
14538  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14539  __ret; \
14540})
14541#endif
14542
14543#ifdef __LITTLE_ENDIAN__
14544#define vmul_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14545  int32x2_t __s0 = __p0; \
14546  int32x2_t __s1 = __p1; \
14547  int32x2_t __ret; \
14548  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
14549  __ret; \
14550})
14551#else
14552#define vmul_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14553  int32x2_t __s0 = __p0; \
14554  int32x2_t __s1 = __p1; \
14555  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14556  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14557  int32x2_t __ret; \
14558  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
14559  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14560  __ret; \
14561})
14562#endif
14563
14564#ifdef __LITTLE_ENDIAN__
14565#define vmul_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14566  int16x4_t __s0 = __p0; \
14567  int16x4_t __s1 = __p1; \
14568  int16x4_t __ret; \
14569  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14570  __ret; \
14571})
14572#else
14573#define vmul_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14574  int16x4_t __s0 = __p0; \
14575  int16x4_t __s1 = __p1; \
14576  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14577  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14578  int16x4_t __ret; \
14579  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14580  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14581  __ret; \
14582})
14583#endif
14584
14585#ifdef __LITTLE_ENDIAN__
14586__ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1) {
14587  uint32x4_t __ret;
14588  __ret = __p0 * (uint32x4_t) {__p1, __p1, __p1, __p1};
14589  return __ret;
14590}
14591#else
14592__ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1) {
14593  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14594  uint32x4_t __ret;
14595  __ret = __rev0 * (uint32x4_t) {__p1, __p1, __p1, __p1};
14596  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14597  return __ret;
14598}
14599#endif
14600
14601#ifdef __LITTLE_ENDIAN__
14602__ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1) {
14603  uint16x8_t __ret;
14604  __ret = __p0 * (uint16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14605  return __ret;
14606}
14607#else
14608__ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1) {
14609  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14610  uint16x8_t __ret;
14611  __ret = __rev0 * (uint16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14612  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14613  return __ret;
14614}
14615#endif
14616
14617#ifdef __LITTLE_ENDIAN__
14618__ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1) {
14619  float32x4_t __ret;
14620  __ret = __p0 * (float32x4_t) {__p1, __p1, __p1, __p1};
14621  return __ret;
14622}
14623#else
14624__ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1) {
14625  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14626  float32x4_t __ret;
14627  __ret = __rev0 * (float32x4_t) {__p1, __p1, __p1, __p1};
14628  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14629  return __ret;
14630}
14631#endif
14632
14633#ifdef __LITTLE_ENDIAN__
14634__ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1) {
14635  int32x4_t __ret;
14636  __ret = __p0 * (int32x4_t) {__p1, __p1, __p1, __p1};
14637  return __ret;
14638}
14639#else
14640__ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1) {
14641  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14642  int32x4_t __ret;
14643  __ret = __rev0 * (int32x4_t) {__p1, __p1, __p1, __p1};
14644  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14645  return __ret;
14646}
14647#endif
14648
14649#ifdef __LITTLE_ENDIAN__
14650__ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1) {
14651  int16x8_t __ret;
14652  __ret = __p0 * (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14653  return __ret;
14654}
14655#else
14656__ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1) {
14657  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14658  int16x8_t __ret;
14659  __ret = __rev0 * (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14660  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14661  return __ret;
14662}
14663#endif
14664
14665#ifdef __LITTLE_ENDIAN__
14666__ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1) {
14667  uint32x2_t __ret;
14668  __ret = __p0 * (uint32x2_t) {__p1, __p1};
14669  return __ret;
14670}
14671#else
14672__ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1) {
14673  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14674  uint32x2_t __ret;
14675  __ret = __rev0 * (uint32x2_t) {__p1, __p1};
14676  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14677  return __ret;
14678}
14679#endif
14680
14681#ifdef __LITTLE_ENDIAN__
14682__ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1) {
14683  uint16x4_t __ret;
14684  __ret = __p0 * (uint16x4_t) {__p1, __p1, __p1, __p1};
14685  return __ret;
14686}
14687#else
14688__ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1) {
14689  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14690  uint16x4_t __ret;
14691  __ret = __rev0 * (uint16x4_t) {__p1, __p1, __p1, __p1};
14692  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14693  return __ret;
14694}
14695#endif
14696
14697#ifdef __LITTLE_ENDIAN__
14698__ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1) {
14699  float32x2_t __ret;
14700  __ret = __p0 * (float32x2_t) {__p1, __p1};
14701  return __ret;
14702}
14703#else
14704__ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1) {
14705  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14706  float32x2_t __ret;
14707  __ret = __rev0 * (float32x2_t) {__p1, __p1};
14708  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14709  return __ret;
14710}
14711#endif
14712
14713#ifdef __LITTLE_ENDIAN__
14714__ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1) {
14715  int32x2_t __ret;
14716  __ret = __p0 * (int32x2_t) {__p1, __p1};
14717  return __ret;
14718}
14719#else
14720__ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1) {
14721  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14722  int32x2_t __ret;
14723  __ret = __rev0 * (int32x2_t) {__p1, __p1};
14724  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14725  return __ret;
14726}
14727#endif
14728
14729#ifdef __LITTLE_ENDIAN__
14730__ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1) {
14731  int16x4_t __ret;
14732  __ret = __p0 * (int16x4_t) {__p1, __p1, __p1, __p1};
14733  return __ret;
14734}
14735#else
14736__ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1) {
14737  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14738  int16x4_t __ret;
14739  __ret = __rev0 * (int16x4_t) {__p1, __p1, __p1, __p1};
14740  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14741  return __ret;
14742}
14743#endif
14744
14745#ifdef __LITTLE_ENDIAN__
14746__ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
14747  poly16x8_t __ret;
14748  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 37);
14749  return __ret;
14750}
14751#else
14752__ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
14753  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14754  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14755  poly16x8_t __ret;
14756  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 37);
14757  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14758  return __ret;
14759}
14760__ai poly16x8_t __noswap_vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
14761  poly16x8_t __ret;
14762  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 37);
14763  return __ret;
14764}
14765#endif
14766
14767#ifdef __LITTLE_ENDIAN__
14768__ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
14769  uint16x8_t __ret;
14770  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 49);
14771  return __ret;
14772}
14773#else
14774__ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
14775  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14776  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14777  uint16x8_t __ret;
14778  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 49);
14779  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14780  return __ret;
14781}
14782__ai uint16x8_t __noswap_vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
14783  uint16x8_t __ret;
14784  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 49);
14785  return __ret;
14786}
14787#endif
14788
14789#ifdef __LITTLE_ENDIAN__
14790__ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
14791  uint64x2_t __ret;
14792  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 51);
14793  return __ret;
14794}
14795#else
14796__ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
14797  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14798  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14799  uint64x2_t __ret;
14800  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 51);
14801  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14802  return __ret;
14803}
14804__ai uint64x2_t __noswap_vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
14805  uint64x2_t __ret;
14806  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 51);
14807  return __ret;
14808}
14809#endif
14810
14811#ifdef __LITTLE_ENDIAN__
14812__ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
14813  uint32x4_t __ret;
14814  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 50);
14815  return __ret;
14816}
14817#else
14818__ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
14819  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14820  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14821  uint32x4_t __ret;
14822  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 50);
14823  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14824  return __ret;
14825}
14826__ai uint32x4_t __noswap_vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
14827  uint32x4_t __ret;
14828  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 50);
14829  return __ret;
14830}
14831#endif
14832
14833#ifdef __LITTLE_ENDIAN__
14834__ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1) {
14835  int16x8_t __ret;
14836  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 33);
14837  return __ret;
14838}
14839#else
14840__ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1) {
14841  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14842  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14843  int16x8_t __ret;
14844  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 33);
14845  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14846  return __ret;
14847}
14848__ai int16x8_t __noswap_vmull_s8(int8x8_t __p0, int8x8_t __p1) {
14849  int16x8_t __ret;
14850  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 33);
14851  return __ret;
14852}
14853#endif
14854
14855#ifdef __LITTLE_ENDIAN__
14856__ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1) {
14857  int64x2_t __ret;
14858  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
14859  return __ret;
14860}
14861#else
14862__ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1) {
14863  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14864  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14865  int64x2_t __ret;
14866  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 35);
14867  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14868  return __ret;
14869}
14870__ai int64x2_t __noswap_vmull_s32(int32x2_t __p0, int32x2_t __p1) {
14871  int64x2_t __ret;
14872  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
14873  return __ret;
14874}
14875#endif
14876
14877#ifdef __LITTLE_ENDIAN__
14878__ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1) {
14879  int32x4_t __ret;
14880  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
14881  return __ret;
14882}
14883#else
14884__ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1) {
14885  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14886  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14887  int32x4_t __ret;
14888  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 34);
14889  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14890  return __ret;
14891}
14892__ai int32x4_t __noswap_vmull_s16(int16x4_t __p0, int16x4_t __p1) {
14893  int32x4_t __ret;
14894  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
14895  return __ret;
14896}
14897#endif
14898
14899#ifdef __LITTLE_ENDIAN__
14900#define vmull_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14901  uint32x2_t __s0 = __p0; \
14902  uint32x2_t __s1 = __p1; \
14903  uint64x2_t __ret; \
14904  __ret = vmull_u32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
14905  __ret; \
14906})
14907#else
14908#define vmull_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14909  uint32x2_t __s0 = __p0; \
14910  uint32x2_t __s1 = __p1; \
14911  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14912  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14913  uint64x2_t __ret; \
14914  __ret = __noswap_vmull_u32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
14915  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14916  __ret; \
14917})
14918#endif
14919
14920#ifdef __LITTLE_ENDIAN__
14921#define vmull_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14922  uint16x4_t __s0 = __p0; \
14923  uint16x4_t __s1 = __p1; \
14924  uint32x4_t __ret; \
14925  __ret = vmull_u16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
14926  __ret; \
14927})
14928#else
14929#define vmull_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14930  uint16x4_t __s0 = __p0; \
14931  uint16x4_t __s1 = __p1; \
14932  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14933  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14934  uint32x4_t __ret; \
14935  __ret = __noswap_vmull_u16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
14936  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14937  __ret; \
14938})
14939#endif
14940
14941#ifdef __LITTLE_ENDIAN__
14942#define vmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14943  int32x2_t __s0 = __p0; \
14944  int32x2_t __s1 = __p1; \
14945  int64x2_t __ret; \
14946  __ret = vmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
14947  __ret; \
14948})
14949#else
14950#define vmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14951  int32x2_t __s0 = __p0; \
14952  int32x2_t __s1 = __p1; \
14953  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14954  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14955  int64x2_t __ret; \
14956  __ret = __noswap_vmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
14957  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14958  __ret; \
14959})
14960#endif
14961
14962#ifdef __LITTLE_ENDIAN__
14963#define vmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14964  int16x4_t __s0 = __p0; \
14965  int16x4_t __s1 = __p1; \
14966  int32x4_t __ret; \
14967  __ret = vmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
14968  __ret; \
14969})
14970#else
14971#define vmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14972  int16x4_t __s0 = __p0; \
14973  int16x4_t __s1 = __p1; \
14974  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14975  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14976  int32x4_t __ret; \
14977  __ret = __noswap_vmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
14978  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14979  __ret; \
14980})
14981#endif
14982
14983#ifdef __LITTLE_ENDIAN__
14984__ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
14985  uint64x2_t __ret;
14986  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
14987  return __ret;
14988}
14989#else
14990__ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
14991  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14992  uint64x2_t __ret;
14993  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
14994  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14995  return __ret;
14996}
14997__ai uint64x2_t __noswap_vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
14998  uint64x2_t __ret;
14999  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
15000  return __ret;
15001}
15002#endif
15003
15004#ifdef __LITTLE_ENDIAN__
15005__ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
15006  uint32x4_t __ret;
15007  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
15008  return __ret;
15009}
15010#else
15011__ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
15012  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15013  uint32x4_t __ret;
15014  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
15015  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15016  return __ret;
15017}
15018__ai uint32x4_t __noswap_vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
15019  uint32x4_t __ret;
15020  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
15021  return __ret;
15022}
15023#endif
15024
15025#ifdef __LITTLE_ENDIAN__
15026__ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1) {
15027  int64x2_t __ret;
15028  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
15029  return __ret;
15030}
15031#else
15032__ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1) {
15033  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15034  int64x2_t __ret;
15035  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
15036  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15037  return __ret;
15038}
15039__ai int64x2_t __noswap_vmull_n_s32(int32x2_t __p0, int32_t __p1) {
15040  int64x2_t __ret;
15041  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
15042  return __ret;
15043}
15044#endif
15045
15046#ifdef __LITTLE_ENDIAN__
15047__ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1) {
15048  int32x4_t __ret;
15049  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
15050  return __ret;
15051}
15052#else
15053__ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1) {
15054  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15055  int32x4_t __ret;
15056  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
15057  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15058  return __ret;
15059}
15060__ai int32x4_t __noswap_vmull_n_s16(int16x4_t __p0, int16_t __p1) {
15061  int32x4_t __ret;
15062  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
15063  return __ret;
15064}
15065#endif
15066
15067#ifdef __LITTLE_ENDIAN__
15068__ai poly8x8_t vmvn_p8(poly8x8_t __p0) {
15069  poly8x8_t __ret;
15070  __ret = ~__p0;
15071  return __ret;
15072}
15073#else
15074__ai poly8x8_t vmvn_p8(poly8x8_t __p0) {
15075  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15076  poly8x8_t __ret;
15077  __ret = ~__rev0;
15078  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15079  return __ret;
15080}
15081#endif
15082
15083#ifdef __LITTLE_ENDIAN__
15084__ai poly8x16_t vmvnq_p8(poly8x16_t __p0) {
15085  poly8x16_t __ret;
15086  __ret = ~__p0;
15087  return __ret;
15088}
15089#else
15090__ai poly8x16_t vmvnq_p8(poly8x16_t __p0) {
15091  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15092  poly8x16_t __ret;
15093  __ret = ~__rev0;
15094  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15095  return __ret;
15096}
15097#endif
15098
15099#ifdef __LITTLE_ENDIAN__
15100__ai uint8x16_t vmvnq_u8(uint8x16_t __p0) {
15101  uint8x16_t __ret;
15102  __ret = ~__p0;
15103  return __ret;
15104}
15105#else
15106__ai uint8x16_t vmvnq_u8(uint8x16_t __p0) {
15107  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15108  uint8x16_t __ret;
15109  __ret = ~__rev0;
15110  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15111  return __ret;
15112}
15113#endif
15114
15115#ifdef __LITTLE_ENDIAN__
15116__ai uint32x4_t vmvnq_u32(uint32x4_t __p0) {
15117  uint32x4_t __ret;
15118  __ret = ~__p0;
15119  return __ret;
15120}
15121#else
15122__ai uint32x4_t vmvnq_u32(uint32x4_t __p0) {
15123  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15124  uint32x4_t __ret;
15125  __ret = ~__rev0;
15126  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15127  return __ret;
15128}
15129#endif
15130
15131#ifdef __LITTLE_ENDIAN__
15132__ai uint16x8_t vmvnq_u16(uint16x8_t __p0) {
15133  uint16x8_t __ret;
15134  __ret = ~__p0;
15135  return __ret;
15136}
15137#else
15138__ai uint16x8_t vmvnq_u16(uint16x8_t __p0) {
15139  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15140  uint16x8_t __ret;
15141  __ret = ~__rev0;
15142  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15143  return __ret;
15144}
15145#endif
15146
15147#ifdef __LITTLE_ENDIAN__
15148__ai int8x16_t vmvnq_s8(int8x16_t __p0) {
15149  int8x16_t __ret;
15150  __ret = ~__p0;
15151  return __ret;
15152}
15153#else
15154__ai int8x16_t vmvnq_s8(int8x16_t __p0) {
15155  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15156  int8x16_t __ret;
15157  __ret = ~__rev0;
15158  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15159  return __ret;
15160}
15161#endif
15162
15163#ifdef __LITTLE_ENDIAN__
15164__ai int32x4_t vmvnq_s32(int32x4_t __p0) {
15165  int32x4_t __ret;
15166  __ret = ~__p0;
15167  return __ret;
15168}
15169#else
15170__ai int32x4_t vmvnq_s32(int32x4_t __p0) {
15171  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15172  int32x4_t __ret;
15173  __ret = ~__rev0;
15174  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15175  return __ret;
15176}
15177#endif
15178
15179#ifdef __LITTLE_ENDIAN__
15180__ai int16x8_t vmvnq_s16(int16x8_t __p0) {
15181  int16x8_t __ret;
15182  __ret = ~__p0;
15183  return __ret;
15184}
15185#else
15186__ai int16x8_t vmvnq_s16(int16x8_t __p0) {
15187  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15188  int16x8_t __ret;
15189  __ret = ~__rev0;
15190  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15191  return __ret;
15192}
15193#endif
15194
15195#ifdef __LITTLE_ENDIAN__
15196__ai uint8x8_t vmvn_u8(uint8x8_t __p0) {
15197  uint8x8_t __ret;
15198  __ret = ~__p0;
15199  return __ret;
15200}
15201#else
15202__ai uint8x8_t vmvn_u8(uint8x8_t __p0) {
15203  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15204  uint8x8_t __ret;
15205  __ret = ~__rev0;
15206  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15207  return __ret;
15208}
15209#endif
15210
15211#ifdef __LITTLE_ENDIAN__
15212__ai uint32x2_t vmvn_u32(uint32x2_t __p0) {
15213  uint32x2_t __ret;
15214  __ret = ~__p0;
15215  return __ret;
15216}
15217#else
15218__ai uint32x2_t vmvn_u32(uint32x2_t __p0) {
15219  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15220  uint32x2_t __ret;
15221  __ret = ~__rev0;
15222  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15223  return __ret;
15224}
15225#endif
15226
15227#ifdef __LITTLE_ENDIAN__
15228__ai uint16x4_t vmvn_u16(uint16x4_t __p0) {
15229  uint16x4_t __ret;
15230  __ret = ~__p0;
15231  return __ret;
15232}
15233#else
15234__ai uint16x4_t vmvn_u16(uint16x4_t __p0) {
15235  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15236  uint16x4_t __ret;
15237  __ret = ~__rev0;
15238  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15239  return __ret;
15240}
15241#endif
15242
15243#ifdef __LITTLE_ENDIAN__
15244__ai int8x8_t vmvn_s8(int8x8_t __p0) {
15245  int8x8_t __ret;
15246  __ret = ~__p0;
15247  return __ret;
15248}
15249#else
15250__ai int8x8_t vmvn_s8(int8x8_t __p0) {
15251  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15252  int8x8_t __ret;
15253  __ret = ~__rev0;
15254  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15255  return __ret;
15256}
15257#endif
15258
15259#ifdef __LITTLE_ENDIAN__
15260__ai int32x2_t vmvn_s32(int32x2_t __p0) {
15261  int32x2_t __ret;
15262  __ret = ~__p0;
15263  return __ret;
15264}
15265#else
15266__ai int32x2_t vmvn_s32(int32x2_t __p0) {
15267  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15268  int32x2_t __ret;
15269  __ret = ~__rev0;
15270  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15271  return __ret;
15272}
15273#endif
15274
15275#ifdef __LITTLE_ENDIAN__
15276__ai int16x4_t vmvn_s16(int16x4_t __p0) {
15277  int16x4_t __ret;
15278  __ret = ~__p0;
15279  return __ret;
15280}
15281#else
15282__ai int16x4_t vmvn_s16(int16x4_t __p0) {
15283  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15284  int16x4_t __ret;
15285  __ret = ~__rev0;
15286  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15287  return __ret;
15288}
15289#endif
15290
15291#ifdef __LITTLE_ENDIAN__
15292__ai int8x16_t vnegq_s8(int8x16_t __p0) {
15293  int8x16_t __ret;
15294  __ret = -__p0;
15295  return __ret;
15296}
15297#else
15298__ai int8x16_t vnegq_s8(int8x16_t __p0) {
15299  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15300  int8x16_t __ret;
15301  __ret = -__rev0;
15302  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15303  return __ret;
15304}
15305#endif
15306
15307#ifdef __LITTLE_ENDIAN__
15308__ai float32x4_t vnegq_f32(float32x4_t __p0) {
15309  float32x4_t __ret;
15310  __ret = -__p0;
15311  return __ret;
15312}
15313#else
15314__ai float32x4_t vnegq_f32(float32x4_t __p0) {
15315  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15316  float32x4_t __ret;
15317  __ret = -__rev0;
15318  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15319  return __ret;
15320}
15321#endif
15322
15323#ifdef __LITTLE_ENDIAN__
15324__ai int32x4_t vnegq_s32(int32x4_t __p0) {
15325  int32x4_t __ret;
15326  __ret = -__p0;
15327  return __ret;
15328}
15329#else
15330__ai int32x4_t vnegq_s32(int32x4_t __p0) {
15331  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15332  int32x4_t __ret;
15333  __ret = -__rev0;
15334  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15335  return __ret;
15336}
15337#endif
15338
15339#ifdef __LITTLE_ENDIAN__
15340__ai int16x8_t vnegq_s16(int16x8_t __p0) {
15341  int16x8_t __ret;
15342  __ret = -__p0;
15343  return __ret;
15344}
15345#else
15346__ai int16x8_t vnegq_s16(int16x8_t __p0) {
15347  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15348  int16x8_t __ret;
15349  __ret = -__rev0;
15350  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15351  return __ret;
15352}
15353#endif
15354
15355#ifdef __LITTLE_ENDIAN__
15356__ai int8x8_t vneg_s8(int8x8_t __p0) {
15357  int8x8_t __ret;
15358  __ret = -__p0;
15359  return __ret;
15360}
15361#else
15362__ai int8x8_t vneg_s8(int8x8_t __p0) {
15363  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15364  int8x8_t __ret;
15365  __ret = -__rev0;
15366  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15367  return __ret;
15368}
15369#endif
15370
15371#ifdef __LITTLE_ENDIAN__
15372__ai float32x2_t vneg_f32(float32x2_t __p0) {
15373  float32x2_t __ret;
15374  __ret = -__p0;
15375  return __ret;
15376}
15377#else
15378__ai float32x2_t vneg_f32(float32x2_t __p0) {
15379  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15380  float32x2_t __ret;
15381  __ret = -__rev0;
15382  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15383  return __ret;
15384}
15385#endif
15386
15387#ifdef __LITTLE_ENDIAN__
15388__ai int32x2_t vneg_s32(int32x2_t __p0) {
15389  int32x2_t __ret;
15390  __ret = -__p0;
15391  return __ret;
15392}
15393#else
15394__ai int32x2_t vneg_s32(int32x2_t __p0) {
15395  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15396  int32x2_t __ret;
15397  __ret = -__rev0;
15398  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15399  return __ret;
15400}
15401#endif
15402
15403#ifdef __LITTLE_ENDIAN__
15404__ai int16x4_t vneg_s16(int16x4_t __p0) {
15405  int16x4_t __ret;
15406  __ret = -__p0;
15407  return __ret;
15408}
15409#else
15410__ai int16x4_t vneg_s16(int16x4_t __p0) {
15411  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15412  int16x4_t __ret;
15413  __ret = -__rev0;
15414  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15415  return __ret;
15416}
15417#endif
15418
15419#ifdef __LITTLE_ENDIAN__
15420__ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15421  uint8x16_t __ret;
15422  __ret = __p0 | ~__p1;
15423  return __ret;
15424}
15425#else
15426__ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15427  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15428  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15429  uint8x16_t __ret;
15430  __ret = __rev0 | ~__rev1;
15431  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15432  return __ret;
15433}
15434#endif
15435
15436#ifdef __LITTLE_ENDIAN__
15437__ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15438  uint32x4_t __ret;
15439  __ret = __p0 | ~__p1;
15440  return __ret;
15441}
15442#else
15443__ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15444  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15445  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15446  uint32x4_t __ret;
15447  __ret = __rev0 | ~__rev1;
15448  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15449  return __ret;
15450}
15451#endif
15452
15453#ifdef __LITTLE_ENDIAN__
15454__ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15455  uint64x2_t __ret;
15456  __ret = __p0 | ~__p1;
15457  return __ret;
15458}
15459#else
15460__ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15461  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15462  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15463  uint64x2_t __ret;
15464  __ret = __rev0 | ~__rev1;
15465  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15466  return __ret;
15467}
15468#endif
15469
15470#ifdef __LITTLE_ENDIAN__
15471__ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15472  uint16x8_t __ret;
15473  __ret = __p0 | ~__p1;
15474  return __ret;
15475}
15476#else
15477__ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15478  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15479  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15480  uint16x8_t __ret;
15481  __ret = __rev0 | ~__rev1;
15482  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15483  return __ret;
15484}
15485#endif
15486
15487#ifdef __LITTLE_ENDIAN__
15488__ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1) {
15489  int8x16_t __ret;
15490  __ret = __p0 | ~__p1;
15491  return __ret;
15492}
15493#else
15494__ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1) {
15495  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15496  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15497  int8x16_t __ret;
15498  __ret = __rev0 | ~__rev1;
15499  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15500  return __ret;
15501}
15502#endif
15503
15504#ifdef __LITTLE_ENDIAN__
15505__ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1) {
15506  int32x4_t __ret;
15507  __ret = __p0 | ~__p1;
15508  return __ret;
15509}
15510#else
15511__ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1) {
15512  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15513  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15514  int32x4_t __ret;
15515  __ret = __rev0 | ~__rev1;
15516  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15517  return __ret;
15518}
15519#endif
15520
15521#ifdef __LITTLE_ENDIAN__
15522__ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1) {
15523  int64x2_t __ret;
15524  __ret = __p0 | ~__p1;
15525  return __ret;
15526}
15527#else
15528__ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1) {
15529  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15530  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15531  int64x2_t __ret;
15532  __ret = __rev0 | ~__rev1;
15533  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15534  return __ret;
15535}
15536#endif
15537
15538#ifdef __LITTLE_ENDIAN__
15539__ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1) {
15540  int16x8_t __ret;
15541  __ret = __p0 | ~__p1;
15542  return __ret;
15543}
15544#else
15545__ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1) {
15546  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15547  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15548  int16x8_t __ret;
15549  __ret = __rev0 | ~__rev1;
15550  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15551  return __ret;
15552}
15553#endif
15554
15555#ifdef __LITTLE_ENDIAN__
15556__ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1) {
15557  uint8x8_t __ret;
15558  __ret = __p0 | ~__p1;
15559  return __ret;
15560}
15561#else
15562__ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1) {
15563  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15564  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15565  uint8x8_t __ret;
15566  __ret = __rev0 | ~__rev1;
15567  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15568  return __ret;
15569}
15570#endif
15571
15572#ifdef __LITTLE_ENDIAN__
15573__ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1) {
15574  uint32x2_t __ret;
15575  __ret = __p0 | ~__p1;
15576  return __ret;
15577}
15578#else
15579__ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1) {
15580  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15581  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15582  uint32x2_t __ret;
15583  __ret = __rev0 | ~__rev1;
15584  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15585  return __ret;
15586}
15587#endif
15588
15589#ifdef __LITTLE_ENDIAN__
15590__ai uint64x1_t vorn_u64(uint64x1_t __p0, uint64x1_t __p1) {
15591  uint64x1_t __ret;
15592  __ret = __p0 | ~__p1;
15593  return __ret;
15594}
15595#else
15596__ai uint64x1_t vorn_u64(uint64x1_t __p0, uint64x1_t __p1) {
15597  uint64x1_t __ret;
15598  __ret = __p0 | ~__p1;
15599  return __ret;
15600}
15601#endif
15602
15603#ifdef __LITTLE_ENDIAN__
15604__ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1) {
15605  uint16x4_t __ret;
15606  __ret = __p0 | ~__p1;
15607  return __ret;
15608}
15609#else
15610__ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1) {
15611  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15612  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15613  uint16x4_t __ret;
15614  __ret = __rev0 | ~__rev1;
15615  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15616  return __ret;
15617}
15618#endif
15619
15620#ifdef __LITTLE_ENDIAN__
15621__ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1) {
15622  int8x8_t __ret;
15623  __ret = __p0 | ~__p1;
15624  return __ret;
15625}
15626#else
15627__ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1) {
15628  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15629  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15630  int8x8_t __ret;
15631  __ret = __rev0 | ~__rev1;
15632  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15633  return __ret;
15634}
15635#endif
15636
15637#ifdef __LITTLE_ENDIAN__
15638__ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1) {
15639  int32x2_t __ret;
15640  __ret = __p0 | ~__p1;
15641  return __ret;
15642}
15643#else
15644__ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1) {
15645  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15646  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15647  int32x2_t __ret;
15648  __ret = __rev0 | ~__rev1;
15649  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15650  return __ret;
15651}
15652#endif
15653
15654#ifdef __LITTLE_ENDIAN__
15655__ai int64x1_t vorn_s64(int64x1_t __p0, int64x1_t __p1) {
15656  int64x1_t __ret;
15657  __ret = __p0 | ~__p1;
15658  return __ret;
15659}
15660#else
15661__ai int64x1_t vorn_s64(int64x1_t __p0, int64x1_t __p1) {
15662  int64x1_t __ret;
15663  __ret = __p0 | ~__p1;
15664  return __ret;
15665}
15666#endif
15667
15668#ifdef __LITTLE_ENDIAN__
15669__ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1) {
15670  int16x4_t __ret;
15671  __ret = __p0 | ~__p1;
15672  return __ret;
15673}
15674#else
15675__ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1) {
15676  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15677  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15678  int16x4_t __ret;
15679  __ret = __rev0 | ~__rev1;
15680  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15681  return __ret;
15682}
15683#endif
15684
15685#ifdef __LITTLE_ENDIAN__
15686__ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15687  uint8x16_t __ret;
15688  __ret = __p0 | __p1;
15689  return __ret;
15690}
15691#else
15692__ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15693  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15694  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15695  uint8x16_t __ret;
15696  __ret = __rev0 | __rev1;
15697  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15698  return __ret;
15699}
15700#endif
15701
15702#ifdef __LITTLE_ENDIAN__
15703__ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15704  uint32x4_t __ret;
15705  __ret = __p0 | __p1;
15706  return __ret;
15707}
15708#else
15709__ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15710  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15711  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15712  uint32x4_t __ret;
15713  __ret = __rev0 | __rev1;
15714  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15715  return __ret;
15716}
15717#endif
15718
15719#ifdef __LITTLE_ENDIAN__
15720__ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15721  uint64x2_t __ret;
15722  __ret = __p0 | __p1;
15723  return __ret;
15724}
15725#else
15726__ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15727  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15728  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15729  uint64x2_t __ret;
15730  __ret = __rev0 | __rev1;
15731  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15732  return __ret;
15733}
15734#endif
15735
15736#ifdef __LITTLE_ENDIAN__
15737__ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15738  uint16x8_t __ret;
15739  __ret = __p0 | __p1;
15740  return __ret;
15741}
15742#else
15743__ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15744  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15745  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15746  uint16x8_t __ret;
15747  __ret = __rev0 | __rev1;
15748  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15749  return __ret;
15750}
15751#endif
15752
15753#ifdef __LITTLE_ENDIAN__
15754__ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1) {
15755  int8x16_t __ret;
15756  __ret = __p0 | __p1;
15757  return __ret;
15758}
15759#else
15760__ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1) {
15761  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15762  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15763  int8x16_t __ret;
15764  __ret = __rev0 | __rev1;
15765  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15766  return __ret;
15767}
15768#endif
15769
15770#ifdef __LITTLE_ENDIAN__
15771__ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1) {
15772  int32x4_t __ret;
15773  __ret = __p0 | __p1;
15774  return __ret;
15775}
15776#else
15777__ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1) {
15778  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15779  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15780  int32x4_t __ret;
15781  __ret = __rev0 | __rev1;
15782  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15783  return __ret;
15784}
15785#endif
15786
15787#ifdef __LITTLE_ENDIAN__
15788__ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1) {
15789  int64x2_t __ret;
15790  __ret = __p0 | __p1;
15791  return __ret;
15792}
15793#else
15794__ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1) {
15795  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15796  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15797  int64x2_t __ret;
15798  __ret = __rev0 | __rev1;
15799  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15800  return __ret;
15801}
15802#endif
15803
15804#ifdef __LITTLE_ENDIAN__
15805__ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1) {
15806  int16x8_t __ret;
15807  __ret = __p0 | __p1;
15808  return __ret;
15809}
15810#else
15811__ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1) {
15812  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15813  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15814  int16x8_t __ret;
15815  __ret = __rev0 | __rev1;
15816  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15817  return __ret;
15818}
15819#endif
15820
15821#ifdef __LITTLE_ENDIAN__
15822__ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1) {
15823  uint8x8_t __ret;
15824  __ret = __p0 | __p1;
15825  return __ret;
15826}
15827#else
15828__ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1) {
15829  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15830  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15831  uint8x8_t __ret;
15832  __ret = __rev0 | __rev1;
15833  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15834  return __ret;
15835}
15836#endif
15837
15838#ifdef __LITTLE_ENDIAN__
15839__ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1) {
15840  uint32x2_t __ret;
15841  __ret = __p0 | __p1;
15842  return __ret;
15843}
15844#else
15845__ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1) {
15846  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15847  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15848  uint32x2_t __ret;
15849  __ret = __rev0 | __rev1;
15850  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15851  return __ret;
15852}
15853#endif
15854
15855#ifdef __LITTLE_ENDIAN__
15856__ai uint64x1_t vorr_u64(uint64x1_t __p0, uint64x1_t __p1) {
15857  uint64x1_t __ret;
15858  __ret = __p0 | __p1;
15859  return __ret;
15860}
15861#else
15862__ai uint64x1_t vorr_u64(uint64x1_t __p0, uint64x1_t __p1) {
15863  uint64x1_t __ret;
15864  __ret = __p0 | __p1;
15865  return __ret;
15866}
15867#endif
15868
15869#ifdef __LITTLE_ENDIAN__
15870__ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1) {
15871  uint16x4_t __ret;
15872  __ret = __p0 | __p1;
15873  return __ret;
15874}
15875#else
15876__ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1) {
15877  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15878  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15879  uint16x4_t __ret;
15880  __ret = __rev0 | __rev1;
15881  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15882  return __ret;
15883}
15884#endif
15885
15886#ifdef __LITTLE_ENDIAN__
15887__ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1) {
15888  int8x8_t __ret;
15889  __ret = __p0 | __p1;
15890  return __ret;
15891}
15892#else
15893__ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1) {
15894  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15895  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15896  int8x8_t __ret;
15897  __ret = __rev0 | __rev1;
15898  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15899  return __ret;
15900}
15901#endif
15902
15903#ifdef __LITTLE_ENDIAN__
15904__ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1) {
15905  int32x2_t __ret;
15906  __ret = __p0 | __p1;
15907  return __ret;
15908}
15909#else
15910__ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1) {
15911  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15912  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15913  int32x2_t __ret;
15914  __ret = __rev0 | __rev1;
15915  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15916  return __ret;
15917}
15918#endif
15919
15920#ifdef __LITTLE_ENDIAN__
15921__ai int64x1_t vorr_s64(int64x1_t __p0, int64x1_t __p1) {
15922  int64x1_t __ret;
15923  __ret = __p0 | __p1;
15924  return __ret;
15925}
15926#else
15927__ai int64x1_t vorr_s64(int64x1_t __p0, int64x1_t __p1) {
15928  int64x1_t __ret;
15929  __ret = __p0 | __p1;
15930  return __ret;
15931}
15932#endif
15933
15934#ifdef __LITTLE_ENDIAN__
15935__ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1) {
15936  int16x4_t __ret;
15937  __ret = __p0 | __p1;
15938  return __ret;
15939}
15940#else
15941__ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1) {
15942  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15943  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15944  int16x4_t __ret;
15945  __ret = __rev0 | __rev1;
15946  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15947  return __ret;
15948}
15949#endif
15950
15951#ifdef __LITTLE_ENDIAN__
15952__ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1) {
15953  uint16x8_t __ret;
15954  __ret = (uint16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
15955  return __ret;
15956}
15957#else
15958__ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1) {
15959  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15960  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15961  uint16x8_t __ret;
15962  __ret = (uint16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
15963  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15964  return __ret;
15965}
15966#endif
15967
15968#ifdef __LITTLE_ENDIAN__
15969__ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1) {
15970  uint64x2_t __ret;
15971  __ret = (uint64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
15972  return __ret;
15973}
15974#else
15975__ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1) {
15976  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15977  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15978  uint64x2_t __ret;
15979  __ret = (uint64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
15980  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15981  return __ret;
15982}
15983#endif
15984
15985#ifdef __LITTLE_ENDIAN__
15986__ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1) {
15987  uint32x4_t __ret;
15988  __ret = (uint32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
15989  return __ret;
15990}
15991#else
15992__ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1) {
15993  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15994  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15995  uint32x4_t __ret;
15996  __ret = (uint32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
15997  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15998  return __ret;
15999}
16000#endif
16001
16002#ifdef __LITTLE_ENDIAN__
16003__ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1) {
16004  int16x8_t __ret;
16005  __ret = (int16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
16006  return __ret;
16007}
16008#else
16009__ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1) {
16010  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16011  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16012  int16x8_t __ret;
16013  __ret = (int16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
16014  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16015  return __ret;
16016}
16017#endif
16018
16019#ifdef __LITTLE_ENDIAN__
16020__ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1) {
16021  int64x2_t __ret;
16022  __ret = (int64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
16023  return __ret;
16024}
16025#else
16026__ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1) {
16027  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16028  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16029  int64x2_t __ret;
16030  __ret = (int64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
16031  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16032  return __ret;
16033}
16034#endif
16035
16036#ifdef __LITTLE_ENDIAN__
16037__ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1) {
16038  int32x4_t __ret;
16039  __ret = (int32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
16040  return __ret;
16041}
16042#else
16043__ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1) {
16044  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16045  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16046  int32x4_t __ret;
16047  __ret = (int32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
16048  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16049  return __ret;
16050}
16051#endif
16052
16053#ifdef __LITTLE_ENDIAN__
16054__ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1) {
16055  uint16x4_t __ret;
16056  __ret = (uint16x4_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16057  return __ret;
16058}
16059#else
16060__ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1) {
16061  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16062  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16063  uint16x4_t __ret;
16064  __ret = (uint16x4_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16065  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16066  return __ret;
16067}
16068#endif
16069
16070#ifdef __LITTLE_ENDIAN__
16071__ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1) {
16072  uint64x1_t __ret;
16073  __ret = (uint64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
16074  return __ret;
16075}
16076#else
16077__ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1) {
16078  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16079  uint64x1_t __ret;
16080  __ret = (uint64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__rev1, 19);
16081  return __ret;
16082}
16083#endif
16084
16085#ifdef __LITTLE_ENDIAN__
16086__ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1) {
16087  uint32x2_t __ret;
16088  __ret = (uint32x2_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16089  return __ret;
16090}
16091#else
16092__ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1) {
16093  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16094  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16095  uint32x2_t __ret;
16096  __ret = (uint32x2_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16097  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16098  return __ret;
16099}
16100#endif
16101
16102#ifdef __LITTLE_ENDIAN__
16103__ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1) {
16104  int16x4_t __ret;
16105  __ret = (int16x4_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16106  return __ret;
16107}
16108#else
16109__ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1) {
16110  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16111  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16112  int16x4_t __ret;
16113  __ret = (int16x4_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16114  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16115  return __ret;
16116}
16117#endif
16118
16119#ifdef __LITTLE_ENDIAN__
16120__ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1) {
16121  int64x1_t __ret;
16122  __ret = (int64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
16123  return __ret;
16124}
16125#else
16126__ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1) {
16127  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16128  int64x1_t __ret;
16129  __ret = (int64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__rev1, 3);
16130  return __ret;
16131}
16132#endif
16133
16134#ifdef __LITTLE_ENDIAN__
16135__ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1) {
16136  int32x2_t __ret;
16137  __ret = (int32x2_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16138  return __ret;
16139}
16140#else
16141__ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1) {
16142  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16143  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16144  int32x2_t __ret;
16145  __ret = (int32x2_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16146  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16147  return __ret;
16148}
16149#endif
16150
16151#ifdef __LITTLE_ENDIAN__
16152__ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16153  uint8x8_t __ret;
16154  __ret = (uint8x8_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16155  return __ret;
16156}
16157#else
16158__ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16159  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16160  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16161  uint8x8_t __ret;
16162  __ret = (uint8x8_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16163  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16164  return __ret;
16165}
16166#endif
16167
16168#ifdef __LITTLE_ENDIAN__
16169__ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16170  uint32x2_t __ret;
16171  __ret = (uint32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16172  return __ret;
16173}
16174#else
16175__ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16176  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16177  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16178  uint32x2_t __ret;
16179  __ret = (uint32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16180  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16181  return __ret;
16182}
16183#endif
16184
16185#ifdef __LITTLE_ENDIAN__
16186__ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16187  uint16x4_t __ret;
16188  __ret = (uint16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16189  return __ret;
16190}
16191#else
16192__ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16193  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16194  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16195  uint16x4_t __ret;
16196  __ret = (uint16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16197  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16198  return __ret;
16199}
16200#endif
16201
16202#ifdef __LITTLE_ENDIAN__
16203__ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1) {
16204  int8x8_t __ret;
16205  __ret = (int8x8_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
16206  return __ret;
16207}
16208#else
16209__ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1) {
16210  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16211  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16212  int8x8_t __ret;
16213  __ret = (int8x8_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
16214  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16215  return __ret;
16216}
16217#endif
16218
16219#ifdef __LITTLE_ENDIAN__
16220__ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1) {
16221  float32x2_t __ret;
16222  __ret = (float32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
16223  return __ret;
16224}
16225#else
16226__ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1) {
16227  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16228  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16229  float32x2_t __ret;
16230  __ret = (float32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
16231  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16232  return __ret;
16233}
16234#endif
16235
16236#ifdef __LITTLE_ENDIAN__
16237__ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1) {
16238  int32x2_t __ret;
16239  __ret = (int32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16240  return __ret;
16241}
16242#else
16243__ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1) {
16244  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16245  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16246  int32x2_t __ret;
16247  __ret = (int32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16248  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16249  return __ret;
16250}
16251#endif
16252
16253#ifdef __LITTLE_ENDIAN__
16254__ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1) {
16255  int16x4_t __ret;
16256  __ret = (int16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16257  return __ret;
16258}
16259#else
16260__ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1) {
16261  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16262  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16263  int16x4_t __ret;
16264  __ret = (int16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16265  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16266  return __ret;
16267}
16268#endif
16269
16270#ifdef __LITTLE_ENDIAN__
16271__ai uint16x8_t vpaddlq_u8(uint8x16_t __p0) {
16272  uint16x8_t __ret;
16273  __ret = (uint16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 49);
16274  return __ret;
16275}
16276#else
16277__ai uint16x8_t vpaddlq_u8(uint8x16_t __p0) {
16278  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16279  uint16x8_t __ret;
16280  __ret = (uint16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 49);
16281  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16282  return __ret;
16283}
16284#endif
16285
16286#ifdef __LITTLE_ENDIAN__
16287__ai uint64x2_t vpaddlq_u32(uint32x4_t __p0) {
16288  uint64x2_t __ret;
16289  __ret = (uint64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 51);
16290  return __ret;
16291}
16292#else
16293__ai uint64x2_t vpaddlq_u32(uint32x4_t __p0) {
16294  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16295  uint64x2_t __ret;
16296  __ret = (uint64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 51);
16297  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16298  return __ret;
16299}
16300#endif
16301
16302#ifdef __LITTLE_ENDIAN__
16303__ai uint32x4_t vpaddlq_u16(uint16x8_t __p0) {
16304  uint32x4_t __ret;
16305  __ret = (uint32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 50);
16306  return __ret;
16307}
16308#else
16309__ai uint32x4_t vpaddlq_u16(uint16x8_t __p0) {
16310  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16311  uint32x4_t __ret;
16312  __ret = (uint32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 50);
16313  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16314  return __ret;
16315}
16316#endif
16317
16318#ifdef __LITTLE_ENDIAN__
16319__ai int16x8_t vpaddlq_s8(int8x16_t __p0) {
16320  int16x8_t __ret;
16321  __ret = (int16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 33);
16322  return __ret;
16323}
16324#else
16325__ai int16x8_t vpaddlq_s8(int8x16_t __p0) {
16326  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16327  int16x8_t __ret;
16328  __ret = (int16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 33);
16329  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16330  return __ret;
16331}
16332#endif
16333
16334#ifdef __LITTLE_ENDIAN__
16335__ai int64x2_t vpaddlq_s32(int32x4_t __p0) {
16336  int64x2_t __ret;
16337  __ret = (int64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 35);
16338  return __ret;
16339}
16340#else
16341__ai int64x2_t vpaddlq_s32(int32x4_t __p0) {
16342  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16343  int64x2_t __ret;
16344  __ret = (int64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 35);
16345  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16346  return __ret;
16347}
16348#endif
16349
16350#ifdef __LITTLE_ENDIAN__
16351__ai int32x4_t vpaddlq_s16(int16x8_t __p0) {
16352  int32x4_t __ret;
16353  __ret = (int32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 34);
16354  return __ret;
16355}
16356#else
16357__ai int32x4_t vpaddlq_s16(int16x8_t __p0) {
16358  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16359  int32x4_t __ret;
16360  __ret = (int32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 34);
16361  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16362  return __ret;
16363}
16364#endif
16365
16366#ifdef __LITTLE_ENDIAN__
16367__ai uint16x4_t vpaddl_u8(uint8x8_t __p0) {
16368  uint16x4_t __ret;
16369  __ret = (uint16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 17);
16370  return __ret;
16371}
16372#else
16373__ai uint16x4_t vpaddl_u8(uint8x8_t __p0) {
16374  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16375  uint16x4_t __ret;
16376  __ret = (uint16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 17);
16377  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16378  return __ret;
16379}
16380#endif
16381
16382#ifdef __LITTLE_ENDIAN__
16383__ai uint64x1_t vpaddl_u32(uint32x2_t __p0) {
16384  uint64x1_t __ret;
16385  __ret = (uint64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 19);
16386  return __ret;
16387}
16388#else
16389__ai uint64x1_t vpaddl_u32(uint32x2_t __p0) {
16390  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16391  uint64x1_t __ret;
16392  __ret = (uint64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 19);
16393  return __ret;
16394}
16395#endif
16396
16397#ifdef __LITTLE_ENDIAN__
16398__ai uint32x2_t vpaddl_u16(uint16x4_t __p0) {
16399  uint32x2_t __ret;
16400  __ret = (uint32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 18);
16401  return __ret;
16402}
16403#else
16404__ai uint32x2_t vpaddl_u16(uint16x4_t __p0) {
16405  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16406  uint32x2_t __ret;
16407  __ret = (uint32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 18);
16408  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16409  return __ret;
16410}
16411#endif
16412
16413#ifdef __LITTLE_ENDIAN__
16414__ai int16x4_t vpaddl_s8(int8x8_t __p0) {
16415  int16x4_t __ret;
16416  __ret = (int16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 1);
16417  return __ret;
16418}
16419#else
16420__ai int16x4_t vpaddl_s8(int8x8_t __p0) {
16421  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16422  int16x4_t __ret;
16423  __ret = (int16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 1);
16424  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16425  return __ret;
16426}
16427#endif
16428
16429#ifdef __LITTLE_ENDIAN__
16430__ai int64x1_t vpaddl_s32(int32x2_t __p0) {
16431  int64x1_t __ret;
16432  __ret = (int64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 3);
16433  return __ret;
16434}
16435#else
16436__ai int64x1_t vpaddl_s32(int32x2_t __p0) {
16437  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16438  int64x1_t __ret;
16439  __ret = (int64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 3);
16440  return __ret;
16441}
16442#endif
16443
16444#ifdef __LITTLE_ENDIAN__
16445__ai int32x2_t vpaddl_s16(int16x4_t __p0) {
16446  int32x2_t __ret;
16447  __ret = (int32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 2);
16448  return __ret;
16449}
16450#else
16451__ai int32x2_t vpaddl_s16(int16x4_t __p0) {
16452  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16453  int32x2_t __ret;
16454  __ret = (int32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 2);
16455  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16456  return __ret;
16457}
16458#endif
16459
16460#ifdef __LITTLE_ENDIAN__
16461__ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
16462  uint8x8_t __ret;
16463  __ret = (uint8x8_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16464  return __ret;
16465}
16466#else
16467__ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
16468  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16469  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16470  uint8x8_t __ret;
16471  __ret = (uint8x8_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16472  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16473  return __ret;
16474}
16475#endif
16476
16477#ifdef __LITTLE_ENDIAN__
16478__ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
16479  uint32x2_t __ret;
16480  __ret = (uint32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16481  return __ret;
16482}
16483#else
16484__ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
16485  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16486  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16487  uint32x2_t __ret;
16488  __ret = (uint32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16489  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16490  return __ret;
16491}
16492#endif
16493
16494#ifdef __LITTLE_ENDIAN__
16495__ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
16496  uint16x4_t __ret;
16497  __ret = (uint16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16498  return __ret;
16499}
16500#else
16501__ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
16502  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16503  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16504  uint16x4_t __ret;
16505  __ret = (uint16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16506  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16507  return __ret;
16508}
16509#endif
16510
16511#ifdef __LITTLE_ENDIAN__
16512__ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1) {
16513  int8x8_t __ret;
16514  __ret = (int8x8_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
16515  return __ret;
16516}
16517#else
16518__ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1) {
16519  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16520  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16521  int8x8_t __ret;
16522  __ret = (int8x8_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
16523  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16524  return __ret;
16525}
16526#endif
16527
16528#ifdef __LITTLE_ENDIAN__
16529__ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1) {
16530  float32x2_t __ret;
16531  __ret = (float32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
16532  return __ret;
16533}
16534#else
16535__ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1) {
16536  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16537  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16538  float32x2_t __ret;
16539  __ret = (float32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
16540  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16541  return __ret;
16542}
16543#endif
16544
16545#ifdef __LITTLE_ENDIAN__
16546__ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1) {
16547  int32x2_t __ret;
16548  __ret = (int32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16549  return __ret;
16550}
16551#else
16552__ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1) {
16553  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16554  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16555  int32x2_t __ret;
16556  __ret = (int32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16557  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16558  return __ret;
16559}
16560#endif
16561
16562#ifdef __LITTLE_ENDIAN__
16563__ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1) {
16564  int16x4_t __ret;
16565  __ret = (int16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16566  return __ret;
16567}
16568#else
16569__ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1) {
16570  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16571  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16572  int16x4_t __ret;
16573  __ret = (int16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16574  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16575  return __ret;
16576}
16577#endif
16578
16579#ifdef __LITTLE_ENDIAN__
16580__ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
16581  uint8x8_t __ret;
16582  __ret = (uint8x8_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16583  return __ret;
16584}
16585#else
16586__ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
16587  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16588  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16589  uint8x8_t __ret;
16590  __ret = (uint8x8_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16591  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16592  return __ret;
16593}
16594#endif
16595
16596#ifdef __LITTLE_ENDIAN__
16597__ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
16598  uint32x2_t __ret;
16599  __ret = (uint32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16600  return __ret;
16601}
16602#else
16603__ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
16604  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16605  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16606  uint32x2_t __ret;
16607  __ret = (uint32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16608  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16609  return __ret;
16610}
16611#endif
16612
16613#ifdef __LITTLE_ENDIAN__
16614__ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
16615  uint16x4_t __ret;
16616  __ret = (uint16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16617  return __ret;
16618}
16619#else
16620__ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
16621  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16622  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16623  uint16x4_t __ret;
16624  __ret = (uint16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16625  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16626  return __ret;
16627}
16628#endif
16629
16630#ifdef __LITTLE_ENDIAN__
16631__ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1) {
16632  int8x8_t __ret;
16633  __ret = (int8x8_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
16634  return __ret;
16635}
16636#else
16637__ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1) {
16638  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16639  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16640  int8x8_t __ret;
16641  __ret = (int8x8_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
16642  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16643  return __ret;
16644}
16645#endif
16646
16647#ifdef __LITTLE_ENDIAN__
16648__ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1) {
16649  float32x2_t __ret;
16650  __ret = (float32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
16651  return __ret;
16652}
16653#else
16654__ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1) {
16655  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16656  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16657  float32x2_t __ret;
16658  __ret = (float32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
16659  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16660  return __ret;
16661}
16662#endif
16663
16664#ifdef __LITTLE_ENDIAN__
16665__ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1) {
16666  int32x2_t __ret;
16667  __ret = (int32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16668  return __ret;
16669}
16670#else
16671__ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1) {
16672  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16673  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16674  int32x2_t __ret;
16675  __ret = (int32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16676  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16677  return __ret;
16678}
16679#endif
16680
16681#ifdef __LITTLE_ENDIAN__
16682__ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1) {
16683  int16x4_t __ret;
16684  __ret = (int16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16685  return __ret;
16686}
16687#else
16688__ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1) {
16689  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16690  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16691  int16x4_t __ret;
16692  __ret = (int16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16693  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16694  return __ret;
16695}
16696#endif
16697
16698#ifdef __LITTLE_ENDIAN__
16699__ai int8x16_t vqabsq_s8(int8x16_t __p0) {
16700  int8x16_t __ret;
16701  __ret = (int8x16_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 32);
16702  return __ret;
16703}
16704#else
16705__ai int8x16_t vqabsq_s8(int8x16_t __p0) {
16706  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16707  int8x16_t __ret;
16708  __ret = (int8x16_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 32);
16709  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16710  return __ret;
16711}
16712#endif
16713
16714#ifdef __LITTLE_ENDIAN__
16715__ai int32x4_t vqabsq_s32(int32x4_t __p0) {
16716  int32x4_t __ret;
16717  __ret = (int32x4_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 34);
16718  return __ret;
16719}
16720#else
16721__ai int32x4_t vqabsq_s32(int32x4_t __p0) {
16722  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16723  int32x4_t __ret;
16724  __ret = (int32x4_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 34);
16725  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16726  return __ret;
16727}
16728#endif
16729
16730#ifdef __LITTLE_ENDIAN__
16731__ai int16x8_t vqabsq_s16(int16x8_t __p0) {
16732  int16x8_t __ret;
16733  __ret = (int16x8_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 33);
16734  return __ret;
16735}
16736#else
16737__ai int16x8_t vqabsq_s16(int16x8_t __p0) {
16738  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16739  int16x8_t __ret;
16740  __ret = (int16x8_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 33);
16741  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16742  return __ret;
16743}
16744#endif
16745
16746#ifdef __LITTLE_ENDIAN__
16747__ai int8x8_t vqabs_s8(int8x8_t __p0) {
16748  int8x8_t __ret;
16749  __ret = (int8x8_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 0);
16750  return __ret;
16751}
16752#else
16753__ai int8x8_t vqabs_s8(int8x8_t __p0) {
16754  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16755  int8x8_t __ret;
16756  __ret = (int8x8_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 0);
16757  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16758  return __ret;
16759}
16760#endif
16761
16762#ifdef __LITTLE_ENDIAN__
16763__ai int32x2_t vqabs_s32(int32x2_t __p0) {
16764  int32x2_t __ret;
16765  __ret = (int32x2_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 2);
16766  return __ret;
16767}
16768#else
16769__ai int32x2_t vqabs_s32(int32x2_t __p0) {
16770  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16771  int32x2_t __ret;
16772  __ret = (int32x2_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 2);
16773  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16774  return __ret;
16775}
16776#endif
16777
16778#ifdef __LITTLE_ENDIAN__
16779__ai int16x4_t vqabs_s16(int16x4_t __p0) {
16780  int16x4_t __ret;
16781  __ret = (int16x4_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 1);
16782  return __ret;
16783}
16784#else
16785__ai int16x4_t vqabs_s16(int16x4_t __p0) {
16786  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16787  int16x4_t __ret;
16788  __ret = (int16x4_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 1);
16789  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16790  return __ret;
16791}
16792#endif
16793
16794#ifdef __LITTLE_ENDIAN__
16795__ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
16796  uint8x16_t __ret;
16797  __ret = (uint8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
16798  return __ret;
16799}
16800#else
16801__ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
16802  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16803  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16804  uint8x16_t __ret;
16805  __ret = (uint8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
16806  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16807  return __ret;
16808}
16809#endif
16810
16811#ifdef __LITTLE_ENDIAN__
16812__ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
16813  uint32x4_t __ret;
16814  __ret = (uint32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
16815  return __ret;
16816}
16817#else
16818__ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
16819  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16820  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16821  uint32x4_t __ret;
16822  __ret = (uint32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
16823  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16824  return __ret;
16825}
16826#endif
16827
16828#ifdef __LITTLE_ENDIAN__
16829__ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
16830  uint64x2_t __ret;
16831  __ret = (uint64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
16832  return __ret;
16833}
16834#else
16835__ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
16836  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16837  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16838  uint64x2_t __ret;
16839  __ret = (uint64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
16840  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16841  return __ret;
16842}
16843#endif
16844
16845#ifdef __LITTLE_ENDIAN__
16846__ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
16847  uint16x8_t __ret;
16848  __ret = (uint16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
16849  return __ret;
16850}
16851#else
16852__ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
16853  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16854  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16855  uint16x8_t __ret;
16856  __ret = (uint16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
16857  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16858  return __ret;
16859}
16860#endif
16861
16862#ifdef __LITTLE_ENDIAN__
16863__ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
16864  int8x16_t __ret;
16865  __ret = (int8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
16866  return __ret;
16867}
16868#else
16869__ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
16870  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16871  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16872  int8x16_t __ret;
16873  __ret = (int8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
16874  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16875  return __ret;
16876}
16877#endif
16878
16879#ifdef __LITTLE_ENDIAN__
16880__ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
16881  int32x4_t __ret;
16882  __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
16883  return __ret;
16884}
16885#else
16886__ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
16887  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16888  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16889  int32x4_t __ret;
16890  __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
16891  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16892  return __ret;
16893}
16894#endif
16895
16896#ifdef __LITTLE_ENDIAN__
16897__ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
16898  int64x2_t __ret;
16899  __ret = (int64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
16900  return __ret;
16901}
16902#else
16903__ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
16904  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16905  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16906  int64x2_t __ret;
16907  __ret = (int64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
16908  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16909  return __ret;
16910}
16911#endif
16912
16913#ifdef __LITTLE_ENDIAN__
16914__ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
16915  int16x8_t __ret;
16916  __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
16917  return __ret;
16918}
16919#else
16920__ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
16921  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16922  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16923  int16x8_t __ret;
16924  __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
16925  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16926  return __ret;
16927}
16928#endif
16929
16930#ifdef __LITTLE_ENDIAN__
16931__ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16932  uint8x8_t __ret;
16933  __ret = (uint8x8_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16934  return __ret;
16935}
16936#else
16937__ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16938  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16939  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16940  uint8x8_t __ret;
16941  __ret = (uint8x8_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16942  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16943  return __ret;
16944}
16945#endif
16946
16947#ifdef __LITTLE_ENDIAN__
16948__ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16949  uint32x2_t __ret;
16950  __ret = (uint32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16951  return __ret;
16952}
16953#else
16954__ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16955  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16956  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16957  uint32x2_t __ret;
16958  __ret = (uint32x2_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16959  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16960  return __ret;
16961}
16962#endif
16963
16964#ifdef __LITTLE_ENDIAN__
16965__ai uint64x1_t vqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
16966  uint64x1_t __ret;
16967  __ret = (uint64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
16968  return __ret;
16969}
16970#else
16971__ai uint64x1_t vqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
16972  uint64x1_t __ret;
16973  __ret = (uint64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
16974  return __ret;
16975}
16976#endif
16977
16978#ifdef __LITTLE_ENDIAN__
16979__ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16980  uint16x4_t __ret;
16981  __ret = (uint16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16982  return __ret;
16983}
16984#else
16985__ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16986  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16987  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16988  uint16x4_t __ret;
16989  __ret = (uint16x4_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16990  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16991  return __ret;
16992}
16993#endif
16994
16995#ifdef __LITTLE_ENDIAN__
16996__ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1) {
16997  int8x8_t __ret;
16998  __ret = (int8x8_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
16999  return __ret;
17000}
17001#else
17002__ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1) {
17003  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17004  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17005  int8x8_t __ret;
17006  __ret = (int8x8_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
17007  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17008  return __ret;
17009}
17010#endif
17011
17012#ifdef __LITTLE_ENDIAN__
17013__ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
17014  int32x2_t __ret;
17015  __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17016  return __ret;
17017}
17018#else
17019__ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
17020  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17021  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17022  int32x2_t __ret;
17023  __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
17024  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17025  return __ret;
17026}
17027#endif
17028
17029#ifdef __LITTLE_ENDIAN__
17030__ai int64x1_t vqadd_s64(int64x1_t __p0, int64x1_t __p1) {
17031  int64x1_t __ret;
17032  __ret = (int64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
17033  return __ret;
17034}
17035#else
17036__ai int64x1_t vqadd_s64(int64x1_t __p0, int64x1_t __p1) {
17037  int64x1_t __ret;
17038  __ret = (int64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
17039  return __ret;
17040}
17041#endif
17042
17043#ifdef __LITTLE_ENDIAN__
17044__ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
17045  int16x4_t __ret;
17046  __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
17047  return __ret;
17048}
17049#else
17050__ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
17051  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17052  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17053  int16x4_t __ret;
17054  __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
17055  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17056  return __ret;
17057}
17058#endif
17059
17060#ifdef __LITTLE_ENDIAN__
17061__ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17062  int64x2_t __ret;
17063  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17064  return __ret;
17065}
17066#else
17067__ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17068  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17069  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17070  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
17071  int64x2_t __ret;
17072  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 35);
17073  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17074  return __ret;
17075}
17076__ai int64x2_t __noswap_vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17077  int64x2_t __ret;
17078  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17079  return __ret;
17080}
17081#endif
17082
17083#ifdef __LITTLE_ENDIAN__
17084__ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17085  int32x4_t __ret;
17086  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17087  return __ret;
17088}
17089#else
17090__ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17091  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17092  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17093  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
17094  int32x4_t __ret;
17095  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 34);
17096  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17097  return __ret;
17098}
17099__ai int32x4_t __noswap_vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17100  int32x4_t __ret;
17101  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17102  return __ret;
17103}
17104#endif
17105
17106#ifdef __LITTLE_ENDIAN__
17107#define vqdmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17108  int64x2_t __s0 = __p0; \
17109  int32x2_t __s1 = __p1; \
17110  int32x2_t __s2 = __p2; \
17111  int64x2_t __ret; \
17112  __ret = vqdmlal_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
17113  __ret; \
17114})
17115#else
17116#define vqdmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17117  int64x2_t __s0 = __p0; \
17118  int32x2_t __s1 = __p1; \
17119  int32x2_t __s2 = __p2; \
17120  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17121  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17122  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
17123  int64x2_t __ret; \
17124  __ret = __noswap_vqdmlal_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
17125  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17126  __ret; \
17127})
17128#endif
17129
17130#ifdef __LITTLE_ENDIAN__
17131#define vqdmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17132  int32x4_t __s0 = __p0; \
17133  int16x4_t __s1 = __p1; \
17134  int16x4_t __s2 = __p2; \
17135  int32x4_t __ret; \
17136  __ret = vqdmlal_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
17137  __ret; \
17138})
17139#else
17140#define vqdmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17141  int32x4_t __s0 = __p0; \
17142  int16x4_t __s1 = __p1; \
17143  int16x4_t __s2 = __p2; \
17144  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17145  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17146  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
17147  int32x4_t __ret; \
17148  __ret = __noswap_vqdmlal_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
17149  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17150  __ret; \
17151})
17152#endif
17153
17154#ifdef __LITTLE_ENDIAN__
17155__ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17156  int64x2_t __ret;
17157  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17158  return __ret;
17159}
17160#else
17161__ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17162  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17163  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17164  int64x2_t __ret;
17165  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17166  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17167  return __ret;
17168}
17169__ai int64x2_t __noswap_vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17170  int64x2_t __ret;
17171  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17172  return __ret;
17173}
17174#endif
17175
17176#ifdef __LITTLE_ENDIAN__
17177__ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17178  int32x4_t __ret;
17179  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17180  return __ret;
17181}
17182#else
17183__ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17184  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17185  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17186  int32x4_t __ret;
17187  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17188  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17189  return __ret;
17190}
17191__ai int32x4_t __noswap_vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17192  int32x4_t __ret;
17193  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17194  return __ret;
17195}
17196#endif
17197
17198#ifdef __LITTLE_ENDIAN__
17199__ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17200  int64x2_t __ret;
17201  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17202  return __ret;
17203}
17204#else
17205__ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17206  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17207  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17208  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
17209  int64x2_t __ret;
17210  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 35);
17211  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17212  return __ret;
17213}
17214__ai int64x2_t __noswap_vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17215  int64x2_t __ret;
17216  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17217  return __ret;
17218}
17219#endif
17220
17221#ifdef __LITTLE_ENDIAN__
17222__ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17223  int32x4_t __ret;
17224  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17225  return __ret;
17226}
17227#else
17228__ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17229  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17230  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17231  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
17232  int32x4_t __ret;
17233  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 34);
17234  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17235  return __ret;
17236}
17237__ai int32x4_t __noswap_vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17238  int32x4_t __ret;
17239  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17240  return __ret;
17241}
17242#endif
17243
17244#ifdef __LITTLE_ENDIAN__
17245#define vqdmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17246  int64x2_t __s0 = __p0; \
17247  int32x2_t __s1 = __p1; \
17248  int32x2_t __s2 = __p2; \
17249  int64x2_t __ret; \
17250  __ret = vqdmlsl_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
17251  __ret; \
17252})
17253#else
17254#define vqdmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17255  int64x2_t __s0 = __p0; \
17256  int32x2_t __s1 = __p1; \
17257  int32x2_t __s2 = __p2; \
17258  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17259  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17260  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
17261  int64x2_t __ret; \
17262  __ret = __noswap_vqdmlsl_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
17263  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17264  __ret; \
17265})
17266#endif
17267
17268#ifdef __LITTLE_ENDIAN__
17269#define vqdmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17270  int32x4_t __s0 = __p0; \
17271  int16x4_t __s1 = __p1; \
17272  int16x4_t __s2 = __p2; \
17273  int32x4_t __ret; \
17274  __ret = vqdmlsl_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
17275  __ret; \
17276})
17277#else
17278#define vqdmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17279  int32x4_t __s0 = __p0; \
17280  int16x4_t __s1 = __p1; \
17281  int16x4_t __s2 = __p2; \
17282  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17283  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17284  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
17285  int32x4_t __ret; \
17286  __ret = __noswap_vqdmlsl_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
17287  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17288  __ret; \
17289})
17290#endif
17291
17292#ifdef __LITTLE_ENDIAN__
17293__ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17294  int64x2_t __ret;
17295  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17296  return __ret;
17297}
17298#else
17299__ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17300  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17301  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17302  int64x2_t __ret;
17303  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17304  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17305  return __ret;
17306}
17307__ai int64x2_t __noswap_vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17308  int64x2_t __ret;
17309  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17310  return __ret;
17311}
17312#endif
17313
17314#ifdef __LITTLE_ENDIAN__
17315__ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17316  int32x4_t __ret;
17317  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17318  return __ret;
17319}
17320#else
17321__ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17322  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17323  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17324  int32x4_t __ret;
17325  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17326  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17327  return __ret;
17328}
17329__ai int32x4_t __noswap_vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17330  int32x4_t __ret;
17331  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17332  return __ret;
17333}
17334#endif
17335
17336#ifdef __LITTLE_ENDIAN__
17337__ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17338  int32x4_t __ret;
17339  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
17340  return __ret;
17341}
17342#else
17343__ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17344  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17345  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17346  int32x4_t __ret;
17347  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
17348  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17349  return __ret;
17350}
17351__ai int32x4_t __noswap_vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17352  int32x4_t __ret;
17353  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
17354  return __ret;
17355}
17356#endif
17357
17358#ifdef __LITTLE_ENDIAN__
17359__ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17360  int16x8_t __ret;
17361  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
17362  return __ret;
17363}
17364#else
17365__ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17366  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17367  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17368  int16x8_t __ret;
17369  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
17370  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17371  return __ret;
17372}
17373__ai int16x8_t __noswap_vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17374  int16x8_t __ret;
17375  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
17376  return __ret;
17377}
17378#endif
17379
17380#ifdef __LITTLE_ENDIAN__
17381__ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17382  int32x2_t __ret;
17383  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17384  return __ret;
17385}
17386#else
17387__ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17388  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17389  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17390  int32x2_t __ret;
17391  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
17392  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17393  return __ret;
17394}
17395__ai int32x2_t __noswap_vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17396  int32x2_t __ret;
17397  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17398  return __ret;
17399}
17400#endif
17401
17402#ifdef __LITTLE_ENDIAN__
17403__ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17404  int16x4_t __ret;
17405  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
17406  return __ret;
17407}
17408#else
17409__ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17410  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17411  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17412  int16x4_t __ret;
17413  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
17414  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17415  return __ret;
17416}
17417__ai int16x4_t __noswap_vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17418  int16x4_t __ret;
17419  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
17420  return __ret;
17421}
17422#endif
17423
17424#ifdef __LITTLE_ENDIAN__
17425#define vqdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17426  int32x4_t __s0 = __p0; \
17427  int32x2_t __s1 = __p1; \
17428  int32x4_t __ret; \
17429  __ret = vqdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
17430  __ret; \
17431})
17432#else
17433#define vqdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17434  int32x4_t __s0 = __p0; \
17435  int32x2_t __s1 = __p1; \
17436  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17437  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17438  int32x4_t __ret; \
17439  __ret = __noswap_vqdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
17440  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17441  __ret; \
17442})
17443#endif
17444
17445#ifdef __LITTLE_ENDIAN__
17446#define vqdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17447  int16x8_t __s0 = __p0; \
17448  int16x4_t __s1 = __p1; \
17449  int16x8_t __ret; \
17450  __ret = vqdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
17451  __ret; \
17452})
17453#else
17454#define vqdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17455  int16x8_t __s0 = __p0; \
17456  int16x4_t __s1 = __p1; \
17457  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
17458  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17459  int16x8_t __ret; \
17460  __ret = __noswap_vqdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
17461  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
17462  __ret; \
17463})
17464#endif
17465
17466#ifdef __LITTLE_ENDIAN__
17467#define vqdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17468  int32x2_t __s0 = __p0; \
17469  int32x2_t __s1 = __p1; \
17470  int32x2_t __ret; \
17471  __ret = vqdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
17472  __ret; \
17473})
17474#else
17475#define vqdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17476  int32x2_t __s0 = __p0; \
17477  int32x2_t __s1 = __p1; \
17478  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17479  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17480  int32x2_t __ret; \
17481  __ret = __noswap_vqdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
17482  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17483  __ret; \
17484})
17485#endif
17486
17487#ifdef __LITTLE_ENDIAN__
17488#define vqdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17489  int16x4_t __s0 = __p0; \
17490  int16x4_t __s1 = __p1; \
17491  int16x4_t __ret; \
17492  __ret = vqdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
17493  __ret; \
17494})
17495#else
17496#define vqdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17497  int16x4_t __s0 = __p0; \
17498  int16x4_t __s1 = __p1; \
17499  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17500  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17501  int16x4_t __ret; \
17502  __ret = __noswap_vqdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
17503  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17504  __ret; \
17505})
17506#endif
17507
17508#ifdef __LITTLE_ENDIAN__
17509__ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
17510  int32x4_t __ret;
17511  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
17512  return __ret;
17513}
17514#else
17515__ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
17516  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17517  int32x4_t __ret;
17518  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
17519  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17520  return __ret;
17521}
17522#endif
17523
17524#ifdef __LITTLE_ENDIAN__
17525__ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
17526  int16x8_t __ret;
17527  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
17528  return __ret;
17529}
17530#else
17531__ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
17532  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17533  int16x8_t __ret;
17534  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
17535  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17536  return __ret;
17537}
17538#endif
17539
17540#ifdef __LITTLE_ENDIAN__
17541__ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
17542  int32x2_t __ret;
17543  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
17544  return __ret;
17545}
17546#else
17547__ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
17548  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17549  int32x2_t __ret;
17550  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
17551  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17552  return __ret;
17553}
17554#endif
17555
17556#ifdef __LITTLE_ENDIAN__
17557__ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
17558  int16x4_t __ret;
17559  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
17560  return __ret;
17561}
17562#else
17563__ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
17564  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17565  int16x4_t __ret;
17566  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
17567  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17568  return __ret;
17569}
17570#endif
17571
17572#ifdef __LITTLE_ENDIAN__
17573__ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
17574  int64x2_t __ret;
17575  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
17576  return __ret;
17577}
17578#else
17579__ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
17580  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17581  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17582  int64x2_t __ret;
17583  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 35);
17584  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17585  return __ret;
17586}
17587__ai int64x2_t __noswap_vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
17588  int64x2_t __ret;
17589  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
17590  return __ret;
17591}
17592#endif
17593
17594#ifdef __LITTLE_ENDIAN__
17595__ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
17596  int32x4_t __ret;
17597  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
17598  return __ret;
17599}
17600#else
17601__ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
17602  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17603  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17604  int32x4_t __ret;
17605  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 34);
17606  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17607  return __ret;
17608}
17609__ai int32x4_t __noswap_vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
17610  int32x4_t __ret;
17611  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
17612  return __ret;
17613}
17614#endif
17615
17616#ifdef __LITTLE_ENDIAN__
17617#define vqdmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17618  int32x2_t __s0 = __p0; \
17619  int32x2_t __s1 = __p1; \
17620  int64x2_t __ret; \
17621  __ret = vqdmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
17622  __ret; \
17623})
17624#else
17625#define vqdmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17626  int32x2_t __s0 = __p0; \
17627  int32x2_t __s1 = __p1; \
17628  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17629  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17630  int64x2_t __ret; \
17631  __ret = __noswap_vqdmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
17632  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17633  __ret; \
17634})
17635#endif
17636
17637#ifdef __LITTLE_ENDIAN__
17638#define vqdmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17639  int16x4_t __s0 = __p0; \
17640  int16x4_t __s1 = __p1; \
17641  int32x4_t __ret; \
17642  __ret = vqdmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
17643  __ret; \
17644})
17645#else
17646#define vqdmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17647  int16x4_t __s0 = __p0; \
17648  int16x4_t __s1 = __p1; \
17649  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17650  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17651  int32x4_t __ret; \
17652  __ret = __noswap_vqdmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
17653  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17654  __ret; \
17655})
17656#endif
17657
17658#ifdef __LITTLE_ENDIAN__
17659__ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
17660  int64x2_t __ret;
17661  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
17662  return __ret;
17663}
17664#else
17665__ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
17666  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17667  int64x2_t __ret;
17668  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
17669  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17670  return __ret;
17671}
17672__ai int64x2_t __noswap_vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
17673  int64x2_t __ret;
17674  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
17675  return __ret;
17676}
17677#endif
17678
17679#ifdef __LITTLE_ENDIAN__
17680__ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
17681  int32x4_t __ret;
17682  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
17683  return __ret;
17684}
17685#else
17686__ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
17687  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17688  int32x4_t __ret;
17689  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
17690  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17691  return __ret;
17692}
17693__ai int32x4_t __noswap_vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
17694  int32x4_t __ret;
17695  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
17696  return __ret;
17697}
17698#endif
17699
17700#ifdef __LITTLE_ENDIAN__
17701__ai uint16x4_t vqmovn_u32(uint32x4_t __p0) {
17702  uint16x4_t __ret;
17703  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 17);
17704  return __ret;
17705}
17706#else
17707__ai uint16x4_t vqmovn_u32(uint32x4_t __p0) {
17708  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17709  uint16x4_t __ret;
17710  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 17);
17711  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17712  return __ret;
17713}
17714__ai uint16x4_t __noswap_vqmovn_u32(uint32x4_t __p0) {
17715  uint16x4_t __ret;
17716  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 17);
17717  return __ret;
17718}
17719#endif
17720
17721#ifdef __LITTLE_ENDIAN__
17722__ai uint32x2_t vqmovn_u64(uint64x2_t __p0) {
17723  uint32x2_t __ret;
17724  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 18);
17725  return __ret;
17726}
17727#else
17728__ai uint32x2_t vqmovn_u64(uint64x2_t __p0) {
17729  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17730  uint32x2_t __ret;
17731  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 18);
17732  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17733  return __ret;
17734}
17735__ai uint32x2_t __noswap_vqmovn_u64(uint64x2_t __p0) {
17736  uint32x2_t __ret;
17737  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 18);
17738  return __ret;
17739}
17740#endif
17741
17742#ifdef __LITTLE_ENDIAN__
17743__ai uint8x8_t vqmovn_u16(uint16x8_t __p0) {
17744  uint8x8_t __ret;
17745  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 16);
17746  return __ret;
17747}
17748#else
17749__ai uint8x8_t vqmovn_u16(uint16x8_t __p0) {
17750  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17751  uint8x8_t __ret;
17752  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 16);
17753  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17754  return __ret;
17755}
17756__ai uint8x8_t __noswap_vqmovn_u16(uint16x8_t __p0) {
17757  uint8x8_t __ret;
17758  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 16);
17759  return __ret;
17760}
17761#endif
17762
17763#ifdef __LITTLE_ENDIAN__
17764__ai int16x4_t vqmovn_s32(int32x4_t __p0) {
17765  int16x4_t __ret;
17766  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 1);
17767  return __ret;
17768}
17769#else
17770__ai int16x4_t vqmovn_s32(int32x4_t __p0) {
17771  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17772  int16x4_t __ret;
17773  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 1);
17774  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17775  return __ret;
17776}
17777__ai int16x4_t __noswap_vqmovn_s32(int32x4_t __p0) {
17778  int16x4_t __ret;
17779  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 1);
17780  return __ret;
17781}
17782#endif
17783
17784#ifdef __LITTLE_ENDIAN__
17785__ai int32x2_t vqmovn_s64(int64x2_t __p0) {
17786  int32x2_t __ret;
17787  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 2);
17788  return __ret;
17789}
17790#else
17791__ai int32x2_t vqmovn_s64(int64x2_t __p0) {
17792  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17793  int32x2_t __ret;
17794  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 2);
17795  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17796  return __ret;
17797}
17798__ai int32x2_t __noswap_vqmovn_s64(int64x2_t __p0) {
17799  int32x2_t __ret;
17800  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 2);
17801  return __ret;
17802}
17803#endif
17804
17805#ifdef __LITTLE_ENDIAN__
17806__ai int8x8_t vqmovn_s16(int16x8_t __p0) {
17807  int8x8_t __ret;
17808  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 0);
17809  return __ret;
17810}
17811#else
17812__ai int8x8_t vqmovn_s16(int16x8_t __p0) {
17813  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17814  int8x8_t __ret;
17815  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 0);
17816  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17817  return __ret;
17818}
17819__ai int8x8_t __noswap_vqmovn_s16(int16x8_t __p0) {
17820  int8x8_t __ret;
17821  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 0);
17822  return __ret;
17823}
17824#endif
17825
17826#ifdef __LITTLE_ENDIAN__
17827__ai uint16x4_t vqmovun_s32(int32x4_t __p0) {
17828  uint16x4_t __ret;
17829  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 17);
17830  return __ret;
17831}
17832#else
17833__ai uint16x4_t vqmovun_s32(int32x4_t __p0) {
17834  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17835  uint16x4_t __ret;
17836  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 17);
17837  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17838  return __ret;
17839}
17840__ai uint16x4_t __noswap_vqmovun_s32(int32x4_t __p0) {
17841  uint16x4_t __ret;
17842  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 17);
17843  return __ret;
17844}
17845#endif
17846
17847#ifdef __LITTLE_ENDIAN__
17848__ai uint32x2_t vqmovun_s64(int64x2_t __p0) {
17849  uint32x2_t __ret;
17850  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 18);
17851  return __ret;
17852}
17853#else
17854__ai uint32x2_t vqmovun_s64(int64x2_t __p0) {
17855  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17856  uint32x2_t __ret;
17857  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 18);
17858  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17859  return __ret;
17860}
17861__ai uint32x2_t __noswap_vqmovun_s64(int64x2_t __p0) {
17862  uint32x2_t __ret;
17863  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 18);
17864  return __ret;
17865}
17866#endif
17867
17868#ifdef __LITTLE_ENDIAN__
17869__ai uint8x8_t vqmovun_s16(int16x8_t __p0) {
17870  uint8x8_t __ret;
17871  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 16);
17872  return __ret;
17873}
17874#else
17875__ai uint8x8_t vqmovun_s16(int16x8_t __p0) {
17876  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17877  uint8x8_t __ret;
17878  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 16);
17879  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17880  return __ret;
17881}
17882__ai uint8x8_t __noswap_vqmovun_s16(int16x8_t __p0) {
17883  uint8x8_t __ret;
17884  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 16);
17885  return __ret;
17886}
17887#endif
17888
17889#ifdef __LITTLE_ENDIAN__
17890__ai int8x16_t vqnegq_s8(int8x16_t __p0) {
17891  int8x16_t __ret;
17892  __ret = (int8x16_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 32);
17893  return __ret;
17894}
17895#else
17896__ai int8x16_t vqnegq_s8(int8x16_t __p0) {
17897  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17898  int8x16_t __ret;
17899  __ret = (int8x16_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 32);
17900  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17901  return __ret;
17902}
17903#endif
17904
17905#ifdef __LITTLE_ENDIAN__
17906__ai int32x4_t vqnegq_s32(int32x4_t __p0) {
17907  int32x4_t __ret;
17908  __ret = (int32x4_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 34);
17909  return __ret;
17910}
17911#else
17912__ai int32x4_t vqnegq_s32(int32x4_t __p0) {
17913  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17914  int32x4_t __ret;
17915  __ret = (int32x4_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 34);
17916  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17917  return __ret;
17918}
17919#endif
17920
17921#ifdef __LITTLE_ENDIAN__
17922__ai int16x8_t vqnegq_s16(int16x8_t __p0) {
17923  int16x8_t __ret;
17924  __ret = (int16x8_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 33);
17925  return __ret;
17926}
17927#else
17928__ai int16x8_t vqnegq_s16(int16x8_t __p0) {
17929  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17930  int16x8_t __ret;
17931  __ret = (int16x8_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 33);
17932  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17933  return __ret;
17934}
17935#endif
17936
17937#ifdef __LITTLE_ENDIAN__
17938__ai int8x8_t vqneg_s8(int8x8_t __p0) {
17939  int8x8_t __ret;
17940  __ret = (int8x8_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 0);
17941  return __ret;
17942}
17943#else
17944__ai int8x8_t vqneg_s8(int8x8_t __p0) {
17945  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17946  int8x8_t __ret;
17947  __ret = (int8x8_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 0);
17948  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17949  return __ret;
17950}
17951#endif
17952
17953#ifdef __LITTLE_ENDIAN__
17954__ai int32x2_t vqneg_s32(int32x2_t __p0) {
17955  int32x2_t __ret;
17956  __ret = (int32x2_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 2);
17957  return __ret;
17958}
17959#else
17960__ai int32x2_t vqneg_s32(int32x2_t __p0) {
17961  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17962  int32x2_t __ret;
17963  __ret = (int32x2_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 2);
17964  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17965  return __ret;
17966}
17967#endif
17968
17969#ifdef __LITTLE_ENDIAN__
17970__ai int16x4_t vqneg_s16(int16x4_t __p0) {
17971  int16x4_t __ret;
17972  __ret = (int16x4_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 1);
17973  return __ret;
17974}
17975#else
17976__ai int16x4_t vqneg_s16(int16x4_t __p0) {
17977  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17978  int16x4_t __ret;
17979  __ret = (int16x4_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 1);
17980  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17981  return __ret;
17982}
17983#endif
17984
17985#ifdef __LITTLE_ENDIAN__
17986__ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17987  int32x4_t __ret;
17988  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
17989  return __ret;
17990}
17991#else
17992__ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17993  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17994  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17995  int32x4_t __ret;
17996  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
17997  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17998  return __ret;
17999}
18000__ai int32x4_t __noswap_vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
18001  int32x4_t __ret;
18002  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18003  return __ret;
18004}
18005#endif
18006
18007#ifdef __LITTLE_ENDIAN__
18008__ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
18009  int16x8_t __ret;
18010  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18011  return __ret;
18012}
18013#else
18014__ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
18015  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18016  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18017  int16x8_t __ret;
18018  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
18019  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18020  return __ret;
18021}
18022__ai int16x8_t __noswap_vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
18023  int16x8_t __ret;
18024  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18025  return __ret;
18026}
18027#endif
18028
18029#ifdef __LITTLE_ENDIAN__
18030__ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
18031  int32x2_t __ret;
18032  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18033  return __ret;
18034}
18035#else
18036__ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
18037  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18038  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18039  int32x2_t __ret;
18040  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18041  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18042  return __ret;
18043}
18044__ai int32x2_t __noswap_vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
18045  int32x2_t __ret;
18046  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18047  return __ret;
18048}
18049#endif
18050
18051#ifdef __LITTLE_ENDIAN__
18052__ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
18053  int16x4_t __ret;
18054  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18055  return __ret;
18056}
18057#else
18058__ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
18059  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18060  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18061  int16x4_t __ret;
18062  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18063  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18064  return __ret;
18065}
18066__ai int16x4_t __noswap_vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
18067  int16x4_t __ret;
18068  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18069  return __ret;
18070}
18071#endif
18072
18073#ifdef __LITTLE_ENDIAN__
18074#define vqrdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18075  int32x4_t __s0 = __p0; \
18076  int32x2_t __s1 = __p1; \
18077  int32x4_t __ret; \
18078  __ret = vqrdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
18079  __ret; \
18080})
18081#else
18082#define vqrdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18083  int32x4_t __s0 = __p0; \
18084  int32x2_t __s1 = __p1; \
18085  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18086  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
18087  int32x4_t __ret; \
18088  __ret = __noswap_vqrdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
18089  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18090  __ret; \
18091})
18092#endif
18093
18094#ifdef __LITTLE_ENDIAN__
18095#define vqrdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18096  int16x8_t __s0 = __p0; \
18097  int16x4_t __s1 = __p1; \
18098  int16x8_t __ret; \
18099  __ret = vqrdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
18100  __ret; \
18101})
18102#else
18103#define vqrdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18104  int16x8_t __s0 = __p0; \
18105  int16x4_t __s1 = __p1; \
18106  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18107  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
18108  int16x8_t __ret; \
18109  __ret = __noswap_vqrdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
18110  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18111  __ret; \
18112})
18113#endif
18114
18115#ifdef __LITTLE_ENDIAN__
18116#define vqrdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18117  int32x2_t __s0 = __p0; \
18118  int32x2_t __s1 = __p1; \
18119  int32x2_t __ret; \
18120  __ret = vqrdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
18121  __ret; \
18122})
18123#else
18124#define vqrdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18125  int32x2_t __s0 = __p0; \
18126  int32x2_t __s1 = __p1; \
18127  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18128  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
18129  int32x2_t __ret; \
18130  __ret = __noswap_vqrdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
18131  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18132  __ret; \
18133})
18134#endif
18135
18136#ifdef __LITTLE_ENDIAN__
18137#define vqrdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18138  int16x4_t __s0 = __p0; \
18139  int16x4_t __s1 = __p1; \
18140  int16x4_t __ret; \
18141  __ret = vqrdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
18142  __ret; \
18143})
18144#else
18145#define vqrdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18146  int16x4_t __s0 = __p0; \
18147  int16x4_t __s1 = __p1; \
18148  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18149  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
18150  int16x4_t __ret; \
18151  __ret = __noswap_vqrdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
18152  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18153  __ret; \
18154})
18155#endif
18156
18157#ifdef __LITTLE_ENDIAN__
18158__ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
18159  int32x4_t __ret;
18160  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
18161  return __ret;
18162}
18163#else
18164__ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
18165  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18166  int32x4_t __ret;
18167  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
18168  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18169  return __ret;
18170}
18171#endif
18172
18173#ifdef __LITTLE_ENDIAN__
18174__ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
18175  int16x8_t __ret;
18176  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
18177  return __ret;
18178}
18179#else
18180__ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
18181  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18182  int16x8_t __ret;
18183  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
18184  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18185  return __ret;
18186}
18187#endif
18188
18189#ifdef __LITTLE_ENDIAN__
18190__ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
18191  int32x2_t __ret;
18192  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
18193  return __ret;
18194}
18195#else
18196__ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
18197  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18198  int32x2_t __ret;
18199  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
18200  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18201  return __ret;
18202}
18203#endif
18204
18205#ifdef __LITTLE_ENDIAN__
18206__ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
18207  int16x4_t __ret;
18208  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
18209  return __ret;
18210}
18211#else
18212__ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
18213  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18214  int16x4_t __ret;
18215  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
18216  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18217  return __ret;
18218}
18219#endif
18220
18221#ifdef __LITTLE_ENDIAN__
18222__ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18223  uint8x16_t __ret;
18224  __ret = (uint8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
18225  return __ret;
18226}
18227#else
18228__ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18229  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18230  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18231  uint8x16_t __ret;
18232  __ret = (uint8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
18233  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18234  return __ret;
18235}
18236#endif
18237
18238#ifdef __LITTLE_ENDIAN__
18239__ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18240  uint32x4_t __ret;
18241  __ret = (uint32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
18242  return __ret;
18243}
18244#else
18245__ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18246  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18247  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18248  uint32x4_t __ret;
18249  __ret = (uint32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
18250  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18251  return __ret;
18252}
18253#endif
18254
18255#ifdef __LITTLE_ENDIAN__
18256__ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18257  uint64x2_t __ret;
18258  __ret = (uint64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
18259  return __ret;
18260}
18261#else
18262__ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18263  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18264  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18265  uint64x2_t __ret;
18266  __ret = (uint64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
18267  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18268  return __ret;
18269}
18270#endif
18271
18272#ifdef __LITTLE_ENDIAN__
18273__ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18274  uint16x8_t __ret;
18275  __ret = (uint16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
18276  return __ret;
18277}
18278#else
18279__ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18280  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18281  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18282  uint16x8_t __ret;
18283  __ret = (uint16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
18284  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18285  return __ret;
18286}
18287#endif
18288
18289#ifdef __LITTLE_ENDIAN__
18290__ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18291  int8x16_t __ret;
18292  __ret = (int8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
18293  return __ret;
18294}
18295#else
18296__ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18297  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18298  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18299  int8x16_t __ret;
18300  __ret = (int8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
18301  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18302  return __ret;
18303}
18304#endif
18305
18306#ifdef __LITTLE_ENDIAN__
18307__ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18308  int32x4_t __ret;
18309  __ret = (int32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18310  return __ret;
18311}
18312#else
18313__ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18314  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18315  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18316  int32x4_t __ret;
18317  __ret = (int32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
18318  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18319  return __ret;
18320}
18321#endif
18322
18323#ifdef __LITTLE_ENDIAN__
18324__ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18325  int64x2_t __ret;
18326  __ret = (int64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
18327  return __ret;
18328}
18329#else
18330__ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18331  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18332  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18333  int64x2_t __ret;
18334  __ret = (int64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
18335  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18336  return __ret;
18337}
18338#endif
18339
18340#ifdef __LITTLE_ENDIAN__
18341__ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18342  int16x8_t __ret;
18343  __ret = (int16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18344  return __ret;
18345}
18346#else
18347__ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18348  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18349  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18350  int16x8_t __ret;
18351  __ret = (int16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
18352  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18353  return __ret;
18354}
18355#endif
18356
18357#ifdef __LITTLE_ENDIAN__
18358__ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18359  uint8x8_t __ret;
18360  __ret = (uint8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18361  return __ret;
18362}
18363#else
18364__ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18365  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18366  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18367  uint8x8_t __ret;
18368  __ret = (uint8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18369  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18370  return __ret;
18371}
18372#endif
18373
18374#ifdef __LITTLE_ENDIAN__
18375__ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18376  uint32x2_t __ret;
18377  __ret = (uint32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18378  return __ret;
18379}
18380#else
18381__ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18382  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18383  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18384  uint32x2_t __ret;
18385  __ret = (uint32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18386  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18387  return __ret;
18388}
18389#endif
18390
18391#ifdef __LITTLE_ENDIAN__
18392__ai uint64x1_t vqrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18393  uint64x1_t __ret;
18394  __ret = (uint64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18395  return __ret;
18396}
18397#else
18398__ai uint64x1_t vqrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18399  uint64x1_t __ret;
18400  __ret = (uint64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18401  return __ret;
18402}
18403#endif
18404
18405#ifdef __LITTLE_ENDIAN__
18406__ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18407  uint16x4_t __ret;
18408  __ret = (uint16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18409  return __ret;
18410}
18411#else
18412__ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18413  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18414  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18415  uint16x4_t __ret;
18416  __ret = (uint16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18417  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18418  return __ret;
18419}
18420#endif
18421
18422#ifdef __LITTLE_ENDIAN__
18423__ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1) {
18424  int8x8_t __ret;
18425  __ret = (int8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18426  return __ret;
18427}
18428#else
18429__ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1) {
18430  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18431  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18432  int8x8_t __ret;
18433  __ret = (int8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18434  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18435  return __ret;
18436}
18437#endif
18438
18439#ifdef __LITTLE_ENDIAN__
18440__ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1) {
18441  int32x2_t __ret;
18442  __ret = (int32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18443  return __ret;
18444}
18445#else
18446__ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1) {
18447  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18448  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18449  int32x2_t __ret;
18450  __ret = (int32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18451  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18452  return __ret;
18453}
18454#endif
18455
18456#ifdef __LITTLE_ENDIAN__
18457__ai int64x1_t vqrshl_s64(int64x1_t __p0, int64x1_t __p1) {
18458  int64x1_t __ret;
18459  __ret = (int64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18460  return __ret;
18461}
18462#else
18463__ai int64x1_t vqrshl_s64(int64x1_t __p0, int64x1_t __p1) {
18464  int64x1_t __ret;
18465  __ret = (int64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18466  return __ret;
18467}
18468#endif
18469
18470#ifdef __LITTLE_ENDIAN__
18471__ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1) {
18472  int16x4_t __ret;
18473  __ret = (int16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18474  return __ret;
18475}
18476#else
18477__ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1) {
18478  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18479  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18480  int16x4_t __ret;
18481  __ret = (int16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18482  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18483  return __ret;
18484}
18485#endif
18486
18487#ifdef __LITTLE_ENDIAN__
18488#define vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
18489  uint32x4_t __s0 = __p0; \
18490  uint16x4_t __ret; \
18491  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 17); \
18492  __ret; \
18493})
18494#else
18495#define vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
18496  uint32x4_t __s0 = __p0; \
18497  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18498  uint16x4_t __ret; \
18499  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 17); \
18500  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18501  __ret; \
18502})
18503#define __noswap_vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
18504  uint32x4_t __s0 = __p0; \
18505  uint16x4_t __ret; \
18506  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 17); \
18507  __ret; \
18508})
18509#endif
18510
18511#ifdef __LITTLE_ENDIAN__
18512#define vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
18513  uint64x2_t __s0 = __p0; \
18514  uint32x2_t __ret; \
18515  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 18); \
18516  __ret; \
18517})
18518#else
18519#define vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
18520  uint64x2_t __s0 = __p0; \
18521  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18522  uint32x2_t __ret; \
18523  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 18); \
18524  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18525  __ret; \
18526})
18527#define __noswap_vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
18528  uint64x2_t __s0 = __p0; \
18529  uint32x2_t __ret; \
18530  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 18); \
18531  __ret; \
18532})
18533#endif
18534
18535#ifdef __LITTLE_ENDIAN__
18536#define vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
18537  uint16x8_t __s0 = __p0; \
18538  uint8x8_t __ret; \
18539  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 16); \
18540  __ret; \
18541})
18542#else
18543#define vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
18544  uint16x8_t __s0 = __p0; \
18545  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18546  uint8x8_t __ret; \
18547  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 16); \
18548  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18549  __ret; \
18550})
18551#define __noswap_vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
18552  uint16x8_t __s0 = __p0; \
18553  uint8x8_t __ret; \
18554  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 16); \
18555  __ret; \
18556})
18557#endif
18558
18559#ifdef __LITTLE_ENDIAN__
18560#define vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
18561  int32x4_t __s0 = __p0; \
18562  int16x4_t __ret; \
18563  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 1); \
18564  __ret; \
18565})
18566#else
18567#define vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
18568  int32x4_t __s0 = __p0; \
18569  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18570  int16x4_t __ret; \
18571  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 1); \
18572  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18573  __ret; \
18574})
18575#define __noswap_vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
18576  int32x4_t __s0 = __p0; \
18577  int16x4_t __ret; \
18578  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 1); \
18579  __ret; \
18580})
18581#endif
18582
18583#ifdef __LITTLE_ENDIAN__
18584#define vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
18585  int64x2_t __s0 = __p0; \
18586  int32x2_t __ret; \
18587  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 2); \
18588  __ret; \
18589})
18590#else
18591#define vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
18592  int64x2_t __s0 = __p0; \
18593  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18594  int32x2_t __ret; \
18595  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 2); \
18596  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18597  __ret; \
18598})
18599#define __noswap_vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
18600  int64x2_t __s0 = __p0; \
18601  int32x2_t __ret; \
18602  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 2); \
18603  __ret; \
18604})
18605#endif
18606
18607#ifdef __LITTLE_ENDIAN__
18608#define vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
18609  int16x8_t __s0 = __p0; \
18610  int8x8_t __ret; \
18611  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 0); \
18612  __ret; \
18613})
18614#else
18615#define vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
18616  int16x8_t __s0 = __p0; \
18617  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18618  int8x8_t __ret; \
18619  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 0); \
18620  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18621  __ret; \
18622})
18623#define __noswap_vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
18624  int16x8_t __s0 = __p0; \
18625  int8x8_t __ret; \
18626  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 0); \
18627  __ret; \
18628})
18629#endif
18630
18631#ifdef __LITTLE_ENDIAN__
18632#define vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
18633  int32x4_t __s0 = __p0; \
18634  uint16x4_t __ret; \
18635  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 17); \
18636  __ret; \
18637})
18638#else
18639#define vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
18640  int32x4_t __s0 = __p0; \
18641  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18642  uint16x4_t __ret; \
18643  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 17); \
18644  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18645  __ret; \
18646})
18647#define __noswap_vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
18648  int32x4_t __s0 = __p0; \
18649  uint16x4_t __ret; \
18650  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 17); \
18651  __ret; \
18652})
18653#endif
18654
18655#ifdef __LITTLE_ENDIAN__
18656#define vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
18657  int64x2_t __s0 = __p0; \
18658  uint32x2_t __ret; \
18659  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 18); \
18660  __ret; \
18661})
18662#else
18663#define vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
18664  int64x2_t __s0 = __p0; \
18665  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18666  uint32x2_t __ret; \
18667  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 18); \
18668  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18669  __ret; \
18670})
18671#define __noswap_vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
18672  int64x2_t __s0 = __p0; \
18673  uint32x2_t __ret; \
18674  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 18); \
18675  __ret; \
18676})
18677#endif
18678
18679#ifdef __LITTLE_ENDIAN__
18680#define vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
18681  int16x8_t __s0 = __p0; \
18682  uint8x8_t __ret; \
18683  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 16); \
18684  __ret; \
18685})
18686#else
18687#define vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
18688  int16x8_t __s0 = __p0; \
18689  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18690  uint8x8_t __ret; \
18691  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 16); \
18692  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18693  __ret; \
18694})
18695#define __noswap_vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
18696  int16x8_t __s0 = __p0; \
18697  uint8x8_t __ret; \
18698  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 16); \
18699  __ret; \
18700})
18701#endif
18702
18703#ifdef __LITTLE_ENDIAN__
18704__ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18705  uint8x16_t __ret;
18706  __ret = (uint8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
18707  return __ret;
18708}
18709#else
18710__ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18711  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18712  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18713  uint8x16_t __ret;
18714  __ret = (uint8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
18715  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18716  return __ret;
18717}
18718#endif
18719
18720#ifdef __LITTLE_ENDIAN__
18721__ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18722  uint32x4_t __ret;
18723  __ret = (uint32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
18724  return __ret;
18725}
18726#else
18727__ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18728  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18729  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18730  uint32x4_t __ret;
18731  __ret = (uint32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
18732  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18733  return __ret;
18734}
18735#endif
18736
18737#ifdef __LITTLE_ENDIAN__
18738__ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18739  uint64x2_t __ret;
18740  __ret = (uint64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
18741  return __ret;
18742}
18743#else
18744__ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18745  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18746  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18747  uint64x2_t __ret;
18748  __ret = (uint64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
18749  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18750  return __ret;
18751}
18752#endif
18753
18754#ifdef __LITTLE_ENDIAN__
18755__ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18756  uint16x8_t __ret;
18757  __ret = (uint16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
18758  return __ret;
18759}
18760#else
18761__ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18762  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18763  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18764  uint16x8_t __ret;
18765  __ret = (uint16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
18766  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18767  return __ret;
18768}
18769#endif
18770
18771#ifdef __LITTLE_ENDIAN__
18772__ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18773  int8x16_t __ret;
18774  __ret = (int8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
18775  return __ret;
18776}
18777#else
18778__ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18779  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18780  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18781  int8x16_t __ret;
18782  __ret = (int8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
18783  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18784  return __ret;
18785}
18786#endif
18787
18788#ifdef __LITTLE_ENDIAN__
18789__ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18790  int32x4_t __ret;
18791  __ret = (int32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18792  return __ret;
18793}
18794#else
18795__ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18796  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18797  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18798  int32x4_t __ret;
18799  __ret = (int32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
18800  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18801  return __ret;
18802}
18803#endif
18804
18805#ifdef __LITTLE_ENDIAN__
18806__ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18807  int64x2_t __ret;
18808  __ret = (int64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
18809  return __ret;
18810}
18811#else
18812__ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18813  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18814  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18815  int64x2_t __ret;
18816  __ret = (int64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
18817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18818  return __ret;
18819}
18820#endif
18821
18822#ifdef __LITTLE_ENDIAN__
18823__ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18824  int16x8_t __ret;
18825  __ret = (int16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18826  return __ret;
18827}
18828#else
18829__ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18830  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18831  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18832  int16x8_t __ret;
18833  __ret = (int16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
18834  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18835  return __ret;
18836}
18837#endif
18838
18839#ifdef __LITTLE_ENDIAN__
18840__ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18841  uint8x8_t __ret;
18842  __ret = (uint8x8_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18843  return __ret;
18844}
18845#else
18846__ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18847  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18848  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18849  uint8x8_t __ret;
18850  __ret = (uint8x8_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18851  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18852  return __ret;
18853}
18854#endif
18855
18856#ifdef __LITTLE_ENDIAN__
18857__ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18858  uint32x2_t __ret;
18859  __ret = (uint32x2_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18860  return __ret;
18861}
18862#else
18863__ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18864  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18865  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18866  uint32x2_t __ret;
18867  __ret = (uint32x2_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18868  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18869  return __ret;
18870}
18871#endif
18872
18873#ifdef __LITTLE_ENDIAN__
18874__ai uint64x1_t vqshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18875  uint64x1_t __ret;
18876  __ret = (uint64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18877  return __ret;
18878}
18879#else
18880__ai uint64x1_t vqshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18881  uint64x1_t __ret;
18882  __ret = (uint64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18883  return __ret;
18884}
18885#endif
18886
18887#ifdef __LITTLE_ENDIAN__
18888__ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18889  uint16x4_t __ret;
18890  __ret = (uint16x4_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18891  return __ret;
18892}
18893#else
18894__ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18895  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18896  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18897  uint16x4_t __ret;
18898  __ret = (uint16x4_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18899  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18900  return __ret;
18901}
18902#endif
18903
18904#ifdef __LITTLE_ENDIAN__
18905__ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1) {
18906  int8x8_t __ret;
18907  __ret = (int8x8_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18908  return __ret;
18909}
18910#else
18911__ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1) {
18912  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18913  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18914  int8x8_t __ret;
18915  __ret = (int8x8_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18916  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18917  return __ret;
18918}
18919#endif
18920
18921#ifdef __LITTLE_ENDIAN__
18922__ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1) {
18923  int32x2_t __ret;
18924  __ret = (int32x2_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18925  return __ret;
18926}
18927#else
18928__ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1) {
18929  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18930  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18931  int32x2_t __ret;
18932  __ret = (int32x2_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18933  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18934  return __ret;
18935}
18936#endif
18937
18938#ifdef __LITTLE_ENDIAN__
18939__ai int64x1_t vqshl_s64(int64x1_t __p0, int64x1_t __p1) {
18940  int64x1_t __ret;
18941  __ret = (int64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18942  return __ret;
18943}
18944#else
18945__ai int64x1_t vqshl_s64(int64x1_t __p0, int64x1_t __p1) {
18946  int64x1_t __ret;
18947  __ret = (int64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18948  return __ret;
18949}
18950#endif
18951
18952#ifdef __LITTLE_ENDIAN__
18953__ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1) {
18954  int16x4_t __ret;
18955  __ret = (int16x4_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18956  return __ret;
18957}
18958#else
18959__ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1) {
18960  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18961  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18962  int16x4_t __ret;
18963  __ret = (int16x4_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18964  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18965  return __ret;
18966}
18967#endif
18968
18969#ifdef __LITTLE_ENDIAN__
18970#define vqshlq_n_u8(__p0, __p1) __extension__ ({ \
18971  uint8x16_t __s0 = __p0; \
18972  uint8x16_t __ret; \
18973  __ret = (uint8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 48); \
18974  __ret; \
18975})
18976#else
18977#define vqshlq_n_u8(__p0, __p1) __extension__ ({ \
18978  uint8x16_t __s0 = __p0; \
18979  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
18980  uint8x16_t __ret; \
18981  __ret = (uint8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 48); \
18982  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
18983  __ret; \
18984})
18985#endif
18986
18987#ifdef __LITTLE_ENDIAN__
18988#define vqshlq_n_u32(__p0, __p1) __extension__ ({ \
18989  uint32x4_t __s0 = __p0; \
18990  uint32x4_t __ret; \
18991  __ret = (uint32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 50); \
18992  __ret; \
18993})
18994#else
18995#define vqshlq_n_u32(__p0, __p1) __extension__ ({ \
18996  uint32x4_t __s0 = __p0; \
18997  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18998  uint32x4_t __ret; \
18999  __ret = (uint32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 50); \
19000  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19001  __ret; \
19002})
19003#endif
19004
19005#ifdef __LITTLE_ENDIAN__
19006#define vqshlq_n_u64(__p0, __p1) __extension__ ({ \
19007  uint64x2_t __s0 = __p0; \
19008  uint64x2_t __ret; \
19009  __ret = (uint64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 51); \
19010  __ret; \
19011})
19012#else
19013#define vqshlq_n_u64(__p0, __p1) __extension__ ({ \
19014  uint64x2_t __s0 = __p0; \
19015  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19016  uint64x2_t __ret; \
19017  __ret = (uint64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 51); \
19018  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19019  __ret; \
19020})
19021#endif
19022
19023#ifdef __LITTLE_ENDIAN__
19024#define vqshlq_n_u16(__p0, __p1) __extension__ ({ \
19025  uint16x8_t __s0 = __p0; \
19026  uint16x8_t __ret; \
19027  __ret = (uint16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 49); \
19028  __ret; \
19029})
19030#else
19031#define vqshlq_n_u16(__p0, __p1) __extension__ ({ \
19032  uint16x8_t __s0 = __p0; \
19033  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19034  uint16x8_t __ret; \
19035  __ret = (uint16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 49); \
19036  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19037  __ret; \
19038})
19039#endif
19040
19041#ifdef __LITTLE_ENDIAN__
19042#define vqshlq_n_s8(__p0, __p1) __extension__ ({ \
19043  int8x16_t __s0 = __p0; \
19044  int8x16_t __ret; \
19045  __ret = (int8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 32); \
19046  __ret; \
19047})
19048#else
19049#define vqshlq_n_s8(__p0, __p1) __extension__ ({ \
19050  int8x16_t __s0 = __p0; \
19051  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19052  int8x16_t __ret; \
19053  __ret = (int8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 32); \
19054  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19055  __ret; \
19056})
19057#endif
19058
19059#ifdef __LITTLE_ENDIAN__
19060#define vqshlq_n_s32(__p0, __p1) __extension__ ({ \
19061  int32x4_t __s0 = __p0; \
19062  int32x4_t __ret; \
19063  __ret = (int32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 34); \
19064  __ret; \
19065})
19066#else
19067#define vqshlq_n_s32(__p0, __p1) __extension__ ({ \
19068  int32x4_t __s0 = __p0; \
19069  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19070  int32x4_t __ret; \
19071  __ret = (int32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 34); \
19072  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19073  __ret; \
19074})
19075#endif
19076
19077#ifdef __LITTLE_ENDIAN__
19078#define vqshlq_n_s64(__p0, __p1) __extension__ ({ \
19079  int64x2_t __s0 = __p0; \
19080  int64x2_t __ret; \
19081  __ret = (int64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 35); \
19082  __ret; \
19083})
19084#else
19085#define vqshlq_n_s64(__p0, __p1) __extension__ ({ \
19086  int64x2_t __s0 = __p0; \
19087  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19088  int64x2_t __ret; \
19089  __ret = (int64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 35); \
19090  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19091  __ret; \
19092})
19093#endif
19094
19095#ifdef __LITTLE_ENDIAN__
19096#define vqshlq_n_s16(__p0, __p1) __extension__ ({ \
19097  int16x8_t __s0 = __p0; \
19098  int16x8_t __ret; \
19099  __ret = (int16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 33); \
19100  __ret; \
19101})
19102#else
19103#define vqshlq_n_s16(__p0, __p1) __extension__ ({ \
19104  int16x8_t __s0 = __p0; \
19105  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19106  int16x8_t __ret; \
19107  __ret = (int16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 33); \
19108  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19109  __ret; \
19110})
19111#endif
19112
19113#ifdef __LITTLE_ENDIAN__
19114#define vqshl_n_u8(__p0, __p1) __extension__ ({ \
19115  uint8x8_t __s0 = __p0; \
19116  uint8x8_t __ret; \
19117  __ret = (uint8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 16); \
19118  __ret; \
19119})
19120#else
19121#define vqshl_n_u8(__p0, __p1) __extension__ ({ \
19122  uint8x8_t __s0 = __p0; \
19123  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19124  uint8x8_t __ret; \
19125  __ret = (uint8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 16); \
19126  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19127  __ret; \
19128})
19129#endif
19130
19131#ifdef __LITTLE_ENDIAN__
19132#define vqshl_n_u32(__p0, __p1) __extension__ ({ \
19133  uint32x2_t __s0 = __p0; \
19134  uint32x2_t __ret; \
19135  __ret = (uint32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 18); \
19136  __ret; \
19137})
19138#else
19139#define vqshl_n_u32(__p0, __p1) __extension__ ({ \
19140  uint32x2_t __s0 = __p0; \
19141  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19142  uint32x2_t __ret; \
19143  __ret = (uint32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 18); \
19144  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19145  __ret; \
19146})
19147#endif
19148
19149#ifdef __LITTLE_ENDIAN__
19150#define vqshl_n_u64(__p0, __p1) __extension__ ({ \
19151  uint64x1_t __s0 = __p0; \
19152  uint64x1_t __ret; \
19153  __ret = (uint64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 19); \
19154  __ret; \
19155})
19156#else
19157#define vqshl_n_u64(__p0, __p1) __extension__ ({ \
19158  uint64x1_t __s0 = __p0; \
19159  uint64x1_t __ret; \
19160  __ret = (uint64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 19); \
19161  __ret; \
19162})
19163#endif
19164
19165#ifdef __LITTLE_ENDIAN__
19166#define vqshl_n_u16(__p0, __p1) __extension__ ({ \
19167  uint16x4_t __s0 = __p0; \
19168  uint16x4_t __ret; \
19169  __ret = (uint16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 17); \
19170  __ret; \
19171})
19172#else
19173#define vqshl_n_u16(__p0, __p1) __extension__ ({ \
19174  uint16x4_t __s0 = __p0; \
19175  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19176  uint16x4_t __ret; \
19177  __ret = (uint16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 17); \
19178  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19179  __ret; \
19180})
19181#endif
19182
19183#ifdef __LITTLE_ENDIAN__
19184#define vqshl_n_s8(__p0, __p1) __extension__ ({ \
19185  int8x8_t __s0 = __p0; \
19186  int8x8_t __ret; \
19187  __ret = (int8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 0); \
19188  __ret; \
19189})
19190#else
19191#define vqshl_n_s8(__p0, __p1) __extension__ ({ \
19192  int8x8_t __s0 = __p0; \
19193  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19194  int8x8_t __ret; \
19195  __ret = (int8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 0); \
19196  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19197  __ret; \
19198})
19199#endif
19200
19201#ifdef __LITTLE_ENDIAN__
19202#define vqshl_n_s32(__p0, __p1) __extension__ ({ \
19203  int32x2_t __s0 = __p0; \
19204  int32x2_t __ret; \
19205  __ret = (int32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 2); \
19206  __ret; \
19207})
19208#else
19209#define vqshl_n_s32(__p0, __p1) __extension__ ({ \
19210  int32x2_t __s0 = __p0; \
19211  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19212  int32x2_t __ret; \
19213  __ret = (int32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 2); \
19214  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19215  __ret; \
19216})
19217#endif
19218
19219#ifdef __LITTLE_ENDIAN__
19220#define vqshl_n_s64(__p0, __p1) __extension__ ({ \
19221  int64x1_t __s0 = __p0; \
19222  int64x1_t __ret; \
19223  __ret = (int64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 3); \
19224  __ret; \
19225})
19226#else
19227#define vqshl_n_s64(__p0, __p1) __extension__ ({ \
19228  int64x1_t __s0 = __p0; \
19229  int64x1_t __ret; \
19230  __ret = (int64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 3); \
19231  __ret; \
19232})
19233#endif
19234
19235#ifdef __LITTLE_ENDIAN__
19236#define vqshl_n_s16(__p0, __p1) __extension__ ({ \
19237  int16x4_t __s0 = __p0; \
19238  int16x4_t __ret; \
19239  __ret = (int16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 1); \
19240  __ret; \
19241})
19242#else
19243#define vqshl_n_s16(__p0, __p1) __extension__ ({ \
19244  int16x4_t __s0 = __p0; \
19245  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19246  int16x4_t __ret; \
19247  __ret = (int16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 1); \
19248  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19249  __ret; \
19250})
19251#endif
19252
19253#ifdef __LITTLE_ENDIAN__
19254#define vqshluq_n_s8(__p0, __p1) __extension__ ({ \
19255  int8x16_t __s0 = __p0; \
19256  uint8x16_t __ret; \
19257  __ret = (uint8x16_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 48); \
19258  __ret; \
19259})
19260#else
19261#define vqshluq_n_s8(__p0, __p1) __extension__ ({ \
19262  int8x16_t __s0 = __p0; \
19263  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19264  uint8x16_t __ret; \
19265  __ret = (uint8x16_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 48); \
19266  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19267  __ret; \
19268})
19269#endif
19270
19271#ifdef __LITTLE_ENDIAN__
19272#define vqshluq_n_s32(__p0, __p1) __extension__ ({ \
19273  int32x4_t __s0 = __p0; \
19274  uint32x4_t __ret; \
19275  __ret = (uint32x4_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 50); \
19276  __ret; \
19277})
19278#else
19279#define vqshluq_n_s32(__p0, __p1) __extension__ ({ \
19280  int32x4_t __s0 = __p0; \
19281  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19282  uint32x4_t __ret; \
19283  __ret = (uint32x4_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 50); \
19284  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19285  __ret; \
19286})
19287#endif
19288
19289#ifdef __LITTLE_ENDIAN__
19290#define vqshluq_n_s64(__p0, __p1) __extension__ ({ \
19291  int64x2_t __s0 = __p0; \
19292  uint64x2_t __ret; \
19293  __ret = (uint64x2_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 51); \
19294  __ret; \
19295})
19296#else
19297#define vqshluq_n_s64(__p0, __p1) __extension__ ({ \
19298  int64x2_t __s0 = __p0; \
19299  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19300  uint64x2_t __ret; \
19301  __ret = (uint64x2_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 51); \
19302  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19303  __ret; \
19304})
19305#endif
19306
19307#ifdef __LITTLE_ENDIAN__
19308#define vqshluq_n_s16(__p0, __p1) __extension__ ({ \
19309  int16x8_t __s0 = __p0; \
19310  uint16x8_t __ret; \
19311  __ret = (uint16x8_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 49); \
19312  __ret; \
19313})
19314#else
19315#define vqshluq_n_s16(__p0, __p1) __extension__ ({ \
19316  int16x8_t __s0 = __p0; \
19317  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19318  uint16x8_t __ret; \
19319  __ret = (uint16x8_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 49); \
19320  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19321  __ret; \
19322})
19323#endif
19324
19325#ifdef __LITTLE_ENDIAN__
19326#define vqshlu_n_s8(__p0, __p1) __extension__ ({ \
19327  int8x8_t __s0 = __p0; \
19328  uint8x8_t __ret; \
19329  __ret = (uint8x8_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 16); \
19330  __ret; \
19331})
19332#else
19333#define vqshlu_n_s8(__p0, __p1) __extension__ ({ \
19334  int8x8_t __s0 = __p0; \
19335  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19336  uint8x8_t __ret; \
19337  __ret = (uint8x8_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 16); \
19338  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19339  __ret; \
19340})
19341#endif
19342
19343#ifdef __LITTLE_ENDIAN__
19344#define vqshlu_n_s32(__p0, __p1) __extension__ ({ \
19345  int32x2_t __s0 = __p0; \
19346  uint32x2_t __ret; \
19347  __ret = (uint32x2_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 18); \
19348  __ret; \
19349})
19350#else
19351#define vqshlu_n_s32(__p0, __p1) __extension__ ({ \
19352  int32x2_t __s0 = __p0; \
19353  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19354  uint32x2_t __ret; \
19355  __ret = (uint32x2_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 18); \
19356  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19357  __ret; \
19358})
19359#endif
19360
19361#ifdef __LITTLE_ENDIAN__
19362#define vqshlu_n_s64(__p0, __p1) __extension__ ({ \
19363  int64x1_t __s0 = __p0; \
19364  uint64x1_t __ret; \
19365  __ret = (uint64x1_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 19); \
19366  __ret; \
19367})
19368#else
19369#define vqshlu_n_s64(__p0, __p1) __extension__ ({ \
19370  int64x1_t __s0 = __p0; \
19371  uint64x1_t __ret; \
19372  __ret = (uint64x1_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 19); \
19373  __ret; \
19374})
19375#endif
19376
19377#ifdef __LITTLE_ENDIAN__
19378#define vqshlu_n_s16(__p0, __p1) __extension__ ({ \
19379  int16x4_t __s0 = __p0; \
19380  uint16x4_t __ret; \
19381  __ret = (uint16x4_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 17); \
19382  __ret; \
19383})
19384#else
19385#define vqshlu_n_s16(__p0, __p1) __extension__ ({ \
19386  int16x4_t __s0 = __p0; \
19387  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19388  uint16x4_t __ret; \
19389  __ret = (uint16x4_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 17); \
19390  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19391  __ret; \
19392})
19393#endif
19394
19395#ifdef __LITTLE_ENDIAN__
19396#define vqshrn_n_u32(__p0, __p1) __extension__ ({ \
19397  uint32x4_t __s0 = __p0; \
19398  uint16x4_t __ret; \
19399  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 17); \
19400  __ret; \
19401})
19402#else
19403#define vqshrn_n_u32(__p0, __p1) __extension__ ({ \
19404  uint32x4_t __s0 = __p0; \
19405  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19406  uint16x4_t __ret; \
19407  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 17); \
19408  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19409  __ret; \
19410})
19411#define __noswap_vqshrn_n_u32(__p0, __p1) __extension__ ({ \
19412  uint32x4_t __s0 = __p0; \
19413  uint16x4_t __ret; \
19414  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 17); \
19415  __ret; \
19416})
19417#endif
19418
19419#ifdef __LITTLE_ENDIAN__
19420#define vqshrn_n_u64(__p0, __p1) __extension__ ({ \
19421  uint64x2_t __s0 = __p0; \
19422  uint32x2_t __ret; \
19423  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 18); \
19424  __ret; \
19425})
19426#else
19427#define vqshrn_n_u64(__p0, __p1) __extension__ ({ \
19428  uint64x2_t __s0 = __p0; \
19429  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19430  uint32x2_t __ret; \
19431  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 18); \
19432  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19433  __ret; \
19434})
19435#define __noswap_vqshrn_n_u64(__p0, __p1) __extension__ ({ \
19436  uint64x2_t __s0 = __p0; \
19437  uint32x2_t __ret; \
19438  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 18); \
19439  __ret; \
19440})
19441#endif
19442
19443#ifdef __LITTLE_ENDIAN__
19444#define vqshrn_n_u16(__p0, __p1) __extension__ ({ \
19445  uint16x8_t __s0 = __p0; \
19446  uint8x8_t __ret; \
19447  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 16); \
19448  __ret; \
19449})
19450#else
19451#define vqshrn_n_u16(__p0, __p1) __extension__ ({ \
19452  uint16x8_t __s0 = __p0; \
19453  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19454  uint8x8_t __ret; \
19455  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 16); \
19456  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19457  __ret; \
19458})
19459#define __noswap_vqshrn_n_u16(__p0, __p1) __extension__ ({ \
19460  uint16x8_t __s0 = __p0; \
19461  uint8x8_t __ret; \
19462  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 16); \
19463  __ret; \
19464})
19465#endif
19466
19467#ifdef __LITTLE_ENDIAN__
19468#define vqshrn_n_s32(__p0, __p1) __extension__ ({ \
19469  int32x4_t __s0 = __p0; \
19470  int16x4_t __ret; \
19471  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 1); \
19472  __ret; \
19473})
19474#else
19475#define vqshrn_n_s32(__p0, __p1) __extension__ ({ \
19476  int32x4_t __s0 = __p0; \
19477  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19478  int16x4_t __ret; \
19479  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 1); \
19480  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19481  __ret; \
19482})
19483#define __noswap_vqshrn_n_s32(__p0, __p1) __extension__ ({ \
19484  int32x4_t __s0 = __p0; \
19485  int16x4_t __ret; \
19486  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 1); \
19487  __ret; \
19488})
19489#endif
19490
19491#ifdef __LITTLE_ENDIAN__
19492#define vqshrn_n_s64(__p0, __p1) __extension__ ({ \
19493  int64x2_t __s0 = __p0; \
19494  int32x2_t __ret; \
19495  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 2); \
19496  __ret; \
19497})
19498#else
19499#define vqshrn_n_s64(__p0, __p1) __extension__ ({ \
19500  int64x2_t __s0 = __p0; \
19501  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19502  int32x2_t __ret; \
19503  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 2); \
19504  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19505  __ret; \
19506})
19507#define __noswap_vqshrn_n_s64(__p0, __p1) __extension__ ({ \
19508  int64x2_t __s0 = __p0; \
19509  int32x2_t __ret; \
19510  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 2); \
19511  __ret; \
19512})
19513#endif
19514
19515#ifdef __LITTLE_ENDIAN__
19516#define vqshrn_n_s16(__p0, __p1) __extension__ ({ \
19517  int16x8_t __s0 = __p0; \
19518  int8x8_t __ret; \
19519  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 0); \
19520  __ret; \
19521})
19522#else
19523#define vqshrn_n_s16(__p0, __p1) __extension__ ({ \
19524  int16x8_t __s0 = __p0; \
19525  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19526  int8x8_t __ret; \
19527  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 0); \
19528  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19529  __ret; \
19530})
19531#define __noswap_vqshrn_n_s16(__p0, __p1) __extension__ ({ \
19532  int16x8_t __s0 = __p0; \
19533  int8x8_t __ret; \
19534  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 0); \
19535  __ret; \
19536})
19537#endif
19538
19539#ifdef __LITTLE_ENDIAN__
19540#define vqshrun_n_s32(__p0, __p1) __extension__ ({ \
19541  int32x4_t __s0 = __p0; \
19542  uint16x4_t __ret; \
19543  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 17); \
19544  __ret; \
19545})
19546#else
19547#define vqshrun_n_s32(__p0, __p1) __extension__ ({ \
19548  int32x4_t __s0 = __p0; \
19549  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19550  uint16x4_t __ret; \
19551  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 17); \
19552  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19553  __ret; \
19554})
19555#define __noswap_vqshrun_n_s32(__p0, __p1) __extension__ ({ \
19556  int32x4_t __s0 = __p0; \
19557  uint16x4_t __ret; \
19558  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 17); \
19559  __ret; \
19560})
19561#endif
19562
19563#ifdef __LITTLE_ENDIAN__
19564#define vqshrun_n_s64(__p0, __p1) __extension__ ({ \
19565  int64x2_t __s0 = __p0; \
19566  uint32x2_t __ret; \
19567  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 18); \
19568  __ret; \
19569})
19570#else
19571#define vqshrun_n_s64(__p0, __p1) __extension__ ({ \
19572  int64x2_t __s0 = __p0; \
19573  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19574  uint32x2_t __ret; \
19575  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 18); \
19576  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19577  __ret; \
19578})
19579#define __noswap_vqshrun_n_s64(__p0, __p1) __extension__ ({ \
19580  int64x2_t __s0 = __p0; \
19581  uint32x2_t __ret; \
19582  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 18); \
19583  __ret; \
19584})
19585#endif
19586
19587#ifdef __LITTLE_ENDIAN__
19588#define vqshrun_n_s16(__p0, __p1) __extension__ ({ \
19589  int16x8_t __s0 = __p0; \
19590  uint8x8_t __ret; \
19591  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 16); \
19592  __ret; \
19593})
19594#else
19595#define vqshrun_n_s16(__p0, __p1) __extension__ ({ \
19596  int16x8_t __s0 = __p0; \
19597  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19598  uint8x8_t __ret; \
19599  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 16); \
19600  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19601  __ret; \
19602})
19603#define __noswap_vqshrun_n_s16(__p0, __p1) __extension__ ({ \
19604  int16x8_t __s0 = __p0; \
19605  uint8x8_t __ret; \
19606  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 16); \
19607  __ret; \
19608})
19609#endif
19610
19611#ifdef __LITTLE_ENDIAN__
19612__ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
19613  uint8x16_t __ret;
19614  __ret = (uint8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
19615  return __ret;
19616}
19617#else
19618__ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
19619  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19620  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19621  uint8x16_t __ret;
19622  __ret = (uint8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
19623  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19624  return __ret;
19625}
19626#endif
19627
19628#ifdef __LITTLE_ENDIAN__
19629__ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
19630  uint32x4_t __ret;
19631  __ret = (uint32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
19632  return __ret;
19633}
19634#else
19635__ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
19636  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19637  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19638  uint32x4_t __ret;
19639  __ret = (uint32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
19640  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19641  return __ret;
19642}
19643#endif
19644
19645#ifdef __LITTLE_ENDIAN__
19646__ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
19647  uint64x2_t __ret;
19648  __ret = (uint64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
19649  return __ret;
19650}
19651#else
19652__ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
19653  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19654  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19655  uint64x2_t __ret;
19656  __ret = (uint64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
19657  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19658  return __ret;
19659}
19660#endif
19661
19662#ifdef __LITTLE_ENDIAN__
19663__ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
19664  uint16x8_t __ret;
19665  __ret = (uint16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
19666  return __ret;
19667}
19668#else
19669__ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
19670  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19671  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19672  uint16x8_t __ret;
19673  __ret = (uint16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
19674  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19675  return __ret;
19676}
19677#endif
19678
19679#ifdef __LITTLE_ENDIAN__
19680__ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1) {
19681  int8x16_t __ret;
19682  __ret = (int8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
19683  return __ret;
19684}
19685#else
19686__ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1) {
19687  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19688  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19689  int8x16_t __ret;
19690  __ret = (int8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
19691  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19692  return __ret;
19693}
19694#endif
19695
19696#ifdef __LITTLE_ENDIAN__
19697__ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
19698  int32x4_t __ret;
19699  __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19700  return __ret;
19701}
19702#else
19703__ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
19704  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19705  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19706  int32x4_t __ret;
19707  __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
19708  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19709  return __ret;
19710}
19711#endif
19712
19713#ifdef __LITTLE_ENDIAN__
19714__ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1) {
19715  int64x2_t __ret;
19716  __ret = (int64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
19717  return __ret;
19718}
19719#else
19720__ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1) {
19721  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19722  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19723  int64x2_t __ret;
19724  __ret = (int64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
19725  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19726  return __ret;
19727}
19728#endif
19729
19730#ifdef __LITTLE_ENDIAN__
19731__ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
19732  int16x8_t __ret;
19733  __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19734  return __ret;
19735}
19736#else
19737__ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
19738  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19739  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19740  int16x8_t __ret;
19741  __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
19742  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19743  return __ret;
19744}
19745#endif
19746
19747#ifdef __LITTLE_ENDIAN__
19748__ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
19749  uint8x8_t __ret;
19750  __ret = (uint8x8_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
19751  return __ret;
19752}
19753#else
19754__ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
19755  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19756  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19757  uint8x8_t __ret;
19758  __ret = (uint8x8_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
19759  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19760  return __ret;
19761}
19762#endif
19763
19764#ifdef __LITTLE_ENDIAN__
19765__ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
19766  uint32x2_t __ret;
19767  __ret = (uint32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
19768  return __ret;
19769}
19770#else
19771__ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
19772  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19773  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19774  uint32x2_t __ret;
19775  __ret = (uint32x2_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
19776  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19777  return __ret;
19778}
19779#endif
19780
19781#ifdef __LITTLE_ENDIAN__
19782__ai uint64x1_t vqsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
19783  uint64x1_t __ret;
19784  __ret = (uint64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
19785  return __ret;
19786}
19787#else
19788__ai uint64x1_t vqsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
19789  uint64x1_t __ret;
19790  __ret = (uint64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
19791  return __ret;
19792}
19793#endif
19794
19795#ifdef __LITTLE_ENDIAN__
19796__ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
19797  uint16x4_t __ret;
19798  __ret = (uint16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
19799  return __ret;
19800}
19801#else
19802__ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
19803  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19804  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19805  uint16x4_t __ret;
19806  __ret = (uint16x4_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
19807  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19808  return __ret;
19809}
19810#endif
19811
19812#ifdef __LITTLE_ENDIAN__
19813__ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1) {
19814  int8x8_t __ret;
19815  __ret = (int8x8_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
19816  return __ret;
19817}
19818#else
19819__ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1) {
19820  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19821  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19822  int8x8_t __ret;
19823  __ret = (int8x8_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
19824  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19825  return __ret;
19826}
19827#endif
19828
19829#ifdef __LITTLE_ENDIAN__
19830__ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
19831  int32x2_t __ret;
19832  __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19833  return __ret;
19834}
19835#else
19836__ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
19837  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19838  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19839  int32x2_t __ret;
19840  __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
19841  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19842  return __ret;
19843}
19844#endif
19845
19846#ifdef __LITTLE_ENDIAN__
19847__ai int64x1_t vqsub_s64(int64x1_t __p0, int64x1_t __p1) {
19848  int64x1_t __ret;
19849  __ret = (int64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
19850  return __ret;
19851}
19852#else
19853__ai int64x1_t vqsub_s64(int64x1_t __p0, int64x1_t __p1) {
19854  int64x1_t __ret;
19855  __ret = (int64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
19856  return __ret;
19857}
19858#endif
19859
19860#ifdef __LITTLE_ENDIAN__
19861__ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
19862  int16x4_t __ret;
19863  __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19864  return __ret;
19865}
19866#else
19867__ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
19868  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19869  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19870  int16x4_t __ret;
19871  __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
19872  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19873  return __ret;
19874}
19875#endif
19876
19877#ifdef __LITTLE_ENDIAN__
19878__ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
19879  uint16x4_t __ret;
19880  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
19881  return __ret;
19882}
19883#else
19884__ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
19885  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19886  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19887  uint16x4_t __ret;
19888  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
19889  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19890  return __ret;
19891}
19892__ai uint16x4_t __noswap_vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
19893  uint16x4_t __ret;
19894  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
19895  return __ret;
19896}
19897#endif
19898
19899#ifdef __LITTLE_ENDIAN__
19900__ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
19901  uint32x2_t __ret;
19902  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
19903  return __ret;
19904}
19905#else
19906__ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
19907  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19908  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19909  uint32x2_t __ret;
19910  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
19911  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19912  return __ret;
19913}
19914__ai uint32x2_t __noswap_vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
19915  uint32x2_t __ret;
19916  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
19917  return __ret;
19918}
19919#endif
19920
19921#ifdef __LITTLE_ENDIAN__
19922__ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
19923  uint8x8_t __ret;
19924  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
19925  return __ret;
19926}
19927#else
19928__ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
19929  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19930  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19931  uint8x8_t __ret;
19932  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
19933  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19934  return __ret;
19935}
19936__ai uint8x8_t __noswap_vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
19937  uint8x8_t __ret;
19938  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
19939  return __ret;
19940}
19941#endif
19942
19943#ifdef __LITTLE_ENDIAN__
19944__ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
19945  int16x4_t __ret;
19946  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
19947  return __ret;
19948}
19949#else
19950__ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
19951  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19952  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19953  int16x4_t __ret;
19954  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
19955  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19956  return __ret;
19957}
19958__ai int16x4_t __noswap_vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
19959  int16x4_t __ret;
19960  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
19961  return __ret;
19962}
19963#endif
19964
19965#ifdef __LITTLE_ENDIAN__
19966__ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
19967  int32x2_t __ret;
19968  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
19969  return __ret;
19970}
19971#else
19972__ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
19973  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19974  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19975  int32x2_t __ret;
19976  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
19977  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19978  return __ret;
19979}
19980__ai int32x2_t __noswap_vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
19981  int32x2_t __ret;
19982  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
19983  return __ret;
19984}
19985#endif
19986
19987#ifdef __LITTLE_ENDIAN__
19988__ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
19989  int8x8_t __ret;
19990  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
19991  return __ret;
19992}
19993#else
19994__ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
19995  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19996  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19997  int8x8_t __ret;
19998  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
19999  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20000  return __ret;
20001}
20002__ai int8x8_t __noswap_vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
20003  int8x8_t __ret;
20004  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
20005  return __ret;
20006}
20007#endif
20008
20009#ifdef __LITTLE_ENDIAN__
20010__ai uint32x4_t vrecpeq_u32(uint32x4_t __p0) {
20011  uint32x4_t __ret;
20012  __ret = (uint32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 50);
20013  return __ret;
20014}
20015#else
20016__ai uint32x4_t vrecpeq_u32(uint32x4_t __p0) {
20017  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20018  uint32x4_t __ret;
20019  __ret = (uint32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 50);
20020  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20021  return __ret;
20022}
20023#endif
20024
20025#ifdef __LITTLE_ENDIAN__
20026__ai float32x4_t vrecpeq_f32(float32x4_t __p0) {
20027  float32x4_t __ret;
20028  __ret = (float32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 41);
20029  return __ret;
20030}
20031#else
20032__ai float32x4_t vrecpeq_f32(float32x4_t __p0) {
20033  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20034  float32x4_t __ret;
20035  __ret = (float32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 41);
20036  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20037  return __ret;
20038}
20039#endif
20040
20041#ifdef __LITTLE_ENDIAN__
20042__ai uint32x2_t vrecpe_u32(uint32x2_t __p0) {
20043  uint32x2_t __ret;
20044  __ret = (uint32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 18);
20045  return __ret;
20046}
20047#else
20048__ai uint32x2_t vrecpe_u32(uint32x2_t __p0) {
20049  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20050  uint32x2_t __ret;
20051  __ret = (uint32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 18);
20052  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20053  return __ret;
20054}
20055#endif
20056
20057#ifdef __LITTLE_ENDIAN__
20058__ai float32x2_t vrecpe_f32(float32x2_t __p0) {
20059  float32x2_t __ret;
20060  __ret = (float32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 9);
20061  return __ret;
20062}
20063#else
20064__ai float32x2_t vrecpe_f32(float32x2_t __p0) {
20065  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20066  float32x2_t __ret;
20067  __ret = (float32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 9);
20068  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20069  return __ret;
20070}
20071#endif
20072
20073#ifdef __LITTLE_ENDIAN__
20074__ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1) {
20075  float32x4_t __ret;
20076  __ret = (float32x4_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
20077  return __ret;
20078}
20079#else
20080__ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1) {
20081  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20082  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20083  float32x4_t __ret;
20084  __ret = (float32x4_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
20085  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20086  return __ret;
20087}
20088#endif
20089
20090#ifdef __LITTLE_ENDIAN__
20091__ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1) {
20092  float32x2_t __ret;
20093  __ret = (float32x2_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
20094  return __ret;
20095}
20096#else
20097__ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1) {
20098  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20099  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20100  float32x2_t __ret;
20101  __ret = (float32x2_t) __builtin_neon_vrecps_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
20102  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20103  return __ret;
20104}
20105#endif
20106
20107#ifdef __LITTLE_ENDIAN__
20108__ai poly8x8_t vrev16_p8(poly8x8_t __p0) {
20109  poly8x8_t __ret;
20110  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20111  return __ret;
20112}
20113#else
20114__ai poly8x8_t vrev16_p8(poly8x8_t __p0) {
20115  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20116  poly8x8_t __ret;
20117  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20118  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20119  return __ret;
20120}
20121#endif
20122
20123#ifdef __LITTLE_ENDIAN__
20124__ai poly8x16_t vrev16q_p8(poly8x16_t __p0) {
20125  poly8x16_t __ret;
20126  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20127  return __ret;
20128}
20129#else
20130__ai poly8x16_t vrev16q_p8(poly8x16_t __p0) {
20131  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20132  poly8x16_t __ret;
20133  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20134  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20135  return __ret;
20136}
20137#endif
20138
20139#ifdef __LITTLE_ENDIAN__
20140__ai uint8x16_t vrev16q_u8(uint8x16_t __p0) {
20141  uint8x16_t __ret;
20142  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20143  return __ret;
20144}
20145#else
20146__ai uint8x16_t vrev16q_u8(uint8x16_t __p0) {
20147  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20148  uint8x16_t __ret;
20149  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20150  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20151  return __ret;
20152}
20153#endif
20154
20155#ifdef __LITTLE_ENDIAN__
20156__ai int8x16_t vrev16q_s8(int8x16_t __p0) {
20157  int8x16_t __ret;
20158  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20159  return __ret;
20160}
20161#else
20162__ai int8x16_t vrev16q_s8(int8x16_t __p0) {
20163  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20164  int8x16_t __ret;
20165  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20166  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20167  return __ret;
20168}
20169#endif
20170
20171#ifdef __LITTLE_ENDIAN__
20172__ai uint8x8_t vrev16_u8(uint8x8_t __p0) {
20173  uint8x8_t __ret;
20174  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20175  return __ret;
20176}
20177#else
20178__ai uint8x8_t vrev16_u8(uint8x8_t __p0) {
20179  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20180  uint8x8_t __ret;
20181  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20182  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20183  return __ret;
20184}
20185#endif
20186
20187#ifdef __LITTLE_ENDIAN__
20188__ai int8x8_t vrev16_s8(int8x8_t __p0) {
20189  int8x8_t __ret;
20190  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20191  return __ret;
20192}
20193#else
20194__ai int8x8_t vrev16_s8(int8x8_t __p0) {
20195  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20196  int8x8_t __ret;
20197  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20198  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20199  return __ret;
20200}
20201#endif
20202
20203#ifdef __LITTLE_ENDIAN__
20204__ai poly8x8_t vrev32_p8(poly8x8_t __p0) {
20205  poly8x8_t __ret;
20206  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20207  return __ret;
20208}
20209#else
20210__ai poly8x8_t vrev32_p8(poly8x8_t __p0) {
20211  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20212  poly8x8_t __ret;
20213  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20214  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20215  return __ret;
20216}
20217#endif
20218
20219#ifdef __LITTLE_ENDIAN__
20220__ai poly16x4_t vrev32_p16(poly16x4_t __p0) {
20221  poly16x4_t __ret;
20222  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20223  return __ret;
20224}
20225#else
20226__ai poly16x4_t vrev32_p16(poly16x4_t __p0) {
20227  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20228  poly16x4_t __ret;
20229  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20230  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20231  return __ret;
20232}
20233#endif
20234
20235#ifdef __LITTLE_ENDIAN__
20236__ai poly8x16_t vrev32q_p8(poly8x16_t __p0) {
20237  poly8x16_t __ret;
20238  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20239  return __ret;
20240}
20241#else
20242__ai poly8x16_t vrev32q_p8(poly8x16_t __p0) {
20243  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20244  poly8x16_t __ret;
20245  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20246  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20247  return __ret;
20248}
20249#endif
20250
20251#ifdef __LITTLE_ENDIAN__
20252__ai poly16x8_t vrev32q_p16(poly16x8_t __p0) {
20253  poly16x8_t __ret;
20254  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20255  return __ret;
20256}
20257#else
20258__ai poly16x8_t vrev32q_p16(poly16x8_t __p0) {
20259  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20260  poly16x8_t __ret;
20261  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20262  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20263  return __ret;
20264}
20265#endif
20266
20267#ifdef __LITTLE_ENDIAN__
20268__ai uint8x16_t vrev32q_u8(uint8x16_t __p0) {
20269  uint8x16_t __ret;
20270  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20271  return __ret;
20272}
20273#else
20274__ai uint8x16_t vrev32q_u8(uint8x16_t __p0) {
20275  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20276  uint8x16_t __ret;
20277  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20278  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20279  return __ret;
20280}
20281#endif
20282
20283#ifdef __LITTLE_ENDIAN__
20284__ai uint16x8_t vrev32q_u16(uint16x8_t __p0) {
20285  uint16x8_t __ret;
20286  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20287  return __ret;
20288}
20289#else
20290__ai uint16x8_t vrev32q_u16(uint16x8_t __p0) {
20291  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20292  uint16x8_t __ret;
20293  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20294  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20295  return __ret;
20296}
20297#endif
20298
20299#ifdef __LITTLE_ENDIAN__
20300__ai int8x16_t vrev32q_s8(int8x16_t __p0) {
20301  int8x16_t __ret;
20302  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20303  return __ret;
20304}
20305#else
20306__ai int8x16_t vrev32q_s8(int8x16_t __p0) {
20307  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20308  int8x16_t __ret;
20309  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20310  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20311  return __ret;
20312}
20313#endif
20314
20315#ifdef __LITTLE_ENDIAN__
20316__ai int16x8_t vrev32q_s16(int16x8_t __p0) {
20317  int16x8_t __ret;
20318  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20319  return __ret;
20320}
20321#else
20322__ai int16x8_t vrev32q_s16(int16x8_t __p0) {
20323  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20324  int16x8_t __ret;
20325  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20326  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20327  return __ret;
20328}
20329#endif
20330
20331#ifdef __LITTLE_ENDIAN__
20332__ai uint8x8_t vrev32_u8(uint8x8_t __p0) {
20333  uint8x8_t __ret;
20334  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20335  return __ret;
20336}
20337#else
20338__ai uint8x8_t vrev32_u8(uint8x8_t __p0) {
20339  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20340  uint8x8_t __ret;
20341  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20342  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20343  return __ret;
20344}
20345#endif
20346
20347#ifdef __LITTLE_ENDIAN__
20348__ai uint16x4_t vrev32_u16(uint16x4_t __p0) {
20349  uint16x4_t __ret;
20350  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20351  return __ret;
20352}
20353#else
20354__ai uint16x4_t vrev32_u16(uint16x4_t __p0) {
20355  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20356  uint16x4_t __ret;
20357  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20358  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20359  return __ret;
20360}
20361#endif
20362
20363#ifdef __LITTLE_ENDIAN__
20364__ai int8x8_t vrev32_s8(int8x8_t __p0) {
20365  int8x8_t __ret;
20366  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20367  return __ret;
20368}
20369#else
20370__ai int8x8_t vrev32_s8(int8x8_t __p0) {
20371  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20372  int8x8_t __ret;
20373  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20374  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20375  return __ret;
20376}
20377#endif
20378
20379#ifdef __LITTLE_ENDIAN__
20380__ai int16x4_t vrev32_s16(int16x4_t __p0) {
20381  int16x4_t __ret;
20382  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20383  return __ret;
20384}
20385#else
20386__ai int16x4_t vrev32_s16(int16x4_t __p0) {
20387  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20388  int16x4_t __ret;
20389  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20390  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20391  return __ret;
20392}
20393#endif
20394
20395#ifdef __LITTLE_ENDIAN__
20396__ai poly8x8_t vrev64_p8(poly8x8_t __p0) {
20397  poly8x8_t __ret;
20398  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20399  return __ret;
20400}
20401#else
20402__ai poly8x8_t vrev64_p8(poly8x8_t __p0) {
20403  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20404  poly8x8_t __ret;
20405  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
20406  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20407  return __ret;
20408}
20409#endif
20410
20411#ifdef __LITTLE_ENDIAN__
20412__ai poly16x4_t vrev64_p16(poly16x4_t __p0) {
20413  poly16x4_t __ret;
20414  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20415  return __ret;
20416}
20417#else
20418__ai poly16x4_t vrev64_p16(poly16x4_t __p0) {
20419  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20420  poly16x4_t __ret;
20421  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
20422  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20423  return __ret;
20424}
20425#endif
20426
20427#ifdef __LITTLE_ENDIAN__
20428__ai poly8x16_t vrev64q_p8(poly8x16_t __p0) {
20429  poly8x16_t __ret;
20430  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20431  return __ret;
20432}
20433#else
20434__ai poly8x16_t vrev64q_p8(poly8x16_t __p0) {
20435  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20436  poly8x16_t __ret;
20437  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20438  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20439  return __ret;
20440}
20441#endif
20442
20443#ifdef __LITTLE_ENDIAN__
20444__ai poly16x8_t vrev64q_p16(poly16x8_t __p0) {
20445  poly16x8_t __ret;
20446  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20447  return __ret;
20448}
20449#else
20450__ai poly16x8_t vrev64q_p16(poly16x8_t __p0) {
20451  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20452  poly16x8_t __ret;
20453  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20454  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20455  return __ret;
20456}
20457#endif
20458
20459#ifdef __LITTLE_ENDIAN__
20460__ai uint8x16_t vrev64q_u8(uint8x16_t __p0) {
20461  uint8x16_t __ret;
20462  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20463  return __ret;
20464}
20465#else
20466__ai uint8x16_t vrev64q_u8(uint8x16_t __p0) {
20467  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20468  uint8x16_t __ret;
20469  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20470  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20471  return __ret;
20472}
20473#endif
20474
20475#ifdef __LITTLE_ENDIAN__
20476__ai uint32x4_t vrev64q_u32(uint32x4_t __p0) {
20477  uint32x4_t __ret;
20478  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20479  return __ret;
20480}
20481#else
20482__ai uint32x4_t vrev64q_u32(uint32x4_t __p0) {
20483  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20484  uint32x4_t __ret;
20485  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20486  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20487  return __ret;
20488}
20489#endif
20490
20491#ifdef __LITTLE_ENDIAN__
20492__ai uint16x8_t vrev64q_u16(uint16x8_t __p0) {
20493  uint16x8_t __ret;
20494  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20495  return __ret;
20496}
20497#else
20498__ai uint16x8_t vrev64q_u16(uint16x8_t __p0) {
20499  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20500  uint16x8_t __ret;
20501  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20502  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20503  return __ret;
20504}
20505#endif
20506
20507#ifdef __LITTLE_ENDIAN__
20508__ai int8x16_t vrev64q_s8(int8x16_t __p0) {
20509  int8x16_t __ret;
20510  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20511  return __ret;
20512}
20513#else
20514__ai int8x16_t vrev64q_s8(int8x16_t __p0) {
20515  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20516  int8x16_t __ret;
20517  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20518  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20519  return __ret;
20520}
20521#endif
20522
20523#ifdef __LITTLE_ENDIAN__
20524__ai float32x4_t vrev64q_f32(float32x4_t __p0) {
20525  float32x4_t __ret;
20526  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20527  return __ret;
20528}
20529#else
20530__ai float32x4_t vrev64q_f32(float32x4_t __p0) {
20531  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20532  float32x4_t __ret;
20533  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20534  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20535  return __ret;
20536}
20537#endif
20538
20539#ifdef __LITTLE_ENDIAN__
20540__ai int32x4_t vrev64q_s32(int32x4_t __p0) {
20541  int32x4_t __ret;
20542  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20543  return __ret;
20544}
20545#else
20546__ai int32x4_t vrev64q_s32(int32x4_t __p0) {
20547  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20548  int32x4_t __ret;
20549  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20550  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20551  return __ret;
20552}
20553#endif
20554
20555#ifdef __LITTLE_ENDIAN__
20556__ai int16x8_t vrev64q_s16(int16x8_t __p0) {
20557  int16x8_t __ret;
20558  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20559  return __ret;
20560}
20561#else
20562__ai int16x8_t vrev64q_s16(int16x8_t __p0) {
20563  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20564  int16x8_t __ret;
20565  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20566  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20567  return __ret;
20568}
20569#endif
20570
20571#ifdef __LITTLE_ENDIAN__
20572__ai uint8x8_t vrev64_u8(uint8x8_t __p0) {
20573  uint8x8_t __ret;
20574  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20575  return __ret;
20576}
20577#else
20578__ai uint8x8_t vrev64_u8(uint8x8_t __p0) {
20579  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20580  uint8x8_t __ret;
20581  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
20582  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20583  return __ret;
20584}
20585#endif
20586
20587#ifdef __LITTLE_ENDIAN__
20588__ai uint32x2_t vrev64_u32(uint32x2_t __p0) {
20589  uint32x2_t __ret;
20590  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
20591  return __ret;
20592}
20593#else
20594__ai uint32x2_t vrev64_u32(uint32x2_t __p0) {
20595  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20596  uint32x2_t __ret;
20597  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
20598  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20599  return __ret;
20600}
20601#endif
20602
20603#ifdef __LITTLE_ENDIAN__
20604__ai uint16x4_t vrev64_u16(uint16x4_t __p0) {
20605  uint16x4_t __ret;
20606  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20607  return __ret;
20608}
20609#else
20610__ai uint16x4_t vrev64_u16(uint16x4_t __p0) {
20611  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20612  uint16x4_t __ret;
20613  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
20614  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20615  return __ret;
20616}
20617#endif
20618
20619#ifdef __LITTLE_ENDIAN__
20620__ai int8x8_t vrev64_s8(int8x8_t __p0) {
20621  int8x8_t __ret;
20622  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20623  return __ret;
20624}
20625#else
20626__ai int8x8_t vrev64_s8(int8x8_t __p0) {
20627  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20628  int8x8_t __ret;
20629  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
20630  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20631  return __ret;
20632}
20633#endif
20634
20635#ifdef __LITTLE_ENDIAN__
20636__ai float32x2_t vrev64_f32(float32x2_t __p0) {
20637  float32x2_t __ret;
20638  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
20639  return __ret;
20640}
20641#else
20642__ai float32x2_t vrev64_f32(float32x2_t __p0) {
20643  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20644  float32x2_t __ret;
20645  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
20646  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20647  return __ret;
20648}
20649#endif
20650
20651#ifdef __LITTLE_ENDIAN__
20652__ai int32x2_t vrev64_s32(int32x2_t __p0) {
20653  int32x2_t __ret;
20654  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
20655  return __ret;
20656}
20657#else
20658__ai int32x2_t vrev64_s32(int32x2_t __p0) {
20659  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20660  int32x2_t __ret;
20661  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
20662  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20663  return __ret;
20664}
20665#endif
20666
20667#ifdef __LITTLE_ENDIAN__
20668__ai int16x4_t vrev64_s16(int16x4_t __p0) {
20669  int16x4_t __ret;
20670  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20671  return __ret;
20672}
20673#else
20674__ai int16x4_t vrev64_s16(int16x4_t __p0) {
20675  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20676  int16x4_t __ret;
20677  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
20678  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20679  return __ret;
20680}
20681#endif
20682
20683#ifdef __LITTLE_ENDIAN__
20684__ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
20685  uint8x16_t __ret;
20686  __ret = (uint8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
20687  return __ret;
20688}
20689#else
20690__ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
20691  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20692  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20693  uint8x16_t __ret;
20694  __ret = (uint8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
20695  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20696  return __ret;
20697}
20698#endif
20699
20700#ifdef __LITTLE_ENDIAN__
20701__ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
20702  uint32x4_t __ret;
20703  __ret = (uint32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
20704  return __ret;
20705}
20706#else
20707__ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
20708  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20709  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20710  uint32x4_t __ret;
20711  __ret = (uint32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
20712  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20713  return __ret;
20714}
20715#endif
20716
20717#ifdef __LITTLE_ENDIAN__
20718__ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
20719  uint16x8_t __ret;
20720  __ret = (uint16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
20721  return __ret;
20722}
20723#else
20724__ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
20725  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20726  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20727  uint16x8_t __ret;
20728  __ret = (uint16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
20729  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20730  return __ret;
20731}
20732#endif
20733
20734#ifdef __LITTLE_ENDIAN__
20735__ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
20736  int8x16_t __ret;
20737  __ret = (int8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
20738  return __ret;
20739}
20740#else
20741__ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
20742  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20743  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20744  int8x16_t __ret;
20745  __ret = (int8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
20746  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20747  return __ret;
20748}
20749#endif
20750
20751#ifdef __LITTLE_ENDIAN__
20752__ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
20753  int32x4_t __ret;
20754  __ret = (int32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
20755  return __ret;
20756}
20757#else
20758__ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
20759  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20760  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20761  int32x4_t __ret;
20762  __ret = (int32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
20763  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20764  return __ret;
20765}
20766#endif
20767
20768#ifdef __LITTLE_ENDIAN__
20769__ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
20770  int16x8_t __ret;
20771  __ret = (int16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
20772  return __ret;
20773}
20774#else
20775__ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
20776  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20777  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20778  int16x8_t __ret;
20779  __ret = (int16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
20780  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20781  return __ret;
20782}
20783#endif
20784
20785#ifdef __LITTLE_ENDIAN__
20786__ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
20787  uint8x8_t __ret;
20788  __ret = (uint8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
20789  return __ret;
20790}
20791#else
20792__ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
20793  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20794  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20795  uint8x8_t __ret;
20796  __ret = (uint8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
20797  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20798  return __ret;
20799}
20800#endif
20801
20802#ifdef __LITTLE_ENDIAN__
20803__ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
20804  uint32x2_t __ret;
20805  __ret = (uint32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
20806  return __ret;
20807}
20808#else
20809__ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
20810  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20811  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20812  uint32x2_t __ret;
20813  __ret = (uint32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
20814  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20815  return __ret;
20816}
20817#endif
20818
20819#ifdef __LITTLE_ENDIAN__
20820__ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
20821  uint16x4_t __ret;
20822  __ret = (uint16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
20823  return __ret;
20824}
20825#else
20826__ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
20827  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20828  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20829  uint16x4_t __ret;
20830  __ret = (uint16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
20831  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20832  return __ret;
20833}
20834#endif
20835
20836#ifdef __LITTLE_ENDIAN__
20837__ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1) {
20838  int8x8_t __ret;
20839  __ret = (int8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
20840  return __ret;
20841}
20842#else
20843__ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1) {
20844  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20845  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20846  int8x8_t __ret;
20847  __ret = (int8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
20848  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20849  return __ret;
20850}
20851#endif
20852
20853#ifdef __LITTLE_ENDIAN__
20854__ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1) {
20855  int32x2_t __ret;
20856  __ret = (int32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
20857  return __ret;
20858}
20859#else
20860__ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1) {
20861  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20862  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20863  int32x2_t __ret;
20864  __ret = (int32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
20865  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20866  return __ret;
20867}
20868#endif
20869
20870#ifdef __LITTLE_ENDIAN__
20871__ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1) {
20872  int16x4_t __ret;
20873  __ret = (int16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
20874  return __ret;
20875}
20876#else
20877__ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1) {
20878  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20879  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20880  int16x4_t __ret;
20881  __ret = (int16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
20882  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20883  return __ret;
20884}
20885#endif
20886
20887#ifdef __LITTLE_ENDIAN__
20888__ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
20889  uint8x16_t __ret;
20890  __ret = (uint8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
20891  return __ret;
20892}
20893#else
20894__ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
20895  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20896  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20897  uint8x16_t __ret;
20898  __ret = (uint8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
20899  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20900  return __ret;
20901}
20902#endif
20903
20904#ifdef __LITTLE_ENDIAN__
20905__ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
20906  uint32x4_t __ret;
20907  __ret = (uint32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
20908  return __ret;
20909}
20910#else
20911__ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
20912  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20913  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20914  uint32x4_t __ret;
20915  __ret = (uint32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
20916  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20917  return __ret;
20918}
20919#endif
20920
20921#ifdef __LITTLE_ENDIAN__
20922__ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
20923  uint64x2_t __ret;
20924  __ret = (uint64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
20925  return __ret;
20926}
20927#else
20928__ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
20929  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20930  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20931  uint64x2_t __ret;
20932  __ret = (uint64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
20933  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20934  return __ret;
20935}
20936#endif
20937
20938#ifdef __LITTLE_ENDIAN__
20939__ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
20940  uint16x8_t __ret;
20941  __ret = (uint16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
20942  return __ret;
20943}
20944#else
20945__ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
20946  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20947  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20948  uint16x8_t __ret;
20949  __ret = (uint16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
20950  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20951  return __ret;
20952}
20953#endif
20954
20955#ifdef __LITTLE_ENDIAN__
20956__ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
20957  int8x16_t __ret;
20958  __ret = (int8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
20959  return __ret;
20960}
20961#else
20962__ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
20963  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20964  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20965  int8x16_t __ret;
20966  __ret = (int8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
20967  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20968  return __ret;
20969}
20970#endif
20971
20972#ifdef __LITTLE_ENDIAN__
20973__ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
20974  int32x4_t __ret;
20975  __ret = (int32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
20976  return __ret;
20977}
20978#else
20979__ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
20980  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20981  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20982  int32x4_t __ret;
20983  __ret = (int32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
20984  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20985  return __ret;
20986}
20987#endif
20988
20989#ifdef __LITTLE_ENDIAN__
20990__ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
20991  int64x2_t __ret;
20992  __ret = (int64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
20993  return __ret;
20994}
20995#else
20996__ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
20997  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20998  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20999  int64x2_t __ret;
21000  __ret = (int64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
21001  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21002  return __ret;
21003}
21004#endif
21005
21006#ifdef __LITTLE_ENDIAN__
21007__ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
21008  int16x8_t __ret;
21009  __ret = (int16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
21010  return __ret;
21011}
21012#else
21013__ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
21014  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21015  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21016  int16x8_t __ret;
21017  __ret = (int16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
21018  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21019  return __ret;
21020}
21021#endif
21022
21023#ifdef __LITTLE_ENDIAN__
21024__ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
21025  uint8x8_t __ret;
21026  __ret = (uint8x8_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
21027  return __ret;
21028}
21029#else
21030__ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
21031  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21032  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21033  uint8x8_t __ret;
21034  __ret = (uint8x8_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
21035  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21036  return __ret;
21037}
21038#endif
21039
21040#ifdef __LITTLE_ENDIAN__
21041__ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
21042  uint32x2_t __ret;
21043  __ret = (uint32x2_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
21044  return __ret;
21045}
21046#else
21047__ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
21048  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21049  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21050  uint32x2_t __ret;
21051  __ret = (uint32x2_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
21052  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21053  return __ret;
21054}
21055#endif
21056
21057#ifdef __LITTLE_ENDIAN__
21058__ai uint64x1_t vrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
21059  uint64x1_t __ret;
21060  __ret = (uint64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
21061  return __ret;
21062}
21063#else
21064__ai uint64x1_t vrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
21065  uint64x1_t __ret;
21066  __ret = (uint64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
21067  return __ret;
21068}
21069#endif
21070
21071#ifdef __LITTLE_ENDIAN__
21072__ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
21073  uint16x4_t __ret;
21074  __ret = (uint16x4_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
21075  return __ret;
21076}
21077#else
21078__ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
21079  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21080  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21081  uint16x4_t __ret;
21082  __ret = (uint16x4_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
21083  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21084  return __ret;
21085}
21086#endif
21087
21088#ifdef __LITTLE_ENDIAN__
21089__ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1) {
21090  int8x8_t __ret;
21091  __ret = (int8x8_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
21092  return __ret;
21093}
21094#else
21095__ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1) {
21096  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21097  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21098  int8x8_t __ret;
21099  __ret = (int8x8_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
21100  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21101  return __ret;
21102}
21103#endif
21104
21105#ifdef __LITTLE_ENDIAN__
21106__ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1) {
21107  int32x2_t __ret;
21108  __ret = (int32x2_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
21109  return __ret;
21110}
21111#else
21112__ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1) {
21113  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21114  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21115  int32x2_t __ret;
21116  __ret = (int32x2_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
21117  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21118  return __ret;
21119}
21120#endif
21121
21122#ifdef __LITTLE_ENDIAN__
21123__ai int64x1_t vrshl_s64(int64x1_t __p0, int64x1_t __p1) {
21124  int64x1_t __ret;
21125  __ret = (int64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
21126  return __ret;
21127}
21128#else
21129__ai int64x1_t vrshl_s64(int64x1_t __p0, int64x1_t __p1) {
21130  int64x1_t __ret;
21131  __ret = (int64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
21132  return __ret;
21133}
21134#endif
21135
21136#ifdef __LITTLE_ENDIAN__
21137__ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1) {
21138  int16x4_t __ret;
21139  __ret = (int16x4_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
21140  return __ret;
21141}
21142#else
21143__ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1) {
21144  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21145  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21146  int16x4_t __ret;
21147  __ret = (int16x4_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
21148  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21149  return __ret;
21150}
21151#endif
21152
21153#ifdef __LITTLE_ENDIAN__
21154#define vrshrq_n_u8(__p0, __p1) __extension__ ({ \
21155  uint8x16_t __s0 = __p0; \
21156  uint8x16_t __ret; \
21157  __ret = (uint8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 48); \
21158  __ret; \
21159})
21160#else
21161#define vrshrq_n_u8(__p0, __p1) __extension__ ({ \
21162  uint8x16_t __s0 = __p0; \
21163  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21164  uint8x16_t __ret; \
21165  __ret = (uint8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 48); \
21166  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21167  __ret; \
21168})
21169#endif
21170
21171#ifdef __LITTLE_ENDIAN__
21172#define vrshrq_n_u32(__p0, __p1) __extension__ ({ \
21173  uint32x4_t __s0 = __p0; \
21174  uint32x4_t __ret; \
21175  __ret = (uint32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 50); \
21176  __ret; \
21177})
21178#else
21179#define vrshrq_n_u32(__p0, __p1) __extension__ ({ \
21180  uint32x4_t __s0 = __p0; \
21181  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21182  uint32x4_t __ret; \
21183  __ret = (uint32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 50); \
21184  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21185  __ret; \
21186})
21187#endif
21188
21189#ifdef __LITTLE_ENDIAN__
21190#define vrshrq_n_u64(__p0, __p1) __extension__ ({ \
21191  uint64x2_t __s0 = __p0; \
21192  uint64x2_t __ret; \
21193  __ret = (uint64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 51); \
21194  __ret; \
21195})
21196#else
21197#define vrshrq_n_u64(__p0, __p1) __extension__ ({ \
21198  uint64x2_t __s0 = __p0; \
21199  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21200  uint64x2_t __ret; \
21201  __ret = (uint64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 51); \
21202  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21203  __ret; \
21204})
21205#endif
21206
21207#ifdef __LITTLE_ENDIAN__
21208#define vrshrq_n_u16(__p0, __p1) __extension__ ({ \
21209  uint16x8_t __s0 = __p0; \
21210  uint16x8_t __ret; \
21211  __ret = (uint16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 49); \
21212  __ret; \
21213})
21214#else
21215#define vrshrq_n_u16(__p0, __p1) __extension__ ({ \
21216  uint16x8_t __s0 = __p0; \
21217  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21218  uint16x8_t __ret; \
21219  __ret = (uint16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 49); \
21220  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21221  __ret; \
21222})
21223#endif
21224
21225#ifdef __LITTLE_ENDIAN__
21226#define vrshrq_n_s8(__p0, __p1) __extension__ ({ \
21227  int8x16_t __s0 = __p0; \
21228  int8x16_t __ret; \
21229  __ret = (int8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 32); \
21230  __ret; \
21231})
21232#else
21233#define vrshrq_n_s8(__p0, __p1) __extension__ ({ \
21234  int8x16_t __s0 = __p0; \
21235  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21236  int8x16_t __ret; \
21237  __ret = (int8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 32); \
21238  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21239  __ret; \
21240})
21241#endif
21242
21243#ifdef __LITTLE_ENDIAN__
21244#define vrshrq_n_s32(__p0, __p1) __extension__ ({ \
21245  int32x4_t __s0 = __p0; \
21246  int32x4_t __ret; \
21247  __ret = (int32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 34); \
21248  __ret; \
21249})
21250#else
21251#define vrshrq_n_s32(__p0, __p1) __extension__ ({ \
21252  int32x4_t __s0 = __p0; \
21253  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21254  int32x4_t __ret; \
21255  __ret = (int32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 34); \
21256  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21257  __ret; \
21258})
21259#endif
21260
21261#ifdef __LITTLE_ENDIAN__
21262#define vrshrq_n_s64(__p0, __p1) __extension__ ({ \
21263  int64x2_t __s0 = __p0; \
21264  int64x2_t __ret; \
21265  __ret = (int64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 35); \
21266  __ret; \
21267})
21268#else
21269#define vrshrq_n_s64(__p0, __p1) __extension__ ({ \
21270  int64x2_t __s0 = __p0; \
21271  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21272  int64x2_t __ret; \
21273  __ret = (int64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 35); \
21274  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21275  __ret; \
21276})
21277#endif
21278
21279#ifdef __LITTLE_ENDIAN__
21280#define vrshrq_n_s16(__p0, __p1) __extension__ ({ \
21281  int16x8_t __s0 = __p0; \
21282  int16x8_t __ret; \
21283  __ret = (int16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 33); \
21284  __ret; \
21285})
21286#else
21287#define vrshrq_n_s16(__p0, __p1) __extension__ ({ \
21288  int16x8_t __s0 = __p0; \
21289  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21290  int16x8_t __ret; \
21291  __ret = (int16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 33); \
21292  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21293  __ret; \
21294})
21295#endif
21296
21297#ifdef __LITTLE_ENDIAN__
21298#define vrshr_n_u8(__p0, __p1) __extension__ ({ \
21299  uint8x8_t __s0 = __p0; \
21300  uint8x8_t __ret; \
21301  __ret = (uint8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 16); \
21302  __ret; \
21303})
21304#else
21305#define vrshr_n_u8(__p0, __p1) __extension__ ({ \
21306  uint8x8_t __s0 = __p0; \
21307  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21308  uint8x8_t __ret; \
21309  __ret = (uint8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 16); \
21310  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21311  __ret; \
21312})
21313#endif
21314
21315#ifdef __LITTLE_ENDIAN__
21316#define vrshr_n_u32(__p0, __p1) __extension__ ({ \
21317  uint32x2_t __s0 = __p0; \
21318  uint32x2_t __ret; \
21319  __ret = (uint32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 18); \
21320  __ret; \
21321})
21322#else
21323#define vrshr_n_u32(__p0, __p1) __extension__ ({ \
21324  uint32x2_t __s0 = __p0; \
21325  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21326  uint32x2_t __ret; \
21327  __ret = (uint32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 18); \
21328  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21329  __ret; \
21330})
21331#endif
21332
21333#ifdef __LITTLE_ENDIAN__
21334#define vrshr_n_u64(__p0, __p1) __extension__ ({ \
21335  uint64x1_t __s0 = __p0; \
21336  uint64x1_t __ret; \
21337  __ret = (uint64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 19); \
21338  __ret; \
21339})
21340#else
21341#define vrshr_n_u64(__p0, __p1) __extension__ ({ \
21342  uint64x1_t __s0 = __p0; \
21343  uint64x1_t __ret; \
21344  __ret = (uint64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 19); \
21345  __ret; \
21346})
21347#endif
21348
21349#ifdef __LITTLE_ENDIAN__
21350#define vrshr_n_u16(__p0, __p1) __extension__ ({ \
21351  uint16x4_t __s0 = __p0; \
21352  uint16x4_t __ret; \
21353  __ret = (uint16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 17); \
21354  __ret; \
21355})
21356#else
21357#define vrshr_n_u16(__p0, __p1) __extension__ ({ \
21358  uint16x4_t __s0 = __p0; \
21359  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21360  uint16x4_t __ret; \
21361  __ret = (uint16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 17); \
21362  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21363  __ret; \
21364})
21365#endif
21366
21367#ifdef __LITTLE_ENDIAN__
21368#define vrshr_n_s8(__p0, __p1) __extension__ ({ \
21369  int8x8_t __s0 = __p0; \
21370  int8x8_t __ret; \
21371  __ret = (int8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 0); \
21372  __ret; \
21373})
21374#else
21375#define vrshr_n_s8(__p0, __p1) __extension__ ({ \
21376  int8x8_t __s0 = __p0; \
21377  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21378  int8x8_t __ret; \
21379  __ret = (int8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 0); \
21380  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21381  __ret; \
21382})
21383#endif
21384
21385#ifdef __LITTLE_ENDIAN__
21386#define vrshr_n_s32(__p0, __p1) __extension__ ({ \
21387  int32x2_t __s0 = __p0; \
21388  int32x2_t __ret; \
21389  __ret = (int32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 2); \
21390  __ret; \
21391})
21392#else
21393#define vrshr_n_s32(__p0, __p1) __extension__ ({ \
21394  int32x2_t __s0 = __p0; \
21395  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21396  int32x2_t __ret; \
21397  __ret = (int32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 2); \
21398  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21399  __ret; \
21400})
21401#endif
21402
21403#ifdef __LITTLE_ENDIAN__
21404#define vrshr_n_s64(__p0, __p1) __extension__ ({ \
21405  int64x1_t __s0 = __p0; \
21406  int64x1_t __ret; \
21407  __ret = (int64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 3); \
21408  __ret; \
21409})
21410#else
21411#define vrshr_n_s64(__p0, __p1) __extension__ ({ \
21412  int64x1_t __s0 = __p0; \
21413  int64x1_t __ret; \
21414  __ret = (int64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 3); \
21415  __ret; \
21416})
21417#endif
21418
21419#ifdef __LITTLE_ENDIAN__
21420#define vrshr_n_s16(__p0, __p1) __extension__ ({ \
21421  int16x4_t __s0 = __p0; \
21422  int16x4_t __ret; \
21423  __ret = (int16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 1); \
21424  __ret; \
21425})
21426#else
21427#define vrshr_n_s16(__p0, __p1) __extension__ ({ \
21428  int16x4_t __s0 = __p0; \
21429  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21430  int16x4_t __ret; \
21431  __ret = (int16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 1); \
21432  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21433  __ret; \
21434})
21435#endif
21436
21437#ifdef __LITTLE_ENDIAN__
21438#define vrshrn_n_u32(__p0, __p1) __extension__ ({ \
21439  uint32x4_t __s0 = __p0; \
21440  uint16x4_t __ret; \
21441  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 17); \
21442  __ret; \
21443})
21444#else
21445#define vrshrn_n_u32(__p0, __p1) __extension__ ({ \
21446  uint32x4_t __s0 = __p0; \
21447  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21448  uint16x4_t __ret; \
21449  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 17); \
21450  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21451  __ret; \
21452})
21453#define __noswap_vrshrn_n_u32(__p0, __p1) __extension__ ({ \
21454  uint32x4_t __s0 = __p0; \
21455  uint16x4_t __ret; \
21456  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 17); \
21457  __ret; \
21458})
21459#endif
21460
21461#ifdef __LITTLE_ENDIAN__
21462#define vrshrn_n_u64(__p0, __p1) __extension__ ({ \
21463  uint64x2_t __s0 = __p0; \
21464  uint32x2_t __ret; \
21465  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 18); \
21466  __ret; \
21467})
21468#else
21469#define vrshrn_n_u64(__p0, __p1) __extension__ ({ \
21470  uint64x2_t __s0 = __p0; \
21471  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21472  uint32x2_t __ret; \
21473  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 18); \
21474  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21475  __ret; \
21476})
21477#define __noswap_vrshrn_n_u64(__p0, __p1) __extension__ ({ \
21478  uint64x2_t __s0 = __p0; \
21479  uint32x2_t __ret; \
21480  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 18); \
21481  __ret; \
21482})
21483#endif
21484
21485#ifdef __LITTLE_ENDIAN__
21486#define vrshrn_n_u16(__p0, __p1) __extension__ ({ \
21487  uint16x8_t __s0 = __p0; \
21488  uint8x8_t __ret; \
21489  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 16); \
21490  __ret; \
21491})
21492#else
21493#define vrshrn_n_u16(__p0, __p1) __extension__ ({ \
21494  uint16x8_t __s0 = __p0; \
21495  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21496  uint8x8_t __ret; \
21497  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 16); \
21498  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21499  __ret; \
21500})
21501#define __noswap_vrshrn_n_u16(__p0, __p1) __extension__ ({ \
21502  uint16x8_t __s0 = __p0; \
21503  uint8x8_t __ret; \
21504  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 16); \
21505  __ret; \
21506})
21507#endif
21508
21509#ifdef __LITTLE_ENDIAN__
21510#define vrshrn_n_s32(__p0, __p1) __extension__ ({ \
21511  int32x4_t __s0 = __p0; \
21512  int16x4_t __ret; \
21513  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 1); \
21514  __ret; \
21515})
21516#else
21517#define vrshrn_n_s32(__p0, __p1) __extension__ ({ \
21518  int32x4_t __s0 = __p0; \
21519  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21520  int16x4_t __ret; \
21521  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 1); \
21522  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21523  __ret; \
21524})
21525#define __noswap_vrshrn_n_s32(__p0, __p1) __extension__ ({ \
21526  int32x4_t __s0 = __p0; \
21527  int16x4_t __ret; \
21528  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 1); \
21529  __ret; \
21530})
21531#endif
21532
21533#ifdef __LITTLE_ENDIAN__
21534#define vrshrn_n_s64(__p0, __p1) __extension__ ({ \
21535  int64x2_t __s0 = __p0; \
21536  int32x2_t __ret; \
21537  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 2); \
21538  __ret; \
21539})
21540#else
21541#define vrshrn_n_s64(__p0, __p1) __extension__ ({ \
21542  int64x2_t __s0 = __p0; \
21543  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21544  int32x2_t __ret; \
21545  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 2); \
21546  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21547  __ret; \
21548})
21549#define __noswap_vrshrn_n_s64(__p0, __p1) __extension__ ({ \
21550  int64x2_t __s0 = __p0; \
21551  int32x2_t __ret; \
21552  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 2); \
21553  __ret; \
21554})
21555#endif
21556
21557#ifdef __LITTLE_ENDIAN__
21558#define vrshrn_n_s16(__p0, __p1) __extension__ ({ \
21559  int16x8_t __s0 = __p0; \
21560  int8x8_t __ret; \
21561  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 0); \
21562  __ret; \
21563})
21564#else
21565#define vrshrn_n_s16(__p0, __p1) __extension__ ({ \
21566  int16x8_t __s0 = __p0; \
21567  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21568  int8x8_t __ret; \
21569  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 0); \
21570  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21571  __ret; \
21572})
21573#define __noswap_vrshrn_n_s16(__p0, __p1) __extension__ ({ \
21574  int16x8_t __s0 = __p0; \
21575  int8x8_t __ret; \
21576  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 0); \
21577  __ret; \
21578})
21579#endif
21580
21581#ifdef __LITTLE_ENDIAN__
21582__ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0) {
21583  uint32x4_t __ret;
21584  __ret = (uint32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 50);
21585  return __ret;
21586}
21587#else
21588__ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0) {
21589  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21590  uint32x4_t __ret;
21591  __ret = (uint32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 50);
21592  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21593  return __ret;
21594}
21595#endif
21596
21597#ifdef __LITTLE_ENDIAN__
21598__ai float32x4_t vrsqrteq_f32(float32x4_t __p0) {
21599  float32x4_t __ret;
21600  __ret = (float32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 41);
21601  return __ret;
21602}
21603#else
21604__ai float32x4_t vrsqrteq_f32(float32x4_t __p0) {
21605  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21606  float32x4_t __ret;
21607  __ret = (float32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 41);
21608  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21609  return __ret;
21610}
21611#endif
21612
21613#ifdef __LITTLE_ENDIAN__
21614__ai uint32x2_t vrsqrte_u32(uint32x2_t __p0) {
21615  uint32x2_t __ret;
21616  __ret = (uint32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 18);
21617  return __ret;
21618}
21619#else
21620__ai uint32x2_t vrsqrte_u32(uint32x2_t __p0) {
21621  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21622  uint32x2_t __ret;
21623  __ret = (uint32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 18);
21624  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21625  return __ret;
21626}
21627#endif
21628
21629#ifdef __LITTLE_ENDIAN__
21630__ai float32x2_t vrsqrte_f32(float32x2_t __p0) {
21631  float32x2_t __ret;
21632  __ret = (float32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 9);
21633  return __ret;
21634}
21635#else
21636__ai float32x2_t vrsqrte_f32(float32x2_t __p0) {
21637  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21638  float32x2_t __ret;
21639  __ret = (float32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 9);
21640  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21641  return __ret;
21642}
21643#endif
21644
21645#ifdef __LITTLE_ENDIAN__
21646__ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1) {
21647  float32x4_t __ret;
21648  __ret = (float32x4_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
21649  return __ret;
21650}
21651#else
21652__ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1) {
21653  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21654  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21655  float32x4_t __ret;
21656  __ret = (float32x4_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
21657  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21658  return __ret;
21659}
21660#endif
21661
21662#ifdef __LITTLE_ENDIAN__
21663__ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1) {
21664  float32x2_t __ret;
21665  __ret = (float32x2_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
21666  return __ret;
21667}
21668#else
21669__ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1) {
21670  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21671  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21672  float32x2_t __ret;
21673  __ret = (float32x2_t) __builtin_neon_vrsqrts_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
21674  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21675  return __ret;
21676}
21677#endif
21678
21679#ifdef __LITTLE_ENDIAN__
21680#define vrsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
21681  uint8x16_t __s0 = __p0; \
21682  uint8x16_t __s1 = __p1; \
21683  uint8x16_t __ret; \
21684  __ret = (uint8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
21685  __ret; \
21686})
21687#else
21688#define vrsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
21689  uint8x16_t __s0 = __p0; \
21690  uint8x16_t __s1 = __p1; \
21691  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21692  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21693  uint8x16_t __ret; \
21694  __ret = (uint8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
21695  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21696  __ret; \
21697})
21698#endif
21699
21700#ifdef __LITTLE_ENDIAN__
21701#define vrsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
21702  uint32x4_t __s0 = __p0; \
21703  uint32x4_t __s1 = __p1; \
21704  uint32x4_t __ret; \
21705  __ret = (uint32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
21706  __ret; \
21707})
21708#else
21709#define vrsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
21710  uint32x4_t __s0 = __p0; \
21711  uint32x4_t __s1 = __p1; \
21712  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21713  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
21714  uint32x4_t __ret; \
21715  __ret = (uint32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
21716  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21717  __ret; \
21718})
21719#endif
21720
21721#ifdef __LITTLE_ENDIAN__
21722#define vrsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
21723  uint64x2_t __s0 = __p0; \
21724  uint64x2_t __s1 = __p1; \
21725  uint64x2_t __ret; \
21726  __ret = (uint64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
21727  __ret; \
21728})
21729#else
21730#define vrsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
21731  uint64x2_t __s0 = __p0; \
21732  uint64x2_t __s1 = __p1; \
21733  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21734  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
21735  uint64x2_t __ret; \
21736  __ret = (uint64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
21737  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21738  __ret; \
21739})
21740#endif
21741
21742#ifdef __LITTLE_ENDIAN__
21743#define vrsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
21744  uint16x8_t __s0 = __p0; \
21745  uint16x8_t __s1 = __p1; \
21746  uint16x8_t __ret; \
21747  __ret = (uint16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
21748  __ret; \
21749})
21750#else
21751#define vrsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
21752  uint16x8_t __s0 = __p0; \
21753  uint16x8_t __s1 = __p1; \
21754  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21755  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21756  uint16x8_t __ret; \
21757  __ret = (uint16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
21758  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21759  __ret; \
21760})
21761#endif
21762
21763#ifdef __LITTLE_ENDIAN__
21764#define vrsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
21765  int8x16_t __s0 = __p0; \
21766  int8x16_t __s1 = __p1; \
21767  int8x16_t __ret; \
21768  __ret = (int8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
21769  __ret; \
21770})
21771#else
21772#define vrsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
21773  int8x16_t __s0 = __p0; \
21774  int8x16_t __s1 = __p1; \
21775  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21776  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21777  int8x16_t __ret; \
21778  __ret = (int8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
21779  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21780  __ret; \
21781})
21782#endif
21783
21784#ifdef __LITTLE_ENDIAN__
21785#define vrsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
21786  int32x4_t __s0 = __p0; \
21787  int32x4_t __s1 = __p1; \
21788  int32x4_t __ret; \
21789  __ret = (int32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
21790  __ret; \
21791})
21792#else
21793#define vrsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
21794  int32x4_t __s0 = __p0; \
21795  int32x4_t __s1 = __p1; \
21796  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21797  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
21798  int32x4_t __ret; \
21799  __ret = (int32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
21800  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21801  __ret; \
21802})
21803#endif
21804
21805#ifdef __LITTLE_ENDIAN__
21806#define vrsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
21807  int64x2_t __s0 = __p0; \
21808  int64x2_t __s1 = __p1; \
21809  int64x2_t __ret; \
21810  __ret = (int64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
21811  __ret; \
21812})
21813#else
21814#define vrsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
21815  int64x2_t __s0 = __p0; \
21816  int64x2_t __s1 = __p1; \
21817  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21818  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
21819  int64x2_t __ret; \
21820  __ret = (int64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
21821  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21822  __ret; \
21823})
21824#endif
21825
21826#ifdef __LITTLE_ENDIAN__
21827#define vrsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
21828  int16x8_t __s0 = __p0; \
21829  int16x8_t __s1 = __p1; \
21830  int16x8_t __ret; \
21831  __ret = (int16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
21832  __ret; \
21833})
21834#else
21835#define vrsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
21836  int16x8_t __s0 = __p0; \
21837  int16x8_t __s1 = __p1; \
21838  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21839  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21840  int16x8_t __ret; \
21841  __ret = (int16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
21842  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21843  __ret; \
21844})
21845#endif
21846
21847#ifdef __LITTLE_ENDIAN__
21848#define vrsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
21849  uint8x8_t __s0 = __p0; \
21850  uint8x8_t __s1 = __p1; \
21851  uint8x8_t __ret; \
21852  __ret = (uint8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
21853  __ret; \
21854})
21855#else
21856#define vrsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
21857  uint8x8_t __s0 = __p0; \
21858  uint8x8_t __s1 = __p1; \
21859  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21860  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21861  uint8x8_t __ret; \
21862  __ret = (uint8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
21863  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21864  __ret; \
21865})
21866#endif
21867
21868#ifdef __LITTLE_ENDIAN__
21869#define vrsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
21870  uint32x2_t __s0 = __p0; \
21871  uint32x2_t __s1 = __p1; \
21872  uint32x2_t __ret; \
21873  __ret = (uint32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
21874  __ret; \
21875})
21876#else
21877#define vrsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
21878  uint32x2_t __s0 = __p0; \
21879  uint32x2_t __s1 = __p1; \
21880  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21881  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
21882  uint32x2_t __ret; \
21883  __ret = (uint32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
21884  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21885  __ret; \
21886})
21887#endif
21888
21889#ifdef __LITTLE_ENDIAN__
21890#define vrsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
21891  uint64x1_t __s0 = __p0; \
21892  uint64x1_t __s1 = __p1; \
21893  uint64x1_t __ret; \
21894  __ret = (uint64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
21895  __ret; \
21896})
21897#else
21898#define vrsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
21899  uint64x1_t __s0 = __p0; \
21900  uint64x1_t __s1 = __p1; \
21901  uint64x1_t __ret; \
21902  __ret = (uint64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
21903  __ret; \
21904})
21905#endif
21906
21907#ifdef __LITTLE_ENDIAN__
21908#define vrsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
21909  uint16x4_t __s0 = __p0; \
21910  uint16x4_t __s1 = __p1; \
21911  uint16x4_t __ret; \
21912  __ret = (uint16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
21913  __ret; \
21914})
21915#else
21916#define vrsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
21917  uint16x4_t __s0 = __p0; \
21918  uint16x4_t __s1 = __p1; \
21919  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21920  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
21921  uint16x4_t __ret; \
21922  __ret = (uint16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
21923  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21924  __ret; \
21925})
21926#endif
21927
21928#ifdef __LITTLE_ENDIAN__
21929#define vrsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
21930  int8x8_t __s0 = __p0; \
21931  int8x8_t __s1 = __p1; \
21932  int8x8_t __ret; \
21933  __ret = (int8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
21934  __ret; \
21935})
21936#else
21937#define vrsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
21938  int8x8_t __s0 = __p0; \
21939  int8x8_t __s1 = __p1; \
21940  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21941  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21942  int8x8_t __ret; \
21943  __ret = (int8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
21944  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21945  __ret; \
21946})
21947#endif
21948
21949#ifdef __LITTLE_ENDIAN__
21950#define vrsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
21951  int32x2_t __s0 = __p0; \
21952  int32x2_t __s1 = __p1; \
21953  int32x2_t __ret; \
21954  __ret = (int32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
21955  __ret; \
21956})
21957#else
21958#define vrsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
21959  int32x2_t __s0 = __p0; \
21960  int32x2_t __s1 = __p1; \
21961  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21962  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
21963  int32x2_t __ret; \
21964  __ret = (int32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
21965  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21966  __ret; \
21967})
21968#endif
21969
21970#ifdef __LITTLE_ENDIAN__
21971#define vrsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
21972  int64x1_t __s0 = __p0; \
21973  int64x1_t __s1 = __p1; \
21974  int64x1_t __ret; \
21975  __ret = (int64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
21976  __ret; \
21977})
21978#else
21979#define vrsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
21980  int64x1_t __s0 = __p0; \
21981  int64x1_t __s1 = __p1; \
21982  int64x1_t __ret; \
21983  __ret = (int64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
21984  __ret; \
21985})
21986#endif
21987
21988#ifdef __LITTLE_ENDIAN__
21989#define vrsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
21990  int16x4_t __s0 = __p0; \
21991  int16x4_t __s1 = __p1; \
21992  int16x4_t __ret; \
21993  __ret = (int16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
21994  __ret; \
21995})
21996#else
21997#define vrsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
21998  int16x4_t __s0 = __p0; \
21999  int16x4_t __s1 = __p1; \
22000  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
22001  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22002  int16x4_t __ret; \
22003  __ret = (int16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
22004  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22005  __ret; \
22006})
22007#endif
22008
22009#ifdef __LITTLE_ENDIAN__
22010__ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
22011  uint16x4_t __ret;
22012  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
22013  return __ret;
22014}
22015#else
22016__ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
22017  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22018  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22019  uint16x4_t __ret;
22020  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
22021  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22022  return __ret;
22023}
22024__ai uint16x4_t __noswap_vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
22025  uint16x4_t __ret;
22026  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
22027  return __ret;
22028}
22029#endif
22030
22031#ifdef __LITTLE_ENDIAN__
22032__ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
22033  uint32x2_t __ret;
22034  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
22035  return __ret;
22036}
22037#else
22038__ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
22039  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22040  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22041  uint32x2_t __ret;
22042  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
22043  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22044  return __ret;
22045}
22046__ai uint32x2_t __noswap_vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
22047  uint32x2_t __ret;
22048  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
22049  return __ret;
22050}
22051#endif
22052
22053#ifdef __LITTLE_ENDIAN__
22054__ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
22055  uint8x8_t __ret;
22056  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
22057  return __ret;
22058}
22059#else
22060__ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
22061  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22062  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22063  uint8x8_t __ret;
22064  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
22065  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22066  return __ret;
22067}
22068__ai uint8x8_t __noswap_vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
22069  uint8x8_t __ret;
22070  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
22071  return __ret;
22072}
22073#endif
22074
22075#ifdef __LITTLE_ENDIAN__
22076__ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
22077  int16x4_t __ret;
22078  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
22079  return __ret;
22080}
22081#else
22082__ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
22083  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22084  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22085  int16x4_t __ret;
22086  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
22087  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22088  return __ret;
22089}
22090__ai int16x4_t __noswap_vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
22091  int16x4_t __ret;
22092  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
22093  return __ret;
22094}
22095#endif
22096
22097#ifdef __LITTLE_ENDIAN__
22098__ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
22099  int32x2_t __ret;
22100  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
22101  return __ret;
22102}
22103#else
22104__ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
22105  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22106  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22107  int32x2_t __ret;
22108  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
22109  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22110  return __ret;
22111}
22112__ai int32x2_t __noswap_vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
22113  int32x2_t __ret;
22114  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
22115  return __ret;
22116}
22117#endif
22118
22119#ifdef __LITTLE_ENDIAN__
22120__ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
22121  int8x8_t __ret;
22122  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
22123  return __ret;
22124}
22125#else
22126__ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
22127  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22128  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22129  int8x8_t __ret;
22130  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
22131  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22132  return __ret;
22133}
22134__ai int8x8_t __noswap_vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
22135  int8x8_t __ret;
22136  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
22137  return __ret;
22138}
22139#endif
22140
22141#ifdef __LITTLE_ENDIAN__
22142#define vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22143  poly8_t __s0 = __p0; \
22144  poly8x8_t __s1 = __p1; \
22145  poly8x8_t __ret; \
22146  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22147  __ret; \
22148})
22149#else
22150#define vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22151  poly8_t __s0 = __p0; \
22152  poly8x8_t __s1 = __p1; \
22153  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22154  poly8x8_t __ret; \
22155  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
22156  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22157  __ret; \
22158})
22159#define __noswap_vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22160  poly8_t __s0 = __p0; \
22161  poly8x8_t __s1 = __p1; \
22162  poly8x8_t __ret; \
22163  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22164  __ret; \
22165})
22166#endif
22167
22168#ifdef __LITTLE_ENDIAN__
22169#define vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22170  poly16_t __s0 = __p0; \
22171  poly16x4_t __s1 = __p1; \
22172  poly16x4_t __ret; \
22173  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22174  __ret; \
22175})
22176#else
22177#define vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22178  poly16_t __s0 = __p0; \
22179  poly16x4_t __s1 = __p1; \
22180  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22181  poly16x4_t __ret; \
22182  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
22183  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22184  __ret; \
22185})
22186#define __noswap_vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22187  poly16_t __s0 = __p0; \
22188  poly16x4_t __s1 = __p1; \
22189  poly16x4_t __ret; \
22190  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22191  __ret; \
22192})
22193#endif
22194
22195#ifdef __LITTLE_ENDIAN__
22196#define vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22197  poly8_t __s0 = __p0; \
22198  poly8x16_t __s1 = __p1; \
22199  poly8x16_t __ret; \
22200  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22201  __ret; \
22202})
22203#else
22204#define vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22205  poly8_t __s0 = __p0; \
22206  poly8x16_t __s1 = __p1; \
22207  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22208  poly8x16_t __ret; \
22209  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
22210  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22211  __ret; \
22212})
22213#define __noswap_vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22214  poly8_t __s0 = __p0; \
22215  poly8x16_t __s1 = __p1; \
22216  poly8x16_t __ret; \
22217  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22218  __ret; \
22219})
22220#endif
22221
22222#ifdef __LITTLE_ENDIAN__
22223#define vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22224  poly16_t __s0 = __p0; \
22225  poly16x8_t __s1 = __p1; \
22226  poly16x8_t __ret; \
22227  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22228  __ret; \
22229})
22230#else
22231#define vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22232  poly16_t __s0 = __p0; \
22233  poly16x8_t __s1 = __p1; \
22234  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22235  poly16x8_t __ret; \
22236  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
22237  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22238  __ret; \
22239})
22240#define __noswap_vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22241  poly16_t __s0 = __p0; \
22242  poly16x8_t __s1 = __p1; \
22243  poly16x8_t __ret; \
22244  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22245  __ret; \
22246})
22247#endif
22248
22249#ifdef __LITTLE_ENDIAN__
22250#define vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22251  uint8_t __s0 = __p0; \
22252  uint8x16_t __s1 = __p1; \
22253  uint8x16_t __ret; \
22254  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22255  __ret; \
22256})
22257#else
22258#define vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22259  uint8_t __s0 = __p0; \
22260  uint8x16_t __s1 = __p1; \
22261  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22262  uint8x16_t __ret; \
22263  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
22264  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22265  __ret; \
22266})
22267#define __noswap_vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22268  uint8_t __s0 = __p0; \
22269  uint8x16_t __s1 = __p1; \
22270  uint8x16_t __ret; \
22271  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22272  __ret; \
22273})
22274#endif
22275
22276#ifdef __LITTLE_ENDIAN__
22277#define vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22278  uint32_t __s0 = __p0; \
22279  uint32x4_t __s1 = __p1; \
22280  uint32x4_t __ret; \
22281  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22282  __ret; \
22283})
22284#else
22285#define vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22286  uint32_t __s0 = __p0; \
22287  uint32x4_t __s1 = __p1; \
22288  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22289  uint32x4_t __ret; \
22290  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__rev1, __p2); \
22291  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22292  __ret; \
22293})
22294#define __noswap_vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22295  uint32_t __s0 = __p0; \
22296  uint32x4_t __s1 = __p1; \
22297  uint32x4_t __ret; \
22298  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22299  __ret; \
22300})
22301#endif
22302
22303#ifdef __LITTLE_ENDIAN__
22304#define vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22305  uint64_t __s0 = __p0; \
22306  uint64x2_t __s1 = __p1; \
22307  uint64x2_t __ret; \
22308  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22309  __ret; \
22310})
22311#else
22312#define vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22313  uint64_t __s0 = __p0; \
22314  uint64x2_t __s1 = __p1; \
22315  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22316  uint64x2_t __ret; \
22317  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
22318  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22319  __ret; \
22320})
22321#define __noswap_vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22322  uint64_t __s0 = __p0; \
22323  uint64x2_t __s1 = __p1; \
22324  uint64x2_t __ret; \
22325  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22326  __ret; \
22327})
22328#endif
22329
22330#ifdef __LITTLE_ENDIAN__
22331#define vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22332  uint16_t __s0 = __p0; \
22333  uint16x8_t __s1 = __p1; \
22334  uint16x8_t __ret; \
22335  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22336  __ret; \
22337})
22338#else
22339#define vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22340  uint16_t __s0 = __p0; \
22341  uint16x8_t __s1 = __p1; \
22342  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22343  uint16x8_t __ret; \
22344  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
22345  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22346  __ret; \
22347})
22348#define __noswap_vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22349  uint16_t __s0 = __p0; \
22350  uint16x8_t __s1 = __p1; \
22351  uint16x8_t __ret; \
22352  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22353  __ret; \
22354})
22355#endif
22356
22357#ifdef __LITTLE_ENDIAN__
22358#define vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22359  int8_t __s0 = __p0; \
22360  int8x16_t __s1 = __p1; \
22361  int8x16_t __ret; \
22362  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22363  __ret; \
22364})
22365#else
22366#define vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22367  int8_t __s0 = __p0; \
22368  int8x16_t __s1 = __p1; \
22369  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22370  int8x16_t __ret; \
22371  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
22372  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22373  __ret; \
22374})
22375#define __noswap_vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22376  int8_t __s0 = __p0; \
22377  int8x16_t __s1 = __p1; \
22378  int8x16_t __ret; \
22379  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22380  __ret; \
22381})
22382#endif
22383
22384#ifdef __LITTLE_ENDIAN__
22385#define vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22386  float32_t __s0 = __p0; \
22387  float32x4_t __s1 = __p1; \
22388  float32x4_t __ret; \
22389  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__s1, __p2); \
22390  __ret; \
22391})
22392#else
22393#define vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22394  float32_t __s0 = __p0; \
22395  float32x4_t __s1 = __p1; \
22396  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22397  float32x4_t __ret; \
22398  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__rev1, __p2); \
22399  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22400  __ret; \
22401})
22402#define __noswap_vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22403  float32_t __s0 = __p0; \
22404  float32x4_t __s1 = __p1; \
22405  float32x4_t __ret; \
22406  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__s1, __p2); \
22407  __ret; \
22408})
22409#endif
22410
22411#ifdef __LITTLE_ENDIAN__
22412#define vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22413  int32_t __s0 = __p0; \
22414  int32x4_t __s1 = __p1; \
22415  int32x4_t __ret; \
22416  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22417  __ret; \
22418})
22419#else
22420#define vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22421  int32_t __s0 = __p0; \
22422  int32x4_t __s1 = __p1; \
22423  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22424  int32x4_t __ret; \
22425  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__rev1, __p2); \
22426  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22427  __ret; \
22428})
22429#define __noswap_vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22430  int32_t __s0 = __p0; \
22431  int32x4_t __s1 = __p1; \
22432  int32x4_t __ret; \
22433  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22434  __ret; \
22435})
22436#endif
22437
22438#ifdef __LITTLE_ENDIAN__
22439#define vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22440  int64_t __s0 = __p0; \
22441  int64x2_t __s1 = __p1; \
22442  int64x2_t __ret; \
22443  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22444  __ret; \
22445})
22446#else
22447#define vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22448  int64_t __s0 = __p0; \
22449  int64x2_t __s1 = __p1; \
22450  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22451  int64x2_t __ret; \
22452  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
22453  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22454  __ret; \
22455})
22456#define __noswap_vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22457  int64_t __s0 = __p0; \
22458  int64x2_t __s1 = __p1; \
22459  int64x2_t __ret; \
22460  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22461  __ret; \
22462})
22463#endif
22464
22465#ifdef __LITTLE_ENDIAN__
22466#define vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22467  int16_t __s0 = __p0; \
22468  int16x8_t __s1 = __p1; \
22469  int16x8_t __ret; \
22470  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22471  __ret; \
22472})
22473#else
22474#define vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22475  int16_t __s0 = __p0; \
22476  int16x8_t __s1 = __p1; \
22477  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22478  int16x8_t __ret; \
22479  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
22480  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22481  __ret; \
22482})
22483#define __noswap_vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22484  int16_t __s0 = __p0; \
22485  int16x8_t __s1 = __p1; \
22486  int16x8_t __ret; \
22487  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22488  __ret; \
22489})
22490#endif
22491
22492#ifdef __LITTLE_ENDIAN__
22493#define vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22494  uint8_t __s0 = __p0; \
22495  uint8x8_t __s1 = __p1; \
22496  uint8x8_t __ret; \
22497  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22498  __ret; \
22499})
22500#else
22501#define vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22502  uint8_t __s0 = __p0; \
22503  uint8x8_t __s1 = __p1; \
22504  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22505  uint8x8_t __ret; \
22506  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
22507  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22508  __ret; \
22509})
22510#define __noswap_vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22511  uint8_t __s0 = __p0; \
22512  uint8x8_t __s1 = __p1; \
22513  uint8x8_t __ret; \
22514  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22515  __ret; \
22516})
22517#endif
22518
22519#ifdef __LITTLE_ENDIAN__
22520#define vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22521  uint32_t __s0 = __p0; \
22522  uint32x2_t __s1 = __p1; \
22523  uint32x2_t __ret; \
22524  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22525  __ret; \
22526})
22527#else
22528#define vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22529  uint32_t __s0 = __p0; \
22530  uint32x2_t __s1 = __p1; \
22531  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22532  uint32x2_t __ret; \
22533  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__rev1, __p2); \
22534  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22535  __ret; \
22536})
22537#define __noswap_vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22538  uint32_t __s0 = __p0; \
22539  uint32x2_t __s1 = __p1; \
22540  uint32x2_t __ret; \
22541  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22542  __ret; \
22543})
22544#endif
22545
22546#ifdef __LITTLE_ENDIAN__
22547#define vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22548  uint64_t __s0 = __p0; \
22549  uint64x1_t __s1 = __p1; \
22550  uint64x1_t __ret; \
22551  __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22552  __ret; \
22553})
22554#else
22555#define vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22556  uint64_t __s0 = __p0; \
22557  uint64x1_t __s1 = __p1; \
22558  uint64x1_t __ret; \
22559  __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22560  __ret; \
22561})
22562#define __noswap_vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22563  uint64_t __s0 = __p0; \
22564  uint64x1_t __s1 = __p1; \
22565  uint64x1_t __ret; \
22566  __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22567  __ret; \
22568})
22569#endif
22570
22571#ifdef __LITTLE_ENDIAN__
22572#define vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22573  uint16_t __s0 = __p0; \
22574  uint16x4_t __s1 = __p1; \
22575  uint16x4_t __ret; \
22576  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22577  __ret; \
22578})
22579#else
22580#define vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22581  uint16_t __s0 = __p0; \
22582  uint16x4_t __s1 = __p1; \
22583  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22584  uint16x4_t __ret; \
22585  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
22586  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22587  __ret; \
22588})
22589#define __noswap_vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22590  uint16_t __s0 = __p0; \
22591  uint16x4_t __s1 = __p1; \
22592  uint16x4_t __ret; \
22593  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22594  __ret; \
22595})
22596#endif
22597
22598#ifdef __LITTLE_ENDIAN__
22599#define vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22600  int8_t __s0 = __p0; \
22601  int8x8_t __s1 = __p1; \
22602  int8x8_t __ret; \
22603  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22604  __ret; \
22605})
22606#else
22607#define vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22608  int8_t __s0 = __p0; \
22609  int8x8_t __s1 = __p1; \
22610  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22611  int8x8_t __ret; \
22612  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
22613  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22614  __ret; \
22615})
22616#define __noswap_vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22617  int8_t __s0 = __p0; \
22618  int8x8_t __s1 = __p1; \
22619  int8x8_t __ret; \
22620  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22621  __ret; \
22622})
22623#endif
22624
22625#ifdef __LITTLE_ENDIAN__
22626#define vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22627  float32_t __s0 = __p0; \
22628  float32x2_t __s1 = __p1; \
22629  float32x2_t __ret; \
22630  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__s1, __p2); \
22631  __ret; \
22632})
22633#else
22634#define vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22635  float32_t __s0 = __p0; \
22636  float32x2_t __s1 = __p1; \
22637  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22638  float32x2_t __ret; \
22639  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__rev1, __p2); \
22640  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22641  __ret; \
22642})
22643#define __noswap_vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22644  float32_t __s0 = __p0; \
22645  float32x2_t __s1 = __p1; \
22646  float32x2_t __ret; \
22647  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__s1, __p2); \
22648  __ret; \
22649})
22650#endif
22651
22652#ifdef __LITTLE_ENDIAN__
22653#define vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22654  int32_t __s0 = __p0; \
22655  int32x2_t __s1 = __p1; \
22656  int32x2_t __ret; \
22657  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22658  __ret; \
22659})
22660#else
22661#define vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22662  int32_t __s0 = __p0; \
22663  int32x2_t __s1 = __p1; \
22664  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22665  int32x2_t __ret; \
22666  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__rev1, __p2); \
22667  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22668  __ret; \
22669})
22670#define __noswap_vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22671  int32_t __s0 = __p0; \
22672  int32x2_t __s1 = __p1; \
22673  int32x2_t __ret; \
22674  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22675  __ret; \
22676})
22677#endif
22678
22679#ifdef __LITTLE_ENDIAN__
22680#define vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22681  int64_t __s0 = __p0; \
22682  int64x1_t __s1 = __p1; \
22683  int64x1_t __ret; \
22684  __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22685  __ret; \
22686})
22687#else
22688#define vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22689  int64_t __s0 = __p0; \
22690  int64x1_t __s1 = __p1; \
22691  int64x1_t __ret; \
22692  __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22693  __ret; \
22694})
22695#define __noswap_vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22696  int64_t __s0 = __p0; \
22697  int64x1_t __s1 = __p1; \
22698  int64x1_t __ret; \
22699  __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22700  __ret; \
22701})
22702#endif
22703
22704#ifdef __LITTLE_ENDIAN__
22705#define vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22706  int16_t __s0 = __p0; \
22707  int16x4_t __s1 = __p1; \
22708  int16x4_t __ret; \
22709  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22710  __ret; \
22711})
22712#else
22713#define vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22714  int16_t __s0 = __p0; \
22715  int16x4_t __s1 = __p1; \
22716  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22717  int16x4_t __ret; \
22718  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
22719  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22720  __ret; \
22721})
22722#define __noswap_vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22723  int16_t __s0 = __p0; \
22724  int16x4_t __s1 = __p1; \
22725  int16x4_t __ret; \
22726  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22727  __ret; \
22728})
22729#endif
22730
22731#ifdef __LITTLE_ENDIAN__
22732__ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
22733  uint8x16_t __ret;
22734  __ret = (uint8x16_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
22735  return __ret;
22736}
22737#else
22738__ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
22739  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22740  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22741  uint8x16_t __ret;
22742  __ret = (uint8x16_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
22743  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22744  return __ret;
22745}
22746#endif
22747
22748#ifdef __LITTLE_ENDIAN__
22749__ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
22750  uint32x4_t __ret;
22751  __ret = (uint32x4_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
22752  return __ret;
22753}
22754#else
22755__ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
22756  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22757  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22758  uint32x4_t __ret;
22759  __ret = (uint32x4_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
22760  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22761  return __ret;
22762}
22763#endif
22764
22765#ifdef __LITTLE_ENDIAN__
22766__ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
22767  uint64x2_t __ret;
22768  __ret = (uint64x2_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
22769  return __ret;
22770}
22771#else
22772__ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
22773  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22774  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22775  uint64x2_t __ret;
22776  __ret = (uint64x2_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
22777  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22778  return __ret;
22779}
22780#endif
22781
22782#ifdef __LITTLE_ENDIAN__
22783__ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
22784  uint16x8_t __ret;
22785  __ret = (uint16x8_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
22786  return __ret;
22787}
22788#else
22789__ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
22790  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22791  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22792  uint16x8_t __ret;
22793  __ret = (uint16x8_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
22794  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22795  return __ret;
22796}
22797#endif
22798
22799#ifdef __LITTLE_ENDIAN__
22800__ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1) {
22801  int8x16_t __ret;
22802  __ret = (int8x16_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
22803  return __ret;
22804}
22805#else
22806__ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1) {
22807  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22808  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22809  int8x16_t __ret;
22810  __ret = (int8x16_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
22811  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22812  return __ret;
22813}
22814#endif
22815
22816#ifdef __LITTLE_ENDIAN__
22817__ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1) {
22818  int32x4_t __ret;
22819  __ret = (int32x4_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
22820  return __ret;
22821}
22822#else
22823__ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1) {
22824  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22825  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22826  int32x4_t __ret;
22827  __ret = (int32x4_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
22828  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22829  return __ret;
22830}
22831#endif
22832
22833#ifdef __LITTLE_ENDIAN__
22834__ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1) {
22835  int64x2_t __ret;
22836  __ret = (int64x2_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
22837  return __ret;
22838}
22839#else
22840__ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1) {
22841  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22842  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22843  int64x2_t __ret;
22844  __ret = (int64x2_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
22845  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22846  return __ret;
22847}
22848#endif
22849
22850#ifdef __LITTLE_ENDIAN__
22851__ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1) {
22852  int16x8_t __ret;
22853  __ret = (int16x8_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
22854  return __ret;
22855}
22856#else
22857__ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1) {
22858  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22859  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22860  int16x8_t __ret;
22861  __ret = (int16x8_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
22862  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22863  return __ret;
22864}
22865#endif
22866
22867#ifdef __LITTLE_ENDIAN__
22868__ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1) {
22869  uint8x8_t __ret;
22870  __ret = (uint8x8_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
22871  return __ret;
22872}
22873#else
22874__ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1) {
22875  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22876  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22877  uint8x8_t __ret;
22878  __ret = (uint8x8_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
22879  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22880  return __ret;
22881}
22882#endif
22883
22884#ifdef __LITTLE_ENDIAN__
22885__ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1) {
22886  uint32x2_t __ret;
22887  __ret = (uint32x2_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
22888  return __ret;
22889}
22890#else
22891__ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1) {
22892  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22893  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22894  uint32x2_t __ret;
22895  __ret = (uint32x2_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
22896  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22897  return __ret;
22898}
22899#endif
22900
22901#ifdef __LITTLE_ENDIAN__
22902__ai uint64x1_t vshl_u64(uint64x1_t __p0, int64x1_t __p1) {
22903  uint64x1_t __ret;
22904  __ret = (uint64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
22905  return __ret;
22906}
22907#else
22908__ai uint64x1_t vshl_u64(uint64x1_t __p0, int64x1_t __p1) {
22909  uint64x1_t __ret;
22910  __ret = (uint64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
22911  return __ret;
22912}
22913#endif
22914
22915#ifdef __LITTLE_ENDIAN__
22916__ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1) {
22917  uint16x4_t __ret;
22918  __ret = (uint16x4_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
22919  return __ret;
22920}
22921#else
22922__ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1) {
22923  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22924  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22925  uint16x4_t __ret;
22926  __ret = (uint16x4_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
22927  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22928  return __ret;
22929}
22930#endif
22931
22932#ifdef __LITTLE_ENDIAN__
22933__ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1) {
22934  int8x8_t __ret;
22935  __ret = (int8x8_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
22936  return __ret;
22937}
22938#else
22939__ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1) {
22940  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22941  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22942  int8x8_t __ret;
22943  __ret = (int8x8_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
22944  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22945  return __ret;
22946}
22947#endif
22948
22949#ifdef __LITTLE_ENDIAN__
22950__ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1) {
22951  int32x2_t __ret;
22952  __ret = (int32x2_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
22953  return __ret;
22954}
22955#else
22956__ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1) {
22957  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22958  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22959  int32x2_t __ret;
22960  __ret = (int32x2_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
22961  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22962  return __ret;
22963}
22964#endif
22965
22966#ifdef __LITTLE_ENDIAN__
22967__ai int64x1_t vshl_s64(int64x1_t __p0, int64x1_t __p1) {
22968  int64x1_t __ret;
22969  __ret = (int64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
22970  return __ret;
22971}
22972#else
22973__ai int64x1_t vshl_s64(int64x1_t __p0, int64x1_t __p1) {
22974  int64x1_t __ret;
22975  __ret = (int64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
22976  return __ret;
22977}
22978#endif
22979
22980#ifdef __LITTLE_ENDIAN__
22981__ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1) {
22982  int16x4_t __ret;
22983  __ret = (int16x4_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
22984  return __ret;
22985}
22986#else
22987__ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1) {
22988  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22989  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22990  int16x4_t __ret;
22991  __ret = (int16x4_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
22992  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22993  return __ret;
22994}
22995#endif
22996
22997#ifdef __LITTLE_ENDIAN__
22998#define vshlq_n_u8(__p0, __p1) __extension__ ({ \
22999  uint8x16_t __s0 = __p0; \
23000  uint8x16_t __ret; \
23001  __ret = (uint8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 48); \
23002  __ret; \
23003})
23004#else
23005#define vshlq_n_u8(__p0, __p1) __extension__ ({ \
23006  uint8x16_t __s0 = __p0; \
23007  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23008  uint8x16_t __ret; \
23009  __ret = (uint8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 48); \
23010  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23011  __ret; \
23012})
23013#endif
23014
23015#ifdef __LITTLE_ENDIAN__
23016#define vshlq_n_u32(__p0, __p1) __extension__ ({ \
23017  uint32x4_t __s0 = __p0; \
23018  uint32x4_t __ret; \
23019  __ret = (uint32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 50); \
23020  __ret; \
23021})
23022#else
23023#define vshlq_n_u32(__p0, __p1) __extension__ ({ \
23024  uint32x4_t __s0 = __p0; \
23025  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23026  uint32x4_t __ret; \
23027  __ret = (uint32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 50); \
23028  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23029  __ret; \
23030})
23031#endif
23032
23033#ifdef __LITTLE_ENDIAN__
23034#define vshlq_n_u64(__p0, __p1) __extension__ ({ \
23035  uint64x2_t __s0 = __p0; \
23036  uint64x2_t __ret; \
23037  __ret = (uint64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 51); \
23038  __ret; \
23039})
23040#else
23041#define vshlq_n_u64(__p0, __p1) __extension__ ({ \
23042  uint64x2_t __s0 = __p0; \
23043  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23044  uint64x2_t __ret; \
23045  __ret = (uint64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 51); \
23046  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23047  __ret; \
23048})
23049#endif
23050
23051#ifdef __LITTLE_ENDIAN__
23052#define vshlq_n_u16(__p0, __p1) __extension__ ({ \
23053  uint16x8_t __s0 = __p0; \
23054  uint16x8_t __ret; \
23055  __ret = (uint16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 49); \
23056  __ret; \
23057})
23058#else
23059#define vshlq_n_u16(__p0, __p1) __extension__ ({ \
23060  uint16x8_t __s0 = __p0; \
23061  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23062  uint16x8_t __ret; \
23063  __ret = (uint16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 49); \
23064  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23065  __ret; \
23066})
23067#endif
23068
23069#ifdef __LITTLE_ENDIAN__
23070#define vshlq_n_s8(__p0, __p1) __extension__ ({ \
23071  int8x16_t __s0 = __p0; \
23072  int8x16_t __ret; \
23073  __ret = (int8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 32); \
23074  __ret; \
23075})
23076#else
23077#define vshlq_n_s8(__p0, __p1) __extension__ ({ \
23078  int8x16_t __s0 = __p0; \
23079  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23080  int8x16_t __ret; \
23081  __ret = (int8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 32); \
23082  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23083  __ret; \
23084})
23085#endif
23086
23087#ifdef __LITTLE_ENDIAN__
23088#define vshlq_n_s32(__p0, __p1) __extension__ ({ \
23089  int32x4_t __s0 = __p0; \
23090  int32x4_t __ret; \
23091  __ret = (int32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 34); \
23092  __ret; \
23093})
23094#else
23095#define vshlq_n_s32(__p0, __p1) __extension__ ({ \
23096  int32x4_t __s0 = __p0; \
23097  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23098  int32x4_t __ret; \
23099  __ret = (int32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 34); \
23100  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23101  __ret; \
23102})
23103#endif
23104
23105#ifdef __LITTLE_ENDIAN__
23106#define vshlq_n_s64(__p0, __p1) __extension__ ({ \
23107  int64x2_t __s0 = __p0; \
23108  int64x2_t __ret; \
23109  __ret = (int64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 35); \
23110  __ret; \
23111})
23112#else
23113#define vshlq_n_s64(__p0, __p1) __extension__ ({ \
23114  int64x2_t __s0 = __p0; \
23115  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23116  int64x2_t __ret; \
23117  __ret = (int64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 35); \
23118  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23119  __ret; \
23120})
23121#endif
23122
23123#ifdef __LITTLE_ENDIAN__
23124#define vshlq_n_s16(__p0, __p1) __extension__ ({ \
23125  int16x8_t __s0 = __p0; \
23126  int16x8_t __ret; \
23127  __ret = (int16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 33); \
23128  __ret; \
23129})
23130#else
23131#define vshlq_n_s16(__p0, __p1) __extension__ ({ \
23132  int16x8_t __s0 = __p0; \
23133  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23134  int16x8_t __ret; \
23135  __ret = (int16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 33); \
23136  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23137  __ret; \
23138})
23139#endif
23140
23141#ifdef __LITTLE_ENDIAN__
23142#define vshl_n_u8(__p0, __p1) __extension__ ({ \
23143  uint8x8_t __s0 = __p0; \
23144  uint8x8_t __ret; \
23145  __ret = (uint8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 16); \
23146  __ret; \
23147})
23148#else
23149#define vshl_n_u8(__p0, __p1) __extension__ ({ \
23150  uint8x8_t __s0 = __p0; \
23151  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23152  uint8x8_t __ret; \
23153  __ret = (uint8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 16); \
23154  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23155  __ret; \
23156})
23157#endif
23158
23159#ifdef __LITTLE_ENDIAN__
23160#define vshl_n_u32(__p0, __p1) __extension__ ({ \
23161  uint32x2_t __s0 = __p0; \
23162  uint32x2_t __ret; \
23163  __ret = (uint32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 18); \
23164  __ret; \
23165})
23166#else
23167#define vshl_n_u32(__p0, __p1) __extension__ ({ \
23168  uint32x2_t __s0 = __p0; \
23169  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23170  uint32x2_t __ret; \
23171  __ret = (uint32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 18); \
23172  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23173  __ret; \
23174})
23175#endif
23176
23177#ifdef __LITTLE_ENDIAN__
23178#define vshl_n_u64(__p0, __p1) __extension__ ({ \
23179  uint64x1_t __s0 = __p0; \
23180  uint64x1_t __ret; \
23181  __ret = (uint64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 19); \
23182  __ret; \
23183})
23184#else
23185#define vshl_n_u64(__p0, __p1) __extension__ ({ \
23186  uint64x1_t __s0 = __p0; \
23187  uint64x1_t __ret; \
23188  __ret = (uint64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 19); \
23189  __ret; \
23190})
23191#endif
23192
23193#ifdef __LITTLE_ENDIAN__
23194#define vshl_n_u16(__p0, __p1) __extension__ ({ \
23195  uint16x4_t __s0 = __p0; \
23196  uint16x4_t __ret; \
23197  __ret = (uint16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 17); \
23198  __ret; \
23199})
23200#else
23201#define vshl_n_u16(__p0, __p1) __extension__ ({ \
23202  uint16x4_t __s0 = __p0; \
23203  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23204  uint16x4_t __ret; \
23205  __ret = (uint16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 17); \
23206  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23207  __ret; \
23208})
23209#endif
23210
23211#ifdef __LITTLE_ENDIAN__
23212#define vshl_n_s8(__p0, __p1) __extension__ ({ \
23213  int8x8_t __s0 = __p0; \
23214  int8x8_t __ret; \
23215  __ret = (int8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 0); \
23216  __ret; \
23217})
23218#else
23219#define vshl_n_s8(__p0, __p1) __extension__ ({ \
23220  int8x8_t __s0 = __p0; \
23221  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23222  int8x8_t __ret; \
23223  __ret = (int8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 0); \
23224  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23225  __ret; \
23226})
23227#endif
23228
23229#ifdef __LITTLE_ENDIAN__
23230#define vshl_n_s32(__p0, __p1) __extension__ ({ \
23231  int32x2_t __s0 = __p0; \
23232  int32x2_t __ret; \
23233  __ret = (int32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 2); \
23234  __ret; \
23235})
23236#else
23237#define vshl_n_s32(__p0, __p1) __extension__ ({ \
23238  int32x2_t __s0 = __p0; \
23239  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23240  int32x2_t __ret; \
23241  __ret = (int32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 2); \
23242  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23243  __ret; \
23244})
23245#endif
23246
23247#ifdef __LITTLE_ENDIAN__
23248#define vshl_n_s64(__p0, __p1) __extension__ ({ \
23249  int64x1_t __s0 = __p0; \
23250  int64x1_t __ret; \
23251  __ret = (int64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 3); \
23252  __ret; \
23253})
23254#else
23255#define vshl_n_s64(__p0, __p1) __extension__ ({ \
23256  int64x1_t __s0 = __p0; \
23257  int64x1_t __ret; \
23258  __ret = (int64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 3); \
23259  __ret; \
23260})
23261#endif
23262
23263#ifdef __LITTLE_ENDIAN__
23264#define vshl_n_s16(__p0, __p1) __extension__ ({ \
23265  int16x4_t __s0 = __p0; \
23266  int16x4_t __ret; \
23267  __ret = (int16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 1); \
23268  __ret; \
23269})
23270#else
23271#define vshl_n_s16(__p0, __p1) __extension__ ({ \
23272  int16x4_t __s0 = __p0; \
23273  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23274  int16x4_t __ret; \
23275  __ret = (int16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 1); \
23276  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23277  __ret; \
23278})
23279#endif
23280
23281#ifdef __LITTLE_ENDIAN__
23282#define vshll_n_u8(__p0, __p1) __extension__ ({ \
23283  uint8x8_t __s0 = __p0; \
23284  uint16x8_t __ret; \
23285  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 49); \
23286  __ret; \
23287})
23288#else
23289#define vshll_n_u8(__p0, __p1) __extension__ ({ \
23290  uint8x8_t __s0 = __p0; \
23291  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23292  uint16x8_t __ret; \
23293  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 49); \
23294  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23295  __ret; \
23296})
23297#define __noswap_vshll_n_u8(__p0, __p1) __extension__ ({ \
23298  uint8x8_t __s0 = __p0; \
23299  uint16x8_t __ret; \
23300  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 49); \
23301  __ret; \
23302})
23303#endif
23304
23305#ifdef __LITTLE_ENDIAN__
23306#define vshll_n_u32(__p0, __p1) __extension__ ({ \
23307  uint32x2_t __s0 = __p0; \
23308  uint64x2_t __ret; \
23309  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 51); \
23310  __ret; \
23311})
23312#else
23313#define vshll_n_u32(__p0, __p1) __extension__ ({ \
23314  uint32x2_t __s0 = __p0; \
23315  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23316  uint64x2_t __ret; \
23317  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 51); \
23318  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23319  __ret; \
23320})
23321#define __noswap_vshll_n_u32(__p0, __p1) __extension__ ({ \
23322  uint32x2_t __s0 = __p0; \
23323  uint64x2_t __ret; \
23324  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 51); \
23325  __ret; \
23326})
23327#endif
23328
23329#ifdef __LITTLE_ENDIAN__
23330#define vshll_n_u16(__p0, __p1) __extension__ ({ \
23331  uint16x4_t __s0 = __p0; \
23332  uint32x4_t __ret; \
23333  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 50); \
23334  __ret; \
23335})
23336#else
23337#define vshll_n_u16(__p0, __p1) __extension__ ({ \
23338  uint16x4_t __s0 = __p0; \
23339  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23340  uint32x4_t __ret; \
23341  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 50); \
23342  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23343  __ret; \
23344})
23345#define __noswap_vshll_n_u16(__p0, __p1) __extension__ ({ \
23346  uint16x4_t __s0 = __p0; \
23347  uint32x4_t __ret; \
23348  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 50); \
23349  __ret; \
23350})
23351#endif
23352
23353#ifdef __LITTLE_ENDIAN__
23354#define vshll_n_s8(__p0, __p1) __extension__ ({ \
23355  int8x8_t __s0 = __p0; \
23356  int16x8_t __ret; \
23357  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 33); \
23358  __ret; \
23359})
23360#else
23361#define vshll_n_s8(__p0, __p1) __extension__ ({ \
23362  int8x8_t __s0 = __p0; \
23363  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23364  int16x8_t __ret; \
23365  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 33); \
23366  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23367  __ret; \
23368})
23369#define __noswap_vshll_n_s8(__p0, __p1) __extension__ ({ \
23370  int8x8_t __s0 = __p0; \
23371  int16x8_t __ret; \
23372  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 33); \
23373  __ret; \
23374})
23375#endif
23376
23377#ifdef __LITTLE_ENDIAN__
23378#define vshll_n_s32(__p0, __p1) __extension__ ({ \
23379  int32x2_t __s0 = __p0; \
23380  int64x2_t __ret; \
23381  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 35); \
23382  __ret; \
23383})
23384#else
23385#define vshll_n_s32(__p0, __p1) __extension__ ({ \
23386  int32x2_t __s0 = __p0; \
23387  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23388  int64x2_t __ret; \
23389  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 35); \
23390  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23391  __ret; \
23392})
23393#define __noswap_vshll_n_s32(__p0, __p1) __extension__ ({ \
23394  int32x2_t __s0 = __p0; \
23395  int64x2_t __ret; \
23396  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 35); \
23397  __ret; \
23398})
23399#endif
23400
23401#ifdef __LITTLE_ENDIAN__
23402#define vshll_n_s16(__p0, __p1) __extension__ ({ \
23403  int16x4_t __s0 = __p0; \
23404  int32x4_t __ret; \
23405  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 34); \
23406  __ret; \
23407})
23408#else
23409#define vshll_n_s16(__p0, __p1) __extension__ ({ \
23410  int16x4_t __s0 = __p0; \
23411  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23412  int32x4_t __ret; \
23413  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 34); \
23414  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23415  __ret; \
23416})
23417#define __noswap_vshll_n_s16(__p0, __p1) __extension__ ({ \
23418  int16x4_t __s0 = __p0; \
23419  int32x4_t __ret; \
23420  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 34); \
23421  __ret; \
23422})
23423#endif
23424
23425#ifdef __LITTLE_ENDIAN__
23426#define vshrq_n_u8(__p0, __p1) __extension__ ({ \
23427  uint8x16_t __s0 = __p0; \
23428  uint8x16_t __ret; \
23429  __ret = (uint8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 48); \
23430  __ret; \
23431})
23432#else
23433#define vshrq_n_u8(__p0, __p1) __extension__ ({ \
23434  uint8x16_t __s0 = __p0; \
23435  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23436  uint8x16_t __ret; \
23437  __ret = (uint8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 48); \
23438  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23439  __ret; \
23440})
23441#endif
23442
23443#ifdef __LITTLE_ENDIAN__
23444#define vshrq_n_u32(__p0, __p1) __extension__ ({ \
23445  uint32x4_t __s0 = __p0; \
23446  uint32x4_t __ret; \
23447  __ret = (uint32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 50); \
23448  __ret; \
23449})
23450#else
23451#define vshrq_n_u32(__p0, __p1) __extension__ ({ \
23452  uint32x4_t __s0 = __p0; \
23453  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23454  uint32x4_t __ret; \
23455  __ret = (uint32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 50); \
23456  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23457  __ret; \
23458})
23459#endif
23460
23461#ifdef __LITTLE_ENDIAN__
23462#define vshrq_n_u64(__p0, __p1) __extension__ ({ \
23463  uint64x2_t __s0 = __p0; \
23464  uint64x2_t __ret; \
23465  __ret = (uint64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 51); \
23466  __ret; \
23467})
23468#else
23469#define vshrq_n_u64(__p0, __p1) __extension__ ({ \
23470  uint64x2_t __s0 = __p0; \
23471  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23472  uint64x2_t __ret; \
23473  __ret = (uint64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 51); \
23474  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23475  __ret; \
23476})
23477#endif
23478
23479#ifdef __LITTLE_ENDIAN__
23480#define vshrq_n_u16(__p0, __p1) __extension__ ({ \
23481  uint16x8_t __s0 = __p0; \
23482  uint16x8_t __ret; \
23483  __ret = (uint16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 49); \
23484  __ret; \
23485})
23486#else
23487#define vshrq_n_u16(__p0, __p1) __extension__ ({ \
23488  uint16x8_t __s0 = __p0; \
23489  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23490  uint16x8_t __ret; \
23491  __ret = (uint16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 49); \
23492  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23493  __ret; \
23494})
23495#endif
23496
23497#ifdef __LITTLE_ENDIAN__
23498#define vshrq_n_s8(__p0, __p1) __extension__ ({ \
23499  int8x16_t __s0 = __p0; \
23500  int8x16_t __ret; \
23501  __ret = (int8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 32); \
23502  __ret; \
23503})
23504#else
23505#define vshrq_n_s8(__p0, __p1) __extension__ ({ \
23506  int8x16_t __s0 = __p0; \
23507  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23508  int8x16_t __ret; \
23509  __ret = (int8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 32); \
23510  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23511  __ret; \
23512})
23513#endif
23514
23515#ifdef __LITTLE_ENDIAN__
23516#define vshrq_n_s32(__p0, __p1) __extension__ ({ \
23517  int32x4_t __s0 = __p0; \
23518  int32x4_t __ret; \
23519  __ret = (int32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 34); \
23520  __ret; \
23521})
23522#else
23523#define vshrq_n_s32(__p0, __p1) __extension__ ({ \
23524  int32x4_t __s0 = __p0; \
23525  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23526  int32x4_t __ret; \
23527  __ret = (int32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 34); \
23528  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23529  __ret; \
23530})
23531#endif
23532
23533#ifdef __LITTLE_ENDIAN__
23534#define vshrq_n_s64(__p0, __p1) __extension__ ({ \
23535  int64x2_t __s0 = __p0; \
23536  int64x2_t __ret; \
23537  __ret = (int64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 35); \
23538  __ret; \
23539})
23540#else
23541#define vshrq_n_s64(__p0, __p1) __extension__ ({ \
23542  int64x2_t __s0 = __p0; \
23543  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23544  int64x2_t __ret; \
23545  __ret = (int64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 35); \
23546  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23547  __ret; \
23548})
23549#endif
23550
23551#ifdef __LITTLE_ENDIAN__
23552#define vshrq_n_s16(__p0, __p1) __extension__ ({ \
23553  int16x8_t __s0 = __p0; \
23554  int16x8_t __ret; \
23555  __ret = (int16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 33); \
23556  __ret; \
23557})
23558#else
23559#define vshrq_n_s16(__p0, __p1) __extension__ ({ \
23560  int16x8_t __s0 = __p0; \
23561  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23562  int16x8_t __ret; \
23563  __ret = (int16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 33); \
23564  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23565  __ret; \
23566})
23567#endif
23568
23569#ifdef __LITTLE_ENDIAN__
23570#define vshr_n_u8(__p0, __p1) __extension__ ({ \
23571  uint8x8_t __s0 = __p0; \
23572  uint8x8_t __ret; \
23573  __ret = (uint8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 16); \
23574  __ret; \
23575})
23576#else
23577#define vshr_n_u8(__p0, __p1) __extension__ ({ \
23578  uint8x8_t __s0 = __p0; \
23579  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23580  uint8x8_t __ret; \
23581  __ret = (uint8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 16); \
23582  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23583  __ret; \
23584})
23585#endif
23586
23587#ifdef __LITTLE_ENDIAN__
23588#define vshr_n_u32(__p0, __p1) __extension__ ({ \
23589  uint32x2_t __s0 = __p0; \
23590  uint32x2_t __ret; \
23591  __ret = (uint32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 18); \
23592  __ret; \
23593})
23594#else
23595#define vshr_n_u32(__p0, __p1) __extension__ ({ \
23596  uint32x2_t __s0 = __p0; \
23597  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23598  uint32x2_t __ret; \
23599  __ret = (uint32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 18); \
23600  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23601  __ret; \
23602})
23603#endif
23604
23605#ifdef __LITTLE_ENDIAN__
23606#define vshr_n_u64(__p0, __p1) __extension__ ({ \
23607  uint64x1_t __s0 = __p0; \
23608  uint64x1_t __ret; \
23609  __ret = (uint64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 19); \
23610  __ret; \
23611})
23612#else
23613#define vshr_n_u64(__p0, __p1) __extension__ ({ \
23614  uint64x1_t __s0 = __p0; \
23615  uint64x1_t __ret; \
23616  __ret = (uint64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 19); \
23617  __ret; \
23618})
23619#endif
23620
23621#ifdef __LITTLE_ENDIAN__
23622#define vshr_n_u16(__p0, __p1) __extension__ ({ \
23623  uint16x4_t __s0 = __p0; \
23624  uint16x4_t __ret; \
23625  __ret = (uint16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 17); \
23626  __ret; \
23627})
23628#else
23629#define vshr_n_u16(__p0, __p1) __extension__ ({ \
23630  uint16x4_t __s0 = __p0; \
23631  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23632  uint16x4_t __ret; \
23633  __ret = (uint16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 17); \
23634  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23635  __ret; \
23636})
23637#endif
23638
23639#ifdef __LITTLE_ENDIAN__
23640#define vshr_n_s8(__p0, __p1) __extension__ ({ \
23641  int8x8_t __s0 = __p0; \
23642  int8x8_t __ret; \
23643  __ret = (int8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 0); \
23644  __ret; \
23645})
23646#else
23647#define vshr_n_s8(__p0, __p1) __extension__ ({ \
23648  int8x8_t __s0 = __p0; \
23649  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23650  int8x8_t __ret; \
23651  __ret = (int8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 0); \
23652  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23653  __ret; \
23654})
23655#endif
23656
23657#ifdef __LITTLE_ENDIAN__
23658#define vshr_n_s32(__p0, __p1) __extension__ ({ \
23659  int32x2_t __s0 = __p0; \
23660  int32x2_t __ret; \
23661  __ret = (int32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 2); \
23662  __ret; \
23663})
23664#else
23665#define vshr_n_s32(__p0, __p1) __extension__ ({ \
23666  int32x2_t __s0 = __p0; \
23667  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23668  int32x2_t __ret; \
23669  __ret = (int32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 2); \
23670  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23671  __ret; \
23672})
23673#endif
23674
23675#ifdef __LITTLE_ENDIAN__
23676#define vshr_n_s64(__p0, __p1) __extension__ ({ \
23677  int64x1_t __s0 = __p0; \
23678  int64x1_t __ret; \
23679  __ret = (int64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 3); \
23680  __ret; \
23681})
23682#else
23683#define vshr_n_s64(__p0, __p1) __extension__ ({ \
23684  int64x1_t __s0 = __p0; \
23685  int64x1_t __ret; \
23686  __ret = (int64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 3); \
23687  __ret; \
23688})
23689#endif
23690
23691#ifdef __LITTLE_ENDIAN__
23692#define vshr_n_s16(__p0, __p1) __extension__ ({ \
23693  int16x4_t __s0 = __p0; \
23694  int16x4_t __ret; \
23695  __ret = (int16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 1); \
23696  __ret; \
23697})
23698#else
23699#define vshr_n_s16(__p0, __p1) __extension__ ({ \
23700  int16x4_t __s0 = __p0; \
23701  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23702  int16x4_t __ret; \
23703  __ret = (int16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 1); \
23704  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23705  __ret; \
23706})
23707#endif
23708
23709#ifdef __LITTLE_ENDIAN__
23710#define vshrn_n_u32(__p0, __p1) __extension__ ({ \
23711  uint32x4_t __s0 = __p0; \
23712  uint16x4_t __ret; \
23713  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 17); \
23714  __ret; \
23715})
23716#else
23717#define vshrn_n_u32(__p0, __p1) __extension__ ({ \
23718  uint32x4_t __s0 = __p0; \
23719  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23720  uint16x4_t __ret; \
23721  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 17); \
23722  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23723  __ret; \
23724})
23725#define __noswap_vshrn_n_u32(__p0, __p1) __extension__ ({ \
23726  uint32x4_t __s0 = __p0; \
23727  uint16x4_t __ret; \
23728  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 17); \
23729  __ret; \
23730})
23731#endif
23732
23733#ifdef __LITTLE_ENDIAN__
23734#define vshrn_n_u64(__p0, __p1) __extension__ ({ \
23735  uint64x2_t __s0 = __p0; \
23736  uint32x2_t __ret; \
23737  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 18); \
23738  __ret; \
23739})
23740#else
23741#define vshrn_n_u64(__p0, __p1) __extension__ ({ \
23742  uint64x2_t __s0 = __p0; \
23743  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23744  uint32x2_t __ret; \
23745  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 18); \
23746  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23747  __ret; \
23748})
23749#define __noswap_vshrn_n_u64(__p0, __p1) __extension__ ({ \
23750  uint64x2_t __s0 = __p0; \
23751  uint32x2_t __ret; \
23752  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 18); \
23753  __ret; \
23754})
23755#endif
23756
23757#ifdef __LITTLE_ENDIAN__
23758#define vshrn_n_u16(__p0, __p1) __extension__ ({ \
23759  uint16x8_t __s0 = __p0; \
23760  uint8x8_t __ret; \
23761  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 16); \
23762  __ret; \
23763})
23764#else
23765#define vshrn_n_u16(__p0, __p1) __extension__ ({ \
23766  uint16x8_t __s0 = __p0; \
23767  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23768  uint8x8_t __ret; \
23769  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 16); \
23770  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23771  __ret; \
23772})
23773#define __noswap_vshrn_n_u16(__p0, __p1) __extension__ ({ \
23774  uint16x8_t __s0 = __p0; \
23775  uint8x8_t __ret; \
23776  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 16); \
23777  __ret; \
23778})
23779#endif
23780
23781#ifdef __LITTLE_ENDIAN__
23782#define vshrn_n_s32(__p0, __p1) __extension__ ({ \
23783  int32x4_t __s0 = __p0; \
23784  int16x4_t __ret; \
23785  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 1); \
23786  __ret; \
23787})
23788#else
23789#define vshrn_n_s32(__p0, __p1) __extension__ ({ \
23790  int32x4_t __s0 = __p0; \
23791  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23792  int16x4_t __ret; \
23793  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 1); \
23794  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23795  __ret; \
23796})
23797#define __noswap_vshrn_n_s32(__p0, __p1) __extension__ ({ \
23798  int32x4_t __s0 = __p0; \
23799  int16x4_t __ret; \
23800  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 1); \
23801  __ret; \
23802})
23803#endif
23804
23805#ifdef __LITTLE_ENDIAN__
23806#define vshrn_n_s64(__p0, __p1) __extension__ ({ \
23807  int64x2_t __s0 = __p0; \
23808  int32x2_t __ret; \
23809  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 2); \
23810  __ret; \
23811})
23812#else
23813#define vshrn_n_s64(__p0, __p1) __extension__ ({ \
23814  int64x2_t __s0 = __p0; \
23815  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23816  int32x2_t __ret; \
23817  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 2); \
23818  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23819  __ret; \
23820})
23821#define __noswap_vshrn_n_s64(__p0, __p1) __extension__ ({ \
23822  int64x2_t __s0 = __p0; \
23823  int32x2_t __ret; \
23824  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 2); \
23825  __ret; \
23826})
23827#endif
23828
23829#ifdef __LITTLE_ENDIAN__
23830#define vshrn_n_s16(__p0, __p1) __extension__ ({ \
23831  int16x8_t __s0 = __p0; \
23832  int8x8_t __ret; \
23833  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 0); \
23834  __ret; \
23835})
23836#else
23837#define vshrn_n_s16(__p0, __p1) __extension__ ({ \
23838  int16x8_t __s0 = __p0; \
23839  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23840  int8x8_t __ret; \
23841  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 0); \
23842  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23843  __ret; \
23844})
23845#define __noswap_vshrn_n_s16(__p0, __p1) __extension__ ({ \
23846  int16x8_t __s0 = __p0; \
23847  int8x8_t __ret; \
23848  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 0); \
23849  __ret; \
23850})
23851#endif
23852
23853#ifdef __LITTLE_ENDIAN__
23854#define vsli_n_p8(__p0, __p1, __p2) __extension__ ({ \
23855  poly8x8_t __s0 = __p0; \
23856  poly8x8_t __s1 = __p1; \
23857  poly8x8_t __ret; \
23858  __ret = (poly8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
23859  __ret; \
23860})
23861#else
23862#define vsli_n_p8(__p0, __p1, __p2) __extension__ ({ \
23863  poly8x8_t __s0 = __p0; \
23864  poly8x8_t __s1 = __p1; \
23865  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23866  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23867  poly8x8_t __ret; \
23868  __ret = (poly8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
23869  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23870  __ret; \
23871})
23872#endif
23873
23874#ifdef __LITTLE_ENDIAN__
23875#define vsli_n_p16(__p0, __p1, __p2) __extension__ ({ \
23876  poly16x4_t __s0 = __p0; \
23877  poly16x4_t __s1 = __p1; \
23878  poly16x4_t __ret; \
23879  __ret = (poly16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
23880  __ret; \
23881})
23882#else
23883#define vsli_n_p16(__p0, __p1, __p2) __extension__ ({ \
23884  poly16x4_t __s0 = __p0; \
23885  poly16x4_t __s1 = __p1; \
23886  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23887  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23888  poly16x4_t __ret; \
23889  __ret = (poly16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
23890  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23891  __ret; \
23892})
23893#endif
23894
23895#ifdef __LITTLE_ENDIAN__
23896#define vsliq_n_p8(__p0, __p1, __p2) __extension__ ({ \
23897  poly8x16_t __s0 = __p0; \
23898  poly8x16_t __s1 = __p1; \
23899  poly8x16_t __ret; \
23900  __ret = (poly8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
23901  __ret; \
23902})
23903#else
23904#define vsliq_n_p8(__p0, __p1, __p2) __extension__ ({ \
23905  poly8x16_t __s0 = __p0; \
23906  poly8x16_t __s1 = __p1; \
23907  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23908  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23909  poly8x16_t __ret; \
23910  __ret = (poly8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
23911  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23912  __ret; \
23913})
23914#endif
23915
23916#ifdef __LITTLE_ENDIAN__
23917#define vsliq_n_p16(__p0, __p1, __p2) __extension__ ({ \
23918  poly16x8_t __s0 = __p0; \
23919  poly16x8_t __s1 = __p1; \
23920  poly16x8_t __ret; \
23921  __ret = (poly16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
23922  __ret; \
23923})
23924#else
23925#define vsliq_n_p16(__p0, __p1, __p2) __extension__ ({ \
23926  poly16x8_t __s0 = __p0; \
23927  poly16x8_t __s1 = __p1; \
23928  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23929  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23930  poly16x8_t __ret; \
23931  __ret = (poly16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
23932  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23933  __ret; \
23934})
23935#endif
23936
23937#ifdef __LITTLE_ENDIAN__
23938#define vsliq_n_u8(__p0, __p1, __p2) __extension__ ({ \
23939  uint8x16_t __s0 = __p0; \
23940  uint8x16_t __s1 = __p1; \
23941  uint8x16_t __ret; \
23942  __ret = (uint8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
23943  __ret; \
23944})
23945#else
23946#define vsliq_n_u8(__p0, __p1, __p2) __extension__ ({ \
23947  uint8x16_t __s0 = __p0; \
23948  uint8x16_t __s1 = __p1; \
23949  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23950  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23951  uint8x16_t __ret; \
23952  __ret = (uint8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
23953  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23954  __ret; \
23955})
23956#endif
23957
23958#ifdef __LITTLE_ENDIAN__
23959#define vsliq_n_u32(__p0, __p1, __p2) __extension__ ({ \
23960  uint32x4_t __s0 = __p0; \
23961  uint32x4_t __s1 = __p1; \
23962  uint32x4_t __ret; \
23963  __ret = (uint32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
23964  __ret; \
23965})
23966#else
23967#define vsliq_n_u32(__p0, __p1, __p2) __extension__ ({ \
23968  uint32x4_t __s0 = __p0; \
23969  uint32x4_t __s1 = __p1; \
23970  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23971  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23972  uint32x4_t __ret; \
23973  __ret = (uint32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
23974  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23975  __ret; \
23976})
23977#endif
23978
23979#ifdef __LITTLE_ENDIAN__
23980#define vsliq_n_u64(__p0, __p1, __p2) __extension__ ({ \
23981  uint64x2_t __s0 = __p0; \
23982  uint64x2_t __s1 = __p1; \
23983  uint64x2_t __ret; \
23984  __ret = (uint64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
23985  __ret; \
23986})
23987#else
23988#define vsliq_n_u64(__p0, __p1, __p2) __extension__ ({ \
23989  uint64x2_t __s0 = __p0; \
23990  uint64x2_t __s1 = __p1; \
23991  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23992  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
23993  uint64x2_t __ret; \
23994  __ret = (uint64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
23995  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23996  __ret; \
23997})
23998#endif
23999
24000#ifdef __LITTLE_ENDIAN__
24001#define vsliq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24002  uint16x8_t __s0 = __p0; \
24003  uint16x8_t __s1 = __p1; \
24004  uint16x8_t __ret; \
24005  __ret = (uint16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
24006  __ret; \
24007})
24008#else
24009#define vsliq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24010  uint16x8_t __s0 = __p0; \
24011  uint16x8_t __s1 = __p1; \
24012  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24013  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24014  uint16x8_t __ret; \
24015  __ret = (uint16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
24016  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24017  __ret; \
24018})
24019#endif
24020
24021#ifdef __LITTLE_ENDIAN__
24022#define vsliq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24023  int8x16_t __s0 = __p0; \
24024  int8x16_t __s1 = __p1; \
24025  int8x16_t __ret; \
24026  __ret = (int8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
24027  __ret; \
24028})
24029#else
24030#define vsliq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24031  int8x16_t __s0 = __p0; \
24032  int8x16_t __s1 = __p1; \
24033  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24034  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24035  int8x16_t __ret; \
24036  __ret = (int8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
24037  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24038  __ret; \
24039})
24040#endif
24041
24042#ifdef __LITTLE_ENDIAN__
24043#define vsliq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24044  int32x4_t __s0 = __p0; \
24045  int32x4_t __s1 = __p1; \
24046  int32x4_t __ret; \
24047  __ret = (int32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
24048  __ret; \
24049})
24050#else
24051#define vsliq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24052  int32x4_t __s0 = __p0; \
24053  int32x4_t __s1 = __p1; \
24054  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24055  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24056  int32x4_t __ret; \
24057  __ret = (int32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
24058  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24059  __ret; \
24060})
24061#endif
24062
24063#ifdef __LITTLE_ENDIAN__
24064#define vsliq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24065  int64x2_t __s0 = __p0; \
24066  int64x2_t __s1 = __p1; \
24067  int64x2_t __ret; \
24068  __ret = (int64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
24069  __ret; \
24070})
24071#else
24072#define vsliq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24073  int64x2_t __s0 = __p0; \
24074  int64x2_t __s1 = __p1; \
24075  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24076  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24077  int64x2_t __ret; \
24078  __ret = (int64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
24079  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24080  __ret; \
24081})
24082#endif
24083
24084#ifdef __LITTLE_ENDIAN__
24085#define vsliq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24086  int16x8_t __s0 = __p0; \
24087  int16x8_t __s1 = __p1; \
24088  int16x8_t __ret; \
24089  __ret = (int16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
24090  __ret; \
24091})
24092#else
24093#define vsliq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24094  int16x8_t __s0 = __p0; \
24095  int16x8_t __s1 = __p1; \
24096  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24097  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24098  int16x8_t __ret; \
24099  __ret = (int16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
24100  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24101  __ret; \
24102})
24103#endif
24104
24105#ifdef __LITTLE_ENDIAN__
24106#define vsli_n_u8(__p0, __p1, __p2) __extension__ ({ \
24107  uint8x8_t __s0 = __p0; \
24108  uint8x8_t __s1 = __p1; \
24109  uint8x8_t __ret; \
24110  __ret = (uint8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
24111  __ret; \
24112})
24113#else
24114#define vsli_n_u8(__p0, __p1, __p2) __extension__ ({ \
24115  uint8x8_t __s0 = __p0; \
24116  uint8x8_t __s1 = __p1; \
24117  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24118  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24119  uint8x8_t __ret; \
24120  __ret = (uint8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
24121  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24122  __ret; \
24123})
24124#endif
24125
24126#ifdef __LITTLE_ENDIAN__
24127#define vsli_n_u32(__p0, __p1, __p2) __extension__ ({ \
24128  uint32x2_t __s0 = __p0; \
24129  uint32x2_t __s1 = __p1; \
24130  uint32x2_t __ret; \
24131  __ret = (uint32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
24132  __ret; \
24133})
24134#else
24135#define vsli_n_u32(__p0, __p1, __p2) __extension__ ({ \
24136  uint32x2_t __s0 = __p0; \
24137  uint32x2_t __s1 = __p1; \
24138  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24139  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24140  uint32x2_t __ret; \
24141  __ret = (uint32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
24142  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24143  __ret; \
24144})
24145#endif
24146
24147#ifdef __LITTLE_ENDIAN__
24148#define vsli_n_u64(__p0, __p1, __p2) __extension__ ({ \
24149  uint64x1_t __s0 = __p0; \
24150  uint64x1_t __s1 = __p1; \
24151  uint64x1_t __ret; \
24152  __ret = (uint64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24153  __ret; \
24154})
24155#else
24156#define vsli_n_u64(__p0, __p1, __p2) __extension__ ({ \
24157  uint64x1_t __s0 = __p0; \
24158  uint64x1_t __s1 = __p1; \
24159  uint64x1_t __ret; \
24160  __ret = (uint64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24161  __ret; \
24162})
24163#endif
24164
24165#ifdef __LITTLE_ENDIAN__
24166#define vsli_n_u16(__p0, __p1, __p2) __extension__ ({ \
24167  uint16x4_t __s0 = __p0; \
24168  uint16x4_t __s1 = __p1; \
24169  uint16x4_t __ret; \
24170  __ret = (uint16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
24171  __ret; \
24172})
24173#else
24174#define vsli_n_u16(__p0, __p1, __p2) __extension__ ({ \
24175  uint16x4_t __s0 = __p0; \
24176  uint16x4_t __s1 = __p1; \
24177  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24178  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24179  uint16x4_t __ret; \
24180  __ret = (uint16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
24181  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24182  __ret; \
24183})
24184#endif
24185
24186#ifdef __LITTLE_ENDIAN__
24187#define vsli_n_s8(__p0, __p1, __p2) __extension__ ({ \
24188  int8x8_t __s0 = __p0; \
24189  int8x8_t __s1 = __p1; \
24190  int8x8_t __ret; \
24191  __ret = (int8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
24192  __ret; \
24193})
24194#else
24195#define vsli_n_s8(__p0, __p1, __p2) __extension__ ({ \
24196  int8x8_t __s0 = __p0; \
24197  int8x8_t __s1 = __p1; \
24198  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24199  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24200  int8x8_t __ret; \
24201  __ret = (int8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
24202  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24203  __ret; \
24204})
24205#endif
24206
24207#ifdef __LITTLE_ENDIAN__
24208#define vsli_n_s32(__p0, __p1, __p2) __extension__ ({ \
24209  int32x2_t __s0 = __p0; \
24210  int32x2_t __s1 = __p1; \
24211  int32x2_t __ret; \
24212  __ret = (int32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
24213  __ret; \
24214})
24215#else
24216#define vsli_n_s32(__p0, __p1, __p2) __extension__ ({ \
24217  int32x2_t __s0 = __p0; \
24218  int32x2_t __s1 = __p1; \
24219  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24220  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24221  int32x2_t __ret; \
24222  __ret = (int32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
24223  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24224  __ret; \
24225})
24226#endif
24227
24228#ifdef __LITTLE_ENDIAN__
24229#define vsli_n_s64(__p0, __p1, __p2) __extension__ ({ \
24230  int64x1_t __s0 = __p0; \
24231  int64x1_t __s1 = __p1; \
24232  int64x1_t __ret; \
24233  __ret = (int64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24234  __ret; \
24235})
24236#else
24237#define vsli_n_s64(__p0, __p1, __p2) __extension__ ({ \
24238  int64x1_t __s0 = __p0; \
24239  int64x1_t __s1 = __p1; \
24240  int64x1_t __ret; \
24241  __ret = (int64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24242  __ret; \
24243})
24244#endif
24245
24246#ifdef __LITTLE_ENDIAN__
24247#define vsli_n_s16(__p0, __p1, __p2) __extension__ ({ \
24248  int16x4_t __s0 = __p0; \
24249  int16x4_t __s1 = __p1; \
24250  int16x4_t __ret; \
24251  __ret = (int16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
24252  __ret; \
24253})
24254#else
24255#define vsli_n_s16(__p0, __p1, __p2) __extension__ ({ \
24256  int16x4_t __s0 = __p0; \
24257  int16x4_t __s1 = __p1; \
24258  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24259  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24260  int16x4_t __ret; \
24261  __ret = (int16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
24262  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24263  __ret; \
24264})
24265#endif
24266
24267#ifdef __LITTLE_ENDIAN__
24268#define vsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24269  uint8x16_t __s0 = __p0; \
24270  uint8x16_t __s1 = __p1; \
24271  uint8x16_t __ret; \
24272  __ret = (uint8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
24273  __ret; \
24274})
24275#else
24276#define vsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24277  uint8x16_t __s0 = __p0; \
24278  uint8x16_t __s1 = __p1; \
24279  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24280  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24281  uint8x16_t __ret; \
24282  __ret = (uint8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
24283  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24284  __ret; \
24285})
24286#endif
24287
24288#ifdef __LITTLE_ENDIAN__
24289#define vsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24290  uint32x4_t __s0 = __p0; \
24291  uint32x4_t __s1 = __p1; \
24292  uint32x4_t __ret; \
24293  __ret = (uint32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
24294  __ret; \
24295})
24296#else
24297#define vsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24298  uint32x4_t __s0 = __p0; \
24299  uint32x4_t __s1 = __p1; \
24300  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24301  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24302  uint32x4_t __ret; \
24303  __ret = (uint32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
24304  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24305  __ret; \
24306})
24307#endif
24308
24309#ifdef __LITTLE_ENDIAN__
24310#define vsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24311  uint64x2_t __s0 = __p0; \
24312  uint64x2_t __s1 = __p1; \
24313  uint64x2_t __ret; \
24314  __ret = (uint64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
24315  __ret; \
24316})
24317#else
24318#define vsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24319  uint64x2_t __s0 = __p0; \
24320  uint64x2_t __s1 = __p1; \
24321  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24322  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24323  uint64x2_t __ret; \
24324  __ret = (uint64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
24325  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24326  __ret; \
24327})
24328#endif
24329
24330#ifdef __LITTLE_ENDIAN__
24331#define vsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24332  uint16x8_t __s0 = __p0; \
24333  uint16x8_t __s1 = __p1; \
24334  uint16x8_t __ret; \
24335  __ret = (uint16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
24336  __ret; \
24337})
24338#else
24339#define vsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24340  uint16x8_t __s0 = __p0; \
24341  uint16x8_t __s1 = __p1; \
24342  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24343  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24344  uint16x8_t __ret; \
24345  __ret = (uint16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
24346  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24347  __ret; \
24348})
24349#endif
24350
24351#ifdef __LITTLE_ENDIAN__
24352#define vsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24353  int8x16_t __s0 = __p0; \
24354  int8x16_t __s1 = __p1; \
24355  int8x16_t __ret; \
24356  __ret = (int8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
24357  __ret; \
24358})
24359#else
24360#define vsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24361  int8x16_t __s0 = __p0; \
24362  int8x16_t __s1 = __p1; \
24363  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24364  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24365  int8x16_t __ret; \
24366  __ret = (int8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
24367  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24368  __ret; \
24369})
24370#endif
24371
24372#ifdef __LITTLE_ENDIAN__
24373#define vsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24374  int32x4_t __s0 = __p0; \
24375  int32x4_t __s1 = __p1; \
24376  int32x4_t __ret; \
24377  __ret = (int32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
24378  __ret; \
24379})
24380#else
24381#define vsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24382  int32x4_t __s0 = __p0; \
24383  int32x4_t __s1 = __p1; \
24384  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24385  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24386  int32x4_t __ret; \
24387  __ret = (int32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
24388  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24389  __ret; \
24390})
24391#endif
24392
24393#ifdef __LITTLE_ENDIAN__
24394#define vsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24395  int64x2_t __s0 = __p0; \
24396  int64x2_t __s1 = __p1; \
24397  int64x2_t __ret; \
24398  __ret = (int64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
24399  __ret; \
24400})
24401#else
24402#define vsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24403  int64x2_t __s0 = __p0; \
24404  int64x2_t __s1 = __p1; \
24405  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24406  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24407  int64x2_t __ret; \
24408  __ret = (int64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
24409  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24410  __ret; \
24411})
24412#endif
24413
24414#ifdef __LITTLE_ENDIAN__
24415#define vsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24416  int16x8_t __s0 = __p0; \
24417  int16x8_t __s1 = __p1; \
24418  int16x8_t __ret; \
24419  __ret = (int16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
24420  __ret; \
24421})
24422#else
24423#define vsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24424  int16x8_t __s0 = __p0; \
24425  int16x8_t __s1 = __p1; \
24426  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24427  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24428  int16x8_t __ret; \
24429  __ret = (int16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
24430  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24431  __ret; \
24432})
24433#endif
24434
24435#ifdef __LITTLE_ENDIAN__
24436#define vsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
24437  uint8x8_t __s0 = __p0; \
24438  uint8x8_t __s1 = __p1; \
24439  uint8x8_t __ret; \
24440  __ret = (uint8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
24441  __ret; \
24442})
24443#else
24444#define vsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
24445  uint8x8_t __s0 = __p0; \
24446  uint8x8_t __s1 = __p1; \
24447  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24448  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24449  uint8x8_t __ret; \
24450  __ret = (uint8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
24451  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24452  __ret; \
24453})
24454#endif
24455
24456#ifdef __LITTLE_ENDIAN__
24457#define vsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
24458  uint32x2_t __s0 = __p0; \
24459  uint32x2_t __s1 = __p1; \
24460  uint32x2_t __ret; \
24461  __ret = (uint32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
24462  __ret; \
24463})
24464#else
24465#define vsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
24466  uint32x2_t __s0 = __p0; \
24467  uint32x2_t __s1 = __p1; \
24468  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24469  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24470  uint32x2_t __ret; \
24471  __ret = (uint32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
24472  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24473  __ret; \
24474})
24475#endif
24476
24477#ifdef __LITTLE_ENDIAN__
24478#define vsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
24479  uint64x1_t __s0 = __p0; \
24480  uint64x1_t __s1 = __p1; \
24481  uint64x1_t __ret; \
24482  __ret = (uint64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24483  __ret; \
24484})
24485#else
24486#define vsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
24487  uint64x1_t __s0 = __p0; \
24488  uint64x1_t __s1 = __p1; \
24489  uint64x1_t __ret; \
24490  __ret = (uint64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24491  __ret; \
24492})
24493#endif
24494
24495#ifdef __LITTLE_ENDIAN__
24496#define vsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
24497  uint16x4_t __s0 = __p0; \
24498  uint16x4_t __s1 = __p1; \
24499  uint16x4_t __ret; \
24500  __ret = (uint16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
24501  __ret; \
24502})
24503#else
24504#define vsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
24505  uint16x4_t __s0 = __p0; \
24506  uint16x4_t __s1 = __p1; \
24507  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24508  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24509  uint16x4_t __ret; \
24510  __ret = (uint16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
24511  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24512  __ret; \
24513})
24514#endif
24515
24516#ifdef __LITTLE_ENDIAN__
24517#define vsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
24518  int8x8_t __s0 = __p0; \
24519  int8x8_t __s1 = __p1; \
24520  int8x8_t __ret; \
24521  __ret = (int8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
24522  __ret; \
24523})
24524#else
24525#define vsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
24526  int8x8_t __s0 = __p0; \
24527  int8x8_t __s1 = __p1; \
24528  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24529  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24530  int8x8_t __ret; \
24531  __ret = (int8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
24532  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24533  __ret; \
24534})
24535#endif
24536
24537#ifdef __LITTLE_ENDIAN__
24538#define vsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
24539  int32x2_t __s0 = __p0; \
24540  int32x2_t __s1 = __p1; \
24541  int32x2_t __ret; \
24542  __ret = (int32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
24543  __ret; \
24544})
24545#else
24546#define vsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
24547  int32x2_t __s0 = __p0; \
24548  int32x2_t __s1 = __p1; \
24549  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24550  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24551  int32x2_t __ret; \
24552  __ret = (int32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
24553  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24554  __ret; \
24555})
24556#endif
24557
24558#ifdef __LITTLE_ENDIAN__
24559#define vsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
24560  int64x1_t __s0 = __p0; \
24561  int64x1_t __s1 = __p1; \
24562  int64x1_t __ret; \
24563  __ret = (int64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24564  __ret; \
24565})
24566#else
24567#define vsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
24568  int64x1_t __s0 = __p0; \
24569  int64x1_t __s1 = __p1; \
24570  int64x1_t __ret; \
24571  __ret = (int64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24572  __ret; \
24573})
24574#endif
24575
24576#ifdef __LITTLE_ENDIAN__
24577#define vsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
24578  int16x4_t __s0 = __p0; \
24579  int16x4_t __s1 = __p1; \
24580  int16x4_t __ret; \
24581  __ret = (int16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
24582  __ret; \
24583})
24584#else
24585#define vsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
24586  int16x4_t __s0 = __p0; \
24587  int16x4_t __s1 = __p1; \
24588  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24589  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24590  int16x4_t __ret; \
24591  __ret = (int16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
24592  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24593  __ret; \
24594})
24595#endif
24596
24597#ifdef __LITTLE_ENDIAN__
24598#define vsri_n_p8(__p0, __p1, __p2) __extension__ ({ \
24599  poly8x8_t __s0 = __p0; \
24600  poly8x8_t __s1 = __p1; \
24601  poly8x8_t __ret; \
24602  __ret = (poly8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
24603  __ret; \
24604})
24605#else
24606#define vsri_n_p8(__p0, __p1, __p2) __extension__ ({ \
24607  poly8x8_t __s0 = __p0; \
24608  poly8x8_t __s1 = __p1; \
24609  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24610  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24611  poly8x8_t __ret; \
24612  __ret = (poly8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
24613  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24614  __ret; \
24615})
24616#endif
24617
24618#ifdef __LITTLE_ENDIAN__
24619#define vsri_n_p16(__p0, __p1, __p2) __extension__ ({ \
24620  poly16x4_t __s0 = __p0; \
24621  poly16x4_t __s1 = __p1; \
24622  poly16x4_t __ret; \
24623  __ret = (poly16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
24624  __ret; \
24625})
24626#else
24627#define vsri_n_p16(__p0, __p1, __p2) __extension__ ({ \
24628  poly16x4_t __s0 = __p0; \
24629  poly16x4_t __s1 = __p1; \
24630  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24631  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24632  poly16x4_t __ret; \
24633  __ret = (poly16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
24634  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24635  __ret; \
24636})
24637#endif
24638
24639#ifdef __LITTLE_ENDIAN__
24640#define vsriq_n_p8(__p0, __p1, __p2) __extension__ ({ \
24641  poly8x16_t __s0 = __p0; \
24642  poly8x16_t __s1 = __p1; \
24643  poly8x16_t __ret; \
24644  __ret = (poly8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
24645  __ret; \
24646})
24647#else
24648#define vsriq_n_p8(__p0, __p1, __p2) __extension__ ({ \
24649  poly8x16_t __s0 = __p0; \
24650  poly8x16_t __s1 = __p1; \
24651  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24652  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24653  poly8x16_t __ret; \
24654  __ret = (poly8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
24655  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24656  __ret; \
24657})
24658#endif
24659
24660#ifdef __LITTLE_ENDIAN__
24661#define vsriq_n_p16(__p0, __p1, __p2) __extension__ ({ \
24662  poly16x8_t __s0 = __p0; \
24663  poly16x8_t __s1 = __p1; \
24664  poly16x8_t __ret; \
24665  __ret = (poly16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
24666  __ret; \
24667})
24668#else
24669#define vsriq_n_p16(__p0, __p1, __p2) __extension__ ({ \
24670  poly16x8_t __s0 = __p0; \
24671  poly16x8_t __s1 = __p1; \
24672  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24673  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24674  poly16x8_t __ret; \
24675  __ret = (poly16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
24676  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24677  __ret; \
24678})
24679#endif
24680
24681#ifdef __LITTLE_ENDIAN__
24682#define vsriq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24683  uint8x16_t __s0 = __p0; \
24684  uint8x16_t __s1 = __p1; \
24685  uint8x16_t __ret; \
24686  __ret = (uint8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
24687  __ret; \
24688})
24689#else
24690#define vsriq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24691  uint8x16_t __s0 = __p0; \
24692  uint8x16_t __s1 = __p1; \
24693  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24694  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24695  uint8x16_t __ret; \
24696  __ret = (uint8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
24697  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24698  __ret; \
24699})
24700#endif
24701
24702#ifdef __LITTLE_ENDIAN__
24703#define vsriq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24704  uint32x4_t __s0 = __p0; \
24705  uint32x4_t __s1 = __p1; \
24706  uint32x4_t __ret; \
24707  __ret = (uint32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
24708  __ret; \
24709})
24710#else
24711#define vsriq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24712  uint32x4_t __s0 = __p0; \
24713  uint32x4_t __s1 = __p1; \
24714  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24715  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24716  uint32x4_t __ret; \
24717  __ret = (uint32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
24718  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24719  __ret; \
24720})
24721#endif
24722
24723#ifdef __LITTLE_ENDIAN__
24724#define vsriq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24725  uint64x2_t __s0 = __p0; \
24726  uint64x2_t __s1 = __p1; \
24727  uint64x2_t __ret; \
24728  __ret = (uint64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
24729  __ret; \
24730})
24731#else
24732#define vsriq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24733  uint64x2_t __s0 = __p0; \
24734  uint64x2_t __s1 = __p1; \
24735  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24736  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24737  uint64x2_t __ret; \
24738  __ret = (uint64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
24739  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24740  __ret; \
24741})
24742#endif
24743
24744#ifdef __LITTLE_ENDIAN__
24745#define vsriq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24746  uint16x8_t __s0 = __p0; \
24747  uint16x8_t __s1 = __p1; \
24748  uint16x8_t __ret; \
24749  __ret = (uint16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
24750  __ret; \
24751})
24752#else
24753#define vsriq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24754  uint16x8_t __s0 = __p0; \
24755  uint16x8_t __s1 = __p1; \
24756  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24757  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24758  uint16x8_t __ret; \
24759  __ret = (uint16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
24760  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24761  __ret; \
24762})
24763#endif
24764
24765#ifdef __LITTLE_ENDIAN__
24766#define vsriq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24767  int8x16_t __s0 = __p0; \
24768  int8x16_t __s1 = __p1; \
24769  int8x16_t __ret; \
24770  __ret = (int8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
24771  __ret; \
24772})
24773#else
24774#define vsriq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24775  int8x16_t __s0 = __p0; \
24776  int8x16_t __s1 = __p1; \
24777  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24778  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24779  int8x16_t __ret; \
24780  __ret = (int8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
24781  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24782  __ret; \
24783})
24784#endif
24785
24786#ifdef __LITTLE_ENDIAN__
24787#define vsriq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24788  int32x4_t __s0 = __p0; \
24789  int32x4_t __s1 = __p1; \
24790  int32x4_t __ret; \
24791  __ret = (int32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
24792  __ret; \
24793})
24794#else
24795#define vsriq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24796  int32x4_t __s0 = __p0; \
24797  int32x4_t __s1 = __p1; \
24798  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24799  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24800  int32x4_t __ret; \
24801  __ret = (int32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
24802  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24803  __ret; \
24804})
24805#endif
24806
24807#ifdef __LITTLE_ENDIAN__
24808#define vsriq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24809  int64x2_t __s0 = __p0; \
24810  int64x2_t __s1 = __p1; \
24811  int64x2_t __ret; \
24812  __ret = (int64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
24813  __ret; \
24814})
24815#else
24816#define vsriq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24817  int64x2_t __s0 = __p0; \
24818  int64x2_t __s1 = __p1; \
24819  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24820  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24821  int64x2_t __ret; \
24822  __ret = (int64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
24823  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24824  __ret; \
24825})
24826#endif
24827
24828#ifdef __LITTLE_ENDIAN__
24829#define vsriq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24830  int16x8_t __s0 = __p0; \
24831  int16x8_t __s1 = __p1; \
24832  int16x8_t __ret; \
24833  __ret = (int16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
24834  __ret; \
24835})
24836#else
24837#define vsriq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24838  int16x8_t __s0 = __p0; \
24839  int16x8_t __s1 = __p1; \
24840  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24841  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24842  int16x8_t __ret; \
24843  __ret = (int16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
24844  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24845  __ret; \
24846})
24847#endif
24848
24849#ifdef __LITTLE_ENDIAN__
24850#define vsri_n_u8(__p0, __p1, __p2) __extension__ ({ \
24851  uint8x8_t __s0 = __p0; \
24852  uint8x8_t __s1 = __p1; \
24853  uint8x8_t __ret; \
24854  __ret = (uint8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
24855  __ret; \
24856})
24857#else
24858#define vsri_n_u8(__p0, __p1, __p2) __extension__ ({ \
24859  uint8x8_t __s0 = __p0; \
24860  uint8x8_t __s1 = __p1; \
24861  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24862  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24863  uint8x8_t __ret; \
24864  __ret = (uint8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
24865  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24866  __ret; \
24867})
24868#endif
24869
24870#ifdef __LITTLE_ENDIAN__
24871#define vsri_n_u32(__p0, __p1, __p2) __extension__ ({ \
24872  uint32x2_t __s0 = __p0; \
24873  uint32x2_t __s1 = __p1; \
24874  uint32x2_t __ret; \
24875  __ret = (uint32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
24876  __ret; \
24877})
24878#else
24879#define vsri_n_u32(__p0, __p1, __p2) __extension__ ({ \
24880  uint32x2_t __s0 = __p0; \
24881  uint32x2_t __s1 = __p1; \
24882  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24883  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24884  uint32x2_t __ret; \
24885  __ret = (uint32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
24886  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24887  __ret; \
24888})
24889#endif
24890
24891#ifdef __LITTLE_ENDIAN__
24892#define vsri_n_u64(__p0, __p1, __p2) __extension__ ({ \
24893  uint64x1_t __s0 = __p0; \
24894  uint64x1_t __s1 = __p1; \
24895  uint64x1_t __ret; \
24896  __ret = (uint64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24897  __ret; \
24898})
24899#else
24900#define vsri_n_u64(__p0, __p1, __p2) __extension__ ({ \
24901  uint64x1_t __s0 = __p0; \
24902  uint64x1_t __s1 = __p1; \
24903  uint64x1_t __ret; \
24904  __ret = (uint64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24905  __ret; \
24906})
24907#endif
24908
24909#ifdef __LITTLE_ENDIAN__
24910#define vsri_n_u16(__p0, __p1, __p2) __extension__ ({ \
24911  uint16x4_t __s0 = __p0; \
24912  uint16x4_t __s1 = __p1; \
24913  uint16x4_t __ret; \
24914  __ret = (uint16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
24915  __ret; \
24916})
24917#else
24918#define vsri_n_u16(__p0, __p1, __p2) __extension__ ({ \
24919  uint16x4_t __s0 = __p0; \
24920  uint16x4_t __s1 = __p1; \
24921  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24922  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24923  uint16x4_t __ret; \
24924  __ret = (uint16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
24925  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24926  __ret; \
24927})
24928#endif
24929
24930#ifdef __LITTLE_ENDIAN__
24931#define vsri_n_s8(__p0, __p1, __p2) __extension__ ({ \
24932  int8x8_t __s0 = __p0; \
24933  int8x8_t __s1 = __p1; \
24934  int8x8_t __ret; \
24935  __ret = (int8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
24936  __ret; \
24937})
24938#else
24939#define vsri_n_s8(__p0, __p1, __p2) __extension__ ({ \
24940  int8x8_t __s0 = __p0; \
24941  int8x8_t __s1 = __p1; \
24942  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24943  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24944  int8x8_t __ret; \
24945  __ret = (int8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
24946  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24947  __ret; \
24948})
24949#endif
24950
24951#ifdef __LITTLE_ENDIAN__
24952#define vsri_n_s32(__p0, __p1, __p2) __extension__ ({ \
24953  int32x2_t __s0 = __p0; \
24954  int32x2_t __s1 = __p1; \
24955  int32x2_t __ret; \
24956  __ret = (int32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
24957  __ret; \
24958})
24959#else
24960#define vsri_n_s32(__p0, __p1, __p2) __extension__ ({ \
24961  int32x2_t __s0 = __p0; \
24962  int32x2_t __s1 = __p1; \
24963  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24964  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24965  int32x2_t __ret; \
24966  __ret = (int32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
24967  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24968  __ret; \
24969})
24970#endif
24971
24972#ifdef __LITTLE_ENDIAN__
24973#define vsri_n_s64(__p0, __p1, __p2) __extension__ ({ \
24974  int64x1_t __s0 = __p0; \
24975  int64x1_t __s1 = __p1; \
24976  int64x1_t __ret; \
24977  __ret = (int64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24978  __ret; \
24979})
24980#else
24981#define vsri_n_s64(__p0, __p1, __p2) __extension__ ({ \
24982  int64x1_t __s0 = __p0; \
24983  int64x1_t __s1 = __p1; \
24984  int64x1_t __ret; \
24985  __ret = (int64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24986  __ret; \
24987})
24988#endif
24989
24990#ifdef __LITTLE_ENDIAN__
24991#define vsri_n_s16(__p0, __p1, __p2) __extension__ ({ \
24992  int16x4_t __s0 = __p0; \
24993  int16x4_t __s1 = __p1; \
24994  int16x4_t __ret; \
24995  __ret = (int16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
24996  __ret; \
24997})
24998#else
24999#define vsri_n_s16(__p0, __p1, __p2) __extension__ ({ \
25000  int16x4_t __s0 = __p0; \
25001  int16x4_t __s1 = __p1; \
25002  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25003  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25004  int16x4_t __ret; \
25005  __ret = (int16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
25006  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25007  __ret; \
25008})
25009#endif
25010
25011#ifdef __LITTLE_ENDIAN__
25012#define vst1_p8(__p0, __p1) __extension__ ({ \
25013  poly8x8_t __s1 = __p1; \
25014  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 4); \
25015})
25016#else
25017#define vst1_p8(__p0, __p1) __extension__ ({ \
25018  poly8x8_t __s1 = __p1; \
25019  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25020  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 4); \
25021})
25022#endif
25023
25024#ifdef __LITTLE_ENDIAN__
25025#define vst1_p16(__p0, __p1) __extension__ ({ \
25026  poly16x4_t __s1 = __p1; \
25027  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 5); \
25028})
25029#else
25030#define vst1_p16(__p0, __p1) __extension__ ({ \
25031  poly16x4_t __s1 = __p1; \
25032  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25033  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 5); \
25034})
25035#endif
25036
25037#ifdef __LITTLE_ENDIAN__
25038#define vst1q_p8(__p0, __p1) __extension__ ({ \
25039  poly8x16_t __s1 = __p1; \
25040  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 36); \
25041})
25042#else
25043#define vst1q_p8(__p0, __p1) __extension__ ({ \
25044  poly8x16_t __s1 = __p1; \
25045  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25046  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 36); \
25047})
25048#endif
25049
25050#ifdef __LITTLE_ENDIAN__
25051#define vst1q_p16(__p0, __p1) __extension__ ({ \
25052  poly16x8_t __s1 = __p1; \
25053  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 37); \
25054})
25055#else
25056#define vst1q_p16(__p0, __p1) __extension__ ({ \
25057  poly16x8_t __s1 = __p1; \
25058  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25059  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 37); \
25060})
25061#endif
25062
25063#ifdef __LITTLE_ENDIAN__
25064#define vst1q_u8(__p0, __p1) __extension__ ({ \
25065  uint8x16_t __s1 = __p1; \
25066  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 48); \
25067})
25068#else
25069#define vst1q_u8(__p0, __p1) __extension__ ({ \
25070  uint8x16_t __s1 = __p1; \
25071  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25072  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 48); \
25073})
25074#endif
25075
25076#ifdef __LITTLE_ENDIAN__
25077#define vst1q_u32(__p0, __p1) __extension__ ({ \
25078  uint32x4_t __s1 = __p1; \
25079  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 50); \
25080})
25081#else
25082#define vst1q_u32(__p0, __p1) __extension__ ({ \
25083  uint32x4_t __s1 = __p1; \
25084  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25085  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 50); \
25086})
25087#endif
25088
25089#ifdef __LITTLE_ENDIAN__
25090#define vst1q_u64(__p0, __p1) __extension__ ({ \
25091  uint64x2_t __s1 = __p1; \
25092  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 51); \
25093})
25094#else
25095#define vst1q_u64(__p0, __p1) __extension__ ({ \
25096  uint64x2_t __s1 = __p1; \
25097  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25098  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 51); \
25099})
25100#endif
25101
25102#ifdef __LITTLE_ENDIAN__
25103#define vst1q_u16(__p0, __p1) __extension__ ({ \
25104  uint16x8_t __s1 = __p1; \
25105  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 49); \
25106})
25107#else
25108#define vst1q_u16(__p0, __p1) __extension__ ({ \
25109  uint16x8_t __s1 = __p1; \
25110  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25111  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 49); \
25112})
25113#endif
25114
25115#ifdef __LITTLE_ENDIAN__
25116#define vst1q_s8(__p0, __p1) __extension__ ({ \
25117  int8x16_t __s1 = __p1; \
25118  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 32); \
25119})
25120#else
25121#define vst1q_s8(__p0, __p1) __extension__ ({ \
25122  int8x16_t __s1 = __p1; \
25123  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25124  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 32); \
25125})
25126#endif
25127
25128#ifdef __LITTLE_ENDIAN__
25129#define vst1q_f32(__p0, __p1) __extension__ ({ \
25130  float32x4_t __s1 = __p1; \
25131  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 41); \
25132})
25133#else
25134#define vst1q_f32(__p0, __p1) __extension__ ({ \
25135  float32x4_t __s1 = __p1; \
25136  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25137  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 41); \
25138})
25139#endif
25140
25141#ifdef __LITTLE_ENDIAN__
25142#define vst1q_f16(__p0, __p1) __extension__ ({ \
25143  float16x8_t __s1 = __p1; \
25144  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 40); \
25145})
25146#else
25147#define vst1q_f16(__p0, __p1) __extension__ ({ \
25148  float16x8_t __s1 = __p1; \
25149  float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25150  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 40); \
25151})
25152#endif
25153
25154#ifdef __LITTLE_ENDIAN__
25155#define vst1q_s32(__p0, __p1) __extension__ ({ \
25156  int32x4_t __s1 = __p1; \
25157  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 34); \
25158})
25159#else
25160#define vst1q_s32(__p0, __p1) __extension__ ({ \
25161  int32x4_t __s1 = __p1; \
25162  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25163  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 34); \
25164})
25165#endif
25166
25167#ifdef __LITTLE_ENDIAN__
25168#define vst1q_s64(__p0, __p1) __extension__ ({ \
25169  int64x2_t __s1 = __p1; \
25170  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 35); \
25171})
25172#else
25173#define vst1q_s64(__p0, __p1) __extension__ ({ \
25174  int64x2_t __s1 = __p1; \
25175  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25176  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 35); \
25177})
25178#endif
25179
25180#ifdef __LITTLE_ENDIAN__
25181#define vst1q_s16(__p0, __p1) __extension__ ({ \
25182  int16x8_t __s1 = __p1; \
25183  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 33); \
25184})
25185#else
25186#define vst1q_s16(__p0, __p1) __extension__ ({ \
25187  int16x8_t __s1 = __p1; \
25188  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25189  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 33); \
25190})
25191#endif
25192
25193#ifdef __LITTLE_ENDIAN__
25194#define vst1_u8(__p0, __p1) __extension__ ({ \
25195  uint8x8_t __s1 = __p1; \
25196  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 16); \
25197})
25198#else
25199#define vst1_u8(__p0, __p1) __extension__ ({ \
25200  uint8x8_t __s1 = __p1; \
25201  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25202  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 16); \
25203})
25204#endif
25205
25206#ifdef __LITTLE_ENDIAN__
25207#define vst1_u32(__p0, __p1) __extension__ ({ \
25208  uint32x2_t __s1 = __p1; \
25209  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 18); \
25210})
25211#else
25212#define vst1_u32(__p0, __p1) __extension__ ({ \
25213  uint32x2_t __s1 = __p1; \
25214  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25215  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 18); \
25216})
25217#endif
25218
25219#ifdef __LITTLE_ENDIAN__
25220#define vst1_u64(__p0, __p1) __extension__ ({ \
25221  uint64x1_t __s1 = __p1; \
25222  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 19); \
25223})
25224#else
25225#define vst1_u64(__p0, __p1) __extension__ ({ \
25226  uint64x1_t __s1 = __p1; \
25227  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 19); \
25228})
25229#endif
25230
25231#ifdef __LITTLE_ENDIAN__
25232#define vst1_u16(__p0, __p1) __extension__ ({ \
25233  uint16x4_t __s1 = __p1; \
25234  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 17); \
25235})
25236#else
25237#define vst1_u16(__p0, __p1) __extension__ ({ \
25238  uint16x4_t __s1 = __p1; \
25239  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25240  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 17); \
25241})
25242#endif
25243
25244#ifdef __LITTLE_ENDIAN__
25245#define vst1_s8(__p0, __p1) __extension__ ({ \
25246  int8x8_t __s1 = __p1; \
25247  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 0); \
25248})
25249#else
25250#define vst1_s8(__p0, __p1) __extension__ ({ \
25251  int8x8_t __s1 = __p1; \
25252  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25253  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 0); \
25254})
25255#endif
25256
25257#ifdef __LITTLE_ENDIAN__
25258#define vst1_f32(__p0, __p1) __extension__ ({ \
25259  float32x2_t __s1 = __p1; \
25260  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 9); \
25261})
25262#else
25263#define vst1_f32(__p0, __p1) __extension__ ({ \
25264  float32x2_t __s1 = __p1; \
25265  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25266  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 9); \
25267})
25268#endif
25269
25270#ifdef __LITTLE_ENDIAN__
25271#define vst1_f16(__p0, __p1) __extension__ ({ \
25272  float16x4_t __s1 = __p1; \
25273  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 8); \
25274})
25275#else
25276#define vst1_f16(__p0, __p1) __extension__ ({ \
25277  float16x4_t __s1 = __p1; \
25278  float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25279  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 8); \
25280})
25281#endif
25282
25283#ifdef __LITTLE_ENDIAN__
25284#define vst1_s32(__p0, __p1) __extension__ ({ \
25285  int32x2_t __s1 = __p1; \
25286  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 2); \
25287})
25288#else
25289#define vst1_s32(__p0, __p1) __extension__ ({ \
25290  int32x2_t __s1 = __p1; \
25291  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25292  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 2); \
25293})
25294#endif
25295
25296#ifdef __LITTLE_ENDIAN__
25297#define vst1_s64(__p0, __p1) __extension__ ({ \
25298  int64x1_t __s1 = __p1; \
25299  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 3); \
25300})
25301#else
25302#define vst1_s64(__p0, __p1) __extension__ ({ \
25303  int64x1_t __s1 = __p1; \
25304  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 3); \
25305})
25306#endif
25307
25308#ifdef __LITTLE_ENDIAN__
25309#define vst1_s16(__p0, __p1) __extension__ ({ \
25310  int16x4_t __s1 = __p1; \
25311  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 1); \
25312})
25313#else
25314#define vst1_s16(__p0, __p1) __extension__ ({ \
25315  int16x4_t __s1 = __p1; \
25316  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25317  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 1); \
25318})
25319#endif
25320
25321#ifdef __LITTLE_ENDIAN__
25322#define vst1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25323  poly8x8_t __s1 = __p1; \
25324  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 4); \
25325})
25326#else
25327#define vst1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25328  poly8x8_t __s1 = __p1; \
25329  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25330  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 4); \
25331})
25332#endif
25333
25334#ifdef __LITTLE_ENDIAN__
25335#define vst1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25336  poly16x4_t __s1 = __p1; \
25337  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 5); \
25338})
25339#else
25340#define vst1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25341  poly16x4_t __s1 = __p1; \
25342  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25343  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 5); \
25344})
25345#endif
25346
25347#ifdef __LITTLE_ENDIAN__
25348#define vst1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25349  poly8x16_t __s1 = __p1; \
25350  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 36); \
25351})
25352#else
25353#define vst1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25354  poly8x16_t __s1 = __p1; \
25355  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25356  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 36); \
25357})
25358#endif
25359
25360#ifdef __LITTLE_ENDIAN__
25361#define vst1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25362  poly16x8_t __s1 = __p1; \
25363  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 37); \
25364})
25365#else
25366#define vst1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25367  poly16x8_t __s1 = __p1; \
25368  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25369  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 37); \
25370})
25371#endif
25372
25373#ifdef __LITTLE_ENDIAN__
25374#define vst1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25375  uint8x16_t __s1 = __p1; \
25376  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 48); \
25377})
25378#else
25379#define vst1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25380  uint8x16_t __s1 = __p1; \
25381  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25382  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 48); \
25383})
25384#endif
25385
25386#ifdef __LITTLE_ENDIAN__
25387#define vst1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25388  uint32x4_t __s1 = __p1; \
25389  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 50); \
25390})
25391#else
25392#define vst1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25393  uint32x4_t __s1 = __p1; \
25394  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25395  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 50); \
25396})
25397#endif
25398
25399#ifdef __LITTLE_ENDIAN__
25400#define vst1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25401  uint64x2_t __s1 = __p1; \
25402  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 51); \
25403})
25404#else
25405#define vst1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25406  uint64x2_t __s1 = __p1; \
25407  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25408  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 51); \
25409})
25410#endif
25411
25412#ifdef __LITTLE_ENDIAN__
25413#define vst1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25414  uint16x8_t __s1 = __p1; \
25415  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 49); \
25416})
25417#else
25418#define vst1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25419  uint16x8_t __s1 = __p1; \
25420  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25421  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 49); \
25422})
25423#endif
25424
25425#ifdef __LITTLE_ENDIAN__
25426#define vst1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25427  int8x16_t __s1 = __p1; \
25428  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 32); \
25429})
25430#else
25431#define vst1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25432  int8x16_t __s1 = __p1; \
25433  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25434  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 32); \
25435})
25436#endif
25437
25438#ifdef __LITTLE_ENDIAN__
25439#define vst1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25440  float32x4_t __s1 = __p1; \
25441  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 41); \
25442})
25443#else
25444#define vst1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25445  float32x4_t __s1 = __p1; \
25446  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25447  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 41); \
25448})
25449#endif
25450
25451#ifdef __LITTLE_ENDIAN__
25452#define vst1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25453  float16x8_t __s1 = __p1; \
25454  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 40); \
25455})
25456#else
25457#define vst1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25458  float16x8_t __s1 = __p1; \
25459  float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25460  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 40); \
25461})
25462#endif
25463
25464#ifdef __LITTLE_ENDIAN__
25465#define vst1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25466  int32x4_t __s1 = __p1; \
25467  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 34); \
25468})
25469#else
25470#define vst1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25471  int32x4_t __s1 = __p1; \
25472  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25473  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 34); \
25474})
25475#endif
25476
25477#ifdef __LITTLE_ENDIAN__
25478#define vst1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25479  int64x2_t __s1 = __p1; \
25480  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 35); \
25481})
25482#else
25483#define vst1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25484  int64x2_t __s1 = __p1; \
25485  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25486  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 35); \
25487})
25488#endif
25489
25490#ifdef __LITTLE_ENDIAN__
25491#define vst1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25492  int16x8_t __s1 = __p1; \
25493  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 33); \
25494})
25495#else
25496#define vst1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25497  int16x8_t __s1 = __p1; \
25498  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25499  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 33); \
25500})
25501#endif
25502
25503#ifdef __LITTLE_ENDIAN__
25504#define vst1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25505  uint8x8_t __s1 = __p1; \
25506  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 16); \
25507})
25508#else
25509#define vst1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25510  uint8x8_t __s1 = __p1; \
25511  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25512  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 16); \
25513})
25514#endif
25515
25516#ifdef __LITTLE_ENDIAN__
25517#define vst1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25518  uint32x2_t __s1 = __p1; \
25519  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 18); \
25520})
25521#else
25522#define vst1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25523  uint32x2_t __s1 = __p1; \
25524  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25525  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 18); \
25526})
25527#endif
25528
25529#ifdef __LITTLE_ENDIAN__
25530#define vst1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25531  uint64x1_t __s1 = __p1; \
25532  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
25533})
25534#else
25535#define vst1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25536  uint64x1_t __s1 = __p1; \
25537  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
25538})
25539#endif
25540
25541#ifdef __LITTLE_ENDIAN__
25542#define vst1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25543  uint16x4_t __s1 = __p1; \
25544  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 17); \
25545})
25546#else
25547#define vst1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25548  uint16x4_t __s1 = __p1; \
25549  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25550  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 17); \
25551})
25552#endif
25553
25554#ifdef __LITTLE_ENDIAN__
25555#define vst1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25556  int8x8_t __s1 = __p1; \
25557  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 0); \
25558})
25559#else
25560#define vst1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25561  int8x8_t __s1 = __p1; \
25562  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25563  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 0); \
25564})
25565#endif
25566
25567#ifdef __LITTLE_ENDIAN__
25568#define vst1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25569  float32x2_t __s1 = __p1; \
25570  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 9); \
25571})
25572#else
25573#define vst1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25574  float32x2_t __s1 = __p1; \
25575  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25576  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 9); \
25577})
25578#endif
25579
25580#ifdef __LITTLE_ENDIAN__
25581#define vst1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25582  float16x4_t __s1 = __p1; \
25583  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 8); \
25584})
25585#else
25586#define vst1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25587  float16x4_t __s1 = __p1; \
25588  float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25589  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 8); \
25590})
25591#endif
25592
25593#ifdef __LITTLE_ENDIAN__
25594#define vst1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25595  int32x2_t __s1 = __p1; \
25596  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 2); \
25597})
25598#else
25599#define vst1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25600  int32x2_t __s1 = __p1; \
25601  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25602  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 2); \
25603})
25604#endif
25605
25606#ifdef __LITTLE_ENDIAN__
25607#define vst1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25608  int64x1_t __s1 = __p1; \
25609  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
25610})
25611#else
25612#define vst1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25613  int64x1_t __s1 = __p1; \
25614  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
25615})
25616#endif
25617
25618#ifdef __LITTLE_ENDIAN__
25619#define vst1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25620  int16x4_t __s1 = __p1; \
25621  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 1); \
25622})
25623#else
25624#define vst1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25625  int16x4_t __s1 = __p1; \
25626  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25627  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 1); \
25628})
25629#endif
25630
25631#ifdef __LITTLE_ENDIAN__
25632#define vst2_p8(__p0, __p1) __extension__ ({ \
25633  poly8x8x2_t __s1 = __p1; \
25634  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 4); \
25635})
25636#else
25637#define vst2_p8(__p0, __p1) __extension__ ({ \
25638  poly8x8x2_t __s1 = __p1; \
25639  poly8x8x2_t __rev1; \
25640  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25641  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25642  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 4); \
25643})
25644#endif
25645
25646#ifdef __LITTLE_ENDIAN__
25647#define vst2_p16(__p0, __p1) __extension__ ({ \
25648  poly16x4x2_t __s1 = __p1; \
25649  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 5); \
25650})
25651#else
25652#define vst2_p16(__p0, __p1) __extension__ ({ \
25653  poly16x4x2_t __s1 = __p1; \
25654  poly16x4x2_t __rev1; \
25655  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25656  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25657  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 5); \
25658})
25659#endif
25660
25661#ifdef __LITTLE_ENDIAN__
25662#define vst2q_p8(__p0, __p1) __extension__ ({ \
25663  poly8x16x2_t __s1 = __p1; \
25664  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 36); \
25665})
25666#else
25667#define vst2q_p8(__p0, __p1) __extension__ ({ \
25668  poly8x16x2_t __s1 = __p1; \
25669  poly8x16x2_t __rev1; \
25670  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25671  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25672  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 36); \
25673})
25674#endif
25675
25676#ifdef __LITTLE_ENDIAN__
25677#define vst2q_p16(__p0, __p1) __extension__ ({ \
25678  poly16x8x2_t __s1 = __p1; \
25679  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 37); \
25680})
25681#else
25682#define vst2q_p16(__p0, __p1) __extension__ ({ \
25683  poly16x8x2_t __s1 = __p1; \
25684  poly16x8x2_t __rev1; \
25685  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25686  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25687  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 37); \
25688})
25689#endif
25690
25691#ifdef __LITTLE_ENDIAN__
25692#define vst2q_u8(__p0, __p1) __extension__ ({ \
25693  uint8x16x2_t __s1 = __p1; \
25694  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 48); \
25695})
25696#else
25697#define vst2q_u8(__p0, __p1) __extension__ ({ \
25698  uint8x16x2_t __s1 = __p1; \
25699  uint8x16x2_t __rev1; \
25700  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25701  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25702  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 48); \
25703})
25704#endif
25705
25706#ifdef __LITTLE_ENDIAN__
25707#define vst2q_u32(__p0, __p1) __extension__ ({ \
25708  uint32x4x2_t __s1 = __p1; \
25709  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 50); \
25710})
25711#else
25712#define vst2q_u32(__p0, __p1) __extension__ ({ \
25713  uint32x4x2_t __s1 = __p1; \
25714  uint32x4x2_t __rev1; \
25715  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25716  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25717  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 50); \
25718})
25719#endif
25720
25721#ifdef __LITTLE_ENDIAN__
25722#define vst2q_u16(__p0, __p1) __extension__ ({ \
25723  uint16x8x2_t __s1 = __p1; \
25724  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 49); \
25725})
25726#else
25727#define vst2q_u16(__p0, __p1) __extension__ ({ \
25728  uint16x8x2_t __s1 = __p1; \
25729  uint16x8x2_t __rev1; \
25730  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25731  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25732  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 49); \
25733})
25734#endif
25735
25736#ifdef __LITTLE_ENDIAN__
25737#define vst2q_s8(__p0, __p1) __extension__ ({ \
25738  int8x16x2_t __s1 = __p1; \
25739  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 32); \
25740})
25741#else
25742#define vst2q_s8(__p0, __p1) __extension__ ({ \
25743  int8x16x2_t __s1 = __p1; \
25744  int8x16x2_t __rev1; \
25745  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25746  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25747  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 32); \
25748})
25749#endif
25750
25751#ifdef __LITTLE_ENDIAN__
25752#define vst2q_f32(__p0, __p1) __extension__ ({ \
25753  float32x4x2_t __s1 = __p1; \
25754  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 41); \
25755})
25756#else
25757#define vst2q_f32(__p0, __p1) __extension__ ({ \
25758  float32x4x2_t __s1 = __p1; \
25759  float32x4x2_t __rev1; \
25760  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25761  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25762  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 41); \
25763})
25764#endif
25765
25766#ifdef __LITTLE_ENDIAN__
25767#define vst2q_f16(__p0, __p1) __extension__ ({ \
25768  float16x8x2_t __s1 = __p1; \
25769  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 40); \
25770})
25771#else
25772#define vst2q_f16(__p0, __p1) __extension__ ({ \
25773  float16x8x2_t __s1 = __p1; \
25774  float16x8x2_t __rev1; \
25775  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25776  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25777  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 40); \
25778})
25779#endif
25780
25781#ifdef __LITTLE_ENDIAN__
25782#define vst2q_s32(__p0, __p1) __extension__ ({ \
25783  int32x4x2_t __s1 = __p1; \
25784  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 34); \
25785})
25786#else
25787#define vst2q_s32(__p0, __p1) __extension__ ({ \
25788  int32x4x2_t __s1 = __p1; \
25789  int32x4x2_t __rev1; \
25790  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25791  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25792  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 34); \
25793})
25794#endif
25795
25796#ifdef __LITTLE_ENDIAN__
25797#define vst2q_s16(__p0, __p1) __extension__ ({ \
25798  int16x8x2_t __s1 = __p1; \
25799  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 33); \
25800})
25801#else
25802#define vst2q_s16(__p0, __p1) __extension__ ({ \
25803  int16x8x2_t __s1 = __p1; \
25804  int16x8x2_t __rev1; \
25805  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25806  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25807  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 33); \
25808})
25809#endif
25810
25811#ifdef __LITTLE_ENDIAN__
25812#define vst2_u8(__p0, __p1) __extension__ ({ \
25813  uint8x8x2_t __s1 = __p1; \
25814  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 16); \
25815})
25816#else
25817#define vst2_u8(__p0, __p1) __extension__ ({ \
25818  uint8x8x2_t __s1 = __p1; \
25819  uint8x8x2_t __rev1; \
25820  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25821  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25822  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 16); \
25823})
25824#endif
25825
25826#ifdef __LITTLE_ENDIAN__
25827#define vst2_u32(__p0, __p1) __extension__ ({ \
25828  uint32x2x2_t __s1 = __p1; \
25829  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 18); \
25830})
25831#else
25832#define vst2_u32(__p0, __p1) __extension__ ({ \
25833  uint32x2x2_t __s1 = __p1; \
25834  uint32x2x2_t __rev1; \
25835  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
25836  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
25837  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 18); \
25838})
25839#endif
25840
25841#ifdef __LITTLE_ENDIAN__
25842#define vst2_u64(__p0, __p1) __extension__ ({ \
25843  uint64x1x2_t __s1 = __p1; \
25844  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
25845})
25846#else
25847#define vst2_u64(__p0, __p1) __extension__ ({ \
25848  uint64x1x2_t __s1 = __p1; \
25849  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
25850})
25851#endif
25852
25853#ifdef __LITTLE_ENDIAN__
25854#define vst2_u16(__p0, __p1) __extension__ ({ \
25855  uint16x4x2_t __s1 = __p1; \
25856  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 17); \
25857})
25858#else
25859#define vst2_u16(__p0, __p1) __extension__ ({ \
25860  uint16x4x2_t __s1 = __p1; \
25861  uint16x4x2_t __rev1; \
25862  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25863  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25864  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 17); \
25865})
25866#endif
25867
25868#ifdef __LITTLE_ENDIAN__
25869#define vst2_s8(__p0, __p1) __extension__ ({ \
25870  int8x8x2_t __s1 = __p1; \
25871  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 0); \
25872})
25873#else
25874#define vst2_s8(__p0, __p1) __extension__ ({ \
25875  int8x8x2_t __s1 = __p1; \
25876  int8x8x2_t __rev1; \
25877  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25878  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25879  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 0); \
25880})
25881#endif
25882
25883#ifdef __LITTLE_ENDIAN__
25884#define vst2_f32(__p0, __p1) __extension__ ({ \
25885  float32x2x2_t __s1 = __p1; \
25886  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 9); \
25887})
25888#else
25889#define vst2_f32(__p0, __p1) __extension__ ({ \
25890  float32x2x2_t __s1 = __p1; \
25891  float32x2x2_t __rev1; \
25892  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
25893  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
25894  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 9); \
25895})
25896#endif
25897
25898#ifdef __LITTLE_ENDIAN__
25899#define vst2_f16(__p0, __p1) __extension__ ({ \
25900  float16x4x2_t __s1 = __p1; \
25901  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 8); \
25902})
25903#else
25904#define vst2_f16(__p0, __p1) __extension__ ({ \
25905  float16x4x2_t __s1 = __p1; \
25906  float16x4x2_t __rev1; \
25907  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25908  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25909  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 8); \
25910})
25911#endif
25912
25913#ifdef __LITTLE_ENDIAN__
25914#define vst2_s32(__p0, __p1) __extension__ ({ \
25915  int32x2x2_t __s1 = __p1; \
25916  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 2); \
25917})
25918#else
25919#define vst2_s32(__p0, __p1) __extension__ ({ \
25920  int32x2x2_t __s1 = __p1; \
25921  int32x2x2_t __rev1; \
25922  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
25923  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
25924  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 2); \
25925})
25926#endif
25927
25928#ifdef __LITTLE_ENDIAN__
25929#define vst2_s64(__p0, __p1) __extension__ ({ \
25930  int64x1x2_t __s1 = __p1; \
25931  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 3); \
25932})
25933#else
25934#define vst2_s64(__p0, __p1) __extension__ ({ \
25935  int64x1x2_t __s1 = __p1; \
25936  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 3); \
25937})
25938#endif
25939
25940#ifdef __LITTLE_ENDIAN__
25941#define vst2_s16(__p0, __p1) __extension__ ({ \
25942  int16x4x2_t __s1 = __p1; \
25943  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 1); \
25944})
25945#else
25946#define vst2_s16(__p0, __p1) __extension__ ({ \
25947  int16x4x2_t __s1 = __p1; \
25948  int16x4x2_t __rev1; \
25949  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25950  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25951  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 1); \
25952})
25953#endif
25954
25955#ifdef __LITTLE_ENDIAN__
25956#define vst2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25957  poly8x8x2_t __s1 = __p1; \
25958  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 4); \
25959})
25960#else
25961#define vst2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25962  poly8x8x2_t __s1 = __p1; \
25963  poly8x8x2_t __rev1; \
25964  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25965  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25966  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 4); \
25967})
25968#endif
25969
25970#ifdef __LITTLE_ENDIAN__
25971#define vst2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25972  poly16x4x2_t __s1 = __p1; \
25973  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 5); \
25974})
25975#else
25976#define vst2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25977  poly16x4x2_t __s1 = __p1; \
25978  poly16x4x2_t __rev1; \
25979  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25980  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25981  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 5); \
25982})
25983#endif
25984
25985#ifdef __LITTLE_ENDIAN__
25986#define vst2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25987  poly16x8x2_t __s1 = __p1; \
25988  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 37); \
25989})
25990#else
25991#define vst2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25992  poly16x8x2_t __s1 = __p1; \
25993  poly16x8x2_t __rev1; \
25994  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25995  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25996  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 37); \
25997})
25998#endif
25999
26000#ifdef __LITTLE_ENDIAN__
26001#define vst2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26002  uint32x4x2_t __s1 = __p1; \
26003  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 50); \
26004})
26005#else
26006#define vst2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26007  uint32x4x2_t __s1 = __p1; \
26008  uint32x4x2_t __rev1; \
26009  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26010  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26011  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 50); \
26012})
26013#endif
26014
26015#ifdef __LITTLE_ENDIAN__
26016#define vst2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26017  uint16x8x2_t __s1 = __p1; \
26018  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 49); \
26019})
26020#else
26021#define vst2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26022  uint16x8x2_t __s1 = __p1; \
26023  uint16x8x2_t __rev1; \
26024  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26025  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26026  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 49); \
26027})
26028#endif
26029
26030#ifdef __LITTLE_ENDIAN__
26031#define vst2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26032  float32x4x2_t __s1 = __p1; \
26033  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 41); \
26034})
26035#else
26036#define vst2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26037  float32x4x2_t __s1 = __p1; \
26038  float32x4x2_t __rev1; \
26039  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26040  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26041  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 41); \
26042})
26043#endif
26044
26045#ifdef __LITTLE_ENDIAN__
26046#define vst2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26047  float16x8x2_t __s1 = __p1; \
26048  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 40); \
26049})
26050#else
26051#define vst2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26052  float16x8x2_t __s1 = __p1; \
26053  float16x8x2_t __rev1; \
26054  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26055  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26056  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 40); \
26057})
26058#endif
26059
26060#ifdef __LITTLE_ENDIAN__
26061#define vst2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26062  int32x4x2_t __s1 = __p1; \
26063  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 34); \
26064})
26065#else
26066#define vst2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26067  int32x4x2_t __s1 = __p1; \
26068  int32x4x2_t __rev1; \
26069  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26070  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26071  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 34); \
26072})
26073#endif
26074
26075#ifdef __LITTLE_ENDIAN__
26076#define vst2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26077  int16x8x2_t __s1 = __p1; \
26078  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 33); \
26079})
26080#else
26081#define vst2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26082  int16x8x2_t __s1 = __p1; \
26083  int16x8x2_t __rev1; \
26084  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26085  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26086  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 33); \
26087})
26088#endif
26089
26090#ifdef __LITTLE_ENDIAN__
26091#define vst2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26092  uint8x8x2_t __s1 = __p1; \
26093  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 16); \
26094})
26095#else
26096#define vst2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26097  uint8x8x2_t __s1 = __p1; \
26098  uint8x8x2_t __rev1; \
26099  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26100  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26101  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 16); \
26102})
26103#endif
26104
26105#ifdef __LITTLE_ENDIAN__
26106#define vst2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26107  uint32x2x2_t __s1 = __p1; \
26108  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 18); \
26109})
26110#else
26111#define vst2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26112  uint32x2x2_t __s1 = __p1; \
26113  uint32x2x2_t __rev1; \
26114  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26115  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26116  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 18); \
26117})
26118#endif
26119
26120#ifdef __LITTLE_ENDIAN__
26121#define vst2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26122  uint16x4x2_t __s1 = __p1; \
26123  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 17); \
26124})
26125#else
26126#define vst2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26127  uint16x4x2_t __s1 = __p1; \
26128  uint16x4x2_t __rev1; \
26129  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26130  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26131  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 17); \
26132})
26133#endif
26134
26135#ifdef __LITTLE_ENDIAN__
26136#define vst2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26137  int8x8x2_t __s1 = __p1; \
26138  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 0); \
26139})
26140#else
26141#define vst2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26142  int8x8x2_t __s1 = __p1; \
26143  int8x8x2_t __rev1; \
26144  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26145  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26146  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 0); \
26147})
26148#endif
26149
26150#ifdef __LITTLE_ENDIAN__
26151#define vst2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26152  float32x2x2_t __s1 = __p1; \
26153  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 9); \
26154})
26155#else
26156#define vst2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26157  float32x2x2_t __s1 = __p1; \
26158  float32x2x2_t __rev1; \
26159  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26160  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26161  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 9); \
26162})
26163#endif
26164
26165#ifdef __LITTLE_ENDIAN__
26166#define vst2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26167  float16x4x2_t __s1 = __p1; \
26168  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 8); \
26169})
26170#else
26171#define vst2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26172  float16x4x2_t __s1 = __p1; \
26173  float16x4x2_t __rev1; \
26174  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26175  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26176  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 8); \
26177})
26178#endif
26179
26180#ifdef __LITTLE_ENDIAN__
26181#define vst2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26182  int32x2x2_t __s1 = __p1; \
26183  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 2); \
26184})
26185#else
26186#define vst2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26187  int32x2x2_t __s1 = __p1; \
26188  int32x2x2_t __rev1; \
26189  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26190  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26191  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 2); \
26192})
26193#endif
26194
26195#ifdef __LITTLE_ENDIAN__
26196#define vst2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26197  int16x4x2_t __s1 = __p1; \
26198  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 1); \
26199})
26200#else
26201#define vst2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26202  int16x4x2_t __s1 = __p1; \
26203  int16x4x2_t __rev1; \
26204  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26205  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26206  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 1); \
26207})
26208#endif
26209
26210#ifdef __LITTLE_ENDIAN__
26211#define vst3_p8(__p0, __p1) __extension__ ({ \
26212  poly8x8x3_t __s1 = __p1; \
26213  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 4); \
26214})
26215#else
26216#define vst3_p8(__p0, __p1) __extension__ ({ \
26217  poly8x8x3_t __s1 = __p1; \
26218  poly8x8x3_t __rev1; \
26219  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26220  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26221  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26222  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 4); \
26223})
26224#endif
26225
26226#ifdef __LITTLE_ENDIAN__
26227#define vst3_p16(__p0, __p1) __extension__ ({ \
26228  poly16x4x3_t __s1 = __p1; \
26229  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 5); \
26230})
26231#else
26232#define vst3_p16(__p0, __p1) __extension__ ({ \
26233  poly16x4x3_t __s1 = __p1; \
26234  poly16x4x3_t __rev1; \
26235  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26236  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26237  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26238  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 5); \
26239})
26240#endif
26241
26242#ifdef __LITTLE_ENDIAN__
26243#define vst3q_p8(__p0, __p1) __extension__ ({ \
26244  poly8x16x3_t __s1 = __p1; \
26245  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 36); \
26246})
26247#else
26248#define vst3q_p8(__p0, __p1) __extension__ ({ \
26249  poly8x16x3_t __s1 = __p1; \
26250  poly8x16x3_t __rev1; \
26251  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26252  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26253  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26254  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 36); \
26255})
26256#endif
26257
26258#ifdef __LITTLE_ENDIAN__
26259#define vst3q_p16(__p0, __p1) __extension__ ({ \
26260  poly16x8x3_t __s1 = __p1; \
26261  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 37); \
26262})
26263#else
26264#define vst3q_p16(__p0, __p1) __extension__ ({ \
26265  poly16x8x3_t __s1 = __p1; \
26266  poly16x8x3_t __rev1; \
26267  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26268  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26269  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26270  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 37); \
26271})
26272#endif
26273
26274#ifdef __LITTLE_ENDIAN__
26275#define vst3q_u8(__p0, __p1) __extension__ ({ \
26276  uint8x16x3_t __s1 = __p1; \
26277  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 48); \
26278})
26279#else
26280#define vst3q_u8(__p0, __p1) __extension__ ({ \
26281  uint8x16x3_t __s1 = __p1; \
26282  uint8x16x3_t __rev1; \
26283  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26284  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26285  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26286  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 48); \
26287})
26288#endif
26289
26290#ifdef __LITTLE_ENDIAN__
26291#define vst3q_u32(__p0, __p1) __extension__ ({ \
26292  uint32x4x3_t __s1 = __p1; \
26293  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 50); \
26294})
26295#else
26296#define vst3q_u32(__p0, __p1) __extension__ ({ \
26297  uint32x4x3_t __s1 = __p1; \
26298  uint32x4x3_t __rev1; \
26299  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26300  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26301  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26302  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 50); \
26303})
26304#endif
26305
26306#ifdef __LITTLE_ENDIAN__
26307#define vst3q_u16(__p0, __p1) __extension__ ({ \
26308  uint16x8x3_t __s1 = __p1; \
26309  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 49); \
26310})
26311#else
26312#define vst3q_u16(__p0, __p1) __extension__ ({ \
26313  uint16x8x3_t __s1 = __p1; \
26314  uint16x8x3_t __rev1; \
26315  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26316  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26317  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26318  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 49); \
26319})
26320#endif
26321
26322#ifdef __LITTLE_ENDIAN__
26323#define vst3q_s8(__p0, __p1) __extension__ ({ \
26324  int8x16x3_t __s1 = __p1; \
26325  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 32); \
26326})
26327#else
26328#define vst3q_s8(__p0, __p1) __extension__ ({ \
26329  int8x16x3_t __s1 = __p1; \
26330  int8x16x3_t __rev1; \
26331  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26332  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26333  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26334  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 32); \
26335})
26336#endif
26337
26338#ifdef __LITTLE_ENDIAN__
26339#define vst3q_f32(__p0, __p1) __extension__ ({ \
26340  float32x4x3_t __s1 = __p1; \
26341  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 41); \
26342})
26343#else
26344#define vst3q_f32(__p0, __p1) __extension__ ({ \
26345  float32x4x3_t __s1 = __p1; \
26346  float32x4x3_t __rev1; \
26347  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26348  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26349  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26350  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 41); \
26351})
26352#endif
26353
26354#ifdef __LITTLE_ENDIAN__
26355#define vst3q_f16(__p0, __p1) __extension__ ({ \
26356  float16x8x3_t __s1 = __p1; \
26357  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 40); \
26358})
26359#else
26360#define vst3q_f16(__p0, __p1) __extension__ ({ \
26361  float16x8x3_t __s1 = __p1; \
26362  float16x8x3_t __rev1; \
26363  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26364  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26365  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26366  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 40); \
26367})
26368#endif
26369
26370#ifdef __LITTLE_ENDIAN__
26371#define vst3q_s32(__p0, __p1) __extension__ ({ \
26372  int32x4x3_t __s1 = __p1; \
26373  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 34); \
26374})
26375#else
26376#define vst3q_s32(__p0, __p1) __extension__ ({ \
26377  int32x4x3_t __s1 = __p1; \
26378  int32x4x3_t __rev1; \
26379  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26380  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26381  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26382  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 34); \
26383})
26384#endif
26385
26386#ifdef __LITTLE_ENDIAN__
26387#define vst3q_s16(__p0, __p1) __extension__ ({ \
26388  int16x8x3_t __s1 = __p1; \
26389  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 33); \
26390})
26391#else
26392#define vst3q_s16(__p0, __p1) __extension__ ({ \
26393  int16x8x3_t __s1 = __p1; \
26394  int16x8x3_t __rev1; \
26395  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26396  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26397  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26398  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 33); \
26399})
26400#endif
26401
26402#ifdef __LITTLE_ENDIAN__
26403#define vst3_u8(__p0, __p1) __extension__ ({ \
26404  uint8x8x3_t __s1 = __p1; \
26405  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 16); \
26406})
26407#else
26408#define vst3_u8(__p0, __p1) __extension__ ({ \
26409  uint8x8x3_t __s1 = __p1; \
26410  uint8x8x3_t __rev1; \
26411  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26412  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26413  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26414  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 16); \
26415})
26416#endif
26417
26418#ifdef __LITTLE_ENDIAN__
26419#define vst3_u32(__p0, __p1) __extension__ ({ \
26420  uint32x2x3_t __s1 = __p1; \
26421  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 18); \
26422})
26423#else
26424#define vst3_u32(__p0, __p1) __extension__ ({ \
26425  uint32x2x3_t __s1 = __p1; \
26426  uint32x2x3_t __rev1; \
26427  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26428  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26429  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26430  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 18); \
26431})
26432#endif
26433
26434#ifdef __LITTLE_ENDIAN__
26435#define vst3_u64(__p0, __p1) __extension__ ({ \
26436  uint64x1x3_t __s1 = __p1; \
26437  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
26438})
26439#else
26440#define vst3_u64(__p0, __p1) __extension__ ({ \
26441  uint64x1x3_t __s1 = __p1; \
26442  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
26443})
26444#endif
26445
26446#ifdef __LITTLE_ENDIAN__
26447#define vst3_u16(__p0, __p1) __extension__ ({ \
26448  uint16x4x3_t __s1 = __p1; \
26449  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 17); \
26450})
26451#else
26452#define vst3_u16(__p0, __p1) __extension__ ({ \
26453  uint16x4x3_t __s1 = __p1; \
26454  uint16x4x3_t __rev1; \
26455  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26456  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26457  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26458  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 17); \
26459})
26460#endif
26461
26462#ifdef __LITTLE_ENDIAN__
26463#define vst3_s8(__p0, __p1) __extension__ ({ \
26464  int8x8x3_t __s1 = __p1; \
26465  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 0); \
26466})
26467#else
26468#define vst3_s8(__p0, __p1) __extension__ ({ \
26469  int8x8x3_t __s1 = __p1; \
26470  int8x8x3_t __rev1; \
26471  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26472  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26473  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26474  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 0); \
26475})
26476#endif
26477
26478#ifdef __LITTLE_ENDIAN__
26479#define vst3_f32(__p0, __p1) __extension__ ({ \
26480  float32x2x3_t __s1 = __p1; \
26481  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 9); \
26482})
26483#else
26484#define vst3_f32(__p0, __p1) __extension__ ({ \
26485  float32x2x3_t __s1 = __p1; \
26486  float32x2x3_t __rev1; \
26487  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26488  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26489  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26490  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 9); \
26491})
26492#endif
26493
26494#ifdef __LITTLE_ENDIAN__
26495#define vst3_f16(__p0, __p1) __extension__ ({ \
26496  float16x4x3_t __s1 = __p1; \
26497  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 8); \
26498})
26499#else
26500#define vst3_f16(__p0, __p1) __extension__ ({ \
26501  float16x4x3_t __s1 = __p1; \
26502  float16x4x3_t __rev1; \
26503  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26504  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26505  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26506  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 8); \
26507})
26508#endif
26509
26510#ifdef __LITTLE_ENDIAN__
26511#define vst3_s32(__p0, __p1) __extension__ ({ \
26512  int32x2x3_t __s1 = __p1; \
26513  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 2); \
26514})
26515#else
26516#define vst3_s32(__p0, __p1) __extension__ ({ \
26517  int32x2x3_t __s1 = __p1; \
26518  int32x2x3_t __rev1; \
26519  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26520  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26521  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26522  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 2); \
26523})
26524#endif
26525
26526#ifdef __LITTLE_ENDIAN__
26527#define vst3_s64(__p0, __p1) __extension__ ({ \
26528  int64x1x3_t __s1 = __p1; \
26529  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
26530})
26531#else
26532#define vst3_s64(__p0, __p1) __extension__ ({ \
26533  int64x1x3_t __s1 = __p1; \
26534  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
26535})
26536#endif
26537
26538#ifdef __LITTLE_ENDIAN__
26539#define vst3_s16(__p0, __p1) __extension__ ({ \
26540  int16x4x3_t __s1 = __p1; \
26541  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 1); \
26542})
26543#else
26544#define vst3_s16(__p0, __p1) __extension__ ({ \
26545  int16x4x3_t __s1 = __p1; \
26546  int16x4x3_t __rev1; \
26547  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26548  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26549  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26550  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 1); \
26551})
26552#endif
26553
26554#ifdef __LITTLE_ENDIAN__
26555#define vst3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26556  poly8x8x3_t __s1 = __p1; \
26557  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 4); \
26558})
26559#else
26560#define vst3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26561  poly8x8x3_t __s1 = __p1; \
26562  poly8x8x3_t __rev1; \
26563  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26564  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26565  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26566  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 4); \
26567})
26568#endif
26569
26570#ifdef __LITTLE_ENDIAN__
26571#define vst3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26572  poly16x4x3_t __s1 = __p1; \
26573  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 5); \
26574})
26575#else
26576#define vst3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26577  poly16x4x3_t __s1 = __p1; \
26578  poly16x4x3_t __rev1; \
26579  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26580  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26581  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26582  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 5); \
26583})
26584#endif
26585
26586#ifdef __LITTLE_ENDIAN__
26587#define vst3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26588  poly16x8x3_t __s1 = __p1; \
26589  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 37); \
26590})
26591#else
26592#define vst3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26593  poly16x8x3_t __s1 = __p1; \
26594  poly16x8x3_t __rev1; \
26595  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26596  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26597  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26598  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 37); \
26599})
26600#endif
26601
26602#ifdef __LITTLE_ENDIAN__
26603#define vst3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26604  uint32x4x3_t __s1 = __p1; \
26605  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 50); \
26606})
26607#else
26608#define vst3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26609  uint32x4x3_t __s1 = __p1; \
26610  uint32x4x3_t __rev1; \
26611  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26612  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26613  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26614  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 50); \
26615})
26616#endif
26617
26618#ifdef __LITTLE_ENDIAN__
26619#define vst3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26620  uint16x8x3_t __s1 = __p1; \
26621  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 49); \
26622})
26623#else
26624#define vst3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26625  uint16x8x3_t __s1 = __p1; \
26626  uint16x8x3_t __rev1; \
26627  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26628  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26629  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26630  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 49); \
26631})
26632#endif
26633
26634#ifdef __LITTLE_ENDIAN__
26635#define vst3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26636  float32x4x3_t __s1 = __p1; \
26637  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 41); \
26638})
26639#else
26640#define vst3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26641  float32x4x3_t __s1 = __p1; \
26642  float32x4x3_t __rev1; \
26643  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26644  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26645  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26646  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 41); \
26647})
26648#endif
26649
26650#ifdef __LITTLE_ENDIAN__
26651#define vst3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26652  float16x8x3_t __s1 = __p1; \
26653  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 40); \
26654})
26655#else
26656#define vst3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26657  float16x8x3_t __s1 = __p1; \
26658  float16x8x3_t __rev1; \
26659  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26660  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26661  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26662  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 40); \
26663})
26664#endif
26665
26666#ifdef __LITTLE_ENDIAN__
26667#define vst3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26668  int32x4x3_t __s1 = __p1; \
26669  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 34); \
26670})
26671#else
26672#define vst3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26673  int32x4x3_t __s1 = __p1; \
26674  int32x4x3_t __rev1; \
26675  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26676  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26677  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26678  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 34); \
26679})
26680#endif
26681
26682#ifdef __LITTLE_ENDIAN__
26683#define vst3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26684  int16x8x3_t __s1 = __p1; \
26685  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 33); \
26686})
26687#else
26688#define vst3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26689  int16x8x3_t __s1 = __p1; \
26690  int16x8x3_t __rev1; \
26691  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26692  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26693  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26694  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 33); \
26695})
26696#endif
26697
26698#ifdef __LITTLE_ENDIAN__
26699#define vst3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26700  uint8x8x3_t __s1 = __p1; \
26701  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 16); \
26702})
26703#else
26704#define vst3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26705  uint8x8x3_t __s1 = __p1; \
26706  uint8x8x3_t __rev1; \
26707  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26708  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26709  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26710  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 16); \
26711})
26712#endif
26713
26714#ifdef __LITTLE_ENDIAN__
26715#define vst3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26716  uint32x2x3_t __s1 = __p1; \
26717  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 18); \
26718})
26719#else
26720#define vst3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26721  uint32x2x3_t __s1 = __p1; \
26722  uint32x2x3_t __rev1; \
26723  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26724  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26725  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26726  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 18); \
26727})
26728#endif
26729
26730#ifdef __LITTLE_ENDIAN__
26731#define vst3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26732  uint16x4x3_t __s1 = __p1; \
26733  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 17); \
26734})
26735#else
26736#define vst3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26737  uint16x4x3_t __s1 = __p1; \
26738  uint16x4x3_t __rev1; \
26739  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26740  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26741  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26742  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 17); \
26743})
26744#endif
26745
26746#ifdef __LITTLE_ENDIAN__
26747#define vst3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26748  int8x8x3_t __s1 = __p1; \
26749  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 0); \
26750})
26751#else
26752#define vst3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26753  int8x8x3_t __s1 = __p1; \
26754  int8x8x3_t __rev1; \
26755  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26756  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26757  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26758  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 0); \
26759})
26760#endif
26761
26762#ifdef __LITTLE_ENDIAN__
26763#define vst3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26764  float32x2x3_t __s1 = __p1; \
26765  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 9); \
26766})
26767#else
26768#define vst3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26769  float32x2x3_t __s1 = __p1; \
26770  float32x2x3_t __rev1; \
26771  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26772  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26773  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26774  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 9); \
26775})
26776#endif
26777
26778#ifdef __LITTLE_ENDIAN__
26779#define vst3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26780  float16x4x3_t __s1 = __p1; \
26781  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 8); \
26782})
26783#else
26784#define vst3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26785  float16x4x3_t __s1 = __p1; \
26786  float16x4x3_t __rev1; \
26787  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26788  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26789  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26790  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 8); \
26791})
26792#endif
26793
26794#ifdef __LITTLE_ENDIAN__
26795#define vst3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26796  int32x2x3_t __s1 = __p1; \
26797  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 2); \
26798})
26799#else
26800#define vst3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26801  int32x2x3_t __s1 = __p1; \
26802  int32x2x3_t __rev1; \
26803  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26804  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26805  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26806  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 2); \
26807})
26808#endif
26809
26810#ifdef __LITTLE_ENDIAN__
26811#define vst3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26812  int16x4x3_t __s1 = __p1; \
26813  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 1); \
26814})
26815#else
26816#define vst3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26817  int16x4x3_t __s1 = __p1; \
26818  int16x4x3_t __rev1; \
26819  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26820  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26821  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26822  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 1); \
26823})
26824#endif
26825
26826#ifdef __LITTLE_ENDIAN__
26827#define vst4_p8(__p0, __p1) __extension__ ({ \
26828  poly8x8x4_t __s1 = __p1; \
26829  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 4); \
26830})
26831#else
26832#define vst4_p8(__p0, __p1) __extension__ ({ \
26833  poly8x8x4_t __s1 = __p1; \
26834  poly8x8x4_t __rev1; \
26835  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26836  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26837  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26838  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26839  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 4); \
26840})
26841#endif
26842
26843#ifdef __LITTLE_ENDIAN__
26844#define vst4_p16(__p0, __p1) __extension__ ({ \
26845  poly16x4x4_t __s1 = __p1; \
26846  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 5); \
26847})
26848#else
26849#define vst4_p16(__p0, __p1) __extension__ ({ \
26850  poly16x4x4_t __s1 = __p1; \
26851  poly16x4x4_t __rev1; \
26852  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26853  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26854  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26855  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
26856  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 5); \
26857})
26858#endif
26859
26860#ifdef __LITTLE_ENDIAN__
26861#define vst4q_p8(__p0, __p1) __extension__ ({ \
26862  poly8x16x4_t __s1 = __p1; \
26863  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 36); \
26864})
26865#else
26866#define vst4q_p8(__p0, __p1) __extension__ ({ \
26867  poly8x16x4_t __s1 = __p1; \
26868  poly8x16x4_t __rev1; \
26869  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26870  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26871  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26872  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26873  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 36); \
26874})
26875#endif
26876
26877#ifdef __LITTLE_ENDIAN__
26878#define vst4q_p16(__p0, __p1) __extension__ ({ \
26879  poly16x8x4_t __s1 = __p1; \
26880  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 37); \
26881})
26882#else
26883#define vst4q_p16(__p0, __p1) __extension__ ({ \
26884  poly16x8x4_t __s1 = __p1; \
26885  poly16x8x4_t __rev1; \
26886  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26887  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26888  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26889  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26890  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 37); \
26891})
26892#endif
26893
26894#ifdef __LITTLE_ENDIAN__
26895#define vst4q_u8(__p0, __p1) __extension__ ({ \
26896  uint8x16x4_t __s1 = __p1; \
26897  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 48); \
26898})
26899#else
26900#define vst4q_u8(__p0, __p1) __extension__ ({ \
26901  uint8x16x4_t __s1 = __p1; \
26902  uint8x16x4_t __rev1; \
26903  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26904  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26905  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26906  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26907  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 48); \
26908})
26909#endif
26910
26911#ifdef __LITTLE_ENDIAN__
26912#define vst4q_u32(__p0, __p1) __extension__ ({ \
26913  uint32x4x4_t __s1 = __p1; \
26914  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 50); \
26915})
26916#else
26917#define vst4q_u32(__p0, __p1) __extension__ ({ \
26918  uint32x4x4_t __s1 = __p1; \
26919  uint32x4x4_t __rev1; \
26920  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26921  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26922  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26923  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
26924  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 50); \
26925})
26926#endif
26927
26928#ifdef __LITTLE_ENDIAN__
26929#define vst4q_u16(__p0, __p1) __extension__ ({ \
26930  uint16x8x4_t __s1 = __p1; \
26931  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 49); \
26932})
26933#else
26934#define vst4q_u16(__p0, __p1) __extension__ ({ \
26935  uint16x8x4_t __s1 = __p1; \
26936  uint16x8x4_t __rev1; \
26937  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26938  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26939  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26940  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26941  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 49); \
26942})
26943#endif
26944
26945#ifdef __LITTLE_ENDIAN__
26946#define vst4q_s8(__p0, __p1) __extension__ ({ \
26947  int8x16x4_t __s1 = __p1; \
26948  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 32); \
26949})
26950#else
26951#define vst4q_s8(__p0, __p1) __extension__ ({ \
26952  int8x16x4_t __s1 = __p1; \
26953  int8x16x4_t __rev1; \
26954  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26955  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26956  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26957  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26958  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 32); \
26959})
26960#endif
26961
26962#ifdef __LITTLE_ENDIAN__
26963#define vst4q_f32(__p0, __p1) __extension__ ({ \
26964  float32x4x4_t __s1 = __p1; \
26965  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 41); \
26966})
26967#else
26968#define vst4q_f32(__p0, __p1) __extension__ ({ \
26969  float32x4x4_t __s1 = __p1; \
26970  float32x4x4_t __rev1; \
26971  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26972  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26973  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26974  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
26975  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 41); \
26976})
26977#endif
26978
26979#ifdef __LITTLE_ENDIAN__
26980#define vst4q_f16(__p0, __p1) __extension__ ({ \
26981  float16x8x4_t __s1 = __p1; \
26982  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 40); \
26983})
26984#else
26985#define vst4q_f16(__p0, __p1) __extension__ ({ \
26986  float16x8x4_t __s1 = __p1; \
26987  float16x8x4_t __rev1; \
26988  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26989  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26990  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26991  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26992  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 40); \
26993})
26994#endif
26995
26996#ifdef __LITTLE_ENDIAN__
26997#define vst4q_s32(__p0, __p1) __extension__ ({ \
26998  int32x4x4_t __s1 = __p1; \
26999  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 34); \
27000})
27001#else
27002#define vst4q_s32(__p0, __p1) __extension__ ({ \
27003  int32x4x4_t __s1 = __p1; \
27004  int32x4x4_t __rev1; \
27005  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27006  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27007  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27008  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27009  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 34); \
27010})
27011#endif
27012
27013#ifdef __LITTLE_ENDIAN__
27014#define vst4q_s16(__p0, __p1) __extension__ ({ \
27015  int16x8x4_t __s1 = __p1; \
27016  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 33); \
27017})
27018#else
27019#define vst4q_s16(__p0, __p1) __extension__ ({ \
27020  int16x8x4_t __s1 = __p1; \
27021  int16x8x4_t __rev1; \
27022  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27023  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27024  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27025  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27026  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 33); \
27027})
27028#endif
27029
27030#ifdef __LITTLE_ENDIAN__
27031#define vst4_u8(__p0, __p1) __extension__ ({ \
27032  uint8x8x4_t __s1 = __p1; \
27033  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 16); \
27034})
27035#else
27036#define vst4_u8(__p0, __p1) __extension__ ({ \
27037  uint8x8x4_t __s1 = __p1; \
27038  uint8x8x4_t __rev1; \
27039  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27040  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27041  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27042  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27043  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 16); \
27044})
27045#endif
27046
27047#ifdef __LITTLE_ENDIAN__
27048#define vst4_u32(__p0, __p1) __extension__ ({ \
27049  uint32x2x4_t __s1 = __p1; \
27050  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 18); \
27051})
27052#else
27053#define vst4_u32(__p0, __p1) __extension__ ({ \
27054  uint32x2x4_t __s1 = __p1; \
27055  uint32x2x4_t __rev1; \
27056  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27057  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27058  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27059  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27060  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 18); \
27061})
27062#endif
27063
27064#ifdef __LITTLE_ENDIAN__
27065#define vst4_u64(__p0, __p1) __extension__ ({ \
27066  uint64x1x4_t __s1 = __p1; \
27067  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
27068})
27069#else
27070#define vst4_u64(__p0, __p1) __extension__ ({ \
27071  uint64x1x4_t __s1 = __p1; \
27072  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
27073})
27074#endif
27075
27076#ifdef __LITTLE_ENDIAN__
27077#define vst4_u16(__p0, __p1) __extension__ ({ \
27078  uint16x4x4_t __s1 = __p1; \
27079  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 17); \
27080})
27081#else
27082#define vst4_u16(__p0, __p1) __extension__ ({ \
27083  uint16x4x4_t __s1 = __p1; \
27084  uint16x4x4_t __rev1; \
27085  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27086  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27087  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27088  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27089  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 17); \
27090})
27091#endif
27092
27093#ifdef __LITTLE_ENDIAN__
27094#define vst4_s8(__p0, __p1) __extension__ ({ \
27095  int8x8x4_t __s1 = __p1; \
27096  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 0); \
27097})
27098#else
27099#define vst4_s8(__p0, __p1) __extension__ ({ \
27100  int8x8x4_t __s1 = __p1; \
27101  int8x8x4_t __rev1; \
27102  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27103  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27104  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27105  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27106  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 0); \
27107})
27108#endif
27109
27110#ifdef __LITTLE_ENDIAN__
27111#define vst4_f32(__p0, __p1) __extension__ ({ \
27112  float32x2x4_t __s1 = __p1; \
27113  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 9); \
27114})
27115#else
27116#define vst4_f32(__p0, __p1) __extension__ ({ \
27117  float32x2x4_t __s1 = __p1; \
27118  float32x2x4_t __rev1; \
27119  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27120  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27121  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27122  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27123  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 9); \
27124})
27125#endif
27126
27127#ifdef __LITTLE_ENDIAN__
27128#define vst4_f16(__p0, __p1) __extension__ ({ \
27129  float16x4x4_t __s1 = __p1; \
27130  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 8); \
27131})
27132#else
27133#define vst4_f16(__p0, __p1) __extension__ ({ \
27134  float16x4x4_t __s1 = __p1; \
27135  float16x4x4_t __rev1; \
27136  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27137  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27138  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27139  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27140  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 8); \
27141})
27142#endif
27143
27144#ifdef __LITTLE_ENDIAN__
27145#define vst4_s32(__p0, __p1) __extension__ ({ \
27146  int32x2x4_t __s1 = __p1; \
27147  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 2); \
27148})
27149#else
27150#define vst4_s32(__p0, __p1) __extension__ ({ \
27151  int32x2x4_t __s1 = __p1; \
27152  int32x2x4_t __rev1; \
27153  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27154  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27155  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27156  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27157  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 2); \
27158})
27159#endif
27160
27161#ifdef __LITTLE_ENDIAN__
27162#define vst4_s64(__p0, __p1) __extension__ ({ \
27163  int64x1x4_t __s1 = __p1; \
27164  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
27165})
27166#else
27167#define vst4_s64(__p0, __p1) __extension__ ({ \
27168  int64x1x4_t __s1 = __p1; \
27169  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
27170})
27171#endif
27172
27173#ifdef __LITTLE_ENDIAN__
27174#define vst4_s16(__p0, __p1) __extension__ ({ \
27175  int16x4x4_t __s1 = __p1; \
27176  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 1); \
27177})
27178#else
27179#define vst4_s16(__p0, __p1) __extension__ ({ \
27180  int16x4x4_t __s1 = __p1; \
27181  int16x4x4_t __rev1; \
27182  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27183  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27184  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27185  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27186  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 1); \
27187})
27188#endif
27189
27190#ifdef __LITTLE_ENDIAN__
27191#define vst4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
27192  poly8x8x4_t __s1 = __p1; \
27193  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 4); \
27194})
27195#else
27196#define vst4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
27197  poly8x8x4_t __s1 = __p1; \
27198  poly8x8x4_t __rev1; \
27199  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27200  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27201  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27202  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27203  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 4); \
27204})
27205#endif
27206
27207#ifdef __LITTLE_ENDIAN__
27208#define vst4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27209  poly16x4x4_t __s1 = __p1; \
27210  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 5); \
27211})
27212#else
27213#define vst4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27214  poly16x4x4_t __s1 = __p1; \
27215  poly16x4x4_t __rev1; \
27216  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27217  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27218  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27219  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27220  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 5); \
27221})
27222#endif
27223
27224#ifdef __LITTLE_ENDIAN__
27225#define vst4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27226  poly16x8x4_t __s1 = __p1; \
27227  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 37); \
27228})
27229#else
27230#define vst4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27231  poly16x8x4_t __s1 = __p1; \
27232  poly16x8x4_t __rev1; \
27233  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27234  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27235  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27236  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27237  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 37); \
27238})
27239#endif
27240
27241#ifdef __LITTLE_ENDIAN__
27242#define vst4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27243  uint32x4x4_t __s1 = __p1; \
27244  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 50); \
27245})
27246#else
27247#define vst4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27248  uint32x4x4_t __s1 = __p1; \
27249  uint32x4x4_t __rev1; \
27250  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27251  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27252  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27253  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27254  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 50); \
27255})
27256#endif
27257
27258#ifdef __LITTLE_ENDIAN__
27259#define vst4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27260  uint16x8x4_t __s1 = __p1; \
27261  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 49); \
27262})
27263#else
27264#define vst4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27265  uint16x8x4_t __s1 = __p1; \
27266  uint16x8x4_t __rev1; \
27267  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27268  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27269  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27270  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27271  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 49); \
27272})
27273#endif
27274
27275#ifdef __LITTLE_ENDIAN__
27276#define vst4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27277  float32x4x4_t __s1 = __p1; \
27278  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 41); \
27279})
27280#else
27281#define vst4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27282  float32x4x4_t __s1 = __p1; \
27283  float32x4x4_t __rev1; \
27284  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27285  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27286  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27287  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27288  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 41); \
27289})
27290#endif
27291
27292#ifdef __LITTLE_ENDIAN__
27293#define vst4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27294  float16x8x4_t __s1 = __p1; \
27295  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 40); \
27296})
27297#else
27298#define vst4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27299  float16x8x4_t __s1 = __p1; \
27300  float16x8x4_t __rev1; \
27301  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27302  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27303  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27304  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27305  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 40); \
27306})
27307#endif
27308
27309#ifdef __LITTLE_ENDIAN__
27310#define vst4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27311  int32x4x4_t __s1 = __p1; \
27312  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 34); \
27313})
27314#else
27315#define vst4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27316  int32x4x4_t __s1 = __p1; \
27317  int32x4x4_t __rev1; \
27318  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27319  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27320  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27321  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27322  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 34); \
27323})
27324#endif
27325
27326#ifdef __LITTLE_ENDIAN__
27327#define vst4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27328  int16x8x4_t __s1 = __p1; \
27329  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 33); \
27330})
27331#else
27332#define vst4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27333  int16x8x4_t __s1 = __p1; \
27334  int16x8x4_t __rev1; \
27335  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27336  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27337  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27338  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27339  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 33); \
27340})
27341#endif
27342
27343#ifdef __LITTLE_ENDIAN__
27344#define vst4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
27345  uint8x8x4_t __s1 = __p1; \
27346  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 16); \
27347})
27348#else
27349#define vst4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
27350  uint8x8x4_t __s1 = __p1; \
27351  uint8x8x4_t __rev1; \
27352  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27353  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27354  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27355  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27356  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 16); \
27357})
27358#endif
27359
27360#ifdef __LITTLE_ENDIAN__
27361#define vst4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27362  uint32x2x4_t __s1 = __p1; \
27363  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 18); \
27364})
27365#else
27366#define vst4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27367  uint32x2x4_t __s1 = __p1; \
27368  uint32x2x4_t __rev1; \
27369  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27370  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27371  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27372  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27373  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 18); \
27374})
27375#endif
27376
27377#ifdef __LITTLE_ENDIAN__
27378#define vst4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27379  uint16x4x4_t __s1 = __p1; \
27380  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 17); \
27381})
27382#else
27383#define vst4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27384  uint16x4x4_t __s1 = __p1; \
27385  uint16x4x4_t __rev1; \
27386  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27387  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27388  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27389  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27390  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 17); \
27391})
27392#endif
27393
27394#ifdef __LITTLE_ENDIAN__
27395#define vst4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
27396  int8x8x4_t __s1 = __p1; \
27397  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 0); \
27398})
27399#else
27400#define vst4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
27401  int8x8x4_t __s1 = __p1; \
27402  int8x8x4_t __rev1; \
27403  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27404  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27405  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27406  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27407  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 0); \
27408})
27409#endif
27410
27411#ifdef __LITTLE_ENDIAN__
27412#define vst4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27413  float32x2x4_t __s1 = __p1; \
27414  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 9); \
27415})
27416#else
27417#define vst4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27418  float32x2x4_t __s1 = __p1; \
27419  float32x2x4_t __rev1; \
27420  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27421  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27422  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27423  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27424  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 9); \
27425})
27426#endif
27427
27428#ifdef __LITTLE_ENDIAN__
27429#define vst4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27430  float16x4x4_t __s1 = __p1; \
27431  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 8); \
27432})
27433#else
27434#define vst4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27435  float16x4x4_t __s1 = __p1; \
27436  float16x4x4_t __rev1; \
27437  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27438  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27439  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27440  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27441  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 8); \
27442})
27443#endif
27444
27445#ifdef __LITTLE_ENDIAN__
27446#define vst4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27447  int32x2x4_t __s1 = __p1; \
27448  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 2); \
27449})
27450#else
27451#define vst4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27452  int32x2x4_t __s1 = __p1; \
27453  int32x2x4_t __rev1; \
27454  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27455  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27456  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27457  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27458  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 2); \
27459})
27460#endif
27461
27462#ifdef __LITTLE_ENDIAN__
27463#define vst4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27464  int16x4x4_t __s1 = __p1; \
27465  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 1); \
27466})
27467#else
27468#define vst4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27469  int16x4x4_t __s1 = __p1; \
27470  int16x4x4_t __rev1; \
27471  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27472  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27473  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27474  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27475  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 1); \
27476})
27477#endif
27478
27479#ifdef __LITTLE_ENDIAN__
27480__ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
27481  uint8x16_t __ret;
27482  __ret = __p0 - __p1;
27483  return __ret;
27484}
27485#else
27486__ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
27487  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27488  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27489  uint8x16_t __ret;
27490  __ret = __rev0 - __rev1;
27491  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27492  return __ret;
27493}
27494#endif
27495
27496#ifdef __LITTLE_ENDIAN__
27497__ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
27498  uint32x4_t __ret;
27499  __ret = __p0 - __p1;
27500  return __ret;
27501}
27502#else
27503__ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
27504  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27505  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27506  uint32x4_t __ret;
27507  __ret = __rev0 - __rev1;
27508  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27509  return __ret;
27510}
27511#endif
27512
27513#ifdef __LITTLE_ENDIAN__
27514__ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
27515  uint64x2_t __ret;
27516  __ret = __p0 - __p1;
27517  return __ret;
27518}
27519#else
27520__ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
27521  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27522  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27523  uint64x2_t __ret;
27524  __ret = __rev0 - __rev1;
27525  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27526  return __ret;
27527}
27528#endif
27529
27530#ifdef __LITTLE_ENDIAN__
27531__ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
27532  uint16x8_t __ret;
27533  __ret = __p0 - __p1;
27534  return __ret;
27535}
27536#else
27537__ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
27538  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27539  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27540  uint16x8_t __ret;
27541  __ret = __rev0 - __rev1;
27542  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27543  return __ret;
27544}
27545#endif
27546
27547#ifdef __LITTLE_ENDIAN__
27548__ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1) {
27549  int8x16_t __ret;
27550  __ret = __p0 - __p1;
27551  return __ret;
27552}
27553#else
27554__ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1) {
27555  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27556  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27557  int8x16_t __ret;
27558  __ret = __rev0 - __rev1;
27559  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27560  return __ret;
27561}
27562#endif
27563
27564#ifdef __LITTLE_ENDIAN__
27565__ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1) {
27566  float32x4_t __ret;
27567  __ret = __p0 - __p1;
27568  return __ret;
27569}
27570#else
27571__ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1) {
27572  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27573  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27574  float32x4_t __ret;
27575  __ret = __rev0 - __rev1;
27576  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27577  return __ret;
27578}
27579#endif
27580
27581#ifdef __LITTLE_ENDIAN__
27582__ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1) {
27583  int32x4_t __ret;
27584  __ret = __p0 - __p1;
27585  return __ret;
27586}
27587#else
27588__ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1) {
27589  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27590  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27591  int32x4_t __ret;
27592  __ret = __rev0 - __rev1;
27593  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27594  return __ret;
27595}
27596#endif
27597
27598#ifdef __LITTLE_ENDIAN__
27599__ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1) {
27600  int64x2_t __ret;
27601  __ret = __p0 - __p1;
27602  return __ret;
27603}
27604#else
27605__ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1) {
27606  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27607  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27608  int64x2_t __ret;
27609  __ret = __rev0 - __rev1;
27610  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27611  return __ret;
27612}
27613#endif
27614
27615#ifdef __LITTLE_ENDIAN__
27616__ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1) {
27617  int16x8_t __ret;
27618  __ret = __p0 - __p1;
27619  return __ret;
27620}
27621#else
27622__ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1) {
27623  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27624  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27625  int16x8_t __ret;
27626  __ret = __rev0 - __rev1;
27627  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27628  return __ret;
27629}
27630#endif
27631
27632#ifdef __LITTLE_ENDIAN__
27633__ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
27634  uint8x8_t __ret;
27635  __ret = __p0 - __p1;
27636  return __ret;
27637}
27638#else
27639__ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
27640  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27641  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27642  uint8x8_t __ret;
27643  __ret = __rev0 - __rev1;
27644  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27645  return __ret;
27646}
27647#endif
27648
27649#ifdef __LITTLE_ENDIAN__
27650__ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
27651  uint32x2_t __ret;
27652  __ret = __p0 - __p1;
27653  return __ret;
27654}
27655#else
27656__ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
27657  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27658  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27659  uint32x2_t __ret;
27660  __ret = __rev0 - __rev1;
27661  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27662  return __ret;
27663}
27664#endif
27665
27666#ifdef __LITTLE_ENDIAN__
27667__ai uint64x1_t vsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
27668  uint64x1_t __ret;
27669  __ret = __p0 - __p1;
27670  return __ret;
27671}
27672#else
27673__ai uint64x1_t vsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
27674  uint64x1_t __ret;
27675  __ret = __p0 - __p1;
27676  return __ret;
27677}
27678#endif
27679
27680#ifdef __LITTLE_ENDIAN__
27681__ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
27682  uint16x4_t __ret;
27683  __ret = __p0 - __p1;
27684  return __ret;
27685}
27686#else
27687__ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
27688  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27689  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27690  uint16x4_t __ret;
27691  __ret = __rev0 - __rev1;
27692  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27693  return __ret;
27694}
27695#endif
27696
27697#ifdef __LITTLE_ENDIAN__
27698__ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1) {
27699  int8x8_t __ret;
27700  __ret = __p0 - __p1;
27701  return __ret;
27702}
27703#else
27704__ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1) {
27705  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27706  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27707  int8x8_t __ret;
27708  __ret = __rev0 - __rev1;
27709  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27710  return __ret;
27711}
27712#endif
27713
27714#ifdef __LITTLE_ENDIAN__
27715__ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1) {
27716  float32x2_t __ret;
27717  __ret = __p0 - __p1;
27718  return __ret;
27719}
27720#else
27721__ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1) {
27722  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27723  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27724  float32x2_t __ret;
27725  __ret = __rev0 - __rev1;
27726  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27727  return __ret;
27728}
27729#endif
27730
27731#ifdef __LITTLE_ENDIAN__
27732__ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1) {
27733  int32x2_t __ret;
27734  __ret = __p0 - __p1;
27735  return __ret;
27736}
27737#else
27738__ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1) {
27739  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27740  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27741  int32x2_t __ret;
27742  __ret = __rev0 - __rev1;
27743  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27744  return __ret;
27745}
27746#endif
27747
27748#ifdef __LITTLE_ENDIAN__
27749__ai int64x1_t vsub_s64(int64x1_t __p0, int64x1_t __p1) {
27750  int64x1_t __ret;
27751  __ret = __p0 - __p1;
27752  return __ret;
27753}
27754#else
27755__ai int64x1_t vsub_s64(int64x1_t __p0, int64x1_t __p1) {
27756  int64x1_t __ret;
27757  __ret = __p0 - __p1;
27758  return __ret;
27759}
27760#endif
27761
27762#ifdef __LITTLE_ENDIAN__
27763__ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1) {
27764  int16x4_t __ret;
27765  __ret = __p0 - __p1;
27766  return __ret;
27767}
27768#else
27769__ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1) {
27770  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27771  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27772  int16x4_t __ret;
27773  __ret = __rev0 - __rev1;
27774  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27775  return __ret;
27776}
27777#endif
27778
27779#ifdef __LITTLE_ENDIAN__
27780__ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
27781  uint16x4_t __ret;
27782  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
27783  return __ret;
27784}
27785#else
27786__ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
27787  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27788  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27789  uint16x4_t __ret;
27790  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
27791  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27792  return __ret;
27793}
27794__ai uint16x4_t __noswap_vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
27795  uint16x4_t __ret;
27796  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
27797  return __ret;
27798}
27799#endif
27800
27801#ifdef __LITTLE_ENDIAN__
27802__ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
27803  uint32x2_t __ret;
27804  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
27805  return __ret;
27806}
27807#else
27808__ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
27809  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27810  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27811  uint32x2_t __ret;
27812  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
27813  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27814  return __ret;
27815}
27816__ai uint32x2_t __noswap_vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
27817  uint32x2_t __ret;
27818  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
27819  return __ret;
27820}
27821#endif
27822
27823#ifdef __LITTLE_ENDIAN__
27824__ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
27825  uint8x8_t __ret;
27826  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
27827  return __ret;
27828}
27829#else
27830__ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
27831  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27832  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27833  uint8x8_t __ret;
27834  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
27835  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27836  return __ret;
27837}
27838__ai uint8x8_t __noswap_vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
27839  uint8x8_t __ret;
27840  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
27841  return __ret;
27842}
27843#endif
27844
27845#ifdef __LITTLE_ENDIAN__
27846__ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
27847  int16x4_t __ret;
27848  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
27849  return __ret;
27850}
27851#else
27852__ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
27853  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27854  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27855  int16x4_t __ret;
27856  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
27857  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27858  return __ret;
27859}
27860__ai int16x4_t __noswap_vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
27861  int16x4_t __ret;
27862  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
27863  return __ret;
27864}
27865#endif
27866
27867#ifdef __LITTLE_ENDIAN__
27868__ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
27869  int32x2_t __ret;
27870  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
27871  return __ret;
27872}
27873#else
27874__ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
27875  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27876  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27877  int32x2_t __ret;
27878  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
27879  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27880  return __ret;
27881}
27882__ai int32x2_t __noswap_vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
27883  int32x2_t __ret;
27884  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
27885  return __ret;
27886}
27887#endif
27888
27889#ifdef __LITTLE_ENDIAN__
27890__ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
27891  int8x8_t __ret;
27892  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
27893  return __ret;
27894}
27895#else
27896__ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
27897  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27898  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27899  int8x8_t __ret;
27900  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
27901  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27902  return __ret;
27903}
27904__ai int8x8_t __noswap_vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
27905  int8x8_t __ret;
27906  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
27907  return __ret;
27908}
27909#endif
27910
27911#ifdef __LITTLE_ENDIAN__
27912__ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1) {
27913  uint16x8_t __ret;
27914  __ret = vmovl_u8(__p0) - vmovl_u8(__p1);
27915  return __ret;
27916}
27917#else
27918__ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1) {
27919  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27920  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27921  uint16x8_t __ret;
27922  __ret = __noswap_vmovl_u8(__rev0) - __noswap_vmovl_u8(__rev1);
27923  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27924  return __ret;
27925}
27926#endif
27927
27928#ifdef __LITTLE_ENDIAN__
27929__ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1) {
27930  uint64x2_t __ret;
27931  __ret = vmovl_u32(__p0) - vmovl_u32(__p1);
27932  return __ret;
27933}
27934#else
27935__ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1) {
27936  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27937  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27938  uint64x2_t __ret;
27939  __ret = __noswap_vmovl_u32(__rev0) - __noswap_vmovl_u32(__rev1);
27940  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27941  return __ret;
27942}
27943#endif
27944
27945#ifdef __LITTLE_ENDIAN__
27946__ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1) {
27947  uint32x4_t __ret;
27948  __ret = vmovl_u16(__p0) - vmovl_u16(__p1);
27949  return __ret;
27950}
27951#else
27952__ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1) {
27953  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27954  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27955  uint32x4_t __ret;
27956  __ret = __noswap_vmovl_u16(__rev0) - __noswap_vmovl_u16(__rev1);
27957  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27958  return __ret;
27959}
27960#endif
27961
27962#ifdef __LITTLE_ENDIAN__
27963__ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1) {
27964  int16x8_t __ret;
27965  __ret = vmovl_s8(__p0) - vmovl_s8(__p1);
27966  return __ret;
27967}
27968#else
27969__ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1) {
27970  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27971  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27972  int16x8_t __ret;
27973  __ret = __noswap_vmovl_s8(__rev0) - __noswap_vmovl_s8(__rev1);
27974  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27975  return __ret;
27976}
27977#endif
27978
27979#ifdef __LITTLE_ENDIAN__
27980__ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1) {
27981  int64x2_t __ret;
27982  __ret = vmovl_s32(__p0) - vmovl_s32(__p1);
27983  return __ret;
27984}
27985#else
27986__ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1) {
27987  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27988  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27989  int64x2_t __ret;
27990  __ret = __noswap_vmovl_s32(__rev0) - __noswap_vmovl_s32(__rev1);
27991  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27992  return __ret;
27993}
27994#endif
27995
27996#ifdef __LITTLE_ENDIAN__
27997__ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1) {
27998  int32x4_t __ret;
27999  __ret = vmovl_s16(__p0) - vmovl_s16(__p1);
28000  return __ret;
28001}
28002#else
28003__ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1) {
28004  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28005  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28006  int32x4_t __ret;
28007  __ret = __noswap_vmovl_s16(__rev0) - __noswap_vmovl_s16(__rev1);
28008  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
28009  return __ret;
28010}
28011#endif
28012
28013#ifdef __LITTLE_ENDIAN__
28014__ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1) {
28015  uint16x8_t __ret;
28016  __ret = __p0 - vmovl_u8(__p1);
28017  return __ret;
28018}
28019#else
28020__ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1) {
28021  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28022  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28023  uint16x8_t __ret;
28024  __ret = __rev0 - __noswap_vmovl_u8(__rev1);
28025  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28026  return __ret;
28027}
28028#endif
28029
28030#ifdef __LITTLE_ENDIAN__
28031__ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1) {
28032  uint64x2_t __ret;
28033  __ret = __p0 - vmovl_u32(__p1);
28034  return __ret;
28035}
28036#else
28037__ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1) {
28038  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28039  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28040  uint64x2_t __ret;
28041  __ret = __rev0 - __noswap_vmovl_u32(__rev1);
28042  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
28043  return __ret;
28044}
28045#endif
28046
28047#ifdef __LITTLE_ENDIAN__
28048__ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1) {
28049  uint32x4_t __ret;
28050  __ret = __p0 - vmovl_u16(__p1);
28051  return __ret;
28052}
28053#else
28054__ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1) {
28055  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28056  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28057  uint32x4_t __ret;
28058  __ret = __rev0 - __noswap_vmovl_u16(__rev1);
28059  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
28060  return __ret;
28061}
28062#endif
28063
28064#ifdef __LITTLE_ENDIAN__
28065__ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1) {
28066  int16x8_t __ret;
28067  __ret = __p0 - vmovl_s8(__p1);
28068  return __ret;
28069}
28070#else
28071__ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1) {
28072  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28073  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28074  int16x8_t __ret;
28075  __ret = __rev0 - __noswap_vmovl_s8(__rev1);
28076  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28077  return __ret;
28078}
28079#endif
28080
28081#ifdef __LITTLE_ENDIAN__
28082__ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1) {
28083  int64x2_t __ret;
28084  __ret = __p0 - vmovl_s32(__p1);
28085  return __ret;
28086}
28087#else
28088__ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1) {
28089  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28090  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28091  int64x2_t __ret;
28092  __ret = __rev0 - __noswap_vmovl_s32(__rev1);
28093  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
28094  return __ret;
28095}
28096#endif
28097
28098#ifdef __LITTLE_ENDIAN__
28099__ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1) {
28100  int32x4_t __ret;
28101  __ret = __p0 - vmovl_s16(__p1);
28102  return __ret;
28103}
28104#else
28105__ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1) {
28106  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28107  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28108  int32x4_t __ret;
28109  __ret = __rev0 - __noswap_vmovl_s16(__rev1);
28110  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
28111  return __ret;
28112}
28113#endif
28114
28115#ifdef __LITTLE_ENDIAN__
28116__ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1) {
28117  poly8x8_t __ret;
28118  __ret = (poly8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
28119  return __ret;
28120}
28121#else
28122__ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1) {
28123  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28124  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28125  poly8x8_t __ret;
28126  __ret = (poly8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
28127  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28128  return __ret;
28129}
28130#endif
28131
28132#ifdef __LITTLE_ENDIAN__
28133__ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1) {
28134  uint8x8_t __ret;
28135  __ret = (uint8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
28136  return __ret;
28137}
28138#else
28139__ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1) {
28140  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28141  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28142  uint8x8_t __ret;
28143  __ret = (uint8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
28144  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28145  return __ret;
28146}
28147#endif
28148
28149#ifdef __LITTLE_ENDIAN__
28150__ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1) {
28151  int8x8_t __ret;
28152  __ret = (int8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
28153  return __ret;
28154}
28155#else
28156__ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1) {
28157  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28158  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28159  int8x8_t __ret;
28160  __ret = (int8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
28161  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28162  return __ret;
28163}
28164#endif
28165
28166#ifdef __LITTLE_ENDIAN__
28167__ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1) {
28168  poly8x8_t __ret;
28169  __ret = (poly8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 4);
28170  return __ret;
28171}
28172#else
28173__ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1) {
28174  poly8x8x2_t __rev0;
28175  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28176  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28177  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28178  poly8x8_t __ret;
28179  __ret = (poly8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 4);
28180  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28181  return __ret;
28182}
28183#endif
28184
28185#ifdef __LITTLE_ENDIAN__
28186__ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1) {
28187  uint8x8_t __ret;
28188  __ret = (uint8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 16);
28189  return __ret;
28190}
28191#else
28192__ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1) {
28193  uint8x8x2_t __rev0;
28194  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28195  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28196  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28197  uint8x8_t __ret;
28198  __ret = (uint8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 16);
28199  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28200  return __ret;
28201}
28202#endif
28203
28204#ifdef __LITTLE_ENDIAN__
28205__ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1) {
28206  int8x8_t __ret;
28207  __ret = (int8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 0);
28208  return __ret;
28209}
28210#else
28211__ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1) {
28212  int8x8x2_t __rev0;
28213  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28214  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28215  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28216  int8x8_t __ret;
28217  __ret = (int8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 0);
28218  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28219  return __ret;
28220}
28221#endif
28222
28223#ifdef __LITTLE_ENDIAN__
28224__ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1) {
28225  poly8x8_t __ret;
28226  __ret = (poly8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 4);
28227  return __ret;
28228}
28229#else
28230__ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1) {
28231  poly8x8x3_t __rev0;
28232  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28233  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28234  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28235  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28236  poly8x8_t __ret;
28237  __ret = (poly8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 4);
28238  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28239  return __ret;
28240}
28241#endif
28242
28243#ifdef __LITTLE_ENDIAN__
28244__ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1) {
28245  uint8x8_t __ret;
28246  __ret = (uint8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 16);
28247  return __ret;
28248}
28249#else
28250__ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1) {
28251  uint8x8x3_t __rev0;
28252  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28253  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28254  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28255  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28256  uint8x8_t __ret;
28257  __ret = (uint8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 16);
28258  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28259  return __ret;
28260}
28261#endif
28262
28263#ifdef __LITTLE_ENDIAN__
28264__ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1) {
28265  int8x8_t __ret;
28266  __ret = (int8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 0);
28267  return __ret;
28268}
28269#else
28270__ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1) {
28271  int8x8x3_t __rev0;
28272  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28273  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28274  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28275  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28276  int8x8_t __ret;
28277  __ret = (int8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 0);
28278  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28279  return __ret;
28280}
28281#endif
28282
28283#ifdef __LITTLE_ENDIAN__
28284__ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1) {
28285  poly8x8_t __ret;
28286  __ret = (poly8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 4);
28287  return __ret;
28288}
28289#else
28290__ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1) {
28291  poly8x8x4_t __rev0;
28292  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28293  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28294  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28295  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28296  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28297  poly8x8_t __ret;
28298  __ret = (poly8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 4);
28299  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28300  return __ret;
28301}
28302#endif
28303
28304#ifdef __LITTLE_ENDIAN__
28305__ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1) {
28306  uint8x8_t __ret;
28307  __ret = (uint8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 16);
28308  return __ret;
28309}
28310#else
28311__ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1) {
28312  uint8x8x4_t __rev0;
28313  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28314  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28315  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28316  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28317  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28318  uint8x8_t __ret;
28319  __ret = (uint8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 16);
28320  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28321  return __ret;
28322}
28323#endif
28324
28325#ifdef __LITTLE_ENDIAN__
28326__ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1) {
28327  int8x8_t __ret;
28328  __ret = (int8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 0);
28329  return __ret;
28330}
28331#else
28332__ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1) {
28333  int8x8x4_t __rev0;
28334  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28335  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28336  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28337  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28338  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28339  int8x8_t __ret;
28340  __ret = (int8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 0);
28341  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28342  return __ret;
28343}
28344#endif
28345
28346#ifdef __LITTLE_ENDIAN__
28347__ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2) {
28348  poly8x8_t __ret;
28349  __ret = (poly8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 4);
28350  return __ret;
28351}
28352#else
28353__ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2) {
28354  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28355  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28356  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28357  poly8x8_t __ret;
28358  __ret = (poly8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 4);
28359  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28360  return __ret;
28361}
28362#endif
28363
28364#ifdef __LITTLE_ENDIAN__
28365__ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
28366  uint8x8_t __ret;
28367  __ret = (uint8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 16);
28368  return __ret;
28369}
28370#else
28371__ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
28372  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28373  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28374  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28375  uint8x8_t __ret;
28376  __ret = (uint8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 16);
28377  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28378  return __ret;
28379}
28380#endif
28381
28382#ifdef __LITTLE_ENDIAN__
28383__ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
28384  int8x8_t __ret;
28385  __ret = (int8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 0);
28386  return __ret;
28387}
28388#else
28389__ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
28390  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28391  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28392  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28393  int8x8_t __ret;
28394  __ret = (int8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 0);
28395  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28396  return __ret;
28397}
28398#endif
28399
28400#ifdef __LITTLE_ENDIAN__
28401__ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2) {
28402  poly8x8_t __ret;
28403  __ret = (poly8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 4);
28404  return __ret;
28405}
28406#else
28407__ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2) {
28408  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28409  poly8x8x2_t __rev1;
28410  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28411  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28412  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28413  poly8x8_t __ret;
28414  __ret = (poly8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 4);
28415  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28416  return __ret;
28417}
28418#endif
28419
28420#ifdef __LITTLE_ENDIAN__
28421__ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2) {
28422  uint8x8_t __ret;
28423  __ret = (uint8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 16);
28424  return __ret;
28425}
28426#else
28427__ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2) {
28428  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28429  uint8x8x2_t __rev1;
28430  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28431  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28432  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28433  uint8x8_t __ret;
28434  __ret = (uint8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 16);
28435  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28436  return __ret;
28437}
28438#endif
28439
28440#ifdef __LITTLE_ENDIAN__
28441__ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2) {
28442  int8x8_t __ret;
28443  __ret = (int8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 0);
28444  return __ret;
28445}
28446#else
28447__ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2) {
28448  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28449  int8x8x2_t __rev1;
28450  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28451  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28452  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28453  int8x8_t __ret;
28454  __ret = (int8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 0);
28455  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28456  return __ret;
28457}
28458#endif
28459
28460#ifdef __LITTLE_ENDIAN__
28461__ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2) {
28462  poly8x8_t __ret;
28463  __ret = (poly8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 4);
28464  return __ret;
28465}
28466#else
28467__ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2) {
28468  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28469  poly8x8x3_t __rev1;
28470  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28471  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28472  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28473  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28474  poly8x8_t __ret;
28475  __ret = (poly8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 4);
28476  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28477  return __ret;
28478}
28479#endif
28480
28481#ifdef __LITTLE_ENDIAN__
28482__ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2) {
28483  uint8x8_t __ret;
28484  __ret = (uint8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 16);
28485  return __ret;
28486}
28487#else
28488__ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2) {
28489  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28490  uint8x8x3_t __rev1;
28491  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28492  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28493  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28494  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28495  uint8x8_t __ret;
28496  __ret = (uint8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 16);
28497  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28498  return __ret;
28499}
28500#endif
28501
28502#ifdef __LITTLE_ENDIAN__
28503__ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2) {
28504  int8x8_t __ret;
28505  __ret = (int8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 0);
28506  return __ret;
28507}
28508#else
28509__ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2) {
28510  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28511  int8x8x3_t __rev1;
28512  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28513  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28514  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28515  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28516  int8x8_t __ret;
28517  __ret = (int8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 0);
28518  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28519  return __ret;
28520}
28521#endif
28522
28523#ifdef __LITTLE_ENDIAN__
28524__ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2) {
28525  poly8x8_t __ret;
28526  __ret = (poly8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 4);
28527  return __ret;
28528}
28529#else
28530__ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2) {
28531  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28532  poly8x8x4_t __rev1;
28533  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28534  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28535  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28536  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28537  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28538  poly8x8_t __ret;
28539  __ret = (poly8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 4);
28540  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28541  return __ret;
28542}
28543#endif
28544
28545#ifdef __LITTLE_ENDIAN__
28546__ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2) {
28547  uint8x8_t __ret;
28548  __ret = (uint8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 16);
28549  return __ret;
28550}
28551#else
28552__ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2) {
28553  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28554  uint8x8x4_t __rev1;
28555  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28556  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28557  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28558  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28559  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28560  uint8x8_t __ret;
28561  __ret = (uint8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 16);
28562  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28563  return __ret;
28564}
28565#endif
28566
28567#ifdef __LITTLE_ENDIAN__
28568__ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2) {
28569  int8x8_t __ret;
28570  __ret = (int8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 0);
28571  return __ret;
28572}
28573#else
28574__ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2) {
28575  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28576  int8x8x4_t __rev1;
28577  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28578  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28579  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28580  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28581  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28582  int8x8_t __ret;
28583  __ret = (int8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 0);
28584  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28585  return __ret;
28586}
28587#endif
28588
28589#ifdef __LITTLE_ENDIAN__
28590__ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1) {
28591  poly8x8x2_t __ret;
28592  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
28593  return __ret;
28594}
28595#else
28596__ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1) {
28597  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28598  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28599  poly8x8x2_t __ret;
28600  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
28601
28602  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28603  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28604  return __ret;
28605}
28606#endif
28607
28608#ifdef __LITTLE_ENDIAN__
28609__ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1) {
28610  poly16x4x2_t __ret;
28611  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
28612  return __ret;
28613}
28614#else
28615__ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1) {
28616  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28617  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28618  poly16x4x2_t __ret;
28619  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
28620
28621  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28622  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28623  return __ret;
28624}
28625#endif
28626
28627#ifdef __LITTLE_ENDIAN__
28628__ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1) {
28629  poly8x16x2_t __ret;
28630  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
28631  return __ret;
28632}
28633#else
28634__ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1) {
28635  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28636  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28637  poly8x16x2_t __ret;
28638  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
28639
28640  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28641  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28642  return __ret;
28643}
28644#endif
28645
28646#ifdef __LITTLE_ENDIAN__
28647__ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1) {
28648  poly16x8x2_t __ret;
28649  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
28650  return __ret;
28651}
28652#else
28653__ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1) {
28654  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28655  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28656  poly16x8x2_t __ret;
28657  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
28658
28659  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28660  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28661  return __ret;
28662}
28663#endif
28664
28665#ifdef __LITTLE_ENDIAN__
28666__ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1) {
28667  uint8x16x2_t __ret;
28668  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
28669  return __ret;
28670}
28671#else
28672__ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1) {
28673  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28674  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28675  uint8x16x2_t __ret;
28676  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
28677
28678  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28679  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28680  return __ret;
28681}
28682#endif
28683
28684#ifdef __LITTLE_ENDIAN__
28685__ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1) {
28686  uint32x4x2_t __ret;
28687  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
28688  return __ret;
28689}
28690#else
28691__ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1) {
28692  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28693  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28694  uint32x4x2_t __ret;
28695  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
28696
28697  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28698  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28699  return __ret;
28700}
28701#endif
28702
28703#ifdef __LITTLE_ENDIAN__
28704__ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1) {
28705  uint16x8x2_t __ret;
28706  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
28707  return __ret;
28708}
28709#else
28710__ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1) {
28711  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28712  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28713  uint16x8x2_t __ret;
28714  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
28715
28716  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28717  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28718  return __ret;
28719}
28720#endif
28721
28722#ifdef __LITTLE_ENDIAN__
28723__ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1) {
28724  int8x16x2_t __ret;
28725  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
28726  return __ret;
28727}
28728#else
28729__ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1) {
28730  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28731  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28732  int8x16x2_t __ret;
28733  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
28734
28735  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28736  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28737  return __ret;
28738}
28739#endif
28740
28741#ifdef __LITTLE_ENDIAN__
28742__ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1) {
28743  float32x4x2_t __ret;
28744  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
28745  return __ret;
28746}
28747#else
28748__ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1) {
28749  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28750  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28751  float32x4x2_t __ret;
28752  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
28753
28754  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28755  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28756  return __ret;
28757}
28758#endif
28759
28760#ifdef __LITTLE_ENDIAN__
28761__ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1) {
28762  int32x4x2_t __ret;
28763  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
28764  return __ret;
28765}
28766#else
28767__ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1) {
28768  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28769  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28770  int32x4x2_t __ret;
28771  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
28772
28773  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28774  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28775  return __ret;
28776}
28777#endif
28778
28779#ifdef __LITTLE_ENDIAN__
28780__ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1) {
28781  int16x8x2_t __ret;
28782  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
28783  return __ret;
28784}
28785#else
28786__ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1) {
28787  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28788  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28789  int16x8x2_t __ret;
28790  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
28791
28792  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28793  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28794  return __ret;
28795}
28796#endif
28797
28798#ifdef __LITTLE_ENDIAN__
28799__ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1) {
28800  uint8x8x2_t __ret;
28801  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
28802  return __ret;
28803}
28804#else
28805__ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1) {
28806  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28807  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28808  uint8x8x2_t __ret;
28809  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
28810
28811  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28812  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28813  return __ret;
28814}
28815#endif
28816
28817#ifdef __LITTLE_ENDIAN__
28818__ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1) {
28819  uint32x2x2_t __ret;
28820  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
28821  return __ret;
28822}
28823#else
28824__ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1) {
28825  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28826  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28827  uint32x2x2_t __ret;
28828  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
28829
28830  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
28831  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
28832  return __ret;
28833}
28834#endif
28835
28836#ifdef __LITTLE_ENDIAN__
28837__ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1) {
28838  uint16x4x2_t __ret;
28839  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
28840  return __ret;
28841}
28842#else
28843__ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1) {
28844  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28845  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28846  uint16x4x2_t __ret;
28847  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
28848
28849  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28850  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28851  return __ret;
28852}
28853#endif
28854
28855#ifdef __LITTLE_ENDIAN__
28856__ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1) {
28857  int8x8x2_t __ret;
28858  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
28859  return __ret;
28860}
28861#else
28862__ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1) {
28863  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28864  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28865  int8x8x2_t __ret;
28866  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
28867
28868  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28869  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28870  return __ret;
28871}
28872#endif
28873
28874#ifdef __LITTLE_ENDIAN__
28875__ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1) {
28876  float32x2x2_t __ret;
28877  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
28878  return __ret;
28879}
28880#else
28881__ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1) {
28882  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28883  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28884  float32x2x2_t __ret;
28885  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
28886
28887  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
28888  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
28889  return __ret;
28890}
28891#endif
28892
28893#ifdef __LITTLE_ENDIAN__
28894__ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1) {
28895  int32x2x2_t __ret;
28896  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
28897  return __ret;
28898}
28899#else
28900__ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1) {
28901  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28902  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28903  int32x2x2_t __ret;
28904  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
28905
28906  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
28907  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
28908  return __ret;
28909}
28910#endif
28911
28912#ifdef __LITTLE_ENDIAN__
28913__ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1) {
28914  int16x4x2_t __ret;
28915  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
28916  return __ret;
28917}
28918#else
28919__ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1) {
28920  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28921  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28922  int16x4x2_t __ret;
28923  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
28924
28925  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28926  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28927  return __ret;
28928}
28929#endif
28930
28931#ifdef __LITTLE_ENDIAN__
28932__ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1) {
28933  uint8x8_t __ret;
28934  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
28935  return __ret;
28936}
28937#else
28938__ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1) {
28939  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28940  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28941  uint8x8_t __ret;
28942  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
28943  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28944  return __ret;
28945}
28946#endif
28947
28948#ifdef __LITTLE_ENDIAN__
28949__ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1) {
28950  uint16x4_t __ret;
28951  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
28952  return __ret;
28953}
28954#else
28955__ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1) {
28956  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28957  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28958  uint16x4_t __ret;
28959  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
28960  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
28961  return __ret;
28962}
28963#endif
28964
28965#ifdef __LITTLE_ENDIAN__
28966__ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1) {
28967  uint8x16_t __ret;
28968  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
28969  return __ret;
28970}
28971#else
28972__ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1) {
28973  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28974  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28975  uint8x16_t __ret;
28976  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
28977  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28978  return __ret;
28979}
28980#endif
28981
28982#ifdef __LITTLE_ENDIAN__
28983__ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1) {
28984  uint16x8_t __ret;
28985  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
28986  return __ret;
28987}
28988#else
28989__ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1) {
28990  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28991  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28992  uint16x8_t __ret;
28993  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
28994  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28995  return __ret;
28996}
28997#endif
28998
28999#ifdef __LITTLE_ENDIAN__
29000__ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29001  uint8x16_t __ret;
29002  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
29003  return __ret;
29004}
29005#else
29006__ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29007  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29008  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29009  uint8x16_t __ret;
29010  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29011  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29012  return __ret;
29013}
29014#endif
29015
29016#ifdef __LITTLE_ENDIAN__
29017__ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29018  uint32x4_t __ret;
29019  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
29020  return __ret;
29021}
29022#else
29023__ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29024  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29025  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29026  uint32x4_t __ret;
29027  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29028  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29029  return __ret;
29030}
29031#endif
29032
29033#ifdef __LITTLE_ENDIAN__
29034__ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29035  uint16x8_t __ret;
29036  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
29037  return __ret;
29038}
29039#else
29040__ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29041  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29042  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29043  uint16x8_t __ret;
29044  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29045  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29046  return __ret;
29047}
29048#endif
29049
29050#ifdef __LITTLE_ENDIAN__
29051__ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1) {
29052  uint8x16_t __ret;
29053  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
29054  return __ret;
29055}
29056#else
29057__ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1) {
29058  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29059  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29060  uint8x16_t __ret;
29061  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29062  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29063  return __ret;
29064}
29065#endif
29066
29067#ifdef __LITTLE_ENDIAN__
29068__ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1) {
29069  uint32x4_t __ret;
29070  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
29071  return __ret;
29072}
29073#else
29074__ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1) {
29075  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29076  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29077  uint32x4_t __ret;
29078  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29079  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29080  return __ret;
29081}
29082#endif
29083
29084#ifdef __LITTLE_ENDIAN__
29085__ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1) {
29086  uint16x8_t __ret;
29087  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
29088  return __ret;
29089}
29090#else
29091__ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1) {
29092  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29093  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29094  uint16x8_t __ret;
29095  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29096  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29097  return __ret;
29098}
29099#endif
29100
29101#ifdef __LITTLE_ENDIAN__
29102__ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1) {
29103  uint8x8_t __ret;
29104  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
29105  return __ret;
29106}
29107#else
29108__ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1) {
29109  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29110  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29111  uint8x8_t __ret;
29112  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29113  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29114  return __ret;
29115}
29116#endif
29117
29118#ifdef __LITTLE_ENDIAN__
29119__ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1) {
29120  uint32x2_t __ret;
29121  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
29122  return __ret;
29123}
29124#else
29125__ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1) {
29126  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29127  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29128  uint32x2_t __ret;
29129  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29130  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29131  return __ret;
29132}
29133#endif
29134
29135#ifdef __LITTLE_ENDIAN__
29136__ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1) {
29137  uint16x4_t __ret;
29138  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
29139  return __ret;
29140}
29141#else
29142__ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1) {
29143  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29144  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29145  uint16x4_t __ret;
29146  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29147  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29148  return __ret;
29149}
29150#endif
29151
29152#ifdef __LITTLE_ENDIAN__
29153__ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1) {
29154  uint8x8_t __ret;
29155  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
29156  return __ret;
29157}
29158#else
29159__ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1) {
29160  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29161  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29162  uint8x8_t __ret;
29163  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29164  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29165  return __ret;
29166}
29167#endif
29168
29169#ifdef __LITTLE_ENDIAN__
29170__ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1) {
29171  uint32x2_t __ret;
29172  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
29173  return __ret;
29174}
29175#else
29176__ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1) {
29177  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29178  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29179  uint32x2_t __ret;
29180  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29181  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29182  return __ret;
29183}
29184#endif
29185
29186#ifdef __LITTLE_ENDIAN__
29187__ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1) {
29188  uint16x4_t __ret;
29189  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
29190  return __ret;
29191}
29192#else
29193__ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1) {
29194  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29195  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29196  uint16x4_t __ret;
29197  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29198  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29199  return __ret;
29200}
29201#endif
29202
29203#ifdef __LITTLE_ENDIAN__
29204__ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1) {
29205  poly8x8x2_t __ret;
29206  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
29207  return __ret;
29208}
29209#else
29210__ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1) {
29211  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29212  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29213  poly8x8x2_t __ret;
29214  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
29215
29216  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29217  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29218  return __ret;
29219}
29220#endif
29221
29222#ifdef __LITTLE_ENDIAN__
29223__ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1) {
29224  poly16x4x2_t __ret;
29225  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
29226  return __ret;
29227}
29228#else
29229__ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1) {
29230  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29231  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29232  poly16x4x2_t __ret;
29233  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
29234
29235  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29236  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29237  return __ret;
29238}
29239#endif
29240
29241#ifdef __LITTLE_ENDIAN__
29242__ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29243  poly8x16x2_t __ret;
29244  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
29245  return __ret;
29246}
29247#else
29248__ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29249  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29250  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29251  poly8x16x2_t __ret;
29252  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
29253
29254  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29255  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29256  return __ret;
29257}
29258#endif
29259
29260#ifdef __LITTLE_ENDIAN__
29261__ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29262  poly16x8x2_t __ret;
29263  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
29264  return __ret;
29265}
29266#else
29267__ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29268  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29269  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29270  poly16x8x2_t __ret;
29271  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
29272
29273  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29274  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29275  return __ret;
29276}
29277#endif
29278
29279#ifdef __LITTLE_ENDIAN__
29280__ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29281  uint8x16x2_t __ret;
29282  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
29283  return __ret;
29284}
29285#else
29286__ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29287  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29288  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29289  uint8x16x2_t __ret;
29290  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29291
29292  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29293  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29294  return __ret;
29295}
29296#endif
29297
29298#ifdef __LITTLE_ENDIAN__
29299__ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29300  uint32x4x2_t __ret;
29301  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
29302  return __ret;
29303}
29304#else
29305__ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29306  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29307  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29308  uint32x4x2_t __ret;
29309  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29310
29311  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29312  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29313  return __ret;
29314}
29315#endif
29316
29317#ifdef __LITTLE_ENDIAN__
29318__ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29319  uint16x8x2_t __ret;
29320  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
29321  return __ret;
29322}
29323#else
29324__ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29325  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29326  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29327  uint16x8x2_t __ret;
29328  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29329
29330  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29331  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29332  return __ret;
29333}
29334#endif
29335
29336#ifdef __LITTLE_ENDIAN__
29337__ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1) {
29338  int8x16x2_t __ret;
29339  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
29340  return __ret;
29341}
29342#else
29343__ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1) {
29344  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29345  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29346  int8x16x2_t __ret;
29347  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
29348
29349  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29350  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29351  return __ret;
29352}
29353#endif
29354
29355#ifdef __LITTLE_ENDIAN__
29356__ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1) {
29357  float32x4x2_t __ret;
29358  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
29359  return __ret;
29360}
29361#else
29362__ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1) {
29363  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29364  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29365  float32x4x2_t __ret;
29366  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
29367
29368  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29369  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29370  return __ret;
29371}
29372#endif
29373
29374#ifdef __LITTLE_ENDIAN__
29375__ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1) {
29376  int32x4x2_t __ret;
29377  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
29378  return __ret;
29379}
29380#else
29381__ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1) {
29382  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29383  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29384  int32x4x2_t __ret;
29385  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
29386
29387  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29388  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29389  return __ret;
29390}
29391#endif
29392
29393#ifdef __LITTLE_ENDIAN__
29394__ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1) {
29395  int16x8x2_t __ret;
29396  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
29397  return __ret;
29398}
29399#else
29400__ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1) {
29401  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29402  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29403  int16x8x2_t __ret;
29404  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
29405
29406  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29407  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29408  return __ret;
29409}
29410#endif
29411
29412#ifdef __LITTLE_ENDIAN__
29413__ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1) {
29414  uint8x8x2_t __ret;
29415  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
29416  return __ret;
29417}
29418#else
29419__ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1) {
29420  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29421  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29422  uint8x8x2_t __ret;
29423  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29424
29425  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29426  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29427  return __ret;
29428}
29429#endif
29430
29431#ifdef __LITTLE_ENDIAN__
29432__ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1) {
29433  uint32x2x2_t __ret;
29434  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
29435  return __ret;
29436}
29437#else
29438__ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1) {
29439  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29440  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29441  uint32x2x2_t __ret;
29442  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29443
29444  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29445  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29446  return __ret;
29447}
29448#endif
29449
29450#ifdef __LITTLE_ENDIAN__
29451__ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1) {
29452  uint16x4x2_t __ret;
29453  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
29454  return __ret;
29455}
29456#else
29457__ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1) {
29458  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29459  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29460  uint16x4x2_t __ret;
29461  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29462
29463  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29464  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29465  return __ret;
29466}
29467#endif
29468
29469#ifdef __LITTLE_ENDIAN__
29470__ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1) {
29471  int8x8x2_t __ret;
29472  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
29473  return __ret;
29474}
29475#else
29476__ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1) {
29477  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29478  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29479  int8x8x2_t __ret;
29480  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
29481
29482  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29483  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29484  return __ret;
29485}
29486#endif
29487
29488#ifdef __LITTLE_ENDIAN__
29489__ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1) {
29490  float32x2x2_t __ret;
29491  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
29492  return __ret;
29493}
29494#else
29495__ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1) {
29496  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29497  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29498  float32x2x2_t __ret;
29499  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
29500
29501  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29502  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29503  return __ret;
29504}
29505#endif
29506
29507#ifdef __LITTLE_ENDIAN__
29508__ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1) {
29509  int32x2x2_t __ret;
29510  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
29511  return __ret;
29512}
29513#else
29514__ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1) {
29515  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29516  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29517  int32x2x2_t __ret;
29518  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
29519
29520  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29521  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29522  return __ret;
29523}
29524#endif
29525
29526#ifdef __LITTLE_ENDIAN__
29527__ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1) {
29528  int16x4x2_t __ret;
29529  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
29530  return __ret;
29531}
29532#else
29533__ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1) {
29534  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29535  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29536  int16x4x2_t __ret;
29537  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
29538
29539  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29540  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29541  return __ret;
29542}
29543#endif
29544
29545#ifdef __LITTLE_ENDIAN__
29546__ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1) {
29547  poly8x8x2_t __ret;
29548  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
29549  return __ret;
29550}
29551#else
29552__ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1) {
29553  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29554  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29555  poly8x8x2_t __ret;
29556  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
29557
29558  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29559  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29560  return __ret;
29561}
29562#endif
29563
29564#ifdef __LITTLE_ENDIAN__
29565__ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1) {
29566  poly16x4x2_t __ret;
29567  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
29568  return __ret;
29569}
29570#else
29571__ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1) {
29572  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29573  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29574  poly16x4x2_t __ret;
29575  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
29576
29577  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29578  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29579  return __ret;
29580}
29581#endif
29582
29583#ifdef __LITTLE_ENDIAN__
29584__ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29585  poly8x16x2_t __ret;
29586  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
29587  return __ret;
29588}
29589#else
29590__ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29591  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29592  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29593  poly8x16x2_t __ret;
29594  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
29595
29596  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29597  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29598  return __ret;
29599}
29600#endif
29601
29602#ifdef __LITTLE_ENDIAN__
29603__ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29604  poly16x8x2_t __ret;
29605  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
29606  return __ret;
29607}
29608#else
29609__ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29610  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29611  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29612  poly16x8x2_t __ret;
29613  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
29614
29615  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29616  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29617  return __ret;
29618}
29619#endif
29620
29621#ifdef __LITTLE_ENDIAN__
29622__ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29623  uint8x16x2_t __ret;
29624  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
29625  return __ret;
29626}
29627#else
29628__ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29629  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29630  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29631  uint8x16x2_t __ret;
29632  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29633
29634  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29635  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29636  return __ret;
29637}
29638#endif
29639
29640#ifdef __LITTLE_ENDIAN__
29641__ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29642  uint32x4x2_t __ret;
29643  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
29644  return __ret;
29645}
29646#else
29647__ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29648  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29649  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29650  uint32x4x2_t __ret;
29651  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29652
29653  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29654  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29655  return __ret;
29656}
29657#endif
29658
29659#ifdef __LITTLE_ENDIAN__
29660__ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29661  uint16x8x2_t __ret;
29662  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
29663  return __ret;
29664}
29665#else
29666__ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29667  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29668  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29669  uint16x8x2_t __ret;
29670  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29671
29672  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29673  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29674  return __ret;
29675}
29676#endif
29677
29678#ifdef __LITTLE_ENDIAN__
29679__ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1) {
29680  int8x16x2_t __ret;
29681  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
29682  return __ret;
29683}
29684#else
29685__ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1) {
29686  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29687  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29688  int8x16x2_t __ret;
29689  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
29690
29691  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29692  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29693  return __ret;
29694}
29695#endif
29696
29697#ifdef __LITTLE_ENDIAN__
29698__ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1) {
29699  float32x4x2_t __ret;
29700  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
29701  return __ret;
29702}
29703#else
29704__ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1) {
29705  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29706  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29707  float32x4x2_t __ret;
29708  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
29709
29710  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29711  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29712  return __ret;
29713}
29714#endif
29715
29716#ifdef __LITTLE_ENDIAN__
29717__ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1) {
29718  int32x4x2_t __ret;
29719  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
29720  return __ret;
29721}
29722#else
29723__ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1) {
29724  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29725  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29726  int32x4x2_t __ret;
29727  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
29728
29729  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29730  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29731  return __ret;
29732}
29733#endif
29734
29735#ifdef __LITTLE_ENDIAN__
29736__ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1) {
29737  int16x8x2_t __ret;
29738  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
29739  return __ret;
29740}
29741#else
29742__ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1) {
29743  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29744  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29745  int16x8x2_t __ret;
29746  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
29747
29748  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29749  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29750  return __ret;
29751}
29752#endif
29753
29754#ifdef __LITTLE_ENDIAN__
29755__ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1) {
29756  uint8x8x2_t __ret;
29757  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
29758  return __ret;
29759}
29760#else
29761__ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1) {
29762  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29763  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29764  uint8x8x2_t __ret;
29765  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29766
29767  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29768  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29769  return __ret;
29770}
29771#endif
29772
29773#ifdef __LITTLE_ENDIAN__
29774__ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1) {
29775  uint32x2x2_t __ret;
29776  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
29777  return __ret;
29778}
29779#else
29780__ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1) {
29781  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29782  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29783  uint32x2x2_t __ret;
29784  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29785
29786  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29787  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29788  return __ret;
29789}
29790#endif
29791
29792#ifdef __LITTLE_ENDIAN__
29793__ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1) {
29794  uint16x4x2_t __ret;
29795  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
29796  return __ret;
29797}
29798#else
29799__ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1) {
29800  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29801  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29802  uint16x4x2_t __ret;
29803  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29804
29805  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29806  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29807  return __ret;
29808}
29809#endif
29810
29811#ifdef __LITTLE_ENDIAN__
29812__ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1) {
29813  int8x8x2_t __ret;
29814  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
29815  return __ret;
29816}
29817#else
29818__ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1) {
29819  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29820  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29821  int8x8x2_t __ret;
29822  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
29823
29824  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29825  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29826  return __ret;
29827}
29828#endif
29829
29830#ifdef __LITTLE_ENDIAN__
29831__ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1) {
29832  float32x2x2_t __ret;
29833  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
29834  return __ret;
29835}
29836#else
29837__ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1) {
29838  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29839  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29840  float32x2x2_t __ret;
29841  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
29842
29843  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29844  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29845  return __ret;
29846}
29847#endif
29848
29849#ifdef __LITTLE_ENDIAN__
29850__ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1) {
29851  int32x2x2_t __ret;
29852  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
29853  return __ret;
29854}
29855#else
29856__ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1) {
29857  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29858  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29859  int32x2x2_t __ret;
29860  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
29861
29862  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29863  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29864  return __ret;
29865}
29866#endif
29867
29868#ifdef __LITTLE_ENDIAN__
29869__ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1) {
29870  int16x4x2_t __ret;
29871  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
29872  return __ret;
29873}
29874#else
29875__ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1) {
29876  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29877  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29878  int16x4x2_t __ret;
29879  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
29880
29881  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29882  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29883  return __ret;
29884}
29885#endif
29886
29887#if !defined(__aarch64__)
29888#ifdef __LITTLE_ENDIAN__
29889__ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
29890  poly8x8_t __ret;
29891  __ret = (poly8x8_t)(__p0);
29892  return __ret;
29893}
29894#else
29895__ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
29896  poly8x8_t __ret;
29897  __ret = (poly8x8_t)(__p0);
29898  return __ret;
29899}
29900#endif
29901
29902#ifdef __LITTLE_ENDIAN__
29903__ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
29904  poly8x8_t __ret;
29905  __ret = (poly8x8_t)(__p0);
29906  return __ret;
29907}
29908#else
29909__ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
29910  poly8x8_t __ret;
29911  __ret = (poly8x8_t)(__p0);
29912  return __ret;
29913}
29914#endif
29915
29916#ifdef __LITTLE_ENDIAN__
29917__ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
29918  poly8x8_t __ret;
29919  __ret = (poly8x8_t)(__p0);
29920  return __ret;
29921}
29922#else
29923__ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
29924  poly8x8_t __ret;
29925  __ret = (poly8x8_t)(__p0);
29926  return __ret;
29927}
29928#endif
29929
29930#ifdef __LITTLE_ENDIAN__
29931__ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
29932  poly8x8_t __ret;
29933  __ret = (poly8x8_t)(__p0);
29934  return __ret;
29935}
29936#else
29937__ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
29938  poly8x8_t __ret;
29939  __ret = (poly8x8_t)(__p0);
29940  return __ret;
29941}
29942#endif
29943
29944#ifdef __LITTLE_ENDIAN__
29945__ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
29946  poly8x8_t __ret;
29947  __ret = (poly8x8_t)(__p0);
29948  return __ret;
29949}
29950#else
29951__ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
29952  poly8x8_t __ret;
29953  __ret = (poly8x8_t)(__p0);
29954  return __ret;
29955}
29956#endif
29957
29958#ifdef __LITTLE_ENDIAN__
29959__ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
29960  poly8x8_t __ret;
29961  __ret = (poly8x8_t)(__p0);
29962  return __ret;
29963}
29964#else
29965__ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
29966  poly8x8_t __ret;
29967  __ret = (poly8x8_t)(__p0);
29968  return __ret;
29969}
29970#endif
29971
29972#ifdef __LITTLE_ENDIAN__
29973__ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
29974  poly8x8_t __ret;
29975  __ret = (poly8x8_t)(__p0);
29976  return __ret;
29977}
29978#else
29979__ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
29980  poly8x8_t __ret;
29981  __ret = (poly8x8_t)(__p0);
29982  return __ret;
29983}
29984#endif
29985
29986#ifdef __LITTLE_ENDIAN__
29987__ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
29988  poly8x8_t __ret;
29989  __ret = (poly8x8_t)(__p0);
29990  return __ret;
29991}
29992#else
29993__ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
29994  poly8x8_t __ret;
29995  __ret = (poly8x8_t)(__p0);
29996  return __ret;
29997}
29998#endif
29999
30000#ifdef __LITTLE_ENDIAN__
30001__ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
30002  poly8x8_t __ret;
30003  __ret = (poly8x8_t)(__p0);
30004  return __ret;
30005}
30006#else
30007__ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
30008  poly8x8_t __ret;
30009  __ret = (poly8x8_t)(__p0);
30010  return __ret;
30011}
30012#endif
30013
30014#ifdef __LITTLE_ENDIAN__
30015__ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
30016  poly8x8_t __ret;
30017  __ret = (poly8x8_t)(__p0);
30018  return __ret;
30019}
30020#else
30021__ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
30022  poly8x8_t __ret;
30023  __ret = (poly8x8_t)(__p0);
30024  return __ret;
30025}
30026#endif
30027
30028#ifdef __LITTLE_ENDIAN__
30029__ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
30030  poly8x8_t __ret;
30031  __ret = (poly8x8_t)(__p0);
30032  return __ret;
30033}
30034#else
30035__ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
30036  poly8x8_t __ret;
30037  __ret = (poly8x8_t)(__p0);
30038  return __ret;
30039}
30040#endif
30041
30042#ifdef __LITTLE_ENDIAN__
30043__ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
30044  poly16x4_t __ret;
30045  __ret = (poly16x4_t)(__p0);
30046  return __ret;
30047}
30048#else
30049__ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
30050  poly16x4_t __ret;
30051  __ret = (poly16x4_t)(__p0);
30052  return __ret;
30053}
30054#endif
30055
30056#ifdef __LITTLE_ENDIAN__
30057__ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
30058  poly16x4_t __ret;
30059  __ret = (poly16x4_t)(__p0);
30060  return __ret;
30061}
30062#else
30063__ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
30064  poly16x4_t __ret;
30065  __ret = (poly16x4_t)(__p0);
30066  return __ret;
30067}
30068#endif
30069
30070#ifdef __LITTLE_ENDIAN__
30071__ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
30072  poly16x4_t __ret;
30073  __ret = (poly16x4_t)(__p0);
30074  return __ret;
30075}
30076#else
30077__ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
30078  poly16x4_t __ret;
30079  __ret = (poly16x4_t)(__p0);
30080  return __ret;
30081}
30082#endif
30083
30084#ifdef __LITTLE_ENDIAN__
30085__ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
30086  poly16x4_t __ret;
30087  __ret = (poly16x4_t)(__p0);
30088  return __ret;
30089}
30090#else
30091__ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
30092  poly16x4_t __ret;
30093  __ret = (poly16x4_t)(__p0);
30094  return __ret;
30095}
30096#endif
30097
30098#ifdef __LITTLE_ENDIAN__
30099__ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
30100  poly16x4_t __ret;
30101  __ret = (poly16x4_t)(__p0);
30102  return __ret;
30103}
30104#else
30105__ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
30106  poly16x4_t __ret;
30107  __ret = (poly16x4_t)(__p0);
30108  return __ret;
30109}
30110#endif
30111
30112#ifdef __LITTLE_ENDIAN__
30113__ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
30114  poly16x4_t __ret;
30115  __ret = (poly16x4_t)(__p0);
30116  return __ret;
30117}
30118#else
30119__ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
30120  poly16x4_t __ret;
30121  __ret = (poly16x4_t)(__p0);
30122  return __ret;
30123}
30124#endif
30125
30126#ifdef __LITTLE_ENDIAN__
30127__ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
30128  poly16x4_t __ret;
30129  __ret = (poly16x4_t)(__p0);
30130  return __ret;
30131}
30132#else
30133__ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
30134  poly16x4_t __ret;
30135  __ret = (poly16x4_t)(__p0);
30136  return __ret;
30137}
30138#endif
30139
30140#ifdef __LITTLE_ENDIAN__
30141__ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
30142  poly16x4_t __ret;
30143  __ret = (poly16x4_t)(__p0);
30144  return __ret;
30145}
30146#else
30147__ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
30148  poly16x4_t __ret;
30149  __ret = (poly16x4_t)(__p0);
30150  return __ret;
30151}
30152#endif
30153
30154#ifdef __LITTLE_ENDIAN__
30155__ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
30156  poly16x4_t __ret;
30157  __ret = (poly16x4_t)(__p0);
30158  return __ret;
30159}
30160#else
30161__ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
30162  poly16x4_t __ret;
30163  __ret = (poly16x4_t)(__p0);
30164  return __ret;
30165}
30166#endif
30167
30168#ifdef __LITTLE_ENDIAN__
30169__ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
30170  poly16x4_t __ret;
30171  __ret = (poly16x4_t)(__p0);
30172  return __ret;
30173}
30174#else
30175__ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
30176  poly16x4_t __ret;
30177  __ret = (poly16x4_t)(__p0);
30178  return __ret;
30179}
30180#endif
30181
30182#ifdef __LITTLE_ENDIAN__
30183__ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
30184  poly16x4_t __ret;
30185  __ret = (poly16x4_t)(__p0);
30186  return __ret;
30187}
30188#else
30189__ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
30190  poly16x4_t __ret;
30191  __ret = (poly16x4_t)(__p0);
30192  return __ret;
30193}
30194#endif
30195
30196#ifdef __LITTLE_ENDIAN__
30197__ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
30198  poly8x16_t __ret;
30199  __ret = (poly8x16_t)(__p0);
30200  return __ret;
30201}
30202#else
30203__ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
30204  poly8x16_t __ret;
30205  __ret = (poly8x16_t)(__p0);
30206  return __ret;
30207}
30208#endif
30209
30210#ifdef __LITTLE_ENDIAN__
30211__ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
30212  poly8x16_t __ret;
30213  __ret = (poly8x16_t)(__p0);
30214  return __ret;
30215}
30216#else
30217__ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
30218  poly8x16_t __ret;
30219  __ret = (poly8x16_t)(__p0);
30220  return __ret;
30221}
30222#endif
30223
30224#ifdef __LITTLE_ENDIAN__
30225__ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
30226  poly8x16_t __ret;
30227  __ret = (poly8x16_t)(__p0);
30228  return __ret;
30229}
30230#else
30231__ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
30232  poly8x16_t __ret;
30233  __ret = (poly8x16_t)(__p0);
30234  return __ret;
30235}
30236#endif
30237
30238#ifdef __LITTLE_ENDIAN__
30239__ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
30240  poly8x16_t __ret;
30241  __ret = (poly8x16_t)(__p0);
30242  return __ret;
30243}
30244#else
30245__ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
30246  poly8x16_t __ret;
30247  __ret = (poly8x16_t)(__p0);
30248  return __ret;
30249}
30250#endif
30251
30252#ifdef __LITTLE_ENDIAN__
30253__ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
30254  poly8x16_t __ret;
30255  __ret = (poly8x16_t)(__p0);
30256  return __ret;
30257}
30258#else
30259__ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
30260  poly8x16_t __ret;
30261  __ret = (poly8x16_t)(__p0);
30262  return __ret;
30263}
30264#endif
30265
30266#ifdef __LITTLE_ENDIAN__
30267__ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
30268  poly8x16_t __ret;
30269  __ret = (poly8x16_t)(__p0);
30270  return __ret;
30271}
30272#else
30273__ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
30274  poly8x16_t __ret;
30275  __ret = (poly8x16_t)(__p0);
30276  return __ret;
30277}
30278#endif
30279
30280#ifdef __LITTLE_ENDIAN__
30281__ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
30282  poly8x16_t __ret;
30283  __ret = (poly8x16_t)(__p0);
30284  return __ret;
30285}
30286#else
30287__ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
30288  poly8x16_t __ret;
30289  __ret = (poly8x16_t)(__p0);
30290  return __ret;
30291}
30292#endif
30293
30294#ifdef __LITTLE_ENDIAN__
30295__ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
30296  poly8x16_t __ret;
30297  __ret = (poly8x16_t)(__p0);
30298  return __ret;
30299}
30300#else
30301__ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
30302  poly8x16_t __ret;
30303  __ret = (poly8x16_t)(__p0);
30304  return __ret;
30305}
30306#endif
30307
30308#ifdef __LITTLE_ENDIAN__
30309__ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
30310  poly8x16_t __ret;
30311  __ret = (poly8x16_t)(__p0);
30312  return __ret;
30313}
30314#else
30315__ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
30316  poly8x16_t __ret;
30317  __ret = (poly8x16_t)(__p0);
30318  return __ret;
30319}
30320#endif
30321
30322#ifdef __LITTLE_ENDIAN__
30323__ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
30324  poly8x16_t __ret;
30325  __ret = (poly8x16_t)(__p0);
30326  return __ret;
30327}
30328#else
30329__ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
30330  poly8x16_t __ret;
30331  __ret = (poly8x16_t)(__p0);
30332  return __ret;
30333}
30334#endif
30335
30336#ifdef __LITTLE_ENDIAN__
30337__ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
30338  poly8x16_t __ret;
30339  __ret = (poly8x16_t)(__p0);
30340  return __ret;
30341}
30342#else
30343__ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
30344  poly8x16_t __ret;
30345  __ret = (poly8x16_t)(__p0);
30346  return __ret;
30347}
30348#endif
30349
30350#ifdef __LITTLE_ENDIAN__
30351__ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
30352  poly16x8_t __ret;
30353  __ret = (poly16x8_t)(__p0);
30354  return __ret;
30355}
30356#else
30357__ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
30358  poly16x8_t __ret;
30359  __ret = (poly16x8_t)(__p0);
30360  return __ret;
30361}
30362#endif
30363
30364#ifdef __LITTLE_ENDIAN__
30365__ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
30366  poly16x8_t __ret;
30367  __ret = (poly16x8_t)(__p0);
30368  return __ret;
30369}
30370#else
30371__ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
30372  poly16x8_t __ret;
30373  __ret = (poly16x8_t)(__p0);
30374  return __ret;
30375}
30376#endif
30377
30378#ifdef __LITTLE_ENDIAN__
30379__ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
30380  poly16x8_t __ret;
30381  __ret = (poly16x8_t)(__p0);
30382  return __ret;
30383}
30384#else
30385__ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
30386  poly16x8_t __ret;
30387  __ret = (poly16x8_t)(__p0);
30388  return __ret;
30389}
30390#endif
30391
30392#ifdef __LITTLE_ENDIAN__
30393__ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
30394  poly16x8_t __ret;
30395  __ret = (poly16x8_t)(__p0);
30396  return __ret;
30397}
30398#else
30399__ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
30400  poly16x8_t __ret;
30401  __ret = (poly16x8_t)(__p0);
30402  return __ret;
30403}
30404#endif
30405
30406#ifdef __LITTLE_ENDIAN__
30407__ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
30408  poly16x8_t __ret;
30409  __ret = (poly16x8_t)(__p0);
30410  return __ret;
30411}
30412#else
30413__ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
30414  poly16x8_t __ret;
30415  __ret = (poly16x8_t)(__p0);
30416  return __ret;
30417}
30418#endif
30419
30420#ifdef __LITTLE_ENDIAN__
30421__ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
30422  poly16x8_t __ret;
30423  __ret = (poly16x8_t)(__p0);
30424  return __ret;
30425}
30426#else
30427__ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
30428  poly16x8_t __ret;
30429  __ret = (poly16x8_t)(__p0);
30430  return __ret;
30431}
30432#endif
30433
30434#ifdef __LITTLE_ENDIAN__
30435__ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
30436  poly16x8_t __ret;
30437  __ret = (poly16x8_t)(__p0);
30438  return __ret;
30439}
30440#else
30441__ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
30442  poly16x8_t __ret;
30443  __ret = (poly16x8_t)(__p0);
30444  return __ret;
30445}
30446#endif
30447
30448#ifdef __LITTLE_ENDIAN__
30449__ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
30450  poly16x8_t __ret;
30451  __ret = (poly16x8_t)(__p0);
30452  return __ret;
30453}
30454#else
30455__ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
30456  poly16x8_t __ret;
30457  __ret = (poly16x8_t)(__p0);
30458  return __ret;
30459}
30460#endif
30461
30462#ifdef __LITTLE_ENDIAN__
30463__ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
30464  poly16x8_t __ret;
30465  __ret = (poly16x8_t)(__p0);
30466  return __ret;
30467}
30468#else
30469__ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
30470  poly16x8_t __ret;
30471  __ret = (poly16x8_t)(__p0);
30472  return __ret;
30473}
30474#endif
30475
30476#ifdef __LITTLE_ENDIAN__
30477__ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
30478  poly16x8_t __ret;
30479  __ret = (poly16x8_t)(__p0);
30480  return __ret;
30481}
30482#else
30483__ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
30484  poly16x8_t __ret;
30485  __ret = (poly16x8_t)(__p0);
30486  return __ret;
30487}
30488#endif
30489
30490#ifdef __LITTLE_ENDIAN__
30491__ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
30492  poly16x8_t __ret;
30493  __ret = (poly16x8_t)(__p0);
30494  return __ret;
30495}
30496#else
30497__ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
30498  poly16x8_t __ret;
30499  __ret = (poly16x8_t)(__p0);
30500  return __ret;
30501}
30502#endif
30503
30504#ifdef __LITTLE_ENDIAN__
30505__ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
30506  uint8x16_t __ret;
30507  __ret = (uint8x16_t)(__p0);
30508  return __ret;
30509}
30510#else
30511__ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
30512  uint8x16_t __ret;
30513  __ret = (uint8x16_t)(__p0);
30514  return __ret;
30515}
30516#endif
30517
30518#ifdef __LITTLE_ENDIAN__
30519__ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
30520  uint8x16_t __ret;
30521  __ret = (uint8x16_t)(__p0);
30522  return __ret;
30523}
30524#else
30525__ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
30526  uint8x16_t __ret;
30527  __ret = (uint8x16_t)(__p0);
30528  return __ret;
30529}
30530#endif
30531
30532#ifdef __LITTLE_ENDIAN__
30533__ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
30534  uint8x16_t __ret;
30535  __ret = (uint8x16_t)(__p0);
30536  return __ret;
30537}
30538#else
30539__ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
30540  uint8x16_t __ret;
30541  __ret = (uint8x16_t)(__p0);
30542  return __ret;
30543}
30544#endif
30545
30546#ifdef __LITTLE_ENDIAN__
30547__ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
30548  uint8x16_t __ret;
30549  __ret = (uint8x16_t)(__p0);
30550  return __ret;
30551}
30552#else
30553__ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
30554  uint8x16_t __ret;
30555  __ret = (uint8x16_t)(__p0);
30556  return __ret;
30557}
30558#endif
30559
30560#ifdef __LITTLE_ENDIAN__
30561__ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
30562  uint8x16_t __ret;
30563  __ret = (uint8x16_t)(__p0);
30564  return __ret;
30565}
30566#else
30567__ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
30568  uint8x16_t __ret;
30569  __ret = (uint8x16_t)(__p0);
30570  return __ret;
30571}
30572#endif
30573
30574#ifdef __LITTLE_ENDIAN__
30575__ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
30576  uint8x16_t __ret;
30577  __ret = (uint8x16_t)(__p0);
30578  return __ret;
30579}
30580#else
30581__ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
30582  uint8x16_t __ret;
30583  __ret = (uint8x16_t)(__p0);
30584  return __ret;
30585}
30586#endif
30587
30588#ifdef __LITTLE_ENDIAN__
30589__ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
30590  uint8x16_t __ret;
30591  __ret = (uint8x16_t)(__p0);
30592  return __ret;
30593}
30594#else
30595__ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
30596  uint8x16_t __ret;
30597  __ret = (uint8x16_t)(__p0);
30598  return __ret;
30599}
30600#endif
30601
30602#ifdef __LITTLE_ENDIAN__
30603__ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
30604  uint8x16_t __ret;
30605  __ret = (uint8x16_t)(__p0);
30606  return __ret;
30607}
30608#else
30609__ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
30610  uint8x16_t __ret;
30611  __ret = (uint8x16_t)(__p0);
30612  return __ret;
30613}
30614#endif
30615
30616#ifdef __LITTLE_ENDIAN__
30617__ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
30618  uint8x16_t __ret;
30619  __ret = (uint8x16_t)(__p0);
30620  return __ret;
30621}
30622#else
30623__ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
30624  uint8x16_t __ret;
30625  __ret = (uint8x16_t)(__p0);
30626  return __ret;
30627}
30628#endif
30629
30630#ifdef __LITTLE_ENDIAN__
30631__ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
30632  uint8x16_t __ret;
30633  __ret = (uint8x16_t)(__p0);
30634  return __ret;
30635}
30636#else
30637__ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
30638  uint8x16_t __ret;
30639  __ret = (uint8x16_t)(__p0);
30640  return __ret;
30641}
30642#endif
30643
30644#ifdef __LITTLE_ENDIAN__
30645__ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
30646  uint8x16_t __ret;
30647  __ret = (uint8x16_t)(__p0);
30648  return __ret;
30649}
30650#else
30651__ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
30652  uint8x16_t __ret;
30653  __ret = (uint8x16_t)(__p0);
30654  return __ret;
30655}
30656#endif
30657
30658#ifdef __LITTLE_ENDIAN__
30659__ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
30660  uint32x4_t __ret;
30661  __ret = (uint32x4_t)(__p0);
30662  return __ret;
30663}
30664#else
30665__ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
30666  uint32x4_t __ret;
30667  __ret = (uint32x4_t)(__p0);
30668  return __ret;
30669}
30670#endif
30671
30672#ifdef __LITTLE_ENDIAN__
30673__ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
30674  uint32x4_t __ret;
30675  __ret = (uint32x4_t)(__p0);
30676  return __ret;
30677}
30678#else
30679__ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
30680  uint32x4_t __ret;
30681  __ret = (uint32x4_t)(__p0);
30682  return __ret;
30683}
30684#endif
30685
30686#ifdef __LITTLE_ENDIAN__
30687__ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
30688  uint32x4_t __ret;
30689  __ret = (uint32x4_t)(__p0);
30690  return __ret;
30691}
30692#else
30693__ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
30694  uint32x4_t __ret;
30695  __ret = (uint32x4_t)(__p0);
30696  return __ret;
30697}
30698#endif
30699
30700#ifdef __LITTLE_ENDIAN__
30701__ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
30702  uint32x4_t __ret;
30703  __ret = (uint32x4_t)(__p0);
30704  return __ret;
30705}
30706#else
30707__ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
30708  uint32x4_t __ret;
30709  __ret = (uint32x4_t)(__p0);
30710  return __ret;
30711}
30712#endif
30713
30714#ifdef __LITTLE_ENDIAN__
30715__ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
30716  uint32x4_t __ret;
30717  __ret = (uint32x4_t)(__p0);
30718  return __ret;
30719}
30720#else
30721__ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
30722  uint32x4_t __ret;
30723  __ret = (uint32x4_t)(__p0);
30724  return __ret;
30725}
30726#endif
30727
30728#ifdef __LITTLE_ENDIAN__
30729__ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
30730  uint32x4_t __ret;
30731  __ret = (uint32x4_t)(__p0);
30732  return __ret;
30733}
30734#else
30735__ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
30736  uint32x4_t __ret;
30737  __ret = (uint32x4_t)(__p0);
30738  return __ret;
30739}
30740#endif
30741
30742#ifdef __LITTLE_ENDIAN__
30743__ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
30744  uint32x4_t __ret;
30745  __ret = (uint32x4_t)(__p0);
30746  return __ret;
30747}
30748#else
30749__ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
30750  uint32x4_t __ret;
30751  __ret = (uint32x4_t)(__p0);
30752  return __ret;
30753}
30754#endif
30755
30756#ifdef __LITTLE_ENDIAN__
30757__ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
30758  uint32x4_t __ret;
30759  __ret = (uint32x4_t)(__p0);
30760  return __ret;
30761}
30762#else
30763__ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
30764  uint32x4_t __ret;
30765  __ret = (uint32x4_t)(__p0);
30766  return __ret;
30767}
30768#endif
30769
30770#ifdef __LITTLE_ENDIAN__
30771__ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
30772  uint32x4_t __ret;
30773  __ret = (uint32x4_t)(__p0);
30774  return __ret;
30775}
30776#else
30777__ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
30778  uint32x4_t __ret;
30779  __ret = (uint32x4_t)(__p0);
30780  return __ret;
30781}
30782#endif
30783
30784#ifdef __LITTLE_ENDIAN__
30785__ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
30786  uint32x4_t __ret;
30787  __ret = (uint32x4_t)(__p0);
30788  return __ret;
30789}
30790#else
30791__ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
30792  uint32x4_t __ret;
30793  __ret = (uint32x4_t)(__p0);
30794  return __ret;
30795}
30796#endif
30797
30798#ifdef __LITTLE_ENDIAN__
30799__ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
30800  uint32x4_t __ret;
30801  __ret = (uint32x4_t)(__p0);
30802  return __ret;
30803}
30804#else
30805__ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
30806  uint32x4_t __ret;
30807  __ret = (uint32x4_t)(__p0);
30808  return __ret;
30809}
30810#endif
30811
30812#ifdef __LITTLE_ENDIAN__
30813__ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
30814  uint64x2_t __ret;
30815  __ret = (uint64x2_t)(__p0);
30816  return __ret;
30817}
30818#else
30819__ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
30820  uint64x2_t __ret;
30821  __ret = (uint64x2_t)(__p0);
30822  return __ret;
30823}
30824#endif
30825
30826#ifdef __LITTLE_ENDIAN__
30827__ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
30828  uint64x2_t __ret;
30829  __ret = (uint64x2_t)(__p0);
30830  return __ret;
30831}
30832#else
30833__ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
30834  uint64x2_t __ret;
30835  __ret = (uint64x2_t)(__p0);
30836  return __ret;
30837}
30838#endif
30839
30840#ifdef __LITTLE_ENDIAN__
30841__ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
30842  uint64x2_t __ret;
30843  __ret = (uint64x2_t)(__p0);
30844  return __ret;
30845}
30846#else
30847__ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
30848  uint64x2_t __ret;
30849  __ret = (uint64x2_t)(__p0);
30850  return __ret;
30851}
30852#endif
30853
30854#ifdef __LITTLE_ENDIAN__
30855__ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
30856  uint64x2_t __ret;
30857  __ret = (uint64x2_t)(__p0);
30858  return __ret;
30859}
30860#else
30861__ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
30862  uint64x2_t __ret;
30863  __ret = (uint64x2_t)(__p0);
30864  return __ret;
30865}
30866#endif
30867
30868#ifdef __LITTLE_ENDIAN__
30869__ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
30870  uint64x2_t __ret;
30871  __ret = (uint64x2_t)(__p0);
30872  return __ret;
30873}
30874#else
30875__ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
30876  uint64x2_t __ret;
30877  __ret = (uint64x2_t)(__p0);
30878  return __ret;
30879}
30880#endif
30881
30882#ifdef __LITTLE_ENDIAN__
30883__ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
30884  uint64x2_t __ret;
30885  __ret = (uint64x2_t)(__p0);
30886  return __ret;
30887}
30888#else
30889__ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
30890  uint64x2_t __ret;
30891  __ret = (uint64x2_t)(__p0);
30892  return __ret;
30893}
30894#endif
30895
30896#ifdef __LITTLE_ENDIAN__
30897__ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
30898  uint64x2_t __ret;
30899  __ret = (uint64x2_t)(__p0);
30900  return __ret;
30901}
30902#else
30903__ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
30904  uint64x2_t __ret;
30905  __ret = (uint64x2_t)(__p0);
30906  return __ret;
30907}
30908#endif
30909
30910#ifdef __LITTLE_ENDIAN__
30911__ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
30912  uint64x2_t __ret;
30913  __ret = (uint64x2_t)(__p0);
30914  return __ret;
30915}
30916#else
30917__ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
30918  uint64x2_t __ret;
30919  __ret = (uint64x2_t)(__p0);
30920  return __ret;
30921}
30922#endif
30923
30924#ifdef __LITTLE_ENDIAN__
30925__ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
30926  uint64x2_t __ret;
30927  __ret = (uint64x2_t)(__p0);
30928  return __ret;
30929}
30930#else
30931__ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
30932  uint64x2_t __ret;
30933  __ret = (uint64x2_t)(__p0);
30934  return __ret;
30935}
30936#endif
30937
30938#ifdef __LITTLE_ENDIAN__
30939__ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
30940  uint64x2_t __ret;
30941  __ret = (uint64x2_t)(__p0);
30942  return __ret;
30943}
30944#else
30945__ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
30946  uint64x2_t __ret;
30947  __ret = (uint64x2_t)(__p0);
30948  return __ret;
30949}
30950#endif
30951
30952#ifdef __LITTLE_ENDIAN__
30953__ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
30954  uint64x2_t __ret;
30955  __ret = (uint64x2_t)(__p0);
30956  return __ret;
30957}
30958#else
30959__ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
30960  uint64x2_t __ret;
30961  __ret = (uint64x2_t)(__p0);
30962  return __ret;
30963}
30964#endif
30965
30966#ifdef __LITTLE_ENDIAN__
30967__ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
30968  uint16x8_t __ret;
30969  __ret = (uint16x8_t)(__p0);
30970  return __ret;
30971}
30972#else
30973__ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
30974  uint16x8_t __ret;
30975  __ret = (uint16x8_t)(__p0);
30976  return __ret;
30977}
30978#endif
30979
30980#ifdef __LITTLE_ENDIAN__
30981__ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
30982  uint16x8_t __ret;
30983  __ret = (uint16x8_t)(__p0);
30984  return __ret;
30985}
30986#else
30987__ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
30988  uint16x8_t __ret;
30989  __ret = (uint16x8_t)(__p0);
30990  return __ret;
30991}
30992#endif
30993
30994#ifdef __LITTLE_ENDIAN__
30995__ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
30996  uint16x8_t __ret;
30997  __ret = (uint16x8_t)(__p0);
30998  return __ret;
30999}
31000#else
31001__ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
31002  uint16x8_t __ret;
31003  __ret = (uint16x8_t)(__p0);
31004  return __ret;
31005}
31006#endif
31007
31008#ifdef __LITTLE_ENDIAN__
31009__ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
31010  uint16x8_t __ret;
31011  __ret = (uint16x8_t)(__p0);
31012  return __ret;
31013}
31014#else
31015__ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
31016  uint16x8_t __ret;
31017  __ret = (uint16x8_t)(__p0);
31018  return __ret;
31019}
31020#endif
31021
31022#ifdef __LITTLE_ENDIAN__
31023__ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
31024  uint16x8_t __ret;
31025  __ret = (uint16x8_t)(__p0);
31026  return __ret;
31027}
31028#else
31029__ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
31030  uint16x8_t __ret;
31031  __ret = (uint16x8_t)(__p0);
31032  return __ret;
31033}
31034#endif
31035
31036#ifdef __LITTLE_ENDIAN__
31037__ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
31038  uint16x8_t __ret;
31039  __ret = (uint16x8_t)(__p0);
31040  return __ret;
31041}
31042#else
31043__ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
31044  uint16x8_t __ret;
31045  __ret = (uint16x8_t)(__p0);
31046  return __ret;
31047}
31048#endif
31049
31050#ifdef __LITTLE_ENDIAN__
31051__ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
31052  uint16x8_t __ret;
31053  __ret = (uint16x8_t)(__p0);
31054  return __ret;
31055}
31056#else
31057__ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
31058  uint16x8_t __ret;
31059  __ret = (uint16x8_t)(__p0);
31060  return __ret;
31061}
31062#endif
31063
31064#ifdef __LITTLE_ENDIAN__
31065__ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
31066  uint16x8_t __ret;
31067  __ret = (uint16x8_t)(__p0);
31068  return __ret;
31069}
31070#else
31071__ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
31072  uint16x8_t __ret;
31073  __ret = (uint16x8_t)(__p0);
31074  return __ret;
31075}
31076#endif
31077
31078#ifdef __LITTLE_ENDIAN__
31079__ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
31080  uint16x8_t __ret;
31081  __ret = (uint16x8_t)(__p0);
31082  return __ret;
31083}
31084#else
31085__ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
31086  uint16x8_t __ret;
31087  __ret = (uint16x8_t)(__p0);
31088  return __ret;
31089}
31090#endif
31091
31092#ifdef __LITTLE_ENDIAN__
31093__ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
31094  uint16x8_t __ret;
31095  __ret = (uint16x8_t)(__p0);
31096  return __ret;
31097}
31098#else
31099__ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
31100  uint16x8_t __ret;
31101  __ret = (uint16x8_t)(__p0);
31102  return __ret;
31103}
31104#endif
31105
31106#ifdef __LITTLE_ENDIAN__
31107__ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
31108  uint16x8_t __ret;
31109  __ret = (uint16x8_t)(__p0);
31110  return __ret;
31111}
31112#else
31113__ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
31114  uint16x8_t __ret;
31115  __ret = (uint16x8_t)(__p0);
31116  return __ret;
31117}
31118#endif
31119
31120#ifdef __LITTLE_ENDIAN__
31121__ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
31122  int8x16_t __ret;
31123  __ret = (int8x16_t)(__p0);
31124  return __ret;
31125}
31126#else
31127__ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
31128  int8x16_t __ret;
31129  __ret = (int8x16_t)(__p0);
31130  return __ret;
31131}
31132#endif
31133
31134#ifdef __LITTLE_ENDIAN__
31135__ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
31136  int8x16_t __ret;
31137  __ret = (int8x16_t)(__p0);
31138  return __ret;
31139}
31140#else
31141__ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
31142  int8x16_t __ret;
31143  __ret = (int8x16_t)(__p0);
31144  return __ret;
31145}
31146#endif
31147
31148#ifdef __LITTLE_ENDIAN__
31149__ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
31150  int8x16_t __ret;
31151  __ret = (int8x16_t)(__p0);
31152  return __ret;
31153}
31154#else
31155__ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
31156  int8x16_t __ret;
31157  __ret = (int8x16_t)(__p0);
31158  return __ret;
31159}
31160#endif
31161
31162#ifdef __LITTLE_ENDIAN__
31163__ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
31164  int8x16_t __ret;
31165  __ret = (int8x16_t)(__p0);
31166  return __ret;
31167}
31168#else
31169__ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
31170  int8x16_t __ret;
31171  __ret = (int8x16_t)(__p0);
31172  return __ret;
31173}
31174#endif
31175
31176#ifdef __LITTLE_ENDIAN__
31177__ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
31178  int8x16_t __ret;
31179  __ret = (int8x16_t)(__p0);
31180  return __ret;
31181}
31182#else
31183__ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
31184  int8x16_t __ret;
31185  __ret = (int8x16_t)(__p0);
31186  return __ret;
31187}
31188#endif
31189
31190#ifdef __LITTLE_ENDIAN__
31191__ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
31192  int8x16_t __ret;
31193  __ret = (int8x16_t)(__p0);
31194  return __ret;
31195}
31196#else
31197__ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
31198  int8x16_t __ret;
31199  __ret = (int8x16_t)(__p0);
31200  return __ret;
31201}
31202#endif
31203
31204#ifdef __LITTLE_ENDIAN__
31205__ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
31206  int8x16_t __ret;
31207  __ret = (int8x16_t)(__p0);
31208  return __ret;
31209}
31210#else
31211__ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
31212  int8x16_t __ret;
31213  __ret = (int8x16_t)(__p0);
31214  return __ret;
31215}
31216#endif
31217
31218#ifdef __LITTLE_ENDIAN__
31219__ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
31220  int8x16_t __ret;
31221  __ret = (int8x16_t)(__p0);
31222  return __ret;
31223}
31224#else
31225__ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
31226  int8x16_t __ret;
31227  __ret = (int8x16_t)(__p0);
31228  return __ret;
31229}
31230#endif
31231
31232#ifdef __LITTLE_ENDIAN__
31233__ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
31234  int8x16_t __ret;
31235  __ret = (int8x16_t)(__p0);
31236  return __ret;
31237}
31238#else
31239__ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
31240  int8x16_t __ret;
31241  __ret = (int8x16_t)(__p0);
31242  return __ret;
31243}
31244#endif
31245
31246#ifdef __LITTLE_ENDIAN__
31247__ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
31248  int8x16_t __ret;
31249  __ret = (int8x16_t)(__p0);
31250  return __ret;
31251}
31252#else
31253__ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
31254  int8x16_t __ret;
31255  __ret = (int8x16_t)(__p0);
31256  return __ret;
31257}
31258#endif
31259
31260#ifdef __LITTLE_ENDIAN__
31261__ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
31262  int8x16_t __ret;
31263  __ret = (int8x16_t)(__p0);
31264  return __ret;
31265}
31266#else
31267__ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
31268  int8x16_t __ret;
31269  __ret = (int8x16_t)(__p0);
31270  return __ret;
31271}
31272#endif
31273
31274#ifdef __LITTLE_ENDIAN__
31275__ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
31276  float32x4_t __ret;
31277  __ret = (float32x4_t)(__p0);
31278  return __ret;
31279}
31280#else
31281__ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
31282  float32x4_t __ret;
31283  __ret = (float32x4_t)(__p0);
31284  return __ret;
31285}
31286#endif
31287
31288#ifdef __LITTLE_ENDIAN__
31289__ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
31290  float32x4_t __ret;
31291  __ret = (float32x4_t)(__p0);
31292  return __ret;
31293}
31294#else
31295__ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
31296  float32x4_t __ret;
31297  __ret = (float32x4_t)(__p0);
31298  return __ret;
31299}
31300#endif
31301
31302#ifdef __LITTLE_ENDIAN__
31303__ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
31304  float32x4_t __ret;
31305  __ret = (float32x4_t)(__p0);
31306  return __ret;
31307}
31308#else
31309__ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
31310  float32x4_t __ret;
31311  __ret = (float32x4_t)(__p0);
31312  return __ret;
31313}
31314#endif
31315
31316#ifdef __LITTLE_ENDIAN__
31317__ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
31318  float32x4_t __ret;
31319  __ret = (float32x4_t)(__p0);
31320  return __ret;
31321}
31322#else
31323__ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
31324  float32x4_t __ret;
31325  __ret = (float32x4_t)(__p0);
31326  return __ret;
31327}
31328#endif
31329
31330#ifdef __LITTLE_ENDIAN__
31331__ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
31332  float32x4_t __ret;
31333  __ret = (float32x4_t)(__p0);
31334  return __ret;
31335}
31336#else
31337__ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
31338  float32x4_t __ret;
31339  __ret = (float32x4_t)(__p0);
31340  return __ret;
31341}
31342#endif
31343
31344#ifdef __LITTLE_ENDIAN__
31345__ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
31346  float32x4_t __ret;
31347  __ret = (float32x4_t)(__p0);
31348  return __ret;
31349}
31350#else
31351__ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
31352  float32x4_t __ret;
31353  __ret = (float32x4_t)(__p0);
31354  return __ret;
31355}
31356#endif
31357
31358#ifdef __LITTLE_ENDIAN__
31359__ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
31360  float32x4_t __ret;
31361  __ret = (float32x4_t)(__p0);
31362  return __ret;
31363}
31364#else
31365__ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
31366  float32x4_t __ret;
31367  __ret = (float32x4_t)(__p0);
31368  return __ret;
31369}
31370#endif
31371
31372#ifdef __LITTLE_ENDIAN__
31373__ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
31374  float32x4_t __ret;
31375  __ret = (float32x4_t)(__p0);
31376  return __ret;
31377}
31378#else
31379__ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
31380  float32x4_t __ret;
31381  __ret = (float32x4_t)(__p0);
31382  return __ret;
31383}
31384#endif
31385
31386#ifdef __LITTLE_ENDIAN__
31387__ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
31388  float32x4_t __ret;
31389  __ret = (float32x4_t)(__p0);
31390  return __ret;
31391}
31392#else
31393__ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
31394  float32x4_t __ret;
31395  __ret = (float32x4_t)(__p0);
31396  return __ret;
31397}
31398#endif
31399
31400#ifdef __LITTLE_ENDIAN__
31401__ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
31402  float32x4_t __ret;
31403  __ret = (float32x4_t)(__p0);
31404  return __ret;
31405}
31406#else
31407__ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
31408  float32x4_t __ret;
31409  __ret = (float32x4_t)(__p0);
31410  return __ret;
31411}
31412#endif
31413
31414#ifdef __LITTLE_ENDIAN__
31415__ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
31416  float32x4_t __ret;
31417  __ret = (float32x4_t)(__p0);
31418  return __ret;
31419}
31420#else
31421__ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
31422  float32x4_t __ret;
31423  __ret = (float32x4_t)(__p0);
31424  return __ret;
31425}
31426#endif
31427
31428#ifdef __LITTLE_ENDIAN__
31429__ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
31430  float16x8_t __ret;
31431  __ret = (float16x8_t)(__p0);
31432  return __ret;
31433}
31434#else
31435__ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
31436  float16x8_t __ret;
31437  __ret = (float16x8_t)(__p0);
31438  return __ret;
31439}
31440#endif
31441
31442#ifdef __LITTLE_ENDIAN__
31443__ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
31444  float16x8_t __ret;
31445  __ret = (float16x8_t)(__p0);
31446  return __ret;
31447}
31448#else
31449__ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
31450  float16x8_t __ret;
31451  __ret = (float16x8_t)(__p0);
31452  return __ret;
31453}
31454#endif
31455
31456#ifdef __LITTLE_ENDIAN__
31457__ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
31458  float16x8_t __ret;
31459  __ret = (float16x8_t)(__p0);
31460  return __ret;
31461}
31462#else
31463__ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
31464  float16x8_t __ret;
31465  __ret = (float16x8_t)(__p0);
31466  return __ret;
31467}
31468#endif
31469
31470#ifdef __LITTLE_ENDIAN__
31471__ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
31472  float16x8_t __ret;
31473  __ret = (float16x8_t)(__p0);
31474  return __ret;
31475}
31476#else
31477__ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
31478  float16x8_t __ret;
31479  __ret = (float16x8_t)(__p0);
31480  return __ret;
31481}
31482#endif
31483
31484#ifdef __LITTLE_ENDIAN__
31485__ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
31486  float16x8_t __ret;
31487  __ret = (float16x8_t)(__p0);
31488  return __ret;
31489}
31490#else
31491__ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
31492  float16x8_t __ret;
31493  __ret = (float16x8_t)(__p0);
31494  return __ret;
31495}
31496#endif
31497
31498#ifdef __LITTLE_ENDIAN__
31499__ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
31500  float16x8_t __ret;
31501  __ret = (float16x8_t)(__p0);
31502  return __ret;
31503}
31504#else
31505__ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
31506  float16x8_t __ret;
31507  __ret = (float16x8_t)(__p0);
31508  return __ret;
31509}
31510#endif
31511
31512#ifdef __LITTLE_ENDIAN__
31513__ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
31514  float16x8_t __ret;
31515  __ret = (float16x8_t)(__p0);
31516  return __ret;
31517}
31518#else
31519__ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
31520  float16x8_t __ret;
31521  __ret = (float16x8_t)(__p0);
31522  return __ret;
31523}
31524#endif
31525
31526#ifdef __LITTLE_ENDIAN__
31527__ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
31528  float16x8_t __ret;
31529  __ret = (float16x8_t)(__p0);
31530  return __ret;
31531}
31532#else
31533__ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
31534  float16x8_t __ret;
31535  __ret = (float16x8_t)(__p0);
31536  return __ret;
31537}
31538#endif
31539
31540#ifdef __LITTLE_ENDIAN__
31541__ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
31542  float16x8_t __ret;
31543  __ret = (float16x8_t)(__p0);
31544  return __ret;
31545}
31546#else
31547__ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
31548  float16x8_t __ret;
31549  __ret = (float16x8_t)(__p0);
31550  return __ret;
31551}
31552#endif
31553
31554#ifdef __LITTLE_ENDIAN__
31555__ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
31556  float16x8_t __ret;
31557  __ret = (float16x8_t)(__p0);
31558  return __ret;
31559}
31560#else
31561__ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
31562  float16x8_t __ret;
31563  __ret = (float16x8_t)(__p0);
31564  return __ret;
31565}
31566#endif
31567
31568#ifdef __LITTLE_ENDIAN__
31569__ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
31570  float16x8_t __ret;
31571  __ret = (float16x8_t)(__p0);
31572  return __ret;
31573}
31574#else
31575__ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
31576  float16x8_t __ret;
31577  __ret = (float16x8_t)(__p0);
31578  return __ret;
31579}
31580#endif
31581
31582#ifdef __LITTLE_ENDIAN__
31583__ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
31584  int32x4_t __ret;
31585  __ret = (int32x4_t)(__p0);
31586  return __ret;
31587}
31588#else
31589__ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
31590  int32x4_t __ret;
31591  __ret = (int32x4_t)(__p0);
31592  return __ret;
31593}
31594#endif
31595
31596#ifdef __LITTLE_ENDIAN__
31597__ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
31598  int32x4_t __ret;
31599  __ret = (int32x4_t)(__p0);
31600  return __ret;
31601}
31602#else
31603__ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
31604  int32x4_t __ret;
31605  __ret = (int32x4_t)(__p0);
31606  return __ret;
31607}
31608#endif
31609
31610#ifdef __LITTLE_ENDIAN__
31611__ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
31612  int32x4_t __ret;
31613  __ret = (int32x4_t)(__p0);
31614  return __ret;
31615}
31616#else
31617__ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
31618  int32x4_t __ret;
31619  __ret = (int32x4_t)(__p0);
31620  return __ret;
31621}
31622#endif
31623
31624#ifdef __LITTLE_ENDIAN__
31625__ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
31626  int32x4_t __ret;
31627  __ret = (int32x4_t)(__p0);
31628  return __ret;
31629}
31630#else
31631__ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
31632  int32x4_t __ret;
31633  __ret = (int32x4_t)(__p0);
31634  return __ret;
31635}
31636#endif
31637
31638#ifdef __LITTLE_ENDIAN__
31639__ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
31640  int32x4_t __ret;
31641  __ret = (int32x4_t)(__p0);
31642  return __ret;
31643}
31644#else
31645__ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
31646  int32x4_t __ret;
31647  __ret = (int32x4_t)(__p0);
31648  return __ret;
31649}
31650#endif
31651
31652#ifdef __LITTLE_ENDIAN__
31653__ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
31654  int32x4_t __ret;
31655  __ret = (int32x4_t)(__p0);
31656  return __ret;
31657}
31658#else
31659__ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
31660  int32x4_t __ret;
31661  __ret = (int32x4_t)(__p0);
31662  return __ret;
31663}
31664#endif
31665
31666#ifdef __LITTLE_ENDIAN__
31667__ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
31668  int32x4_t __ret;
31669  __ret = (int32x4_t)(__p0);
31670  return __ret;
31671}
31672#else
31673__ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
31674  int32x4_t __ret;
31675  __ret = (int32x4_t)(__p0);
31676  return __ret;
31677}
31678#endif
31679
31680#ifdef __LITTLE_ENDIAN__
31681__ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
31682  int32x4_t __ret;
31683  __ret = (int32x4_t)(__p0);
31684  return __ret;
31685}
31686#else
31687__ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
31688  int32x4_t __ret;
31689  __ret = (int32x4_t)(__p0);
31690  return __ret;
31691}
31692#endif
31693
31694#ifdef __LITTLE_ENDIAN__
31695__ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
31696  int32x4_t __ret;
31697  __ret = (int32x4_t)(__p0);
31698  return __ret;
31699}
31700#else
31701__ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
31702  int32x4_t __ret;
31703  __ret = (int32x4_t)(__p0);
31704  return __ret;
31705}
31706#endif
31707
31708#ifdef __LITTLE_ENDIAN__
31709__ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
31710  int32x4_t __ret;
31711  __ret = (int32x4_t)(__p0);
31712  return __ret;
31713}
31714#else
31715__ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
31716  int32x4_t __ret;
31717  __ret = (int32x4_t)(__p0);
31718  return __ret;
31719}
31720#endif
31721
31722#ifdef __LITTLE_ENDIAN__
31723__ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
31724  int32x4_t __ret;
31725  __ret = (int32x4_t)(__p0);
31726  return __ret;
31727}
31728#else
31729__ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
31730  int32x4_t __ret;
31731  __ret = (int32x4_t)(__p0);
31732  return __ret;
31733}
31734#endif
31735
31736#ifdef __LITTLE_ENDIAN__
31737__ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
31738  int64x2_t __ret;
31739  __ret = (int64x2_t)(__p0);
31740  return __ret;
31741}
31742#else
31743__ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
31744  int64x2_t __ret;
31745  __ret = (int64x2_t)(__p0);
31746  return __ret;
31747}
31748#endif
31749
31750#ifdef __LITTLE_ENDIAN__
31751__ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
31752  int64x2_t __ret;
31753  __ret = (int64x2_t)(__p0);
31754  return __ret;
31755}
31756#else
31757__ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
31758  int64x2_t __ret;
31759  __ret = (int64x2_t)(__p0);
31760  return __ret;
31761}
31762#endif
31763
31764#ifdef __LITTLE_ENDIAN__
31765__ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
31766  int64x2_t __ret;
31767  __ret = (int64x2_t)(__p0);
31768  return __ret;
31769}
31770#else
31771__ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
31772  int64x2_t __ret;
31773  __ret = (int64x2_t)(__p0);
31774  return __ret;
31775}
31776#endif
31777
31778#ifdef __LITTLE_ENDIAN__
31779__ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
31780  int64x2_t __ret;
31781  __ret = (int64x2_t)(__p0);
31782  return __ret;
31783}
31784#else
31785__ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
31786  int64x2_t __ret;
31787  __ret = (int64x2_t)(__p0);
31788  return __ret;
31789}
31790#endif
31791
31792#ifdef __LITTLE_ENDIAN__
31793__ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
31794  int64x2_t __ret;
31795  __ret = (int64x2_t)(__p0);
31796  return __ret;
31797}
31798#else
31799__ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
31800  int64x2_t __ret;
31801  __ret = (int64x2_t)(__p0);
31802  return __ret;
31803}
31804#endif
31805
31806#ifdef __LITTLE_ENDIAN__
31807__ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
31808  int64x2_t __ret;
31809  __ret = (int64x2_t)(__p0);
31810  return __ret;
31811}
31812#else
31813__ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
31814  int64x2_t __ret;
31815  __ret = (int64x2_t)(__p0);
31816  return __ret;
31817}
31818#endif
31819
31820#ifdef __LITTLE_ENDIAN__
31821__ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
31822  int64x2_t __ret;
31823  __ret = (int64x2_t)(__p0);
31824  return __ret;
31825}
31826#else
31827__ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
31828  int64x2_t __ret;
31829  __ret = (int64x2_t)(__p0);
31830  return __ret;
31831}
31832#endif
31833
31834#ifdef __LITTLE_ENDIAN__
31835__ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
31836  int64x2_t __ret;
31837  __ret = (int64x2_t)(__p0);
31838  return __ret;
31839}
31840#else
31841__ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
31842  int64x2_t __ret;
31843  __ret = (int64x2_t)(__p0);
31844  return __ret;
31845}
31846#endif
31847
31848#ifdef __LITTLE_ENDIAN__
31849__ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
31850  int64x2_t __ret;
31851  __ret = (int64x2_t)(__p0);
31852  return __ret;
31853}
31854#else
31855__ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
31856  int64x2_t __ret;
31857  __ret = (int64x2_t)(__p0);
31858  return __ret;
31859}
31860#endif
31861
31862#ifdef __LITTLE_ENDIAN__
31863__ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
31864  int64x2_t __ret;
31865  __ret = (int64x2_t)(__p0);
31866  return __ret;
31867}
31868#else
31869__ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
31870  int64x2_t __ret;
31871  __ret = (int64x2_t)(__p0);
31872  return __ret;
31873}
31874#endif
31875
31876#ifdef __LITTLE_ENDIAN__
31877__ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
31878  int64x2_t __ret;
31879  __ret = (int64x2_t)(__p0);
31880  return __ret;
31881}
31882#else
31883__ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
31884  int64x2_t __ret;
31885  __ret = (int64x2_t)(__p0);
31886  return __ret;
31887}
31888#endif
31889
31890#ifdef __LITTLE_ENDIAN__
31891__ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
31892  int16x8_t __ret;
31893  __ret = (int16x8_t)(__p0);
31894  return __ret;
31895}
31896#else
31897__ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
31898  int16x8_t __ret;
31899  __ret = (int16x8_t)(__p0);
31900  return __ret;
31901}
31902#endif
31903
31904#ifdef __LITTLE_ENDIAN__
31905__ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
31906  int16x8_t __ret;
31907  __ret = (int16x8_t)(__p0);
31908  return __ret;
31909}
31910#else
31911__ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
31912  int16x8_t __ret;
31913  __ret = (int16x8_t)(__p0);
31914  return __ret;
31915}
31916#endif
31917
31918#ifdef __LITTLE_ENDIAN__
31919__ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
31920  int16x8_t __ret;
31921  __ret = (int16x8_t)(__p0);
31922  return __ret;
31923}
31924#else
31925__ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
31926  int16x8_t __ret;
31927  __ret = (int16x8_t)(__p0);
31928  return __ret;
31929}
31930#endif
31931
31932#ifdef __LITTLE_ENDIAN__
31933__ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
31934  int16x8_t __ret;
31935  __ret = (int16x8_t)(__p0);
31936  return __ret;
31937}
31938#else
31939__ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
31940  int16x8_t __ret;
31941  __ret = (int16x8_t)(__p0);
31942  return __ret;
31943}
31944#endif
31945
31946#ifdef __LITTLE_ENDIAN__
31947__ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
31948  int16x8_t __ret;
31949  __ret = (int16x8_t)(__p0);
31950  return __ret;
31951}
31952#else
31953__ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
31954  int16x8_t __ret;
31955  __ret = (int16x8_t)(__p0);
31956  return __ret;
31957}
31958#endif
31959
31960#ifdef __LITTLE_ENDIAN__
31961__ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
31962  int16x8_t __ret;
31963  __ret = (int16x8_t)(__p0);
31964  return __ret;
31965}
31966#else
31967__ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
31968  int16x8_t __ret;
31969  __ret = (int16x8_t)(__p0);
31970  return __ret;
31971}
31972#endif
31973
31974#ifdef __LITTLE_ENDIAN__
31975__ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
31976  int16x8_t __ret;
31977  __ret = (int16x8_t)(__p0);
31978  return __ret;
31979}
31980#else
31981__ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
31982  int16x8_t __ret;
31983  __ret = (int16x8_t)(__p0);
31984  return __ret;
31985}
31986#endif
31987
31988#ifdef __LITTLE_ENDIAN__
31989__ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
31990  int16x8_t __ret;
31991  __ret = (int16x8_t)(__p0);
31992  return __ret;
31993}
31994#else
31995__ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
31996  int16x8_t __ret;
31997  __ret = (int16x8_t)(__p0);
31998  return __ret;
31999}
32000#endif
32001
32002#ifdef __LITTLE_ENDIAN__
32003__ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
32004  int16x8_t __ret;
32005  __ret = (int16x8_t)(__p0);
32006  return __ret;
32007}
32008#else
32009__ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
32010  int16x8_t __ret;
32011  __ret = (int16x8_t)(__p0);
32012  return __ret;
32013}
32014#endif
32015
32016#ifdef __LITTLE_ENDIAN__
32017__ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
32018  int16x8_t __ret;
32019  __ret = (int16x8_t)(__p0);
32020  return __ret;
32021}
32022#else
32023__ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
32024  int16x8_t __ret;
32025  __ret = (int16x8_t)(__p0);
32026  return __ret;
32027}
32028#endif
32029
32030#ifdef __LITTLE_ENDIAN__
32031__ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
32032  int16x8_t __ret;
32033  __ret = (int16x8_t)(__p0);
32034  return __ret;
32035}
32036#else
32037__ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
32038  int16x8_t __ret;
32039  __ret = (int16x8_t)(__p0);
32040  return __ret;
32041}
32042#endif
32043
32044#ifdef __LITTLE_ENDIAN__
32045__ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
32046  uint8x8_t __ret;
32047  __ret = (uint8x8_t)(__p0);
32048  return __ret;
32049}
32050#else
32051__ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
32052  uint8x8_t __ret;
32053  __ret = (uint8x8_t)(__p0);
32054  return __ret;
32055}
32056#endif
32057
32058#ifdef __LITTLE_ENDIAN__
32059__ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
32060  uint8x8_t __ret;
32061  __ret = (uint8x8_t)(__p0);
32062  return __ret;
32063}
32064#else
32065__ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
32066  uint8x8_t __ret;
32067  __ret = (uint8x8_t)(__p0);
32068  return __ret;
32069}
32070#endif
32071
32072#ifdef __LITTLE_ENDIAN__
32073__ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
32074  uint8x8_t __ret;
32075  __ret = (uint8x8_t)(__p0);
32076  return __ret;
32077}
32078#else
32079__ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
32080  uint8x8_t __ret;
32081  __ret = (uint8x8_t)(__p0);
32082  return __ret;
32083}
32084#endif
32085
32086#ifdef __LITTLE_ENDIAN__
32087__ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
32088  uint8x8_t __ret;
32089  __ret = (uint8x8_t)(__p0);
32090  return __ret;
32091}
32092#else
32093__ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
32094  uint8x8_t __ret;
32095  __ret = (uint8x8_t)(__p0);
32096  return __ret;
32097}
32098#endif
32099
32100#ifdef __LITTLE_ENDIAN__
32101__ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
32102  uint8x8_t __ret;
32103  __ret = (uint8x8_t)(__p0);
32104  return __ret;
32105}
32106#else
32107__ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
32108  uint8x8_t __ret;
32109  __ret = (uint8x8_t)(__p0);
32110  return __ret;
32111}
32112#endif
32113
32114#ifdef __LITTLE_ENDIAN__
32115__ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
32116  uint8x8_t __ret;
32117  __ret = (uint8x8_t)(__p0);
32118  return __ret;
32119}
32120#else
32121__ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
32122  uint8x8_t __ret;
32123  __ret = (uint8x8_t)(__p0);
32124  return __ret;
32125}
32126#endif
32127
32128#ifdef __LITTLE_ENDIAN__
32129__ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
32130  uint8x8_t __ret;
32131  __ret = (uint8x8_t)(__p0);
32132  return __ret;
32133}
32134#else
32135__ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
32136  uint8x8_t __ret;
32137  __ret = (uint8x8_t)(__p0);
32138  return __ret;
32139}
32140#endif
32141
32142#ifdef __LITTLE_ENDIAN__
32143__ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
32144  uint8x8_t __ret;
32145  __ret = (uint8x8_t)(__p0);
32146  return __ret;
32147}
32148#else
32149__ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
32150  uint8x8_t __ret;
32151  __ret = (uint8x8_t)(__p0);
32152  return __ret;
32153}
32154#endif
32155
32156#ifdef __LITTLE_ENDIAN__
32157__ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
32158  uint8x8_t __ret;
32159  __ret = (uint8x8_t)(__p0);
32160  return __ret;
32161}
32162#else
32163__ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
32164  uint8x8_t __ret;
32165  __ret = (uint8x8_t)(__p0);
32166  return __ret;
32167}
32168#endif
32169
32170#ifdef __LITTLE_ENDIAN__
32171__ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
32172  uint8x8_t __ret;
32173  __ret = (uint8x8_t)(__p0);
32174  return __ret;
32175}
32176#else
32177__ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
32178  uint8x8_t __ret;
32179  __ret = (uint8x8_t)(__p0);
32180  return __ret;
32181}
32182#endif
32183
32184#ifdef __LITTLE_ENDIAN__
32185__ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
32186  uint8x8_t __ret;
32187  __ret = (uint8x8_t)(__p0);
32188  return __ret;
32189}
32190#else
32191__ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
32192  uint8x8_t __ret;
32193  __ret = (uint8x8_t)(__p0);
32194  return __ret;
32195}
32196#endif
32197
32198#ifdef __LITTLE_ENDIAN__
32199__ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
32200  uint32x2_t __ret;
32201  __ret = (uint32x2_t)(__p0);
32202  return __ret;
32203}
32204#else
32205__ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
32206  uint32x2_t __ret;
32207  __ret = (uint32x2_t)(__p0);
32208  return __ret;
32209}
32210#endif
32211
32212#ifdef __LITTLE_ENDIAN__
32213__ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
32214  uint32x2_t __ret;
32215  __ret = (uint32x2_t)(__p0);
32216  return __ret;
32217}
32218#else
32219__ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
32220  uint32x2_t __ret;
32221  __ret = (uint32x2_t)(__p0);
32222  return __ret;
32223}
32224#endif
32225
32226#ifdef __LITTLE_ENDIAN__
32227__ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
32228  uint32x2_t __ret;
32229  __ret = (uint32x2_t)(__p0);
32230  return __ret;
32231}
32232#else
32233__ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
32234  uint32x2_t __ret;
32235  __ret = (uint32x2_t)(__p0);
32236  return __ret;
32237}
32238#endif
32239
32240#ifdef __LITTLE_ENDIAN__
32241__ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
32242  uint32x2_t __ret;
32243  __ret = (uint32x2_t)(__p0);
32244  return __ret;
32245}
32246#else
32247__ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
32248  uint32x2_t __ret;
32249  __ret = (uint32x2_t)(__p0);
32250  return __ret;
32251}
32252#endif
32253
32254#ifdef __LITTLE_ENDIAN__
32255__ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
32256  uint32x2_t __ret;
32257  __ret = (uint32x2_t)(__p0);
32258  return __ret;
32259}
32260#else
32261__ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
32262  uint32x2_t __ret;
32263  __ret = (uint32x2_t)(__p0);
32264  return __ret;
32265}
32266#endif
32267
32268#ifdef __LITTLE_ENDIAN__
32269__ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
32270  uint32x2_t __ret;
32271  __ret = (uint32x2_t)(__p0);
32272  return __ret;
32273}
32274#else
32275__ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
32276  uint32x2_t __ret;
32277  __ret = (uint32x2_t)(__p0);
32278  return __ret;
32279}
32280#endif
32281
32282#ifdef __LITTLE_ENDIAN__
32283__ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
32284  uint32x2_t __ret;
32285  __ret = (uint32x2_t)(__p0);
32286  return __ret;
32287}
32288#else
32289__ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
32290  uint32x2_t __ret;
32291  __ret = (uint32x2_t)(__p0);
32292  return __ret;
32293}
32294#endif
32295
32296#ifdef __LITTLE_ENDIAN__
32297__ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
32298  uint32x2_t __ret;
32299  __ret = (uint32x2_t)(__p0);
32300  return __ret;
32301}
32302#else
32303__ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
32304  uint32x2_t __ret;
32305  __ret = (uint32x2_t)(__p0);
32306  return __ret;
32307}
32308#endif
32309
32310#ifdef __LITTLE_ENDIAN__
32311__ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
32312  uint32x2_t __ret;
32313  __ret = (uint32x2_t)(__p0);
32314  return __ret;
32315}
32316#else
32317__ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
32318  uint32x2_t __ret;
32319  __ret = (uint32x2_t)(__p0);
32320  return __ret;
32321}
32322#endif
32323
32324#ifdef __LITTLE_ENDIAN__
32325__ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
32326  uint32x2_t __ret;
32327  __ret = (uint32x2_t)(__p0);
32328  return __ret;
32329}
32330#else
32331__ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
32332  uint32x2_t __ret;
32333  __ret = (uint32x2_t)(__p0);
32334  return __ret;
32335}
32336#endif
32337
32338#ifdef __LITTLE_ENDIAN__
32339__ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
32340  uint32x2_t __ret;
32341  __ret = (uint32x2_t)(__p0);
32342  return __ret;
32343}
32344#else
32345__ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
32346  uint32x2_t __ret;
32347  __ret = (uint32x2_t)(__p0);
32348  return __ret;
32349}
32350#endif
32351
32352#ifdef __LITTLE_ENDIAN__
32353__ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
32354  uint64x1_t __ret;
32355  __ret = (uint64x1_t)(__p0);
32356  return __ret;
32357}
32358#else
32359__ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
32360  uint64x1_t __ret;
32361  __ret = (uint64x1_t)(__p0);
32362  return __ret;
32363}
32364#endif
32365
32366#ifdef __LITTLE_ENDIAN__
32367__ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
32368  uint64x1_t __ret;
32369  __ret = (uint64x1_t)(__p0);
32370  return __ret;
32371}
32372#else
32373__ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
32374  uint64x1_t __ret;
32375  __ret = (uint64x1_t)(__p0);
32376  return __ret;
32377}
32378#endif
32379
32380#ifdef __LITTLE_ENDIAN__
32381__ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
32382  uint64x1_t __ret;
32383  __ret = (uint64x1_t)(__p0);
32384  return __ret;
32385}
32386#else
32387__ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
32388  uint64x1_t __ret;
32389  __ret = (uint64x1_t)(__p0);
32390  return __ret;
32391}
32392#endif
32393
32394#ifdef __LITTLE_ENDIAN__
32395__ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
32396  uint64x1_t __ret;
32397  __ret = (uint64x1_t)(__p0);
32398  return __ret;
32399}
32400#else
32401__ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
32402  uint64x1_t __ret;
32403  __ret = (uint64x1_t)(__p0);
32404  return __ret;
32405}
32406#endif
32407
32408#ifdef __LITTLE_ENDIAN__
32409__ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
32410  uint64x1_t __ret;
32411  __ret = (uint64x1_t)(__p0);
32412  return __ret;
32413}
32414#else
32415__ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
32416  uint64x1_t __ret;
32417  __ret = (uint64x1_t)(__p0);
32418  return __ret;
32419}
32420#endif
32421
32422#ifdef __LITTLE_ENDIAN__
32423__ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
32424  uint64x1_t __ret;
32425  __ret = (uint64x1_t)(__p0);
32426  return __ret;
32427}
32428#else
32429__ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
32430  uint64x1_t __ret;
32431  __ret = (uint64x1_t)(__p0);
32432  return __ret;
32433}
32434#endif
32435
32436#ifdef __LITTLE_ENDIAN__
32437__ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
32438  uint64x1_t __ret;
32439  __ret = (uint64x1_t)(__p0);
32440  return __ret;
32441}
32442#else
32443__ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
32444  uint64x1_t __ret;
32445  __ret = (uint64x1_t)(__p0);
32446  return __ret;
32447}
32448#endif
32449
32450#ifdef __LITTLE_ENDIAN__
32451__ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
32452  uint64x1_t __ret;
32453  __ret = (uint64x1_t)(__p0);
32454  return __ret;
32455}
32456#else
32457__ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
32458  uint64x1_t __ret;
32459  __ret = (uint64x1_t)(__p0);
32460  return __ret;
32461}
32462#endif
32463
32464#ifdef __LITTLE_ENDIAN__
32465__ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
32466  uint64x1_t __ret;
32467  __ret = (uint64x1_t)(__p0);
32468  return __ret;
32469}
32470#else
32471__ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
32472  uint64x1_t __ret;
32473  __ret = (uint64x1_t)(__p0);
32474  return __ret;
32475}
32476#endif
32477
32478#ifdef __LITTLE_ENDIAN__
32479__ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
32480  uint64x1_t __ret;
32481  __ret = (uint64x1_t)(__p0);
32482  return __ret;
32483}
32484#else
32485__ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
32486  uint64x1_t __ret;
32487  __ret = (uint64x1_t)(__p0);
32488  return __ret;
32489}
32490#endif
32491
32492#ifdef __LITTLE_ENDIAN__
32493__ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
32494  uint64x1_t __ret;
32495  __ret = (uint64x1_t)(__p0);
32496  return __ret;
32497}
32498#else
32499__ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
32500  uint64x1_t __ret;
32501  __ret = (uint64x1_t)(__p0);
32502  return __ret;
32503}
32504#endif
32505
32506#ifdef __LITTLE_ENDIAN__
32507__ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
32508  uint16x4_t __ret;
32509  __ret = (uint16x4_t)(__p0);
32510  return __ret;
32511}
32512#else
32513__ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
32514  uint16x4_t __ret;
32515  __ret = (uint16x4_t)(__p0);
32516  return __ret;
32517}
32518#endif
32519
32520#ifdef __LITTLE_ENDIAN__
32521__ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
32522  uint16x4_t __ret;
32523  __ret = (uint16x4_t)(__p0);
32524  return __ret;
32525}
32526#else
32527__ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
32528  uint16x4_t __ret;
32529  __ret = (uint16x4_t)(__p0);
32530  return __ret;
32531}
32532#endif
32533
32534#ifdef __LITTLE_ENDIAN__
32535__ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
32536  uint16x4_t __ret;
32537  __ret = (uint16x4_t)(__p0);
32538  return __ret;
32539}
32540#else
32541__ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
32542  uint16x4_t __ret;
32543  __ret = (uint16x4_t)(__p0);
32544  return __ret;
32545}
32546#endif
32547
32548#ifdef __LITTLE_ENDIAN__
32549__ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
32550  uint16x4_t __ret;
32551  __ret = (uint16x4_t)(__p0);
32552  return __ret;
32553}
32554#else
32555__ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
32556  uint16x4_t __ret;
32557  __ret = (uint16x4_t)(__p0);
32558  return __ret;
32559}
32560#endif
32561
32562#ifdef __LITTLE_ENDIAN__
32563__ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
32564  uint16x4_t __ret;
32565  __ret = (uint16x4_t)(__p0);
32566  return __ret;
32567}
32568#else
32569__ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
32570  uint16x4_t __ret;
32571  __ret = (uint16x4_t)(__p0);
32572  return __ret;
32573}
32574#endif
32575
32576#ifdef __LITTLE_ENDIAN__
32577__ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
32578  uint16x4_t __ret;
32579  __ret = (uint16x4_t)(__p0);
32580  return __ret;
32581}
32582#else
32583__ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
32584  uint16x4_t __ret;
32585  __ret = (uint16x4_t)(__p0);
32586  return __ret;
32587}
32588#endif
32589
32590#ifdef __LITTLE_ENDIAN__
32591__ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
32592  uint16x4_t __ret;
32593  __ret = (uint16x4_t)(__p0);
32594  return __ret;
32595}
32596#else
32597__ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
32598  uint16x4_t __ret;
32599  __ret = (uint16x4_t)(__p0);
32600  return __ret;
32601}
32602#endif
32603
32604#ifdef __LITTLE_ENDIAN__
32605__ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
32606  uint16x4_t __ret;
32607  __ret = (uint16x4_t)(__p0);
32608  return __ret;
32609}
32610#else
32611__ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
32612  uint16x4_t __ret;
32613  __ret = (uint16x4_t)(__p0);
32614  return __ret;
32615}
32616#endif
32617
32618#ifdef __LITTLE_ENDIAN__
32619__ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
32620  uint16x4_t __ret;
32621  __ret = (uint16x4_t)(__p0);
32622  return __ret;
32623}
32624#else
32625__ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
32626  uint16x4_t __ret;
32627  __ret = (uint16x4_t)(__p0);
32628  return __ret;
32629}
32630#endif
32631
32632#ifdef __LITTLE_ENDIAN__
32633__ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
32634  uint16x4_t __ret;
32635  __ret = (uint16x4_t)(__p0);
32636  return __ret;
32637}
32638#else
32639__ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
32640  uint16x4_t __ret;
32641  __ret = (uint16x4_t)(__p0);
32642  return __ret;
32643}
32644#endif
32645
32646#ifdef __LITTLE_ENDIAN__
32647__ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
32648  uint16x4_t __ret;
32649  __ret = (uint16x4_t)(__p0);
32650  return __ret;
32651}
32652#else
32653__ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
32654  uint16x4_t __ret;
32655  __ret = (uint16x4_t)(__p0);
32656  return __ret;
32657}
32658#endif
32659
32660#ifdef __LITTLE_ENDIAN__
32661__ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
32662  int8x8_t __ret;
32663  __ret = (int8x8_t)(__p0);
32664  return __ret;
32665}
32666#else
32667__ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
32668  int8x8_t __ret;
32669  __ret = (int8x8_t)(__p0);
32670  return __ret;
32671}
32672#endif
32673
32674#ifdef __LITTLE_ENDIAN__
32675__ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
32676  int8x8_t __ret;
32677  __ret = (int8x8_t)(__p0);
32678  return __ret;
32679}
32680#else
32681__ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
32682  int8x8_t __ret;
32683  __ret = (int8x8_t)(__p0);
32684  return __ret;
32685}
32686#endif
32687
32688#ifdef __LITTLE_ENDIAN__
32689__ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
32690  int8x8_t __ret;
32691  __ret = (int8x8_t)(__p0);
32692  return __ret;
32693}
32694#else
32695__ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
32696  int8x8_t __ret;
32697  __ret = (int8x8_t)(__p0);
32698  return __ret;
32699}
32700#endif
32701
32702#ifdef __LITTLE_ENDIAN__
32703__ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
32704  int8x8_t __ret;
32705  __ret = (int8x8_t)(__p0);
32706  return __ret;
32707}
32708#else
32709__ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
32710  int8x8_t __ret;
32711  __ret = (int8x8_t)(__p0);
32712  return __ret;
32713}
32714#endif
32715
32716#ifdef __LITTLE_ENDIAN__
32717__ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
32718  int8x8_t __ret;
32719  __ret = (int8x8_t)(__p0);
32720  return __ret;
32721}
32722#else
32723__ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
32724  int8x8_t __ret;
32725  __ret = (int8x8_t)(__p0);
32726  return __ret;
32727}
32728#endif
32729
32730#ifdef __LITTLE_ENDIAN__
32731__ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
32732  int8x8_t __ret;
32733  __ret = (int8x8_t)(__p0);
32734  return __ret;
32735}
32736#else
32737__ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
32738  int8x8_t __ret;
32739  __ret = (int8x8_t)(__p0);
32740  return __ret;
32741}
32742#endif
32743
32744#ifdef __LITTLE_ENDIAN__
32745__ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
32746  int8x8_t __ret;
32747  __ret = (int8x8_t)(__p0);
32748  return __ret;
32749}
32750#else
32751__ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
32752  int8x8_t __ret;
32753  __ret = (int8x8_t)(__p0);
32754  return __ret;
32755}
32756#endif
32757
32758#ifdef __LITTLE_ENDIAN__
32759__ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
32760  int8x8_t __ret;
32761  __ret = (int8x8_t)(__p0);
32762  return __ret;
32763}
32764#else
32765__ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
32766  int8x8_t __ret;
32767  __ret = (int8x8_t)(__p0);
32768  return __ret;
32769}
32770#endif
32771
32772#ifdef __LITTLE_ENDIAN__
32773__ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
32774  int8x8_t __ret;
32775  __ret = (int8x8_t)(__p0);
32776  return __ret;
32777}
32778#else
32779__ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
32780  int8x8_t __ret;
32781  __ret = (int8x8_t)(__p0);
32782  return __ret;
32783}
32784#endif
32785
32786#ifdef __LITTLE_ENDIAN__
32787__ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
32788  int8x8_t __ret;
32789  __ret = (int8x8_t)(__p0);
32790  return __ret;
32791}
32792#else
32793__ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
32794  int8x8_t __ret;
32795  __ret = (int8x8_t)(__p0);
32796  return __ret;
32797}
32798#endif
32799
32800#ifdef __LITTLE_ENDIAN__
32801__ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
32802  int8x8_t __ret;
32803  __ret = (int8x8_t)(__p0);
32804  return __ret;
32805}
32806#else
32807__ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
32808  int8x8_t __ret;
32809  __ret = (int8x8_t)(__p0);
32810  return __ret;
32811}
32812#endif
32813
32814#ifdef __LITTLE_ENDIAN__
32815__ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
32816  float32x2_t __ret;
32817  __ret = (float32x2_t)(__p0);
32818  return __ret;
32819}
32820#else
32821__ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
32822  float32x2_t __ret;
32823  __ret = (float32x2_t)(__p0);
32824  return __ret;
32825}
32826#endif
32827
32828#ifdef __LITTLE_ENDIAN__
32829__ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
32830  float32x2_t __ret;
32831  __ret = (float32x2_t)(__p0);
32832  return __ret;
32833}
32834#else
32835__ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
32836  float32x2_t __ret;
32837  __ret = (float32x2_t)(__p0);
32838  return __ret;
32839}
32840#endif
32841
32842#ifdef __LITTLE_ENDIAN__
32843__ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
32844  float32x2_t __ret;
32845  __ret = (float32x2_t)(__p0);
32846  return __ret;
32847}
32848#else
32849__ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
32850  float32x2_t __ret;
32851  __ret = (float32x2_t)(__p0);
32852  return __ret;
32853}
32854#endif
32855
32856#ifdef __LITTLE_ENDIAN__
32857__ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
32858  float32x2_t __ret;
32859  __ret = (float32x2_t)(__p0);
32860  return __ret;
32861}
32862#else
32863__ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
32864  float32x2_t __ret;
32865  __ret = (float32x2_t)(__p0);
32866  return __ret;
32867}
32868#endif
32869
32870#ifdef __LITTLE_ENDIAN__
32871__ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
32872  float32x2_t __ret;
32873  __ret = (float32x2_t)(__p0);
32874  return __ret;
32875}
32876#else
32877__ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
32878  float32x2_t __ret;
32879  __ret = (float32x2_t)(__p0);
32880  return __ret;
32881}
32882#endif
32883
32884#ifdef __LITTLE_ENDIAN__
32885__ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
32886  float32x2_t __ret;
32887  __ret = (float32x2_t)(__p0);
32888  return __ret;
32889}
32890#else
32891__ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
32892  float32x2_t __ret;
32893  __ret = (float32x2_t)(__p0);
32894  return __ret;
32895}
32896#endif
32897
32898#ifdef __LITTLE_ENDIAN__
32899__ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
32900  float32x2_t __ret;
32901  __ret = (float32x2_t)(__p0);
32902  return __ret;
32903}
32904#else
32905__ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
32906  float32x2_t __ret;
32907  __ret = (float32x2_t)(__p0);
32908  return __ret;
32909}
32910#endif
32911
32912#ifdef __LITTLE_ENDIAN__
32913__ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
32914  float32x2_t __ret;
32915  __ret = (float32x2_t)(__p0);
32916  return __ret;
32917}
32918#else
32919__ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
32920  float32x2_t __ret;
32921  __ret = (float32x2_t)(__p0);
32922  return __ret;
32923}
32924#endif
32925
32926#ifdef __LITTLE_ENDIAN__
32927__ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
32928  float32x2_t __ret;
32929  __ret = (float32x2_t)(__p0);
32930  return __ret;
32931}
32932#else
32933__ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
32934  float32x2_t __ret;
32935  __ret = (float32x2_t)(__p0);
32936  return __ret;
32937}
32938#endif
32939
32940#ifdef __LITTLE_ENDIAN__
32941__ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
32942  float32x2_t __ret;
32943  __ret = (float32x2_t)(__p0);
32944  return __ret;
32945}
32946#else
32947__ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
32948  float32x2_t __ret;
32949  __ret = (float32x2_t)(__p0);
32950  return __ret;
32951}
32952#endif
32953
32954#ifdef __LITTLE_ENDIAN__
32955__ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
32956  float32x2_t __ret;
32957  __ret = (float32x2_t)(__p0);
32958  return __ret;
32959}
32960#else
32961__ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
32962  float32x2_t __ret;
32963  __ret = (float32x2_t)(__p0);
32964  return __ret;
32965}
32966#endif
32967
32968#ifdef __LITTLE_ENDIAN__
32969__ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
32970  float16x4_t __ret;
32971  __ret = (float16x4_t)(__p0);
32972  return __ret;
32973}
32974#else
32975__ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
32976  float16x4_t __ret;
32977  __ret = (float16x4_t)(__p0);
32978  return __ret;
32979}
32980#endif
32981
32982#ifdef __LITTLE_ENDIAN__
32983__ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
32984  float16x4_t __ret;
32985  __ret = (float16x4_t)(__p0);
32986  return __ret;
32987}
32988#else
32989__ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
32990  float16x4_t __ret;
32991  __ret = (float16x4_t)(__p0);
32992  return __ret;
32993}
32994#endif
32995
32996#ifdef __LITTLE_ENDIAN__
32997__ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
32998  float16x4_t __ret;
32999  __ret = (float16x4_t)(__p0);
33000  return __ret;
33001}
33002#else
33003__ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
33004  float16x4_t __ret;
33005  __ret = (float16x4_t)(__p0);
33006  return __ret;
33007}
33008#endif
33009
33010#ifdef __LITTLE_ENDIAN__
33011__ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
33012  float16x4_t __ret;
33013  __ret = (float16x4_t)(__p0);
33014  return __ret;
33015}
33016#else
33017__ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
33018  float16x4_t __ret;
33019  __ret = (float16x4_t)(__p0);
33020  return __ret;
33021}
33022#endif
33023
33024#ifdef __LITTLE_ENDIAN__
33025__ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
33026  float16x4_t __ret;
33027  __ret = (float16x4_t)(__p0);
33028  return __ret;
33029}
33030#else
33031__ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
33032  float16x4_t __ret;
33033  __ret = (float16x4_t)(__p0);
33034  return __ret;
33035}
33036#endif
33037
33038#ifdef __LITTLE_ENDIAN__
33039__ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
33040  float16x4_t __ret;
33041  __ret = (float16x4_t)(__p0);
33042  return __ret;
33043}
33044#else
33045__ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
33046  float16x4_t __ret;
33047  __ret = (float16x4_t)(__p0);
33048  return __ret;
33049}
33050#endif
33051
33052#ifdef __LITTLE_ENDIAN__
33053__ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
33054  float16x4_t __ret;
33055  __ret = (float16x4_t)(__p0);
33056  return __ret;
33057}
33058#else
33059__ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
33060  float16x4_t __ret;
33061  __ret = (float16x4_t)(__p0);
33062  return __ret;
33063}
33064#endif
33065
33066#ifdef __LITTLE_ENDIAN__
33067__ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
33068  float16x4_t __ret;
33069  __ret = (float16x4_t)(__p0);
33070  return __ret;
33071}
33072#else
33073__ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
33074  float16x4_t __ret;
33075  __ret = (float16x4_t)(__p0);
33076  return __ret;
33077}
33078#endif
33079
33080#ifdef __LITTLE_ENDIAN__
33081__ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
33082  float16x4_t __ret;
33083  __ret = (float16x4_t)(__p0);
33084  return __ret;
33085}
33086#else
33087__ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
33088  float16x4_t __ret;
33089  __ret = (float16x4_t)(__p0);
33090  return __ret;
33091}
33092#endif
33093
33094#ifdef __LITTLE_ENDIAN__
33095__ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
33096  float16x4_t __ret;
33097  __ret = (float16x4_t)(__p0);
33098  return __ret;
33099}
33100#else
33101__ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
33102  float16x4_t __ret;
33103  __ret = (float16x4_t)(__p0);
33104  return __ret;
33105}
33106#endif
33107
33108#ifdef __LITTLE_ENDIAN__
33109__ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
33110  float16x4_t __ret;
33111  __ret = (float16x4_t)(__p0);
33112  return __ret;
33113}
33114#else
33115__ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
33116  float16x4_t __ret;
33117  __ret = (float16x4_t)(__p0);
33118  return __ret;
33119}
33120#endif
33121
33122#ifdef __LITTLE_ENDIAN__
33123__ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
33124  int32x2_t __ret;
33125  __ret = (int32x2_t)(__p0);
33126  return __ret;
33127}
33128#else
33129__ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
33130  int32x2_t __ret;
33131  __ret = (int32x2_t)(__p0);
33132  return __ret;
33133}
33134#endif
33135
33136#ifdef __LITTLE_ENDIAN__
33137__ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
33138  int32x2_t __ret;
33139  __ret = (int32x2_t)(__p0);
33140  return __ret;
33141}
33142#else
33143__ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
33144  int32x2_t __ret;
33145  __ret = (int32x2_t)(__p0);
33146  return __ret;
33147}
33148#endif
33149
33150#ifdef __LITTLE_ENDIAN__
33151__ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
33152  int32x2_t __ret;
33153  __ret = (int32x2_t)(__p0);
33154  return __ret;
33155}
33156#else
33157__ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
33158  int32x2_t __ret;
33159  __ret = (int32x2_t)(__p0);
33160  return __ret;
33161}
33162#endif
33163
33164#ifdef __LITTLE_ENDIAN__
33165__ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
33166  int32x2_t __ret;
33167  __ret = (int32x2_t)(__p0);
33168  return __ret;
33169}
33170#else
33171__ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
33172  int32x2_t __ret;
33173  __ret = (int32x2_t)(__p0);
33174  return __ret;
33175}
33176#endif
33177
33178#ifdef __LITTLE_ENDIAN__
33179__ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
33180  int32x2_t __ret;
33181  __ret = (int32x2_t)(__p0);
33182  return __ret;
33183}
33184#else
33185__ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
33186  int32x2_t __ret;
33187  __ret = (int32x2_t)(__p0);
33188  return __ret;
33189}
33190#endif
33191
33192#ifdef __LITTLE_ENDIAN__
33193__ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
33194  int32x2_t __ret;
33195  __ret = (int32x2_t)(__p0);
33196  return __ret;
33197}
33198#else
33199__ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
33200  int32x2_t __ret;
33201  __ret = (int32x2_t)(__p0);
33202  return __ret;
33203}
33204#endif
33205
33206#ifdef __LITTLE_ENDIAN__
33207__ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
33208  int32x2_t __ret;
33209  __ret = (int32x2_t)(__p0);
33210  return __ret;
33211}
33212#else
33213__ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
33214  int32x2_t __ret;
33215  __ret = (int32x2_t)(__p0);
33216  return __ret;
33217}
33218#endif
33219
33220#ifdef __LITTLE_ENDIAN__
33221__ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
33222  int32x2_t __ret;
33223  __ret = (int32x2_t)(__p0);
33224  return __ret;
33225}
33226#else
33227__ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
33228  int32x2_t __ret;
33229  __ret = (int32x2_t)(__p0);
33230  return __ret;
33231}
33232#endif
33233
33234#ifdef __LITTLE_ENDIAN__
33235__ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
33236  int32x2_t __ret;
33237  __ret = (int32x2_t)(__p0);
33238  return __ret;
33239}
33240#else
33241__ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
33242  int32x2_t __ret;
33243  __ret = (int32x2_t)(__p0);
33244  return __ret;
33245}
33246#endif
33247
33248#ifdef __LITTLE_ENDIAN__
33249__ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
33250  int32x2_t __ret;
33251  __ret = (int32x2_t)(__p0);
33252  return __ret;
33253}
33254#else
33255__ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
33256  int32x2_t __ret;
33257  __ret = (int32x2_t)(__p0);
33258  return __ret;
33259}
33260#endif
33261
33262#ifdef __LITTLE_ENDIAN__
33263__ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
33264  int32x2_t __ret;
33265  __ret = (int32x2_t)(__p0);
33266  return __ret;
33267}
33268#else
33269__ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
33270  int32x2_t __ret;
33271  __ret = (int32x2_t)(__p0);
33272  return __ret;
33273}
33274#endif
33275
33276#ifdef __LITTLE_ENDIAN__
33277__ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
33278  int64x1_t __ret;
33279  __ret = (int64x1_t)(__p0);
33280  return __ret;
33281}
33282#else
33283__ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
33284  int64x1_t __ret;
33285  __ret = (int64x1_t)(__p0);
33286  return __ret;
33287}
33288#endif
33289
33290#ifdef __LITTLE_ENDIAN__
33291__ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
33292  int64x1_t __ret;
33293  __ret = (int64x1_t)(__p0);
33294  return __ret;
33295}
33296#else
33297__ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
33298  int64x1_t __ret;
33299  __ret = (int64x1_t)(__p0);
33300  return __ret;
33301}
33302#endif
33303
33304#ifdef __LITTLE_ENDIAN__
33305__ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
33306  int64x1_t __ret;
33307  __ret = (int64x1_t)(__p0);
33308  return __ret;
33309}
33310#else
33311__ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
33312  int64x1_t __ret;
33313  __ret = (int64x1_t)(__p0);
33314  return __ret;
33315}
33316#endif
33317
33318#ifdef __LITTLE_ENDIAN__
33319__ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
33320  int64x1_t __ret;
33321  __ret = (int64x1_t)(__p0);
33322  return __ret;
33323}
33324#else
33325__ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
33326  int64x1_t __ret;
33327  __ret = (int64x1_t)(__p0);
33328  return __ret;
33329}
33330#endif
33331
33332#ifdef __LITTLE_ENDIAN__
33333__ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
33334  int64x1_t __ret;
33335  __ret = (int64x1_t)(__p0);
33336  return __ret;
33337}
33338#else
33339__ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
33340  int64x1_t __ret;
33341  __ret = (int64x1_t)(__p0);
33342  return __ret;
33343}
33344#endif
33345
33346#ifdef __LITTLE_ENDIAN__
33347__ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
33348  int64x1_t __ret;
33349  __ret = (int64x1_t)(__p0);
33350  return __ret;
33351}
33352#else
33353__ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
33354  int64x1_t __ret;
33355  __ret = (int64x1_t)(__p0);
33356  return __ret;
33357}
33358#endif
33359
33360#ifdef __LITTLE_ENDIAN__
33361__ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
33362  int64x1_t __ret;
33363  __ret = (int64x1_t)(__p0);
33364  return __ret;
33365}
33366#else
33367__ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
33368  int64x1_t __ret;
33369  __ret = (int64x1_t)(__p0);
33370  return __ret;
33371}
33372#endif
33373
33374#ifdef __LITTLE_ENDIAN__
33375__ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
33376  int64x1_t __ret;
33377  __ret = (int64x1_t)(__p0);
33378  return __ret;
33379}
33380#else
33381__ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
33382  int64x1_t __ret;
33383  __ret = (int64x1_t)(__p0);
33384  return __ret;
33385}
33386#endif
33387
33388#ifdef __LITTLE_ENDIAN__
33389__ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
33390  int64x1_t __ret;
33391  __ret = (int64x1_t)(__p0);
33392  return __ret;
33393}
33394#else
33395__ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
33396  int64x1_t __ret;
33397  __ret = (int64x1_t)(__p0);
33398  return __ret;
33399}
33400#endif
33401
33402#ifdef __LITTLE_ENDIAN__
33403__ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
33404  int64x1_t __ret;
33405  __ret = (int64x1_t)(__p0);
33406  return __ret;
33407}
33408#else
33409__ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
33410  int64x1_t __ret;
33411  __ret = (int64x1_t)(__p0);
33412  return __ret;
33413}
33414#endif
33415
33416#ifdef __LITTLE_ENDIAN__
33417__ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
33418  int64x1_t __ret;
33419  __ret = (int64x1_t)(__p0);
33420  return __ret;
33421}
33422#else
33423__ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
33424  int64x1_t __ret;
33425  __ret = (int64x1_t)(__p0);
33426  return __ret;
33427}
33428#endif
33429
33430#ifdef __LITTLE_ENDIAN__
33431__ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
33432  int16x4_t __ret;
33433  __ret = (int16x4_t)(__p0);
33434  return __ret;
33435}
33436#else
33437__ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
33438  int16x4_t __ret;
33439  __ret = (int16x4_t)(__p0);
33440  return __ret;
33441}
33442#endif
33443
33444#ifdef __LITTLE_ENDIAN__
33445__ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
33446  int16x4_t __ret;
33447  __ret = (int16x4_t)(__p0);
33448  return __ret;
33449}
33450#else
33451__ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
33452  int16x4_t __ret;
33453  __ret = (int16x4_t)(__p0);
33454  return __ret;
33455}
33456#endif
33457
33458#ifdef __LITTLE_ENDIAN__
33459__ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
33460  int16x4_t __ret;
33461  __ret = (int16x4_t)(__p0);
33462  return __ret;
33463}
33464#else
33465__ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
33466  int16x4_t __ret;
33467  __ret = (int16x4_t)(__p0);
33468  return __ret;
33469}
33470#endif
33471
33472#ifdef __LITTLE_ENDIAN__
33473__ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
33474  int16x4_t __ret;
33475  __ret = (int16x4_t)(__p0);
33476  return __ret;
33477}
33478#else
33479__ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
33480  int16x4_t __ret;
33481  __ret = (int16x4_t)(__p0);
33482  return __ret;
33483}
33484#endif
33485
33486#ifdef __LITTLE_ENDIAN__
33487__ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
33488  int16x4_t __ret;
33489  __ret = (int16x4_t)(__p0);
33490  return __ret;
33491}
33492#else
33493__ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
33494  int16x4_t __ret;
33495  __ret = (int16x4_t)(__p0);
33496  return __ret;
33497}
33498#endif
33499
33500#ifdef __LITTLE_ENDIAN__
33501__ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
33502  int16x4_t __ret;
33503  __ret = (int16x4_t)(__p0);
33504  return __ret;
33505}
33506#else
33507__ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
33508  int16x4_t __ret;
33509  __ret = (int16x4_t)(__p0);
33510  return __ret;
33511}
33512#endif
33513
33514#ifdef __LITTLE_ENDIAN__
33515__ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
33516  int16x4_t __ret;
33517  __ret = (int16x4_t)(__p0);
33518  return __ret;
33519}
33520#else
33521__ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
33522  int16x4_t __ret;
33523  __ret = (int16x4_t)(__p0);
33524  return __ret;
33525}
33526#endif
33527
33528#ifdef __LITTLE_ENDIAN__
33529__ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
33530  int16x4_t __ret;
33531  __ret = (int16x4_t)(__p0);
33532  return __ret;
33533}
33534#else
33535__ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
33536  int16x4_t __ret;
33537  __ret = (int16x4_t)(__p0);
33538  return __ret;
33539}
33540#endif
33541
33542#ifdef __LITTLE_ENDIAN__
33543__ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
33544  int16x4_t __ret;
33545  __ret = (int16x4_t)(__p0);
33546  return __ret;
33547}
33548#else
33549__ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
33550  int16x4_t __ret;
33551  __ret = (int16x4_t)(__p0);
33552  return __ret;
33553}
33554#endif
33555
33556#ifdef __LITTLE_ENDIAN__
33557__ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
33558  int16x4_t __ret;
33559  __ret = (int16x4_t)(__p0);
33560  return __ret;
33561}
33562#else
33563__ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
33564  int16x4_t __ret;
33565  __ret = (int16x4_t)(__p0);
33566  return __ret;
33567}
33568#endif
33569
33570#ifdef __LITTLE_ENDIAN__
33571__ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
33572  int16x4_t __ret;
33573  __ret = (int16x4_t)(__p0);
33574  return __ret;
33575}
33576#else
33577__ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
33578  int16x4_t __ret;
33579  __ret = (int16x4_t)(__p0);
33580  return __ret;
33581}
33582#endif
33583
33584#endif
33585#if __ARM_ARCH >= 8
33586#ifdef __LITTLE_ENDIAN__
33587__ai int32x4_t vcvtaq_s32_f32(float32x4_t __p0) {
33588  int32x4_t __ret;
33589  __ret = (int32x4_t) __builtin_neon_vcvtaq_s32_v((int8x16_t)__p0, 34);
33590  return __ret;
33591}
33592#else
33593__ai int32x4_t vcvtaq_s32_f32(float32x4_t __p0) {
33594  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33595  int32x4_t __ret;
33596  __ret = (int32x4_t) __builtin_neon_vcvtaq_s32_v((int8x16_t)__rev0, 34);
33597  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33598  return __ret;
33599}
33600#endif
33601
33602#ifdef __LITTLE_ENDIAN__
33603__ai int32x2_t vcvta_s32_f32(float32x2_t __p0) {
33604  int32x2_t __ret;
33605  __ret = (int32x2_t) __builtin_neon_vcvta_s32_v((int8x8_t)__p0, 2);
33606  return __ret;
33607}
33608#else
33609__ai int32x2_t vcvta_s32_f32(float32x2_t __p0) {
33610  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33611  int32x2_t __ret;
33612  __ret = (int32x2_t) __builtin_neon_vcvta_s32_v((int8x8_t)__rev0, 2);
33613  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33614  return __ret;
33615}
33616#endif
33617
33618#ifdef __LITTLE_ENDIAN__
33619__ai uint32x4_t vcvtaq_u32_f32(float32x4_t __p0) {
33620  uint32x4_t __ret;
33621  __ret = (uint32x4_t) __builtin_neon_vcvtaq_u32_v((int8x16_t)__p0, 50);
33622  return __ret;
33623}
33624#else
33625__ai uint32x4_t vcvtaq_u32_f32(float32x4_t __p0) {
33626  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33627  uint32x4_t __ret;
33628  __ret = (uint32x4_t) __builtin_neon_vcvtaq_u32_v((int8x16_t)__rev0, 50);
33629  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33630  return __ret;
33631}
33632#endif
33633
33634#ifdef __LITTLE_ENDIAN__
33635__ai uint32x2_t vcvta_u32_f32(float32x2_t __p0) {
33636  uint32x2_t __ret;
33637  __ret = (uint32x2_t) __builtin_neon_vcvta_u32_v((int8x8_t)__p0, 18);
33638  return __ret;
33639}
33640#else
33641__ai uint32x2_t vcvta_u32_f32(float32x2_t __p0) {
33642  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33643  uint32x2_t __ret;
33644  __ret = (uint32x2_t) __builtin_neon_vcvta_u32_v((int8x8_t)__rev0, 18);
33645  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33646  return __ret;
33647}
33648#endif
33649
33650#ifdef __LITTLE_ENDIAN__
33651__ai int32x4_t vcvtmq_s32_f32(float32x4_t __p0) {
33652  int32x4_t __ret;
33653  __ret = (int32x4_t) __builtin_neon_vcvtmq_s32_v((int8x16_t)__p0, 34);
33654  return __ret;
33655}
33656#else
33657__ai int32x4_t vcvtmq_s32_f32(float32x4_t __p0) {
33658  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33659  int32x4_t __ret;
33660  __ret = (int32x4_t) __builtin_neon_vcvtmq_s32_v((int8x16_t)__rev0, 34);
33661  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33662  return __ret;
33663}
33664#endif
33665
33666#ifdef __LITTLE_ENDIAN__
33667__ai int32x2_t vcvtm_s32_f32(float32x2_t __p0) {
33668  int32x2_t __ret;
33669  __ret = (int32x2_t) __builtin_neon_vcvtm_s32_v((int8x8_t)__p0, 2);
33670  return __ret;
33671}
33672#else
33673__ai int32x2_t vcvtm_s32_f32(float32x2_t __p0) {
33674  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33675  int32x2_t __ret;
33676  __ret = (int32x2_t) __builtin_neon_vcvtm_s32_v((int8x8_t)__rev0, 2);
33677  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33678  return __ret;
33679}
33680#endif
33681
33682#ifdef __LITTLE_ENDIAN__
33683__ai uint32x4_t vcvtmq_u32_f32(float32x4_t __p0) {
33684  uint32x4_t __ret;
33685  __ret = (uint32x4_t) __builtin_neon_vcvtmq_u32_v((int8x16_t)__p0, 50);
33686  return __ret;
33687}
33688#else
33689__ai uint32x4_t vcvtmq_u32_f32(float32x4_t __p0) {
33690  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33691  uint32x4_t __ret;
33692  __ret = (uint32x4_t) __builtin_neon_vcvtmq_u32_v((int8x16_t)__rev0, 50);
33693  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33694  return __ret;
33695}
33696#endif
33697
33698#ifdef __LITTLE_ENDIAN__
33699__ai uint32x2_t vcvtm_u32_f32(float32x2_t __p0) {
33700  uint32x2_t __ret;
33701  __ret = (uint32x2_t) __builtin_neon_vcvtm_u32_v((int8x8_t)__p0, 18);
33702  return __ret;
33703}
33704#else
33705__ai uint32x2_t vcvtm_u32_f32(float32x2_t __p0) {
33706  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33707  uint32x2_t __ret;
33708  __ret = (uint32x2_t) __builtin_neon_vcvtm_u32_v((int8x8_t)__rev0, 18);
33709  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33710  return __ret;
33711}
33712#endif
33713
33714#ifdef __LITTLE_ENDIAN__
33715__ai int32x4_t vcvtnq_s32_f32(float32x4_t __p0) {
33716  int32x4_t __ret;
33717  __ret = (int32x4_t) __builtin_neon_vcvtnq_s32_v((int8x16_t)__p0, 34);
33718  return __ret;
33719}
33720#else
33721__ai int32x4_t vcvtnq_s32_f32(float32x4_t __p0) {
33722  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33723  int32x4_t __ret;
33724  __ret = (int32x4_t) __builtin_neon_vcvtnq_s32_v((int8x16_t)__rev0, 34);
33725  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33726  return __ret;
33727}
33728#endif
33729
33730#ifdef __LITTLE_ENDIAN__
33731__ai int32x2_t vcvtn_s32_f32(float32x2_t __p0) {
33732  int32x2_t __ret;
33733  __ret = (int32x2_t) __builtin_neon_vcvtn_s32_v((int8x8_t)__p0, 2);
33734  return __ret;
33735}
33736#else
33737__ai int32x2_t vcvtn_s32_f32(float32x2_t __p0) {
33738  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33739  int32x2_t __ret;
33740  __ret = (int32x2_t) __builtin_neon_vcvtn_s32_v((int8x8_t)__rev0, 2);
33741  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33742  return __ret;
33743}
33744#endif
33745
33746#ifdef __LITTLE_ENDIAN__
33747__ai uint32x4_t vcvtnq_u32_f32(float32x4_t __p0) {
33748  uint32x4_t __ret;
33749  __ret = (uint32x4_t) __builtin_neon_vcvtnq_u32_v((int8x16_t)__p0, 50);
33750  return __ret;
33751}
33752#else
33753__ai uint32x4_t vcvtnq_u32_f32(float32x4_t __p0) {
33754  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33755  uint32x4_t __ret;
33756  __ret = (uint32x4_t) __builtin_neon_vcvtnq_u32_v((int8x16_t)__rev0, 50);
33757  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33758  return __ret;
33759}
33760#endif
33761
33762#ifdef __LITTLE_ENDIAN__
33763__ai uint32x2_t vcvtn_u32_f32(float32x2_t __p0) {
33764  uint32x2_t __ret;
33765  __ret = (uint32x2_t) __builtin_neon_vcvtn_u32_v((int8x8_t)__p0, 18);
33766  return __ret;
33767}
33768#else
33769__ai uint32x2_t vcvtn_u32_f32(float32x2_t __p0) {
33770  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33771  uint32x2_t __ret;
33772  __ret = (uint32x2_t) __builtin_neon_vcvtn_u32_v((int8x8_t)__rev0, 18);
33773  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33774  return __ret;
33775}
33776#endif
33777
33778#ifdef __LITTLE_ENDIAN__
33779__ai int32x4_t vcvtpq_s32_f32(float32x4_t __p0) {
33780  int32x4_t __ret;
33781  __ret = (int32x4_t) __builtin_neon_vcvtpq_s32_v((int8x16_t)__p0, 34);
33782  return __ret;
33783}
33784#else
33785__ai int32x4_t vcvtpq_s32_f32(float32x4_t __p0) {
33786  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33787  int32x4_t __ret;
33788  __ret = (int32x4_t) __builtin_neon_vcvtpq_s32_v((int8x16_t)__rev0, 34);
33789  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33790  return __ret;
33791}
33792#endif
33793
33794#ifdef __LITTLE_ENDIAN__
33795__ai int32x2_t vcvtp_s32_f32(float32x2_t __p0) {
33796  int32x2_t __ret;
33797  __ret = (int32x2_t) __builtin_neon_vcvtp_s32_v((int8x8_t)__p0, 2);
33798  return __ret;
33799}
33800#else
33801__ai int32x2_t vcvtp_s32_f32(float32x2_t __p0) {
33802  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33803  int32x2_t __ret;
33804  __ret = (int32x2_t) __builtin_neon_vcvtp_s32_v((int8x8_t)__rev0, 2);
33805  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33806  return __ret;
33807}
33808#endif
33809
33810#ifdef __LITTLE_ENDIAN__
33811__ai uint32x4_t vcvtpq_u32_f32(float32x4_t __p0) {
33812  uint32x4_t __ret;
33813  __ret = (uint32x4_t) __builtin_neon_vcvtpq_u32_v((int8x16_t)__p0, 50);
33814  return __ret;
33815}
33816#else
33817__ai uint32x4_t vcvtpq_u32_f32(float32x4_t __p0) {
33818  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33819  uint32x4_t __ret;
33820  __ret = (uint32x4_t) __builtin_neon_vcvtpq_u32_v((int8x16_t)__rev0, 50);
33821  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33822  return __ret;
33823}
33824#endif
33825
33826#ifdef __LITTLE_ENDIAN__
33827__ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) {
33828  uint32x2_t __ret;
33829  __ret = (uint32x2_t) __builtin_neon_vcvtp_u32_v((int8x8_t)__p0, 18);
33830  return __ret;
33831}
33832#else
33833__ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) {
33834  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33835  uint32x2_t __ret;
33836  __ret = (uint32x2_t) __builtin_neon_vcvtp_u32_v((int8x8_t)__rev0, 18);
33837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33838  return __ret;
33839}
33840#endif
33841
33842#endif
33843#if __ARM_ARCH >= 8 && defined(__aarch64__)
33844#ifdef __LITTLE_ENDIAN__
33845__ai int64x2_t vcvtaq_s64_f64(float64x2_t __p0) {
33846  int64x2_t __ret;
33847  __ret = (int64x2_t) __builtin_neon_vcvtaq_s64_v((int8x16_t)__p0, 35);
33848  return __ret;
33849}
33850#else
33851__ai int64x2_t vcvtaq_s64_f64(float64x2_t __p0) {
33852  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33853  int64x2_t __ret;
33854  __ret = (int64x2_t) __builtin_neon_vcvtaq_s64_v((int8x16_t)__rev0, 35);
33855  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33856  return __ret;
33857}
33858#endif
33859
33860#ifdef __LITTLE_ENDIAN__
33861__ai int64x1_t vcvta_s64_f64(float64x1_t __p0) {
33862  int64x1_t __ret;
33863  __ret = (int64x1_t) __builtin_neon_vcvta_s64_v((int8x8_t)__p0, 3);
33864  return __ret;
33865}
33866#else
33867__ai int64x1_t vcvta_s64_f64(float64x1_t __p0) {
33868  int64x1_t __ret;
33869  __ret = (int64x1_t) __builtin_neon_vcvta_s64_v((int8x8_t)__p0, 3);
33870  return __ret;
33871}
33872#endif
33873
33874#ifdef __LITTLE_ENDIAN__
33875__ai uint64x2_t vcvtaq_u64_f64(float64x2_t __p0) {
33876  uint64x2_t __ret;
33877  __ret = (uint64x2_t) __builtin_neon_vcvtaq_u64_v((int8x16_t)__p0, 51);
33878  return __ret;
33879}
33880#else
33881__ai uint64x2_t vcvtaq_u64_f64(float64x2_t __p0) {
33882  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33883  uint64x2_t __ret;
33884  __ret = (uint64x2_t) __builtin_neon_vcvtaq_u64_v((int8x16_t)__rev0, 51);
33885  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33886  return __ret;
33887}
33888#endif
33889
33890#ifdef __LITTLE_ENDIAN__
33891__ai uint64x1_t vcvta_u64_f64(float64x1_t __p0) {
33892  uint64x1_t __ret;
33893  __ret = (uint64x1_t) __builtin_neon_vcvta_u64_v((int8x8_t)__p0, 19);
33894  return __ret;
33895}
33896#else
33897__ai uint64x1_t vcvta_u64_f64(float64x1_t __p0) {
33898  uint64x1_t __ret;
33899  __ret = (uint64x1_t) __builtin_neon_vcvta_u64_v((int8x8_t)__p0, 19);
33900  return __ret;
33901}
33902#endif
33903
33904#ifdef __LITTLE_ENDIAN__
33905__ai int64x2_t vcvtmq_s64_f64(float64x2_t __p0) {
33906  int64x2_t __ret;
33907  __ret = (int64x2_t) __builtin_neon_vcvtmq_s64_v((int8x16_t)__p0, 35);
33908  return __ret;
33909}
33910#else
33911__ai int64x2_t vcvtmq_s64_f64(float64x2_t __p0) {
33912  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33913  int64x2_t __ret;
33914  __ret = (int64x2_t) __builtin_neon_vcvtmq_s64_v((int8x16_t)__rev0, 35);
33915  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33916  return __ret;
33917}
33918#endif
33919
33920#ifdef __LITTLE_ENDIAN__
33921__ai int64x1_t vcvtm_s64_f64(float64x1_t __p0) {
33922  int64x1_t __ret;
33923  __ret = (int64x1_t) __builtin_neon_vcvtm_s64_v((int8x8_t)__p0, 3);
33924  return __ret;
33925}
33926#else
33927__ai int64x1_t vcvtm_s64_f64(float64x1_t __p0) {
33928  int64x1_t __ret;
33929  __ret = (int64x1_t) __builtin_neon_vcvtm_s64_v((int8x8_t)__p0, 3);
33930  return __ret;
33931}
33932#endif
33933
33934#ifdef __LITTLE_ENDIAN__
33935__ai uint64x2_t vcvtmq_u64_f64(float64x2_t __p0) {
33936  uint64x2_t __ret;
33937  __ret = (uint64x2_t) __builtin_neon_vcvtmq_u64_v((int8x16_t)__p0, 51);
33938  return __ret;
33939}
33940#else
33941__ai uint64x2_t vcvtmq_u64_f64(float64x2_t __p0) {
33942  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33943  uint64x2_t __ret;
33944  __ret = (uint64x2_t) __builtin_neon_vcvtmq_u64_v((int8x16_t)__rev0, 51);
33945  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33946  return __ret;
33947}
33948#endif
33949
33950#ifdef __LITTLE_ENDIAN__
33951__ai uint64x1_t vcvtm_u64_f64(float64x1_t __p0) {
33952  uint64x1_t __ret;
33953  __ret = (uint64x1_t) __builtin_neon_vcvtm_u64_v((int8x8_t)__p0, 19);
33954  return __ret;
33955}
33956#else
33957__ai uint64x1_t vcvtm_u64_f64(float64x1_t __p0) {
33958  uint64x1_t __ret;
33959  __ret = (uint64x1_t) __builtin_neon_vcvtm_u64_v((int8x8_t)__p0, 19);
33960  return __ret;
33961}
33962#endif
33963
33964#ifdef __LITTLE_ENDIAN__
33965__ai int64x2_t vcvtnq_s64_f64(float64x2_t __p0) {
33966  int64x2_t __ret;
33967  __ret = (int64x2_t) __builtin_neon_vcvtnq_s64_v((int8x16_t)__p0, 35);
33968  return __ret;
33969}
33970#else
33971__ai int64x2_t vcvtnq_s64_f64(float64x2_t __p0) {
33972  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33973  int64x2_t __ret;
33974  __ret = (int64x2_t) __builtin_neon_vcvtnq_s64_v((int8x16_t)__rev0, 35);
33975  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33976  return __ret;
33977}
33978#endif
33979
33980#ifdef __LITTLE_ENDIAN__
33981__ai int64x1_t vcvtn_s64_f64(float64x1_t __p0) {
33982  int64x1_t __ret;
33983  __ret = (int64x1_t) __builtin_neon_vcvtn_s64_v((int8x8_t)__p0, 3);
33984  return __ret;
33985}
33986#else
33987__ai int64x1_t vcvtn_s64_f64(float64x1_t __p0) {
33988  int64x1_t __ret;
33989  __ret = (int64x1_t) __builtin_neon_vcvtn_s64_v((int8x8_t)__p0, 3);
33990  return __ret;
33991}
33992#endif
33993
33994#ifdef __LITTLE_ENDIAN__
33995__ai uint64x2_t vcvtnq_u64_f64(float64x2_t __p0) {
33996  uint64x2_t __ret;
33997  __ret = (uint64x2_t) __builtin_neon_vcvtnq_u64_v((int8x16_t)__p0, 51);
33998  return __ret;
33999}
34000#else
34001__ai uint64x2_t vcvtnq_u64_f64(float64x2_t __p0) {
34002  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34003  uint64x2_t __ret;
34004  __ret = (uint64x2_t) __builtin_neon_vcvtnq_u64_v((int8x16_t)__rev0, 51);
34005  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34006  return __ret;
34007}
34008#endif
34009
34010#ifdef __LITTLE_ENDIAN__
34011__ai uint64x1_t vcvtn_u64_f64(float64x1_t __p0) {
34012  uint64x1_t __ret;
34013  __ret = (uint64x1_t) __builtin_neon_vcvtn_u64_v((int8x8_t)__p0, 19);
34014  return __ret;
34015}
34016#else
34017__ai uint64x1_t vcvtn_u64_f64(float64x1_t __p0) {
34018  uint64x1_t __ret;
34019  __ret = (uint64x1_t) __builtin_neon_vcvtn_u64_v((int8x8_t)__p0, 19);
34020  return __ret;
34021}
34022#endif
34023
34024#ifdef __LITTLE_ENDIAN__
34025__ai int64x2_t vcvtpq_s64_f64(float64x2_t __p0) {
34026  int64x2_t __ret;
34027  __ret = (int64x2_t) __builtin_neon_vcvtpq_s64_v((int8x16_t)__p0, 35);
34028  return __ret;
34029}
34030#else
34031__ai int64x2_t vcvtpq_s64_f64(float64x2_t __p0) {
34032  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34033  int64x2_t __ret;
34034  __ret = (int64x2_t) __builtin_neon_vcvtpq_s64_v((int8x16_t)__rev0, 35);
34035  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34036  return __ret;
34037}
34038#endif
34039
34040#ifdef __LITTLE_ENDIAN__
34041__ai int64x1_t vcvtp_s64_f64(float64x1_t __p0) {
34042  int64x1_t __ret;
34043  __ret = (int64x1_t) __builtin_neon_vcvtp_s64_v((int8x8_t)__p0, 3);
34044  return __ret;
34045}
34046#else
34047__ai int64x1_t vcvtp_s64_f64(float64x1_t __p0) {
34048  int64x1_t __ret;
34049  __ret = (int64x1_t) __builtin_neon_vcvtp_s64_v((int8x8_t)__p0, 3);
34050  return __ret;
34051}
34052#endif
34053
34054#ifdef __LITTLE_ENDIAN__
34055__ai uint64x2_t vcvtpq_u64_f64(float64x2_t __p0) {
34056  uint64x2_t __ret;
34057  __ret = (uint64x2_t) __builtin_neon_vcvtpq_u64_v((int8x16_t)__p0, 51);
34058  return __ret;
34059}
34060#else
34061__ai uint64x2_t vcvtpq_u64_f64(float64x2_t __p0) {
34062  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34063  uint64x2_t __ret;
34064  __ret = (uint64x2_t) __builtin_neon_vcvtpq_u64_v((int8x16_t)__rev0, 51);
34065  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34066  return __ret;
34067}
34068#endif
34069
34070#ifdef __LITTLE_ENDIAN__
34071__ai uint64x1_t vcvtp_u64_f64(float64x1_t __p0) {
34072  uint64x1_t __ret;
34073  __ret = (uint64x1_t) __builtin_neon_vcvtp_u64_v((int8x8_t)__p0, 19);
34074  return __ret;
34075}
34076#else
34077__ai uint64x1_t vcvtp_u64_f64(float64x1_t __p0) {
34078  uint64x1_t __ret;
34079  __ret = (uint64x1_t) __builtin_neon_vcvtp_u64_v((int8x8_t)__p0, 19);
34080  return __ret;
34081}
34082#endif
34083
34084#ifdef __LITTLE_ENDIAN__
34085__ai poly8x8_t vreinterpret_p8_p64(poly64x1_t __p0) {
34086  poly8x8_t __ret;
34087  __ret = (poly8x8_t)(__p0);
34088  return __ret;
34089}
34090#else
34091__ai poly8x8_t vreinterpret_p8_p64(poly64x1_t __p0) {
34092  poly8x8_t __ret;
34093  __ret = (poly8x8_t)(__p0);
34094  return __ret;
34095}
34096#endif
34097
34098#ifdef __LITTLE_ENDIAN__
34099__ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
34100  poly8x8_t __ret;
34101  __ret = (poly8x8_t)(__p0);
34102  return __ret;
34103}
34104#else
34105__ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
34106  poly8x8_t __ret;
34107  __ret = (poly8x8_t)(__p0);
34108  return __ret;
34109}
34110#endif
34111
34112#ifdef __LITTLE_ENDIAN__
34113__ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
34114  poly8x8_t __ret;
34115  __ret = (poly8x8_t)(__p0);
34116  return __ret;
34117}
34118#else
34119__ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
34120  poly8x8_t __ret;
34121  __ret = (poly8x8_t)(__p0);
34122  return __ret;
34123}
34124#endif
34125
34126#ifdef __LITTLE_ENDIAN__
34127__ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
34128  poly8x8_t __ret;
34129  __ret = (poly8x8_t)(__p0);
34130  return __ret;
34131}
34132#else
34133__ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
34134  poly8x8_t __ret;
34135  __ret = (poly8x8_t)(__p0);
34136  return __ret;
34137}
34138#endif
34139
34140#ifdef __LITTLE_ENDIAN__
34141__ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
34142  poly8x8_t __ret;
34143  __ret = (poly8x8_t)(__p0);
34144  return __ret;
34145}
34146#else
34147__ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
34148  poly8x8_t __ret;
34149  __ret = (poly8x8_t)(__p0);
34150  return __ret;
34151}
34152#endif
34153
34154#ifdef __LITTLE_ENDIAN__
34155__ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
34156  poly8x8_t __ret;
34157  __ret = (poly8x8_t)(__p0);
34158  return __ret;
34159}
34160#else
34161__ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
34162  poly8x8_t __ret;
34163  __ret = (poly8x8_t)(__p0);
34164  return __ret;
34165}
34166#endif
34167
34168#ifdef __LITTLE_ENDIAN__
34169__ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
34170  poly8x8_t __ret;
34171  __ret = (poly8x8_t)(__p0);
34172  return __ret;
34173}
34174#else
34175__ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
34176  poly8x8_t __ret;
34177  __ret = (poly8x8_t)(__p0);
34178  return __ret;
34179}
34180#endif
34181
34182#ifdef __LITTLE_ENDIAN__
34183__ai poly8x8_t vreinterpret_p8_f64(float64x1_t __p0) {
34184  poly8x8_t __ret;
34185  __ret = (poly8x8_t)(__p0);
34186  return __ret;
34187}
34188#else
34189__ai poly8x8_t vreinterpret_p8_f64(float64x1_t __p0) {
34190  poly8x8_t __ret;
34191  __ret = (poly8x8_t)(__p0);
34192  return __ret;
34193}
34194#endif
34195
34196#ifdef __LITTLE_ENDIAN__
34197__ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
34198  poly8x8_t __ret;
34199  __ret = (poly8x8_t)(__p0);
34200  return __ret;
34201}
34202#else
34203__ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
34204  poly8x8_t __ret;
34205  __ret = (poly8x8_t)(__p0);
34206  return __ret;
34207}
34208#endif
34209
34210#ifdef __LITTLE_ENDIAN__
34211__ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
34212  poly8x8_t __ret;
34213  __ret = (poly8x8_t)(__p0);
34214  return __ret;
34215}
34216#else
34217__ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
34218  poly8x8_t __ret;
34219  __ret = (poly8x8_t)(__p0);
34220  return __ret;
34221}
34222#endif
34223
34224#ifdef __LITTLE_ENDIAN__
34225__ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
34226  poly8x8_t __ret;
34227  __ret = (poly8x8_t)(__p0);
34228  return __ret;
34229}
34230#else
34231__ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
34232  poly8x8_t __ret;
34233  __ret = (poly8x8_t)(__p0);
34234  return __ret;
34235}
34236#endif
34237
34238#ifdef __LITTLE_ENDIAN__
34239__ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
34240  poly8x8_t __ret;
34241  __ret = (poly8x8_t)(__p0);
34242  return __ret;
34243}
34244#else
34245__ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
34246  poly8x8_t __ret;
34247  __ret = (poly8x8_t)(__p0);
34248  return __ret;
34249}
34250#endif
34251
34252#ifdef __LITTLE_ENDIAN__
34253__ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
34254  poly8x8_t __ret;
34255  __ret = (poly8x8_t)(__p0);
34256  return __ret;
34257}
34258#else
34259__ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
34260  poly8x8_t __ret;
34261  __ret = (poly8x8_t)(__p0);
34262  return __ret;
34263}
34264#endif
34265
34266#ifdef __LITTLE_ENDIAN__
34267__ai poly64x1_t vreinterpret_p64_p8(poly8x8_t __p0) {
34268  poly64x1_t __ret;
34269  __ret = (poly64x1_t)(__p0);
34270  return __ret;
34271}
34272#else
34273__ai poly64x1_t vreinterpret_p64_p8(poly8x8_t __p0) {
34274  poly64x1_t __ret;
34275  __ret = (poly64x1_t)(__p0);
34276  return __ret;
34277}
34278#endif
34279
34280#ifdef __LITTLE_ENDIAN__
34281__ai poly64x1_t vreinterpret_p64_p16(poly16x4_t __p0) {
34282  poly64x1_t __ret;
34283  __ret = (poly64x1_t)(__p0);
34284  return __ret;
34285}
34286#else
34287__ai poly64x1_t vreinterpret_p64_p16(poly16x4_t __p0) {
34288  poly64x1_t __ret;
34289  __ret = (poly64x1_t)(__p0);
34290  return __ret;
34291}
34292#endif
34293
34294#ifdef __LITTLE_ENDIAN__
34295__ai poly64x1_t vreinterpret_p64_u8(uint8x8_t __p0) {
34296  poly64x1_t __ret;
34297  __ret = (poly64x1_t)(__p0);
34298  return __ret;
34299}
34300#else
34301__ai poly64x1_t vreinterpret_p64_u8(uint8x8_t __p0) {
34302  poly64x1_t __ret;
34303  __ret = (poly64x1_t)(__p0);
34304  return __ret;
34305}
34306#endif
34307
34308#ifdef __LITTLE_ENDIAN__
34309__ai poly64x1_t vreinterpret_p64_u32(uint32x2_t __p0) {
34310  poly64x1_t __ret;
34311  __ret = (poly64x1_t)(__p0);
34312  return __ret;
34313}
34314#else
34315__ai poly64x1_t vreinterpret_p64_u32(uint32x2_t __p0) {
34316  poly64x1_t __ret;
34317  __ret = (poly64x1_t)(__p0);
34318  return __ret;
34319}
34320#endif
34321
34322#ifdef __LITTLE_ENDIAN__
34323__ai poly64x1_t vreinterpret_p64_u64(uint64x1_t __p0) {
34324  poly64x1_t __ret;
34325  __ret = (poly64x1_t)(__p0);
34326  return __ret;
34327}
34328#else
34329__ai poly64x1_t vreinterpret_p64_u64(uint64x1_t __p0) {
34330  poly64x1_t __ret;
34331  __ret = (poly64x1_t)(__p0);
34332  return __ret;
34333}
34334#endif
34335
34336#ifdef __LITTLE_ENDIAN__
34337__ai poly64x1_t vreinterpret_p64_u16(uint16x4_t __p0) {
34338  poly64x1_t __ret;
34339  __ret = (poly64x1_t)(__p0);
34340  return __ret;
34341}
34342#else
34343__ai poly64x1_t vreinterpret_p64_u16(uint16x4_t __p0) {
34344  poly64x1_t __ret;
34345  __ret = (poly64x1_t)(__p0);
34346  return __ret;
34347}
34348#endif
34349
34350#ifdef __LITTLE_ENDIAN__
34351__ai poly64x1_t vreinterpret_p64_s8(int8x8_t __p0) {
34352  poly64x1_t __ret;
34353  __ret = (poly64x1_t)(__p0);
34354  return __ret;
34355}
34356#else
34357__ai poly64x1_t vreinterpret_p64_s8(int8x8_t __p0) {
34358  poly64x1_t __ret;
34359  __ret = (poly64x1_t)(__p0);
34360  return __ret;
34361}
34362#endif
34363
34364#ifdef __LITTLE_ENDIAN__
34365__ai poly64x1_t vreinterpret_p64_f64(float64x1_t __p0) {
34366  poly64x1_t __ret;
34367  __ret = (poly64x1_t)(__p0);
34368  return __ret;
34369}
34370#else
34371__ai poly64x1_t vreinterpret_p64_f64(float64x1_t __p0) {
34372  poly64x1_t __ret;
34373  __ret = (poly64x1_t)(__p0);
34374  return __ret;
34375}
34376#endif
34377
34378#ifdef __LITTLE_ENDIAN__
34379__ai poly64x1_t vreinterpret_p64_f32(float32x2_t __p0) {
34380  poly64x1_t __ret;
34381  __ret = (poly64x1_t)(__p0);
34382  return __ret;
34383}
34384#else
34385__ai poly64x1_t vreinterpret_p64_f32(float32x2_t __p0) {
34386  poly64x1_t __ret;
34387  __ret = (poly64x1_t)(__p0);
34388  return __ret;
34389}
34390#endif
34391
34392#ifdef __LITTLE_ENDIAN__
34393__ai poly64x1_t vreinterpret_p64_f16(float16x4_t __p0) {
34394  poly64x1_t __ret;
34395  __ret = (poly64x1_t)(__p0);
34396  return __ret;
34397}
34398#else
34399__ai poly64x1_t vreinterpret_p64_f16(float16x4_t __p0) {
34400  poly64x1_t __ret;
34401  __ret = (poly64x1_t)(__p0);
34402  return __ret;
34403}
34404#endif
34405
34406#ifdef __LITTLE_ENDIAN__
34407__ai poly64x1_t vreinterpret_p64_s32(int32x2_t __p0) {
34408  poly64x1_t __ret;
34409  __ret = (poly64x1_t)(__p0);
34410  return __ret;
34411}
34412#else
34413__ai poly64x1_t vreinterpret_p64_s32(int32x2_t __p0) {
34414  poly64x1_t __ret;
34415  __ret = (poly64x1_t)(__p0);
34416  return __ret;
34417}
34418#endif
34419
34420#ifdef __LITTLE_ENDIAN__
34421__ai poly64x1_t vreinterpret_p64_s64(int64x1_t __p0) {
34422  poly64x1_t __ret;
34423  __ret = (poly64x1_t)(__p0);
34424  return __ret;
34425}
34426#else
34427__ai poly64x1_t vreinterpret_p64_s64(int64x1_t __p0) {
34428  poly64x1_t __ret;
34429  __ret = (poly64x1_t)(__p0);
34430  return __ret;
34431}
34432#endif
34433
34434#ifdef __LITTLE_ENDIAN__
34435__ai poly64x1_t vreinterpret_p64_s16(int16x4_t __p0) {
34436  poly64x1_t __ret;
34437  __ret = (poly64x1_t)(__p0);
34438  return __ret;
34439}
34440#else
34441__ai poly64x1_t vreinterpret_p64_s16(int16x4_t __p0) {
34442  poly64x1_t __ret;
34443  __ret = (poly64x1_t)(__p0);
34444  return __ret;
34445}
34446#endif
34447
34448#ifdef __LITTLE_ENDIAN__
34449__ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
34450  poly16x4_t __ret;
34451  __ret = (poly16x4_t)(__p0);
34452  return __ret;
34453}
34454#else
34455__ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
34456  poly16x4_t __ret;
34457  __ret = (poly16x4_t)(__p0);
34458  return __ret;
34459}
34460#endif
34461
34462#ifdef __LITTLE_ENDIAN__
34463__ai poly16x4_t vreinterpret_p16_p64(poly64x1_t __p0) {
34464  poly16x4_t __ret;
34465  __ret = (poly16x4_t)(__p0);
34466  return __ret;
34467}
34468#else
34469__ai poly16x4_t vreinterpret_p16_p64(poly64x1_t __p0) {
34470  poly16x4_t __ret;
34471  __ret = (poly16x4_t)(__p0);
34472  return __ret;
34473}
34474#endif
34475
34476#ifdef __LITTLE_ENDIAN__
34477__ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
34478  poly16x4_t __ret;
34479  __ret = (poly16x4_t)(__p0);
34480  return __ret;
34481}
34482#else
34483__ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
34484  poly16x4_t __ret;
34485  __ret = (poly16x4_t)(__p0);
34486  return __ret;
34487}
34488#endif
34489
34490#ifdef __LITTLE_ENDIAN__
34491__ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
34492  poly16x4_t __ret;
34493  __ret = (poly16x4_t)(__p0);
34494  return __ret;
34495}
34496#else
34497__ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
34498  poly16x4_t __ret;
34499  __ret = (poly16x4_t)(__p0);
34500  return __ret;
34501}
34502#endif
34503
34504#ifdef __LITTLE_ENDIAN__
34505__ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
34506  poly16x4_t __ret;
34507  __ret = (poly16x4_t)(__p0);
34508  return __ret;
34509}
34510#else
34511__ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
34512  poly16x4_t __ret;
34513  __ret = (poly16x4_t)(__p0);
34514  return __ret;
34515}
34516#endif
34517
34518#ifdef __LITTLE_ENDIAN__
34519__ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
34520  poly16x4_t __ret;
34521  __ret = (poly16x4_t)(__p0);
34522  return __ret;
34523}
34524#else
34525__ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
34526  poly16x4_t __ret;
34527  __ret = (poly16x4_t)(__p0);
34528  return __ret;
34529}
34530#endif
34531
34532#ifdef __LITTLE_ENDIAN__
34533__ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
34534  poly16x4_t __ret;
34535  __ret = (poly16x4_t)(__p0);
34536  return __ret;
34537}
34538#else
34539__ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
34540  poly16x4_t __ret;
34541  __ret = (poly16x4_t)(__p0);
34542  return __ret;
34543}
34544#endif
34545
34546#ifdef __LITTLE_ENDIAN__
34547__ai poly16x4_t vreinterpret_p16_f64(float64x1_t __p0) {
34548  poly16x4_t __ret;
34549  __ret = (poly16x4_t)(__p0);
34550  return __ret;
34551}
34552#else
34553__ai poly16x4_t vreinterpret_p16_f64(float64x1_t __p0) {
34554  poly16x4_t __ret;
34555  __ret = (poly16x4_t)(__p0);
34556  return __ret;
34557}
34558#endif
34559
34560#ifdef __LITTLE_ENDIAN__
34561__ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
34562  poly16x4_t __ret;
34563  __ret = (poly16x4_t)(__p0);
34564  return __ret;
34565}
34566#else
34567__ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
34568  poly16x4_t __ret;
34569  __ret = (poly16x4_t)(__p0);
34570  return __ret;
34571}
34572#endif
34573
34574#ifdef __LITTLE_ENDIAN__
34575__ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
34576  poly16x4_t __ret;
34577  __ret = (poly16x4_t)(__p0);
34578  return __ret;
34579}
34580#else
34581__ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
34582  poly16x4_t __ret;
34583  __ret = (poly16x4_t)(__p0);
34584  return __ret;
34585}
34586#endif
34587
34588#ifdef __LITTLE_ENDIAN__
34589__ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
34590  poly16x4_t __ret;
34591  __ret = (poly16x4_t)(__p0);
34592  return __ret;
34593}
34594#else
34595__ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
34596  poly16x4_t __ret;
34597  __ret = (poly16x4_t)(__p0);
34598  return __ret;
34599}
34600#endif
34601
34602#ifdef __LITTLE_ENDIAN__
34603__ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
34604  poly16x4_t __ret;
34605  __ret = (poly16x4_t)(__p0);
34606  return __ret;
34607}
34608#else
34609__ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
34610  poly16x4_t __ret;
34611  __ret = (poly16x4_t)(__p0);
34612  return __ret;
34613}
34614#endif
34615
34616#ifdef __LITTLE_ENDIAN__
34617__ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
34618  poly16x4_t __ret;
34619  __ret = (poly16x4_t)(__p0);
34620  return __ret;
34621}
34622#else
34623__ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
34624  poly16x4_t __ret;
34625  __ret = (poly16x4_t)(__p0);
34626  return __ret;
34627}
34628#endif
34629
34630#ifdef __LITTLE_ENDIAN__
34631__ai poly8x16_t vreinterpretq_p8_p128(poly128_t __p0) {
34632  poly8x16_t __ret;
34633  __ret = (poly8x16_t)(__p0);
34634  return __ret;
34635}
34636#else
34637__ai poly8x16_t vreinterpretq_p8_p128(poly128_t __p0) {
34638  poly8x16_t __ret;
34639  __ret = (poly8x16_t)(__p0);
34640  return __ret;
34641}
34642#endif
34643
34644#ifdef __LITTLE_ENDIAN__
34645__ai poly8x16_t vreinterpretq_p8_p64(poly64x2_t __p0) {
34646  poly8x16_t __ret;
34647  __ret = (poly8x16_t)(__p0);
34648  return __ret;
34649}
34650#else
34651__ai poly8x16_t vreinterpretq_p8_p64(poly64x2_t __p0) {
34652  poly8x16_t __ret;
34653  __ret = (poly8x16_t)(__p0);
34654  return __ret;
34655}
34656#endif
34657
34658#ifdef __LITTLE_ENDIAN__
34659__ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
34660  poly8x16_t __ret;
34661  __ret = (poly8x16_t)(__p0);
34662  return __ret;
34663}
34664#else
34665__ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
34666  poly8x16_t __ret;
34667  __ret = (poly8x16_t)(__p0);
34668  return __ret;
34669}
34670#endif
34671
34672#ifdef __LITTLE_ENDIAN__
34673__ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
34674  poly8x16_t __ret;
34675  __ret = (poly8x16_t)(__p0);
34676  return __ret;
34677}
34678#else
34679__ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
34680  poly8x16_t __ret;
34681  __ret = (poly8x16_t)(__p0);
34682  return __ret;
34683}
34684#endif
34685
34686#ifdef __LITTLE_ENDIAN__
34687__ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
34688  poly8x16_t __ret;
34689  __ret = (poly8x16_t)(__p0);
34690  return __ret;
34691}
34692#else
34693__ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
34694  poly8x16_t __ret;
34695  __ret = (poly8x16_t)(__p0);
34696  return __ret;
34697}
34698#endif
34699
34700#ifdef __LITTLE_ENDIAN__
34701__ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
34702  poly8x16_t __ret;
34703  __ret = (poly8x16_t)(__p0);
34704  return __ret;
34705}
34706#else
34707__ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
34708  poly8x16_t __ret;
34709  __ret = (poly8x16_t)(__p0);
34710  return __ret;
34711}
34712#endif
34713
34714#ifdef __LITTLE_ENDIAN__
34715__ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
34716  poly8x16_t __ret;
34717  __ret = (poly8x16_t)(__p0);
34718  return __ret;
34719}
34720#else
34721__ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
34722  poly8x16_t __ret;
34723  __ret = (poly8x16_t)(__p0);
34724  return __ret;
34725}
34726#endif
34727
34728#ifdef __LITTLE_ENDIAN__
34729__ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
34730  poly8x16_t __ret;
34731  __ret = (poly8x16_t)(__p0);
34732  return __ret;
34733}
34734#else
34735__ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
34736  poly8x16_t __ret;
34737  __ret = (poly8x16_t)(__p0);
34738  return __ret;
34739}
34740#endif
34741
34742#ifdef __LITTLE_ENDIAN__
34743__ai poly8x16_t vreinterpretq_p8_f64(float64x2_t __p0) {
34744  poly8x16_t __ret;
34745  __ret = (poly8x16_t)(__p0);
34746  return __ret;
34747}
34748#else
34749__ai poly8x16_t vreinterpretq_p8_f64(float64x2_t __p0) {
34750  poly8x16_t __ret;
34751  __ret = (poly8x16_t)(__p0);
34752  return __ret;
34753}
34754#endif
34755
34756#ifdef __LITTLE_ENDIAN__
34757__ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
34758  poly8x16_t __ret;
34759  __ret = (poly8x16_t)(__p0);
34760  return __ret;
34761}
34762#else
34763__ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
34764  poly8x16_t __ret;
34765  __ret = (poly8x16_t)(__p0);
34766  return __ret;
34767}
34768#endif
34769
34770#ifdef __LITTLE_ENDIAN__
34771__ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
34772  poly8x16_t __ret;
34773  __ret = (poly8x16_t)(__p0);
34774  return __ret;
34775}
34776#else
34777__ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
34778  poly8x16_t __ret;
34779  __ret = (poly8x16_t)(__p0);
34780  return __ret;
34781}
34782#endif
34783
34784#ifdef __LITTLE_ENDIAN__
34785__ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
34786  poly8x16_t __ret;
34787  __ret = (poly8x16_t)(__p0);
34788  return __ret;
34789}
34790#else
34791__ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
34792  poly8x16_t __ret;
34793  __ret = (poly8x16_t)(__p0);
34794  return __ret;
34795}
34796#endif
34797
34798#ifdef __LITTLE_ENDIAN__
34799__ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
34800  poly8x16_t __ret;
34801  __ret = (poly8x16_t)(__p0);
34802  return __ret;
34803}
34804#else
34805__ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
34806  poly8x16_t __ret;
34807  __ret = (poly8x16_t)(__p0);
34808  return __ret;
34809}
34810#endif
34811
34812#ifdef __LITTLE_ENDIAN__
34813__ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
34814  poly8x16_t __ret;
34815  __ret = (poly8x16_t)(__p0);
34816  return __ret;
34817}
34818#else
34819__ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
34820  poly8x16_t __ret;
34821  __ret = (poly8x16_t)(__p0);
34822  return __ret;
34823}
34824#endif
34825
34826#ifdef __LITTLE_ENDIAN__
34827__ai poly128_t vreinterpretq_p128_p8(poly8x16_t __p0) {
34828  poly128_t __ret;
34829  __ret = (poly128_t)(__p0);
34830  return __ret;
34831}
34832#else
34833__ai poly128_t vreinterpretq_p128_p8(poly8x16_t __p0) {
34834  poly128_t __ret;
34835  __ret = (poly128_t)(__p0);
34836  return __ret;
34837}
34838#endif
34839
34840#ifdef __LITTLE_ENDIAN__
34841__ai poly128_t vreinterpretq_p128_p64(poly64x2_t __p0) {
34842  poly128_t __ret;
34843  __ret = (poly128_t)(__p0);
34844  return __ret;
34845}
34846#else
34847__ai poly128_t vreinterpretq_p128_p64(poly64x2_t __p0) {
34848  poly128_t __ret;
34849  __ret = (poly128_t)(__p0);
34850  return __ret;
34851}
34852#endif
34853
34854#ifdef __LITTLE_ENDIAN__
34855__ai poly128_t vreinterpretq_p128_p16(poly16x8_t __p0) {
34856  poly128_t __ret;
34857  __ret = (poly128_t)(__p0);
34858  return __ret;
34859}
34860#else
34861__ai poly128_t vreinterpretq_p128_p16(poly16x8_t __p0) {
34862  poly128_t __ret;
34863  __ret = (poly128_t)(__p0);
34864  return __ret;
34865}
34866#endif
34867
34868#ifdef __LITTLE_ENDIAN__
34869__ai poly128_t vreinterpretq_p128_u8(uint8x16_t __p0) {
34870  poly128_t __ret;
34871  __ret = (poly128_t)(__p0);
34872  return __ret;
34873}
34874#else
34875__ai poly128_t vreinterpretq_p128_u8(uint8x16_t __p0) {
34876  poly128_t __ret;
34877  __ret = (poly128_t)(__p0);
34878  return __ret;
34879}
34880#endif
34881
34882#ifdef __LITTLE_ENDIAN__
34883__ai poly128_t vreinterpretq_p128_u32(uint32x4_t __p0) {
34884  poly128_t __ret;
34885  __ret = (poly128_t)(__p0);
34886  return __ret;
34887}
34888#else
34889__ai poly128_t vreinterpretq_p128_u32(uint32x4_t __p0) {
34890  poly128_t __ret;
34891  __ret = (poly128_t)(__p0);
34892  return __ret;
34893}
34894#endif
34895
34896#ifdef __LITTLE_ENDIAN__
34897__ai poly128_t vreinterpretq_p128_u64(uint64x2_t __p0) {
34898  poly128_t __ret;
34899  __ret = (poly128_t)(__p0);
34900  return __ret;
34901}
34902#else
34903__ai poly128_t vreinterpretq_p128_u64(uint64x2_t __p0) {
34904  poly128_t __ret;
34905  __ret = (poly128_t)(__p0);
34906  return __ret;
34907}
34908#endif
34909
34910#ifdef __LITTLE_ENDIAN__
34911__ai poly128_t vreinterpretq_p128_u16(uint16x8_t __p0) {
34912  poly128_t __ret;
34913  __ret = (poly128_t)(__p0);
34914  return __ret;
34915}
34916#else
34917__ai poly128_t vreinterpretq_p128_u16(uint16x8_t __p0) {
34918  poly128_t __ret;
34919  __ret = (poly128_t)(__p0);
34920  return __ret;
34921}
34922#endif
34923
34924#ifdef __LITTLE_ENDIAN__
34925__ai poly128_t vreinterpretq_p128_s8(int8x16_t __p0) {
34926  poly128_t __ret;
34927  __ret = (poly128_t)(__p0);
34928  return __ret;
34929}
34930#else
34931__ai poly128_t vreinterpretq_p128_s8(int8x16_t __p0) {
34932  poly128_t __ret;
34933  __ret = (poly128_t)(__p0);
34934  return __ret;
34935}
34936#endif
34937
34938#ifdef __LITTLE_ENDIAN__
34939__ai poly128_t vreinterpretq_p128_f64(float64x2_t __p0) {
34940  poly128_t __ret;
34941  __ret = (poly128_t)(__p0);
34942  return __ret;
34943}
34944#else
34945__ai poly128_t vreinterpretq_p128_f64(float64x2_t __p0) {
34946  poly128_t __ret;
34947  __ret = (poly128_t)(__p0);
34948  return __ret;
34949}
34950#endif
34951
34952#ifdef __LITTLE_ENDIAN__
34953__ai poly128_t vreinterpretq_p128_f32(float32x4_t __p0) {
34954  poly128_t __ret;
34955  __ret = (poly128_t)(__p0);
34956  return __ret;
34957}
34958#else
34959__ai poly128_t vreinterpretq_p128_f32(float32x4_t __p0) {
34960  poly128_t __ret;
34961  __ret = (poly128_t)(__p0);
34962  return __ret;
34963}
34964#endif
34965
34966#ifdef __LITTLE_ENDIAN__
34967__ai poly128_t vreinterpretq_p128_f16(float16x8_t __p0) {
34968  poly128_t __ret;
34969  __ret = (poly128_t)(__p0);
34970  return __ret;
34971}
34972#else
34973__ai poly128_t vreinterpretq_p128_f16(float16x8_t __p0) {
34974  poly128_t __ret;
34975  __ret = (poly128_t)(__p0);
34976  return __ret;
34977}
34978#endif
34979
34980#ifdef __LITTLE_ENDIAN__
34981__ai poly128_t vreinterpretq_p128_s32(int32x4_t __p0) {
34982  poly128_t __ret;
34983  __ret = (poly128_t)(__p0);
34984  return __ret;
34985}
34986#else
34987__ai poly128_t vreinterpretq_p128_s32(int32x4_t __p0) {
34988  poly128_t __ret;
34989  __ret = (poly128_t)(__p0);
34990  return __ret;
34991}
34992#endif
34993
34994#ifdef __LITTLE_ENDIAN__
34995__ai poly128_t vreinterpretq_p128_s64(int64x2_t __p0) {
34996  poly128_t __ret;
34997  __ret = (poly128_t)(__p0);
34998  return __ret;
34999}
35000#else
35001__ai poly128_t vreinterpretq_p128_s64(int64x2_t __p0) {
35002  poly128_t __ret;
35003  __ret = (poly128_t)(__p0);
35004  return __ret;
35005}
35006#endif
35007
35008#ifdef __LITTLE_ENDIAN__
35009__ai poly128_t vreinterpretq_p128_s16(int16x8_t __p0) {
35010  poly128_t __ret;
35011  __ret = (poly128_t)(__p0);
35012  return __ret;
35013}
35014#else
35015__ai poly128_t vreinterpretq_p128_s16(int16x8_t __p0) {
35016  poly128_t __ret;
35017  __ret = (poly128_t)(__p0);
35018  return __ret;
35019}
35020#endif
35021
35022#ifdef __LITTLE_ENDIAN__
35023__ai poly64x2_t vreinterpretq_p64_p8(poly8x16_t __p0) {
35024  poly64x2_t __ret;
35025  __ret = (poly64x2_t)(__p0);
35026  return __ret;
35027}
35028#else
35029__ai poly64x2_t vreinterpretq_p64_p8(poly8x16_t __p0) {
35030  poly64x2_t __ret;
35031  __ret = (poly64x2_t)(__p0);
35032  return __ret;
35033}
35034#endif
35035
35036#ifdef __LITTLE_ENDIAN__
35037__ai poly64x2_t vreinterpretq_p64_p128(poly128_t __p0) {
35038  poly64x2_t __ret;
35039  __ret = (poly64x2_t)(__p0);
35040  return __ret;
35041}
35042#else
35043__ai poly64x2_t vreinterpretq_p64_p128(poly128_t __p0) {
35044  poly64x2_t __ret;
35045  __ret = (poly64x2_t)(__p0);
35046  return __ret;
35047}
35048#endif
35049
35050#ifdef __LITTLE_ENDIAN__
35051__ai poly64x2_t vreinterpretq_p64_p16(poly16x8_t __p0) {
35052  poly64x2_t __ret;
35053  __ret = (poly64x2_t)(__p0);
35054  return __ret;
35055}
35056#else
35057__ai poly64x2_t vreinterpretq_p64_p16(poly16x8_t __p0) {
35058  poly64x2_t __ret;
35059  __ret = (poly64x2_t)(__p0);
35060  return __ret;
35061}
35062#endif
35063
35064#ifdef __LITTLE_ENDIAN__
35065__ai poly64x2_t vreinterpretq_p64_u8(uint8x16_t __p0) {
35066  poly64x2_t __ret;
35067  __ret = (poly64x2_t)(__p0);
35068  return __ret;
35069}
35070#else
35071__ai poly64x2_t vreinterpretq_p64_u8(uint8x16_t __p0) {
35072  poly64x2_t __ret;
35073  __ret = (poly64x2_t)(__p0);
35074  return __ret;
35075}
35076#endif
35077
35078#ifdef __LITTLE_ENDIAN__
35079__ai poly64x2_t vreinterpretq_p64_u32(uint32x4_t __p0) {
35080  poly64x2_t __ret;
35081  __ret = (poly64x2_t)(__p0);
35082  return __ret;
35083}
35084#else
35085__ai poly64x2_t vreinterpretq_p64_u32(uint32x4_t __p0) {
35086  poly64x2_t __ret;
35087  __ret = (poly64x2_t)(__p0);
35088  return __ret;
35089}
35090#endif
35091
35092#ifdef __LITTLE_ENDIAN__
35093__ai poly64x2_t vreinterpretq_p64_u64(uint64x2_t __p0) {
35094  poly64x2_t __ret;
35095  __ret = (poly64x2_t)(__p0);
35096  return __ret;
35097}
35098#else
35099__ai poly64x2_t vreinterpretq_p64_u64(uint64x2_t __p0) {
35100  poly64x2_t __ret;
35101  __ret = (poly64x2_t)(__p0);
35102  return __ret;
35103}
35104#endif
35105
35106#ifdef __LITTLE_ENDIAN__
35107__ai poly64x2_t vreinterpretq_p64_u16(uint16x8_t __p0) {
35108  poly64x2_t __ret;
35109  __ret = (poly64x2_t)(__p0);
35110  return __ret;
35111}
35112#else
35113__ai poly64x2_t vreinterpretq_p64_u16(uint16x8_t __p0) {
35114  poly64x2_t __ret;
35115  __ret = (poly64x2_t)(__p0);
35116  return __ret;
35117}
35118#endif
35119
35120#ifdef __LITTLE_ENDIAN__
35121__ai poly64x2_t vreinterpretq_p64_s8(int8x16_t __p0) {
35122  poly64x2_t __ret;
35123  __ret = (poly64x2_t)(__p0);
35124  return __ret;
35125}
35126#else
35127__ai poly64x2_t vreinterpretq_p64_s8(int8x16_t __p0) {
35128  poly64x2_t __ret;
35129  __ret = (poly64x2_t)(__p0);
35130  return __ret;
35131}
35132#endif
35133
35134#ifdef __LITTLE_ENDIAN__
35135__ai poly64x2_t vreinterpretq_p64_f64(float64x2_t __p0) {
35136  poly64x2_t __ret;
35137  __ret = (poly64x2_t)(__p0);
35138  return __ret;
35139}
35140#else
35141__ai poly64x2_t vreinterpretq_p64_f64(float64x2_t __p0) {
35142  poly64x2_t __ret;
35143  __ret = (poly64x2_t)(__p0);
35144  return __ret;
35145}
35146#endif
35147
35148#ifdef __LITTLE_ENDIAN__
35149__ai poly64x2_t vreinterpretq_p64_f32(float32x4_t __p0) {
35150  poly64x2_t __ret;
35151  __ret = (poly64x2_t)(__p0);
35152  return __ret;
35153}
35154#else
35155__ai poly64x2_t vreinterpretq_p64_f32(float32x4_t __p0) {
35156  poly64x2_t __ret;
35157  __ret = (poly64x2_t)(__p0);
35158  return __ret;
35159}
35160#endif
35161
35162#ifdef __LITTLE_ENDIAN__
35163__ai poly64x2_t vreinterpretq_p64_f16(float16x8_t __p0) {
35164  poly64x2_t __ret;
35165  __ret = (poly64x2_t)(__p0);
35166  return __ret;
35167}
35168#else
35169__ai poly64x2_t vreinterpretq_p64_f16(float16x8_t __p0) {
35170  poly64x2_t __ret;
35171  __ret = (poly64x2_t)(__p0);
35172  return __ret;
35173}
35174#endif
35175
35176#ifdef __LITTLE_ENDIAN__
35177__ai poly64x2_t vreinterpretq_p64_s32(int32x4_t __p0) {
35178  poly64x2_t __ret;
35179  __ret = (poly64x2_t)(__p0);
35180  return __ret;
35181}
35182#else
35183__ai poly64x2_t vreinterpretq_p64_s32(int32x4_t __p0) {
35184  poly64x2_t __ret;
35185  __ret = (poly64x2_t)(__p0);
35186  return __ret;
35187}
35188#endif
35189
35190#ifdef __LITTLE_ENDIAN__
35191__ai poly64x2_t vreinterpretq_p64_s64(int64x2_t __p0) {
35192  poly64x2_t __ret;
35193  __ret = (poly64x2_t)(__p0);
35194  return __ret;
35195}
35196#else
35197__ai poly64x2_t vreinterpretq_p64_s64(int64x2_t __p0) {
35198  poly64x2_t __ret;
35199  __ret = (poly64x2_t)(__p0);
35200  return __ret;
35201}
35202#endif
35203
35204#ifdef __LITTLE_ENDIAN__
35205__ai poly64x2_t vreinterpretq_p64_s16(int16x8_t __p0) {
35206  poly64x2_t __ret;
35207  __ret = (poly64x2_t)(__p0);
35208  return __ret;
35209}
35210#else
35211__ai poly64x2_t vreinterpretq_p64_s16(int16x8_t __p0) {
35212  poly64x2_t __ret;
35213  __ret = (poly64x2_t)(__p0);
35214  return __ret;
35215}
35216#endif
35217
35218#ifdef __LITTLE_ENDIAN__
35219__ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
35220  poly16x8_t __ret;
35221  __ret = (poly16x8_t)(__p0);
35222  return __ret;
35223}
35224#else
35225__ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
35226  poly16x8_t __ret;
35227  __ret = (poly16x8_t)(__p0);
35228  return __ret;
35229}
35230#endif
35231
35232#ifdef __LITTLE_ENDIAN__
35233__ai poly16x8_t vreinterpretq_p16_p128(poly128_t __p0) {
35234  poly16x8_t __ret;
35235  __ret = (poly16x8_t)(__p0);
35236  return __ret;
35237}
35238#else
35239__ai poly16x8_t vreinterpretq_p16_p128(poly128_t __p0) {
35240  poly16x8_t __ret;
35241  __ret = (poly16x8_t)(__p0);
35242  return __ret;
35243}
35244#endif
35245
35246#ifdef __LITTLE_ENDIAN__
35247__ai poly16x8_t vreinterpretq_p16_p64(poly64x2_t __p0) {
35248  poly16x8_t __ret;
35249  __ret = (poly16x8_t)(__p0);
35250  return __ret;
35251}
35252#else
35253__ai poly16x8_t vreinterpretq_p16_p64(poly64x2_t __p0) {
35254  poly16x8_t __ret;
35255  __ret = (poly16x8_t)(__p0);
35256  return __ret;
35257}
35258#endif
35259
35260#ifdef __LITTLE_ENDIAN__
35261__ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
35262  poly16x8_t __ret;
35263  __ret = (poly16x8_t)(__p0);
35264  return __ret;
35265}
35266#else
35267__ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
35268  poly16x8_t __ret;
35269  __ret = (poly16x8_t)(__p0);
35270  return __ret;
35271}
35272#endif
35273
35274#ifdef __LITTLE_ENDIAN__
35275__ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
35276  poly16x8_t __ret;
35277  __ret = (poly16x8_t)(__p0);
35278  return __ret;
35279}
35280#else
35281__ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
35282  poly16x8_t __ret;
35283  __ret = (poly16x8_t)(__p0);
35284  return __ret;
35285}
35286#endif
35287
35288#ifdef __LITTLE_ENDIAN__
35289__ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
35290  poly16x8_t __ret;
35291  __ret = (poly16x8_t)(__p0);
35292  return __ret;
35293}
35294#else
35295__ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
35296  poly16x8_t __ret;
35297  __ret = (poly16x8_t)(__p0);
35298  return __ret;
35299}
35300#endif
35301
35302#ifdef __LITTLE_ENDIAN__
35303__ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
35304  poly16x8_t __ret;
35305  __ret = (poly16x8_t)(__p0);
35306  return __ret;
35307}
35308#else
35309__ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
35310  poly16x8_t __ret;
35311  __ret = (poly16x8_t)(__p0);
35312  return __ret;
35313}
35314#endif
35315
35316#ifdef __LITTLE_ENDIAN__
35317__ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
35318  poly16x8_t __ret;
35319  __ret = (poly16x8_t)(__p0);
35320  return __ret;
35321}
35322#else
35323__ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
35324  poly16x8_t __ret;
35325  __ret = (poly16x8_t)(__p0);
35326  return __ret;
35327}
35328#endif
35329
35330#ifdef __LITTLE_ENDIAN__
35331__ai poly16x8_t vreinterpretq_p16_f64(float64x2_t __p0) {
35332  poly16x8_t __ret;
35333  __ret = (poly16x8_t)(__p0);
35334  return __ret;
35335}
35336#else
35337__ai poly16x8_t vreinterpretq_p16_f64(float64x2_t __p0) {
35338  poly16x8_t __ret;
35339  __ret = (poly16x8_t)(__p0);
35340  return __ret;
35341}
35342#endif
35343
35344#ifdef __LITTLE_ENDIAN__
35345__ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
35346  poly16x8_t __ret;
35347  __ret = (poly16x8_t)(__p0);
35348  return __ret;
35349}
35350#else
35351__ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
35352  poly16x8_t __ret;
35353  __ret = (poly16x8_t)(__p0);
35354  return __ret;
35355}
35356#endif
35357
35358#ifdef __LITTLE_ENDIAN__
35359__ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
35360  poly16x8_t __ret;
35361  __ret = (poly16x8_t)(__p0);
35362  return __ret;
35363}
35364#else
35365__ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
35366  poly16x8_t __ret;
35367  __ret = (poly16x8_t)(__p0);
35368  return __ret;
35369}
35370#endif
35371
35372#ifdef __LITTLE_ENDIAN__
35373__ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
35374  poly16x8_t __ret;
35375  __ret = (poly16x8_t)(__p0);
35376  return __ret;
35377}
35378#else
35379__ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
35380  poly16x8_t __ret;
35381  __ret = (poly16x8_t)(__p0);
35382  return __ret;
35383}
35384#endif
35385
35386#ifdef __LITTLE_ENDIAN__
35387__ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
35388  poly16x8_t __ret;
35389  __ret = (poly16x8_t)(__p0);
35390  return __ret;
35391}
35392#else
35393__ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
35394  poly16x8_t __ret;
35395  __ret = (poly16x8_t)(__p0);
35396  return __ret;
35397}
35398#endif
35399
35400#ifdef __LITTLE_ENDIAN__
35401__ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
35402  poly16x8_t __ret;
35403  __ret = (poly16x8_t)(__p0);
35404  return __ret;
35405}
35406#else
35407__ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
35408  poly16x8_t __ret;
35409  __ret = (poly16x8_t)(__p0);
35410  return __ret;
35411}
35412#endif
35413
35414#ifdef __LITTLE_ENDIAN__
35415__ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
35416  uint8x16_t __ret;
35417  __ret = (uint8x16_t)(__p0);
35418  return __ret;
35419}
35420#else
35421__ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
35422  uint8x16_t __ret;
35423  __ret = (uint8x16_t)(__p0);
35424  return __ret;
35425}
35426#endif
35427
35428#ifdef __LITTLE_ENDIAN__
35429__ai uint8x16_t vreinterpretq_u8_p128(poly128_t __p0) {
35430  uint8x16_t __ret;
35431  __ret = (uint8x16_t)(__p0);
35432  return __ret;
35433}
35434#else
35435__ai uint8x16_t vreinterpretq_u8_p128(poly128_t __p0) {
35436  uint8x16_t __ret;
35437  __ret = (uint8x16_t)(__p0);
35438  return __ret;
35439}
35440#endif
35441
35442#ifdef __LITTLE_ENDIAN__
35443__ai uint8x16_t vreinterpretq_u8_p64(poly64x2_t __p0) {
35444  uint8x16_t __ret;
35445  __ret = (uint8x16_t)(__p0);
35446  return __ret;
35447}
35448#else
35449__ai uint8x16_t vreinterpretq_u8_p64(poly64x2_t __p0) {
35450  uint8x16_t __ret;
35451  __ret = (uint8x16_t)(__p0);
35452  return __ret;
35453}
35454#endif
35455
35456#ifdef __LITTLE_ENDIAN__
35457__ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
35458  uint8x16_t __ret;
35459  __ret = (uint8x16_t)(__p0);
35460  return __ret;
35461}
35462#else
35463__ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
35464  uint8x16_t __ret;
35465  __ret = (uint8x16_t)(__p0);
35466  return __ret;
35467}
35468#endif
35469
35470#ifdef __LITTLE_ENDIAN__
35471__ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
35472  uint8x16_t __ret;
35473  __ret = (uint8x16_t)(__p0);
35474  return __ret;
35475}
35476#else
35477__ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
35478  uint8x16_t __ret;
35479  __ret = (uint8x16_t)(__p0);
35480  return __ret;
35481}
35482#endif
35483
35484#ifdef __LITTLE_ENDIAN__
35485__ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
35486  uint8x16_t __ret;
35487  __ret = (uint8x16_t)(__p0);
35488  return __ret;
35489}
35490#else
35491__ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
35492  uint8x16_t __ret;
35493  __ret = (uint8x16_t)(__p0);
35494  return __ret;
35495}
35496#endif
35497
35498#ifdef __LITTLE_ENDIAN__
35499__ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
35500  uint8x16_t __ret;
35501  __ret = (uint8x16_t)(__p0);
35502  return __ret;
35503}
35504#else
35505__ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
35506  uint8x16_t __ret;
35507  __ret = (uint8x16_t)(__p0);
35508  return __ret;
35509}
35510#endif
35511
35512#ifdef __LITTLE_ENDIAN__
35513__ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
35514  uint8x16_t __ret;
35515  __ret = (uint8x16_t)(__p0);
35516  return __ret;
35517}
35518#else
35519__ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
35520  uint8x16_t __ret;
35521  __ret = (uint8x16_t)(__p0);
35522  return __ret;
35523}
35524#endif
35525
35526#ifdef __LITTLE_ENDIAN__
35527__ai uint8x16_t vreinterpretq_u8_f64(float64x2_t __p0) {
35528  uint8x16_t __ret;
35529  __ret = (uint8x16_t)(__p0);
35530  return __ret;
35531}
35532#else
35533__ai uint8x16_t vreinterpretq_u8_f64(float64x2_t __p0) {
35534  uint8x16_t __ret;
35535  __ret = (uint8x16_t)(__p0);
35536  return __ret;
35537}
35538#endif
35539
35540#ifdef __LITTLE_ENDIAN__
35541__ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
35542  uint8x16_t __ret;
35543  __ret = (uint8x16_t)(__p0);
35544  return __ret;
35545}
35546#else
35547__ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
35548  uint8x16_t __ret;
35549  __ret = (uint8x16_t)(__p0);
35550  return __ret;
35551}
35552#endif
35553
35554#ifdef __LITTLE_ENDIAN__
35555__ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
35556  uint8x16_t __ret;
35557  __ret = (uint8x16_t)(__p0);
35558  return __ret;
35559}
35560#else
35561__ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
35562  uint8x16_t __ret;
35563  __ret = (uint8x16_t)(__p0);
35564  return __ret;
35565}
35566#endif
35567
35568#ifdef __LITTLE_ENDIAN__
35569__ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
35570  uint8x16_t __ret;
35571  __ret = (uint8x16_t)(__p0);
35572  return __ret;
35573}
35574#else
35575__ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
35576  uint8x16_t __ret;
35577  __ret = (uint8x16_t)(__p0);
35578  return __ret;
35579}
35580#endif
35581
35582#ifdef __LITTLE_ENDIAN__
35583__ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
35584  uint8x16_t __ret;
35585  __ret = (uint8x16_t)(__p0);
35586  return __ret;
35587}
35588#else
35589__ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
35590  uint8x16_t __ret;
35591  __ret = (uint8x16_t)(__p0);
35592  return __ret;
35593}
35594#endif
35595
35596#ifdef __LITTLE_ENDIAN__
35597__ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
35598  uint8x16_t __ret;
35599  __ret = (uint8x16_t)(__p0);
35600  return __ret;
35601}
35602#else
35603__ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
35604  uint8x16_t __ret;
35605  __ret = (uint8x16_t)(__p0);
35606  return __ret;
35607}
35608#endif
35609
35610#ifdef __LITTLE_ENDIAN__
35611__ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
35612  uint32x4_t __ret;
35613  __ret = (uint32x4_t)(__p0);
35614  return __ret;
35615}
35616#else
35617__ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
35618  uint32x4_t __ret;
35619  __ret = (uint32x4_t)(__p0);
35620  return __ret;
35621}
35622#endif
35623
35624#ifdef __LITTLE_ENDIAN__
35625__ai uint32x4_t vreinterpretq_u32_p128(poly128_t __p0) {
35626  uint32x4_t __ret;
35627  __ret = (uint32x4_t)(__p0);
35628  return __ret;
35629}
35630#else
35631__ai uint32x4_t vreinterpretq_u32_p128(poly128_t __p0) {
35632  uint32x4_t __ret;
35633  __ret = (uint32x4_t)(__p0);
35634  return __ret;
35635}
35636#endif
35637
35638#ifdef __LITTLE_ENDIAN__
35639__ai uint32x4_t vreinterpretq_u32_p64(poly64x2_t __p0) {
35640  uint32x4_t __ret;
35641  __ret = (uint32x4_t)(__p0);
35642  return __ret;
35643}
35644#else
35645__ai uint32x4_t vreinterpretq_u32_p64(poly64x2_t __p0) {
35646  uint32x4_t __ret;
35647  __ret = (uint32x4_t)(__p0);
35648  return __ret;
35649}
35650#endif
35651
35652#ifdef __LITTLE_ENDIAN__
35653__ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
35654  uint32x4_t __ret;
35655  __ret = (uint32x4_t)(__p0);
35656  return __ret;
35657}
35658#else
35659__ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
35660  uint32x4_t __ret;
35661  __ret = (uint32x4_t)(__p0);
35662  return __ret;
35663}
35664#endif
35665
35666#ifdef __LITTLE_ENDIAN__
35667__ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
35668  uint32x4_t __ret;
35669  __ret = (uint32x4_t)(__p0);
35670  return __ret;
35671}
35672#else
35673__ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
35674  uint32x4_t __ret;
35675  __ret = (uint32x4_t)(__p0);
35676  return __ret;
35677}
35678#endif
35679
35680#ifdef __LITTLE_ENDIAN__
35681__ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
35682  uint32x4_t __ret;
35683  __ret = (uint32x4_t)(__p0);
35684  return __ret;
35685}
35686#else
35687__ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
35688  uint32x4_t __ret;
35689  __ret = (uint32x4_t)(__p0);
35690  return __ret;
35691}
35692#endif
35693
35694#ifdef __LITTLE_ENDIAN__
35695__ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
35696  uint32x4_t __ret;
35697  __ret = (uint32x4_t)(__p0);
35698  return __ret;
35699}
35700#else
35701__ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
35702  uint32x4_t __ret;
35703  __ret = (uint32x4_t)(__p0);
35704  return __ret;
35705}
35706#endif
35707
35708#ifdef __LITTLE_ENDIAN__
35709__ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
35710  uint32x4_t __ret;
35711  __ret = (uint32x4_t)(__p0);
35712  return __ret;
35713}
35714#else
35715__ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
35716  uint32x4_t __ret;
35717  __ret = (uint32x4_t)(__p0);
35718  return __ret;
35719}
35720#endif
35721
35722#ifdef __LITTLE_ENDIAN__
35723__ai uint32x4_t vreinterpretq_u32_f64(float64x2_t __p0) {
35724  uint32x4_t __ret;
35725  __ret = (uint32x4_t)(__p0);
35726  return __ret;
35727}
35728#else
35729__ai uint32x4_t vreinterpretq_u32_f64(float64x2_t __p0) {
35730  uint32x4_t __ret;
35731  __ret = (uint32x4_t)(__p0);
35732  return __ret;
35733}
35734#endif
35735
35736#ifdef __LITTLE_ENDIAN__
35737__ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
35738  uint32x4_t __ret;
35739  __ret = (uint32x4_t)(__p0);
35740  return __ret;
35741}
35742#else
35743__ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
35744  uint32x4_t __ret;
35745  __ret = (uint32x4_t)(__p0);
35746  return __ret;
35747}
35748#endif
35749
35750#ifdef __LITTLE_ENDIAN__
35751__ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
35752  uint32x4_t __ret;
35753  __ret = (uint32x4_t)(__p0);
35754  return __ret;
35755}
35756#else
35757__ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
35758  uint32x4_t __ret;
35759  __ret = (uint32x4_t)(__p0);
35760  return __ret;
35761}
35762#endif
35763
35764#ifdef __LITTLE_ENDIAN__
35765__ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
35766  uint32x4_t __ret;
35767  __ret = (uint32x4_t)(__p0);
35768  return __ret;
35769}
35770#else
35771__ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
35772  uint32x4_t __ret;
35773  __ret = (uint32x4_t)(__p0);
35774  return __ret;
35775}
35776#endif
35777
35778#ifdef __LITTLE_ENDIAN__
35779__ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
35780  uint32x4_t __ret;
35781  __ret = (uint32x4_t)(__p0);
35782  return __ret;
35783}
35784#else
35785__ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
35786  uint32x4_t __ret;
35787  __ret = (uint32x4_t)(__p0);
35788  return __ret;
35789}
35790#endif
35791
35792#ifdef __LITTLE_ENDIAN__
35793__ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
35794  uint32x4_t __ret;
35795  __ret = (uint32x4_t)(__p0);
35796  return __ret;
35797}
35798#else
35799__ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
35800  uint32x4_t __ret;
35801  __ret = (uint32x4_t)(__p0);
35802  return __ret;
35803}
35804#endif
35805
35806#ifdef __LITTLE_ENDIAN__
35807__ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
35808  uint64x2_t __ret;
35809  __ret = (uint64x2_t)(__p0);
35810  return __ret;
35811}
35812#else
35813__ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
35814  uint64x2_t __ret;
35815  __ret = (uint64x2_t)(__p0);
35816  return __ret;
35817}
35818#endif
35819
35820#ifdef __LITTLE_ENDIAN__
35821__ai uint64x2_t vreinterpretq_u64_p128(poly128_t __p0) {
35822  uint64x2_t __ret;
35823  __ret = (uint64x2_t)(__p0);
35824  return __ret;
35825}
35826#else
35827__ai uint64x2_t vreinterpretq_u64_p128(poly128_t __p0) {
35828  uint64x2_t __ret;
35829  __ret = (uint64x2_t)(__p0);
35830  return __ret;
35831}
35832#endif
35833
35834#ifdef __LITTLE_ENDIAN__
35835__ai uint64x2_t vreinterpretq_u64_p64(poly64x2_t __p0) {
35836  uint64x2_t __ret;
35837  __ret = (uint64x2_t)(__p0);
35838  return __ret;
35839}
35840#else
35841__ai uint64x2_t vreinterpretq_u64_p64(poly64x2_t __p0) {
35842  uint64x2_t __ret;
35843  __ret = (uint64x2_t)(__p0);
35844  return __ret;
35845}
35846#endif
35847
35848#ifdef __LITTLE_ENDIAN__
35849__ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
35850  uint64x2_t __ret;
35851  __ret = (uint64x2_t)(__p0);
35852  return __ret;
35853}
35854#else
35855__ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
35856  uint64x2_t __ret;
35857  __ret = (uint64x2_t)(__p0);
35858  return __ret;
35859}
35860#endif
35861
35862#ifdef __LITTLE_ENDIAN__
35863__ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
35864  uint64x2_t __ret;
35865  __ret = (uint64x2_t)(__p0);
35866  return __ret;
35867}
35868#else
35869__ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
35870  uint64x2_t __ret;
35871  __ret = (uint64x2_t)(__p0);
35872  return __ret;
35873}
35874#endif
35875
35876#ifdef __LITTLE_ENDIAN__
35877__ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
35878  uint64x2_t __ret;
35879  __ret = (uint64x2_t)(__p0);
35880  return __ret;
35881}
35882#else
35883__ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
35884  uint64x2_t __ret;
35885  __ret = (uint64x2_t)(__p0);
35886  return __ret;
35887}
35888#endif
35889
35890#ifdef __LITTLE_ENDIAN__
35891__ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
35892  uint64x2_t __ret;
35893  __ret = (uint64x2_t)(__p0);
35894  return __ret;
35895}
35896#else
35897__ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
35898  uint64x2_t __ret;
35899  __ret = (uint64x2_t)(__p0);
35900  return __ret;
35901}
35902#endif
35903
35904#ifdef __LITTLE_ENDIAN__
35905__ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
35906  uint64x2_t __ret;
35907  __ret = (uint64x2_t)(__p0);
35908  return __ret;
35909}
35910#else
35911__ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
35912  uint64x2_t __ret;
35913  __ret = (uint64x2_t)(__p0);
35914  return __ret;
35915}
35916#endif
35917
35918#ifdef __LITTLE_ENDIAN__
35919__ai uint64x2_t vreinterpretq_u64_f64(float64x2_t __p0) {
35920  uint64x2_t __ret;
35921  __ret = (uint64x2_t)(__p0);
35922  return __ret;
35923}
35924#else
35925__ai uint64x2_t vreinterpretq_u64_f64(float64x2_t __p0) {
35926  uint64x2_t __ret;
35927  __ret = (uint64x2_t)(__p0);
35928  return __ret;
35929}
35930#endif
35931
35932#ifdef __LITTLE_ENDIAN__
35933__ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
35934  uint64x2_t __ret;
35935  __ret = (uint64x2_t)(__p0);
35936  return __ret;
35937}
35938#else
35939__ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
35940  uint64x2_t __ret;
35941  __ret = (uint64x2_t)(__p0);
35942  return __ret;
35943}
35944#endif
35945
35946#ifdef __LITTLE_ENDIAN__
35947__ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
35948  uint64x2_t __ret;
35949  __ret = (uint64x2_t)(__p0);
35950  return __ret;
35951}
35952#else
35953__ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
35954  uint64x2_t __ret;
35955  __ret = (uint64x2_t)(__p0);
35956  return __ret;
35957}
35958#endif
35959
35960#ifdef __LITTLE_ENDIAN__
35961__ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
35962  uint64x2_t __ret;
35963  __ret = (uint64x2_t)(__p0);
35964  return __ret;
35965}
35966#else
35967__ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
35968  uint64x2_t __ret;
35969  __ret = (uint64x2_t)(__p0);
35970  return __ret;
35971}
35972#endif
35973
35974#ifdef __LITTLE_ENDIAN__
35975__ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
35976  uint64x2_t __ret;
35977  __ret = (uint64x2_t)(__p0);
35978  return __ret;
35979}
35980#else
35981__ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
35982  uint64x2_t __ret;
35983  __ret = (uint64x2_t)(__p0);
35984  return __ret;
35985}
35986#endif
35987
35988#ifdef __LITTLE_ENDIAN__
35989__ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
35990  uint64x2_t __ret;
35991  __ret = (uint64x2_t)(__p0);
35992  return __ret;
35993}
35994#else
35995__ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
35996  uint64x2_t __ret;
35997  __ret = (uint64x2_t)(__p0);
35998  return __ret;
35999}
36000#endif
36001
36002#ifdef __LITTLE_ENDIAN__
36003__ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
36004  uint16x8_t __ret;
36005  __ret = (uint16x8_t)(__p0);
36006  return __ret;
36007}
36008#else
36009__ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
36010  uint16x8_t __ret;
36011  __ret = (uint16x8_t)(__p0);
36012  return __ret;
36013}
36014#endif
36015
36016#ifdef __LITTLE_ENDIAN__
36017__ai uint16x8_t vreinterpretq_u16_p128(poly128_t __p0) {
36018  uint16x8_t __ret;
36019  __ret = (uint16x8_t)(__p0);
36020  return __ret;
36021}
36022#else
36023__ai uint16x8_t vreinterpretq_u16_p128(poly128_t __p0) {
36024  uint16x8_t __ret;
36025  __ret = (uint16x8_t)(__p0);
36026  return __ret;
36027}
36028#endif
36029
36030#ifdef __LITTLE_ENDIAN__
36031__ai uint16x8_t vreinterpretq_u16_p64(poly64x2_t __p0) {
36032  uint16x8_t __ret;
36033  __ret = (uint16x8_t)(__p0);
36034  return __ret;
36035}
36036#else
36037__ai uint16x8_t vreinterpretq_u16_p64(poly64x2_t __p0) {
36038  uint16x8_t __ret;
36039  __ret = (uint16x8_t)(__p0);
36040  return __ret;
36041}
36042#endif
36043
36044#ifdef __LITTLE_ENDIAN__
36045__ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
36046  uint16x8_t __ret;
36047  __ret = (uint16x8_t)(__p0);
36048  return __ret;
36049}
36050#else
36051__ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
36052  uint16x8_t __ret;
36053  __ret = (uint16x8_t)(__p0);
36054  return __ret;
36055}
36056#endif
36057
36058#ifdef __LITTLE_ENDIAN__
36059__ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
36060  uint16x8_t __ret;
36061  __ret = (uint16x8_t)(__p0);
36062  return __ret;
36063}
36064#else
36065__ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
36066  uint16x8_t __ret;
36067  __ret = (uint16x8_t)(__p0);
36068  return __ret;
36069}
36070#endif
36071
36072#ifdef __LITTLE_ENDIAN__
36073__ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
36074  uint16x8_t __ret;
36075  __ret = (uint16x8_t)(__p0);
36076  return __ret;
36077}
36078#else
36079__ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
36080  uint16x8_t __ret;
36081  __ret = (uint16x8_t)(__p0);
36082  return __ret;
36083}
36084#endif
36085
36086#ifdef __LITTLE_ENDIAN__
36087__ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
36088  uint16x8_t __ret;
36089  __ret = (uint16x8_t)(__p0);
36090  return __ret;
36091}
36092#else
36093__ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
36094  uint16x8_t __ret;
36095  __ret = (uint16x8_t)(__p0);
36096  return __ret;
36097}
36098#endif
36099
36100#ifdef __LITTLE_ENDIAN__
36101__ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
36102  uint16x8_t __ret;
36103  __ret = (uint16x8_t)(__p0);
36104  return __ret;
36105}
36106#else
36107__ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
36108  uint16x8_t __ret;
36109  __ret = (uint16x8_t)(__p0);
36110  return __ret;
36111}
36112#endif
36113
36114#ifdef __LITTLE_ENDIAN__
36115__ai uint16x8_t vreinterpretq_u16_f64(float64x2_t __p0) {
36116  uint16x8_t __ret;
36117  __ret = (uint16x8_t)(__p0);
36118  return __ret;
36119}
36120#else
36121__ai uint16x8_t vreinterpretq_u16_f64(float64x2_t __p0) {
36122  uint16x8_t __ret;
36123  __ret = (uint16x8_t)(__p0);
36124  return __ret;
36125}
36126#endif
36127
36128#ifdef __LITTLE_ENDIAN__
36129__ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
36130  uint16x8_t __ret;
36131  __ret = (uint16x8_t)(__p0);
36132  return __ret;
36133}
36134#else
36135__ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
36136  uint16x8_t __ret;
36137  __ret = (uint16x8_t)(__p0);
36138  return __ret;
36139}
36140#endif
36141
36142#ifdef __LITTLE_ENDIAN__
36143__ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
36144  uint16x8_t __ret;
36145  __ret = (uint16x8_t)(__p0);
36146  return __ret;
36147}
36148#else
36149__ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
36150  uint16x8_t __ret;
36151  __ret = (uint16x8_t)(__p0);
36152  return __ret;
36153}
36154#endif
36155
36156#ifdef __LITTLE_ENDIAN__
36157__ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
36158  uint16x8_t __ret;
36159  __ret = (uint16x8_t)(__p0);
36160  return __ret;
36161}
36162#else
36163__ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
36164  uint16x8_t __ret;
36165  __ret = (uint16x8_t)(__p0);
36166  return __ret;
36167}
36168#endif
36169
36170#ifdef __LITTLE_ENDIAN__
36171__ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
36172  uint16x8_t __ret;
36173  __ret = (uint16x8_t)(__p0);
36174  return __ret;
36175}
36176#else
36177__ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
36178  uint16x8_t __ret;
36179  __ret = (uint16x8_t)(__p0);
36180  return __ret;
36181}
36182#endif
36183
36184#ifdef __LITTLE_ENDIAN__
36185__ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
36186  uint16x8_t __ret;
36187  __ret = (uint16x8_t)(__p0);
36188  return __ret;
36189}
36190#else
36191__ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
36192  uint16x8_t __ret;
36193  __ret = (uint16x8_t)(__p0);
36194  return __ret;
36195}
36196#endif
36197
36198#ifdef __LITTLE_ENDIAN__
36199__ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
36200  int8x16_t __ret;
36201  __ret = (int8x16_t)(__p0);
36202  return __ret;
36203}
36204#else
36205__ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
36206  int8x16_t __ret;
36207  __ret = (int8x16_t)(__p0);
36208  return __ret;
36209}
36210#endif
36211
36212#ifdef __LITTLE_ENDIAN__
36213__ai int8x16_t vreinterpretq_s8_p128(poly128_t __p0) {
36214  int8x16_t __ret;
36215  __ret = (int8x16_t)(__p0);
36216  return __ret;
36217}
36218#else
36219__ai int8x16_t vreinterpretq_s8_p128(poly128_t __p0) {
36220  int8x16_t __ret;
36221  __ret = (int8x16_t)(__p0);
36222  return __ret;
36223}
36224#endif
36225
36226#ifdef __LITTLE_ENDIAN__
36227__ai int8x16_t vreinterpretq_s8_p64(poly64x2_t __p0) {
36228  int8x16_t __ret;
36229  __ret = (int8x16_t)(__p0);
36230  return __ret;
36231}
36232#else
36233__ai int8x16_t vreinterpretq_s8_p64(poly64x2_t __p0) {
36234  int8x16_t __ret;
36235  __ret = (int8x16_t)(__p0);
36236  return __ret;
36237}
36238#endif
36239
36240#ifdef __LITTLE_ENDIAN__
36241__ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
36242  int8x16_t __ret;
36243  __ret = (int8x16_t)(__p0);
36244  return __ret;
36245}
36246#else
36247__ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
36248  int8x16_t __ret;
36249  __ret = (int8x16_t)(__p0);
36250  return __ret;
36251}
36252#endif
36253
36254#ifdef __LITTLE_ENDIAN__
36255__ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
36256  int8x16_t __ret;
36257  __ret = (int8x16_t)(__p0);
36258  return __ret;
36259}
36260#else
36261__ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
36262  int8x16_t __ret;
36263  __ret = (int8x16_t)(__p0);
36264  return __ret;
36265}
36266#endif
36267
36268#ifdef __LITTLE_ENDIAN__
36269__ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
36270  int8x16_t __ret;
36271  __ret = (int8x16_t)(__p0);
36272  return __ret;
36273}
36274#else
36275__ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
36276  int8x16_t __ret;
36277  __ret = (int8x16_t)(__p0);
36278  return __ret;
36279}
36280#endif
36281
36282#ifdef __LITTLE_ENDIAN__
36283__ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
36284  int8x16_t __ret;
36285  __ret = (int8x16_t)(__p0);
36286  return __ret;
36287}
36288#else
36289__ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
36290  int8x16_t __ret;
36291  __ret = (int8x16_t)(__p0);
36292  return __ret;
36293}
36294#endif
36295
36296#ifdef __LITTLE_ENDIAN__
36297__ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
36298  int8x16_t __ret;
36299  __ret = (int8x16_t)(__p0);
36300  return __ret;
36301}
36302#else
36303__ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
36304  int8x16_t __ret;
36305  __ret = (int8x16_t)(__p0);
36306  return __ret;
36307}
36308#endif
36309
36310#ifdef __LITTLE_ENDIAN__
36311__ai int8x16_t vreinterpretq_s8_f64(float64x2_t __p0) {
36312  int8x16_t __ret;
36313  __ret = (int8x16_t)(__p0);
36314  return __ret;
36315}
36316#else
36317__ai int8x16_t vreinterpretq_s8_f64(float64x2_t __p0) {
36318  int8x16_t __ret;
36319  __ret = (int8x16_t)(__p0);
36320  return __ret;
36321}
36322#endif
36323
36324#ifdef __LITTLE_ENDIAN__
36325__ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
36326  int8x16_t __ret;
36327  __ret = (int8x16_t)(__p0);
36328  return __ret;
36329}
36330#else
36331__ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
36332  int8x16_t __ret;
36333  __ret = (int8x16_t)(__p0);
36334  return __ret;
36335}
36336#endif
36337
36338#ifdef __LITTLE_ENDIAN__
36339__ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
36340  int8x16_t __ret;
36341  __ret = (int8x16_t)(__p0);
36342  return __ret;
36343}
36344#else
36345__ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
36346  int8x16_t __ret;
36347  __ret = (int8x16_t)(__p0);
36348  return __ret;
36349}
36350#endif
36351
36352#ifdef __LITTLE_ENDIAN__
36353__ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
36354  int8x16_t __ret;
36355  __ret = (int8x16_t)(__p0);
36356  return __ret;
36357}
36358#else
36359__ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
36360  int8x16_t __ret;
36361  __ret = (int8x16_t)(__p0);
36362  return __ret;
36363}
36364#endif
36365
36366#ifdef __LITTLE_ENDIAN__
36367__ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
36368  int8x16_t __ret;
36369  __ret = (int8x16_t)(__p0);
36370  return __ret;
36371}
36372#else
36373__ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
36374  int8x16_t __ret;
36375  __ret = (int8x16_t)(__p0);
36376  return __ret;
36377}
36378#endif
36379
36380#ifdef __LITTLE_ENDIAN__
36381__ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
36382  int8x16_t __ret;
36383  __ret = (int8x16_t)(__p0);
36384  return __ret;
36385}
36386#else
36387__ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
36388  int8x16_t __ret;
36389  __ret = (int8x16_t)(__p0);
36390  return __ret;
36391}
36392#endif
36393
36394#ifdef __LITTLE_ENDIAN__
36395__ai float64x2_t vreinterpretq_f64_p8(poly8x16_t __p0) {
36396  float64x2_t __ret;
36397  __ret = (float64x2_t)(__p0);
36398  return __ret;
36399}
36400#else
36401__ai float64x2_t vreinterpretq_f64_p8(poly8x16_t __p0) {
36402  float64x2_t __ret;
36403  __ret = (float64x2_t)(__p0);
36404  return __ret;
36405}
36406#endif
36407
36408#ifdef __LITTLE_ENDIAN__
36409__ai float64x2_t vreinterpretq_f64_p128(poly128_t __p0) {
36410  float64x2_t __ret;
36411  __ret = (float64x2_t)(__p0);
36412  return __ret;
36413}
36414#else
36415__ai float64x2_t vreinterpretq_f64_p128(poly128_t __p0) {
36416  float64x2_t __ret;
36417  __ret = (float64x2_t)(__p0);
36418  return __ret;
36419}
36420#endif
36421
36422#ifdef __LITTLE_ENDIAN__
36423__ai float64x2_t vreinterpretq_f64_p64(poly64x2_t __p0) {
36424  float64x2_t __ret;
36425  __ret = (float64x2_t)(__p0);
36426  return __ret;
36427}
36428#else
36429__ai float64x2_t vreinterpretq_f64_p64(poly64x2_t __p0) {
36430  float64x2_t __ret;
36431  __ret = (float64x2_t)(__p0);
36432  return __ret;
36433}
36434#endif
36435
36436#ifdef __LITTLE_ENDIAN__
36437__ai float64x2_t vreinterpretq_f64_p16(poly16x8_t __p0) {
36438  float64x2_t __ret;
36439  __ret = (float64x2_t)(__p0);
36440  return __ret;
36441}
36442#else
36443__ai float64x2_t vreinterpretq_f64_p16(poly16x8_t __p0) {
36444  float64x2_t __ret;
36445  __ret = (float64x2_t)(__p0);
36446  return __ret;
36447}
36448#endif
36449
36450#ifdef __LITTLE_ENDIAN__
36451__ai float64x2_t vreinterpretq_f64_u8(uint8x16_t __p0) {
36452  float64x2_t __ret;
36453  __ret = (float64x2_t)(__p0);
36454  return __ret;
36455}
36456#else
36457__ai float64x2_t vreinterpretq_f64_u8(uint8x16_t __p0) {
36458  float64x2_t __ret;
36459  __ret = (float64x2_t)(__p0);
36460  return __ret;
36461}
36462#endif
36463
36464#ifdef __LITTLE_ENDIAN__
36465__ai float64x2_t vreinterpretq_f64_u32(uint32x4_t __p0) {
36466  float64x2_t __ret;
36467  __ret = (float64x2_t)(__p0);
36468  return __ret;
36469}
36470#else
36471__ai float64x2_t vreinterpretq_f64_u32(uint32x4_t __p0) {
36472  float64x2_t __ret;
36473  __ret = (float64x2_t)(__p0);
36474  return __ret;
36475}
36476#endif
36477
36478#ifdef __LITTLE_ENDIAN__
36479__ai float64x2_t vreinterpretq_f64_u64(uint64x2_t __p0) {
36480  float64x2_t __ret;
36481  __ret = (float64x2_t)(__p0);
36482  return __ret;
36483}
36484#else
36485__ai float64x2_t vreinterpretq_f64_u64(uint64x2_t __p0) {
36486  float64x2_t __ret;
36487  __ret = (float64x2_t)(__p0);
36488  return __ret;
36489}
36490#endif
36491
36492#ifdef __LITTLE_ENDIAN__
36493__ai float64x2_t vreinterpretq_f64_u16(uint16x8_t __p0) {
36494  float64x2_t __ret;
36495  __ret = (float64x2_t)(__p0);
36496  return __ret;
36497}
36498#else
36499__ai float64x2_t vreinterpretq_f64_u16(uint16x8_t __p0) {
36500  float64x2_t __ret;
36501  __ret = (float64x2_t)(__p0);
36502  return __ret;
36503}
36504#endif
36505
36506#ifdef __LITTLE_ENDIAN__
36507__ai float64x2_t vreinterpretq_f64_s8(int8x16_t __p0) {
36508  float64x2_t __ret;
36509  __ret = (float64x2_t)(__p0);
36510  return __ret;
36511}
36512#else
36513__ai float64x2_t vreinterpretq_f64_s8(int8x16_t __p0) {
36514  float64x2_t __ret;
36515  __ret = (float64x2_t)(__p0);
36516  return __ret;
36517}
36518#endif
36519
36520#ifdef __LITTLE_ENDIAN__
36521__ai float64x2_t vreinterpretq_f64_f32(float32x4_t __p0) {
36522  float64x2_t __ret;
36523  __ret = (float64x2_t)(__p0);
36524  return __ret;
36525}
36526#else
36527__ai float64x2_t vreinterpretq_f64_f32(float32x4_t __p0) {
36528  float64x2_t __ret;
36529  __ret = (float64x2_t)(__p0);
36530  return __ret;
36531}
36532#endif
36533
36534#ifdef __LITTLE_ENDIAN__
36535__ai float64x2_t vreinterpretq_f64_f16(float16x8_t __p0) {
36536  float64x2_t __ret;
36537  __ret = (float64x2_t)(__p0);
36538  return __ret;
36539}
36540#else
36541__ai float64x2_t vreinterpretq_f64_f16(float16x8_t __p0) {
36542  float64x2_t __ret;
36543  __ret = (float64x2_t)(__p0);
36544  return __ret;
36545}
36546#endif
36547
36548#ifdef __LITTLE_ENDIAN__
36549__ai float64x2_t vreinterpretq_f64_s32(int32x4_t __p0) {
36550  float64x2_t __ret;
36551  __ret = (float64x2_t)(__p0);
36552  return __ret;
36553}
36554#else
36555__ai float64x2_t vreinterpretq_f64_s32(int32x4_t __p0) {
36556  float64x2_t __ret;
36557  __ret = (float64x2_t)(__p0);
36558  return __ret;
36559}
36560#endif
36561
36562#ifdef __LITTLE_ENDIAN__
36563__ai float64x2_t vreinterpretq_f64_s64(int64x2_t __p0) {
36564  float64x2_t __ret;
36565  __ret = (float64x2_t)(__p0);
36566  return __ret;
36567}
36568#else
36569__ai float64x2_t vreinterpretq_f64_s64(int64x2_t __p0) {
36570  float64x2_t __ret;
36571  __ret = (float64x2_t)(__p0);
36572  return __ret;
36573}
36574#endif
36575
36576#ifdef __LITTLE_ENDIAN__
36577__ai float64x2_t vreinterpretq_f64_s16(int16x8_t __p0) {
36578  float64x2_t __ret;
36579  __ret = (float64x2_t)(__p0);
36580  return __ret;
36581}
36582#else
36583__ai float64x2_t vreinterpretq_f64_s16(int16x8_t __p0) {
36584  float64x2_t __ret;
36585  __ret = (float64x2_t)(__p0);
36586  return __ret;
36587}
36588#endif
36589
36590#ifdef __LITTLE_ENDIAN__
36591__ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
36592  float32x4_t __ret;
36593  __ret = (float32x4_t)(__p0);
36594  return __ret;
36595}
36596#else
36597__ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
36598  float32x4_t __ret;
36599  __ret = (float32x4_t)(__p0);
36600  return __ret;
36601}
36602#endif
36603
36604#ifdef __LITTLE_ENDIAN__
36605__ai float32x4_t vreinterpretq_f32_p128(poly128_t __p0) {
36606  float32x4_t __ret;
36607  __ret = (float32x4_t)(__p0);
36608  return __ret;
36609}
36610#else
36611__ai float32x4_t vreinterpretq_f32_p128(poly128_t __p0) {
36612  float32x4_t __ret;
36613  __ret = (float32x4_t)(__p0);
36614  return __ret;
36615}
36616#endif
36617
36618#ifdef __LITTLE_ENDIAN__
36619__ai float32x4_t vreinterpretq_f32_p64(poly64x2_t __p0) {
36620  float32x4_t __ret;
36621  __ret = (float32x4_t)(__p0);
36622  return __ret;
36623}
36624#else
36625__ai float32x4_t vreinterpretq_f32_p64(poly64x2_t __p0) {
36626  float32x4_t __ret;
36627  __ret = (float32x4_t)(__p0);
36628  return __ret;
36629}
36630#endif
36631
36632#ifdef __LITTLE_ENDIAN__
36633__ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
36634  float32x4_t __ret;
36635  __ret = (float32x4_t)(__p0);
36636  return __ret;
36637}
36638#else
36639__ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
36640  float32x4_t __ret;
36641  __ret = (float32x4_t)(__p0);
36642  return __ret;
36643}
36644#endif
36645
36646#ifdef __LITTLE_ENDIAN__
36647__ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
36648  float32x4_t __ret;
36649  __ret = (float32x4_t)(__p0);
36650  return __ret;
36651}
36652#else
36653__ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
36654  float32x4_t __ret;
36655  __ret = (float32x4_t)(__p0);
36656  return __ret;
36657}
36658#endif
36659
36660#ifdef __LITTLE_ENDIAN__
36661__ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
36662  float32x4_t __ret;
36663  __ret = (float32x4_t)(__p0);
36664  return __ret;
36665}
36666#else
36667__ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
36668  float32x4_t __ret;
36669  __ret = (float32x4_t)(__p0);
36670  return __ret;
36671}
36672#endif
36673
36674#ifdef __LITTLE_ENDIAN__
36675__ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
36676  float32x4_t __ret;
36677  __ret = (float32x4_t)(__p0);
36678  return __ret;
36679}
36680#else
36681__ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
36682  float32x4_t __ret;
36683  __ret = (float32x4_t)(__p0);
36684  return __ret;
36685}
36686#endif
36687
36688#ifdef __LITTLE_ENDIAN__
36689__ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
36690  float32x4_t __ret;
36691  __ret = (float32x4_t)(__p0);
36692  return __ret;
36693}
36694#else
36695__ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
36696  float32x4_t __ret;
36697  __ret = (float32x4_t)(__p0);
36698  return __ret;
36699}
36700#endif
36701
36702#ifdef __LITTLE_ENDIAN__
36703__ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
36704  float32x4_t __ret;
36705  __ret = (float32x4_t)(__p0);
36706  return __ret;
36707}
36708#else
36709__ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
36710  float32x4_t __ret;
36711  __ret = (float32x4_t)(__p0);
36712  return __ret;
36713}
36714#endif
36715
36716#ifdef __LITTLE_ENDIAN__
36717__ai float32x4_t vreinterpretq_f32_f64(float64x2_t __p0) {
36718  float32x4_t __ret;
36719  __ret = (float32x4_t)(__p0);
36720  return __ret;
36721}
36722#else
36723__ai float32x4_t vreinterpretq_f32_f64(float64x2_t __p0) {
36724  float32x4_t __ret;
36725  __ret = (float32x4_t)(__p0);
36726  return __ret;
36727}
36728#endif
36729
36730#ifdef __LITTLE_ENDIAN__
36731__ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
36732  float32x4_t __ret;
36733  __ret = (float32x4_t)(__p0);
36734  return __ret;
36735}
36736#else
36737__ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
36738  float32x4_t __ret;
36739  __ret = (float32x4_t)(__p0);
36740  return __ret;
36741}
36742#endif
36743
36744#ifdef __LITTLE_ENDIAN__
36745__ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
36746  float32x4_t __ret;
36747  __ret = (float32x4_t)(__p0);
36748  return __ret;
36749}
36750#else
36751__ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
36752  float32x4_t __ret;
36753  __ret = (float32x4_t)(__p0);
36754  return __ret;
36755}
36756#endif
36757
36758#ifdef __LITTLE_ENDIAN__
36759__ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
36760  float32x4_t __ret;
36761  __ret = (float32x4_t)(__p0);
36762  return __ret;
36763}
36764#else
36765__ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
36766  float32x4_t __ret;
36767  __ret = (float32x4_t)(__p0);
36768  return __ret;
36769}
36770#endif
36771
36772#ifdef __LITTLE_ENDIAN__
36773__ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
36774  float32x4_t __ret;
36775  __ret = (float32x4_t)(__p0);
36776  return __ret;
36777}
36778#else
36779__ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
36780  float32x4_t __ret;
36781  __ret = (float32x4_t)(__p0);
36782  return __ret;
36783}
36784#endif
36785
36786#ifdef __LITTLE_ENDIAN__
36787__ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
36788  float16x8_t __ret;
36789  __ret = (float16x8_t)(__p0);
36790  return __ret;
36791}
36792#else
36793__ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
36794  float16x8_t __ret;
36795  __ret = (float16x8_t)(__p0);
36796  return __ret;
36797}
36798#endif
36799
36800#ifdef __LITTLE_ENDIAN__
36801__ai float16x8_t vreinterpretq_f16_p128(poly128_t __p0) {
36802  float16x8_t __ret;
36803  __ret = (float16x8_t)(__p0);
36804  return __ret;
36805}
36806#else
36807__ai float16x8_t vreinterpretq_f16_p128(poly128_t __p0) {
36808  float16x8_t __ret;
36809  __ret = (float16x8_t)(__p0);
36810  return __ret;
36811}
36812#endif
36813
36814#ifdef __LITTLE_ENDIAN__
36815__ai float16x8_t vreinterpretq_f16_p64(poly64x2_t __p0) {
36816  float16x8_t __ret;
36817  __ret = (float16x8_t)(__p0);
36818  return __ret;
36819}
36820#else
36821__ai float16x8_t vreinterpretq_f16_p64(poly64x2_t __p0) {
36822  float16x8_t __ret;
36823  __ret = (float16x8_t)(__p0);
36824  return __ret;
36825}
36826#endif
36827
36828#ifdef __LITTLE_ENDIAN__
36829__ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
36830  float16x8_t __ret;
36831  __ret = (float16x8_t)(__p0);
36832  return __ret;
36833}
36834#else
36835__ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
36836  float16x8_t __ret;
36837  __ret = (float16x8_t)(__p0);
36838  return __ret;
36839}
36840#endif
36841
36842#ifdef __LITTLE_ENDIAN__
36843__ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
36844  float16x8_t __ret;
36845  __ret = (float16x8_t)(__p0);
36846  return __ret;
36847}
36848#else
36849__ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
36850  float16x8_t __ret;
36851  __ret = (float16x8_t)(__p0);
36852  return __ret;
36853}
36854#endif
36855
36856#ifdef __LITTLE_ENDIAN__
36857__ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
36858  float16x8_t __ret;
36859  __ret = (float16x8_t)(__p0);
36860  return __ret;
36861}
36862#else
36863__ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
36864  float16x8_t __ret;
36865  __ret = (float16x8_t)(__p0);
36866  return __ret;
36867}
36868#endif
36869
36870#ifdef __LITTLE_ENDIAN__
36871__ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
36872  float16x8_t __ret;
36873  __ret = (float16x8_t)(__p0);
36874  return __ret;
36875}
36876#else
36877__ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
36878  float16x8_t __ret;
36879  __ret = (float16x8_t)(__p0);
36880  return __ret;
36881}
36882#endif
36883
36884#ifdef __LITTLE_ENDIAN__
36885__ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
36886  float16x8_t __ret;
36887  __ret = (float16x8_t)(__p0);
36888  return __ret;
36889}
36890#else
36891__ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
36892  float16x8_t __ret;
36893  __ret = (float16x8_t)(__p0);
36894  return __ret;
36895}
36896#endif
36897
36898#ifdef __LITTLE_ENDIAN__
36899__ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
36900  float16x8_t __ret;
36901  __ret = (float16x8_t)(__p0);
36902  return __ret;
36903}
36904#else
36905__ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
36906  float16x8_t __ret;
36907  __ret = (float16x8_t)(__p0);
36908  return __ret;
36909}
36910#endif
36911
36912#ifdef __LITTLE_ENDIAN__
36913__ai float16x8_t vreinterpretq_f16_f64(float64x2_t __p0) {
36914  float16x8_t __ret;
36915  __ret = (float16x8_t)(__p0);
36916  return __ret;
36917}
36918#else
36919__ai float16x8_t vreinterpretq_f16_f64(float64x2_t __p0) {
36920  float16x8_t __ret;
36921  __ret = (float16x8_t)(__p0);
36922  return __ret;
36923}
36924#endif
36925
36926#ifdef __LITTLE_ENDIAN__
36927__ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
36928  float16x8_t __ret;
36929  __ret = (float16x8_t)(__p0);
36930  return __ret;
36931}
36932#else
36933__ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
36934  float16x8_t __ret;
36935  __ret = (float16x8_t)(__p0);
36936  return __ret;
36937}
36938#endif
36939
36940#ifdef __LITTLE_ENDIAN__
36941__ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
36942  float16x8_t __ret;
36943  __ret = (float16x8_t)(__p0);
36944  return __ret;
36945}
36946#else
36947__ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
36948  float16x8_t __ret;
36949  __ret = (float16x8_t)(__p0);
36950  return __ret;
36951}
36952#endif
36953
36954#ifdef __LITTLE_ENDIAN__
36955__ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
36956  float16x8_t __ret;
36957  __ret = (float16x8_t)(__p0);
36958  return __ret;
36959}
36960#else
36961__ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
36962  float16x8_t __ret;
36963  __ret = (float16x8_t)(__p0);
36964  return __ret;
36965}
36966#endif
36967
36968#ifdef __LITTLE_ENDIAN__
36969__ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
36970  float16x8_t __ret;
36971  __ret = (float16x8_t)(__p0);
36972  return __ret;
36973}
36974#else
36975__ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
36976  float16x8_t __ret;
36977  __ret = (float16x8_t)(__p0);
36978  return __ret;
36979}
36980#endif
36981
36982#ifdef __LITTLE_ENDIAN__
36983__ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
36984  int32x4_t __ret;
36985  __ret = (int32x4_t)(__p0);
36986  return __ret;
36987}
36988#else
36989__ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
36990  int32x4_t __ret;
36991  __ret = (int32x4_t)(__p0);
36992  return __ret;
36993}
36994#endif
36995
36996#ifdef __LITTLE_ENDIAN__
36997__ai int32x4_t vreinterpretq_s32_p128(poly128_t __p0) {
36998  int32x4_t __ret;
36999  __ret = (int32x4_t)(__p0);
37000  return __ret;
37001}
37002#else
37003__ai int32x4_t vreinterpretq_s32_p128(poly128_t __p0) {
37004  int32x4_t __ret;
37005  __ret = (int32x4_t)(__p0);
37006  return __ret;
37007}
37008#endif
37009
37010#ifdef __LITTLE_ENDIAN__
37011__ai int32x4_t vreinterpretq_s32_p64(poly64x2_t __p0) {
37012  int32x4_t __ret;
37013  __ret = (int32x4_t)(__p0);
37014  return __ret;
37015}
37016#else
37017__ai int32x4_t vreinterpretq_s32_p64(poly64x2_t __p0) {
37018  int32x4_t __ret;
37019  __ret = (int32x4_t)(__p0);
37020  return __ret;
37021}
37022#endif
37023
37024#ifdef __LITTLE_ENDIAN__
37025__ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
37026  int32x4_t __ret;
37027  __ret = (int32x4_t)(__p0);
37028  return __ret;
37029}
37030#else
37031__ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
37032  int32x4_t __ret;
37033  __ret = (int32x4_t)(__p0);
37034  return __ret;
37035}
37036#endif
37037
37038#ifdef __LITTLE_ENDIAN__
37039__ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
37040  int32x4_t __ret;
37041  __ret = (int32x4_t)(__p0);
37042  return __ret;
37043}
37044#else
37045__ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
37046  int32x4_t __ret;
37047  __ret = (int32x4_t)(__p0);
37048  return __ret;
37049}
37050#endif
37051
37052#ifdef __LITTLE_ENDIAN__
37053__ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
37054  int32x4_t __ret;
37055  __ret = (int32x4_t)(__p0);
37056  return __ret;
37057}
37058#else
37059__ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
37060  int32x4_t __ret;
37061  __ret = (int32x4_t)(__p0);
37062  return __ret;
37063}
37064#endif
37065
37066#ifdef __LITTLE_ENDIAN__
37067__ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
37068  int32x4_t __ret;
37069  __ret = (int32x4_t)(__p0);
37070  return __ret;
37071}
37072#else
37073__ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
37074  int32x4_t __ret;
37075  __ret = (int32x4_t)(__p0);
37076  return __ret;
37077}
37078#endif
37079
37080#ifdef __LITTLE_ENDIAN__
37081__ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
37082  int32x4_t __ret;
37083  __ret = (int32x4_t)(__p0);
37084  return __ret;
37085}
37086#else
37087__ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
37088  int32x4_t __ret;
37089  __ret = (int32x4_t)(__p0);
37090  return __ret;
37091}
37092#endif
37093
37094#ifdef __LITTLE_ENDIAN__
37095__ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
37096  int32x4_t __ret;
37097  __ret = (int32x4_t)(__p0);
37098  return __ret;
37099}
37100#else
37101__ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
37102  int32x4_t __ret;
37103  __ret = (int32x4_t)(__p0);
37104  return __ret;
37105}
37106#endif
37107
37108#ifdef __LITTLE_ENDIAN__
37109__ai int32x4_t vreinterpretq_s32_f64(float64x2_t __p0) {
37110  int32x4_t __ret;
37111  __ret = (int32x4_t)(__p0);
37112  return __ret;
37113}
37114#else
37115__ai int32x4_t vreinterpretq_s32_f64(float64x2_t __p0) {
37116  int32x4_t __ret;
37117  __ret = (int32x4_t)(__p0);
37118  return __ret;
37119}
37120#endif
37121
37122#ifdef __LITTLE_ENDIAN__
37123__ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
37124  int32x4_t __ret;
37125  __ret = (int32x4_t)(__p0);
37126  return __ret;
37127}
37128#else
37129__ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
37130  int32x4_t __ret;
37131  __ret = (int32x4_t)(__p0);
37132  return __ret;
37133}
37134#endif
37135
37136#ifdef __LITTLE_ENDIAN__
37137__ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
37138  int32x4_t __ret;
37139  __ret = (int32x4_t)(__p0);
37140  return __ret;
37141}
37142#else
37143__ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
37144  int32x4_t __ret;
37145  __ret = (int32x4_t)(__p0);
37146  return __ret;
37147}
37148#endif
37149
37150#ifdef __LITTLE_ENDIAN__
37151__ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
37152  int32x4_t __ret;
37153  __ret = (int32x4_t)(__p0);
37154  return __ret;
37155}
37156#else
37157__ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
37158  int32x4_t __ret;
37159  __ret = (int32x4_t)(__p0);
37160  return __ret;
37161}
37162#endif
37163
37164#ifdef __LITTLE_ENDIAN__
37165__ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
37166  int32x4_t __ret;
37167  __ret = (int32x4_t)(__p0);
37168  return __ret;
37169}
37170#else
37171__ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
37172  int32x4_t __ret;
37173  __ret = (int32x4_t)(__p0);
37174  return __ret;
37175}
37176#endif
37177
37178#ifdef __LITTLE_ENDIAN__
37179__ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
37180  int64x2_t __ret;
37181  __ret = (int64x2_t)(__p0);
37182  return __ret;
37183}
37184#else
37185__ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
37186  int64x2_t __ret;
37187  __ret = (int64x2_t)(__p0);
37188  return __ret;
37189}
37190#endif
37191
37192#ifdef __LITTLE_ENDIAN__
37193__ai int64x2_t vreinterpretq_s64_p128(poly128_t __p0) {
37194  int64x2_t __ret;
37195  __ret = (int64x2_t)(__p0);
37196  return __ret;
37197}
37198#else
37199__ai int64x2_t vreinterpretq_s64_p128(poly128_t __p0) {
37200  int64x2_t __ret;
37201  __ret = (int64x2_t)(__p0);
37202  return __ret;
37203}
37204#endif
37205
37206#ifdef __LITTLE_ENDIAN__
37207__ai int64x2_t vreinterpretq_s64_p64(poly64x2_t __p0) {
37208  int64x2_t __ret;
37209  __ret = (int64x2_t)(__p0);
37210  return __ret;
37211}
37212#else
37213__ai int64x2_t vreinterpretq_s64_p64(poly64x2_t __p0) {
37214  int64x2_t __ret;
37215  __ret = (int64x2_t)(__p0);
37216  return __ret;
37217}
37218#endif
37219
37220#ifdef __LITTLE_ENDIAN__
37221__ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
37222  int64x2_t __ret;
37223  __ret = (int64x2_t)(__p0);
37224  return __ret;
37225}
37226#else
37227__ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
37228  int64x2_t __ret;
37229  __ret = (int64x2_t)(__p0);
37230  return __ret;
37231}
37232#endif
37233
37234#ifdef __LITTLE_ENDIAN__
37235__ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
37236  int64x2_t __ret;
37237  __ret = (int64x2_t)(__p0);
37238  return __ret;
37239}
37240#else
37241__ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
37242  int64x2_t __ret;
37243  __ret = (int64x2_t)(__p0);
37244  return __ret;
37245}
37246#endif
37247
37248#ifdef __LITTLE_ENDIAN__
37249__ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
37250  int64x2_t __ret;
37251  __ret = (int64x2_t)(__p0);
37252  return __ret;
37253}
37254#else
37255__ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
37256  int64x2_t __ret;
37257  __ret = (int64x2_t)(__p0);
37258  return __ret;
37259}
37260#endif
37261
37262#ifdef __LITTLE_ENDIAN__
37263__ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
37264  int64x2_t __ret;
37265  __ret = (int64x2_t)(__p0);
37266  return __ret;
37267}
37268#else
37269__ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
37270  int64x2_t __ret;
37271  __ret = (int64x2_t)(__p0);
37272  return __ret;
37273}
37274#endif
37275
37276#ifdef __LITTLE_ENDIAN__
37277__ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
37278  int64x2_t __ret;
37279  __ret = (int64x2_t)(__p0);
37280  return __ret;
37281}
37282#else
37283__ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
37284  int64x2_t __ret;
37285  __ret = (int64x2_t)(__p0);
37286  return __ret;
37287}
37288#endif
37289
37290#ifdef __LITTLE_ENDIAN__
37291__ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
37292  int64x2_t __ret;
37293  __ret = (int64x2_t)(__p0);
37294  return __ret;
37295}
37296#else
37297__ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
37298  int64x2_t __ret;
37299  __ret = (int64x2_t)(__p0);
37300  return __ret;
37301}
37302#endif
37303
37304#ifdef __LITTLE_ENDIAN__
37305__ai int64x2_t vreinterpretq_s64_f64(float64x2_t __p0) {
37306  int64x2_t __ret;
37307  __ret = (int64x2_t)(__p0);
37308  return __ret;
37309}
37310#else
37311__ai int64x2_t vreinterpretq_s64_f64(float64x2_t __p0) {
37312  int64x2_t __ret;
37313  __ret = (int64x2_t)(__p0);
37314  return __ret;
37315}
37316#endif
37317
37318#ifdef __LITTLE_ENDIAN__
37319__ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
37320  int64x2_t __ret;
37321  __ret = (int64x2_t)(__p0);
37322  return __ret;
37323}
37324#else
37325__ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
37326  int64x2_t __ret;
37327  __ret = (int64x2_t)(__p0);
37328  return __ret;
37329}
37330#endif
37331
37332#ifdef __LITTLE_ENDIAN__
37333__ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
37334  int64x2_t __ret;
37335  __ret = (int64x2_t)(__p0);
37336  return __ret;
37337}
37338#else
37339__ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
37340  int64x2_t __ret;
37341  __ret = (int64x2_t)(__p0);
37342  return __ret;
37343}
37344#endif
37345
37346#ifdef __LITTLE_ENDIAN__
37347__ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
37348  int64x2_t __ret;
37349  __ret = (int64x2_t)(__p0);
37350  return __ret;
37351}
37352#else
37353__ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
37354  int64x2_t __ret;
37355  __ret = (int64x2_t)(__p0);
37356  return __ret;
37357}
37358#endif
37359
37360#ifdef __LITTLE_ENDIAN__
37361__ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
37362  int64x2_t __ret;
37363  __ret = (int64x2_t)(__p0);
37364  return __ret;
37365}
37366#else
37367__ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
37368  int64x2_t __ret;
37369  __ret = (int64x2_t)(__p0);
37370  return __ret;
37371}
37372#endif
37373
37374#ifdef __LITTLE_ENDIAN__
37375__ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
37376  int16x8_t __ret;
37377  __ret = (int16x8_t)(__p0);
37378  return __ret;
37379}
37380#else
37381__ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
37382  int16x8_t __ret;
37383  __ret = (int16x8_t)(__p0);
37384  return __ret;
37385}
37386#endif
37387
37388#ifdef __LITTLE_ENDIAN__
37389__ai int16x8_t vreinterpretq_s16_p128(poly128_t __p0) {
37390  int16x8_t __ret;
37391  __ret = (int16x8_t)(__p0);
37392  return __ret;
37393}
37394#else
37395__ai int16x8_t vreinterpretq_s16_p128(poly128_t __p0) {
37396  int16x8_t __ret;
37397  __ret = (int16x8_t)(__p0);
37398  return __ret;
37399}
37400#endif
37401
37402#ifdef __LITTLE_ENDIAN__
37403__ai int16x8_t vreinterpretq_s16_p64(poly64x2_t __p0) {
37404  int16x8_t __ret;
37405  __ret = (int16x8_t)(__p0);
37406  return __ret;
37407}
37408#else
37409__ai int16x8_t vreinterpretq_s16_p64(poly64x2_t __p0) {
37410  int16x8_t __ret;
37411  __ret = (int16x8_t)(__p0);
37412  return __ret;
37413}
37414#endif
37415
37416#ifdef __LITTLE_ENDIAN__
37417__ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
37418  int16x8_t __ret;
37419  __ret = (int16x8_t)(__p0);
37420  return __ret;
37421}
37422#else
37423__ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
37424  int16x8_t __ret;
37425  __ret = (int16x8_t)(__p0);
37426  return __ret;
37427}
37428#endif
37429
37430#ifdef __LITTLE_ENDIAN__
37431__ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
37432  int16x8_t __ret;
37433  __ret = (int16x8_t)(__p0);
37434  return __ret;
37435}
37436#else
37437__ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
37438  int16x8_t __ret;
37439  __ret = (int16x8_t)(__p0);
37440  return __ret;
37441}
37442#endif
37443
37444#ifdef __LITTLE_ENDIAN__
37445__ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
37446  int16x8_t __ret;
37447  __ret = (int16x8_t)(__p0);
37448  return __ret;
37449}
37450#else
37451__ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
37452  int16x8_t __ret;
37453  __ret = (int16x8_t)(__p0);
37454  return __ret;
37455}
37456#endif
37457
37458#ifdef __LITTLE_ENDIAN__
37459__ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
37460  int16x8_t __ret;
37461  __ret = (int16x8_t)(__p0);
37462  return __ret;
37463}
37464#else
37465__ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
37466  int16x8_t __ret;
37467  __ret = (int16x8_t)(__p0);
37468  return __ret;
37469}
37470#endif
37471
37472#ifdef __LITTLE_ENDIAN__
37473__ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
37474  int16x8_t __ret;
37475  __ret = (int16x8_t)(__p0);
37476  return __ret;
37477}
37478#else
37479__ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
37480  int16x8_t __ret;
37481  __ret = (int16x8_t)(__p0);
37482  return __ret;
37483}
37484#endif
37485
37486#ifdef __LITTLE_ENDIAN__
37487__ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
37488  int16x8_t __ret;
37489  __ret = (int16x8_t)(__p0);
37490  return __ret;
37491}
37492#else
37493__ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
37494  int16x8_t __ret;
37495  __ret = (int16x8_t)(__p0);
37496  return __ret;
37497}
37498#endif
37499
37500#ifdef __LITTLE_ENDIAN__
37501__ai int16x8_t vreinterpretq_s16_f64(float64x2_t __p0) {
37502  int16x8_t __ret;
37503  __ret = (int16x8_t)(__p0);
37504  return __ret;
37505}
37506#else
37507__ai int16x8_t vreinterpretq_s16_f64(float64x2_t __p0) {
37508  int16x8_t __ret;
37509  __ret = (int16x8_t)(__p0);
37510  return __ret;
37511}
37512#endif
37513
37514#ifdef __LITTLE_ENDIAN__
37515__ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
37516  int16x8_t __ret;
37517  __ret = (int16x8_t)(__p0);
37518  return __ret;
37519}
37520#else
37521__ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
37522  int16x8_t __ret;
37523  __ret = (int16x8_t)(__p0);
37524  return __ret;
37525}
37526#endif
37527
37528#ifdef __LITTLE_ENDIAN__
37529__ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
37530  int16x8_t __ret;
37531  __ret = (int16x8_t)(__p0);
37532  return __ret;
37533}
37534#else
37535__ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
37536  int16x8_t __ret;
37537  __ret = (int16x8_t)(__p0);
37538  return __ret;
37539}
37540#endif
37541
37542#ifdef __LITTLE_ENDIAN__
37543__ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
37544  int16x8_t __ret;
37545  __ret = (int16x8_t)(__p0);
37546  return __ret;
37547}
37548#else
37549__ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
37550  int16x8_t __ret;
37551  __ret = (int16x8_t)(__p0);
37552  return __ret;
37553}
37554#endif
37555
37556#ifdef __LITTLE_ENDIAN__
37557__ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
37558  int16x8_t __ret;
37559  __ret = (int16x8_t)(__p0);
37560  return __ret;
37561}
37562#else
37563__ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
37564  int16x8_t __ret;
37565  __ret = (int16x8_t)(__p0);
37566  return __ret;
37567}
37568#endif
37569
37570#ifdef __LITTLE_ENDIAN__
37571__ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
37572  uint8x8_t __ret;
37573  __ret = (uint8x8_t)(__p0);
37574  return __ret;
37575}
37576#else
37577__ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
37578  uint8x8_t __ret;
37579  __ret = (uint8x8_t)(__p0);
37580  return __ret;
37581}
37582#endif
37583
37584#ifdef __LITTLE_ENDIAN__
37585__ai uint8x8_t vreinterpret_u8_p64(poly64x1_t __p0) {
37586  uint8x8_t __ret;
37587  __ret = (uint8x8_t)(__p0);
37588  return __ret;
37589}
37590#else
37591__ai uint8x8_t vreinterpret_u8_p64(poly64x1_t __p0) {
37592  uint8x8_t __ret;
37593  __ret = (uint8x8_t)(__p0);
37594  return __ret;
37595}
37596#endif
37597
37598#ifdef __LITTLE_ENDIAN__
37599__ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
37600  uint8x8_t __ret;
37601  __ret = (uint8x8_t)(__p0);
37602  return __ret;
37603}
37604#else
37605__ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
37606  uint8x8_t __ret;
37607  __ret = (uint8x8_t)(__p0);
37608  return __ret;
37609}
37610#endif
37611
37612#ifdef __LITTLE_ENDIAN__
37613__ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
37614  uint8x8_t __ret;
37615  __ret = (uint8x8_t)(__p0);
37616  return __ret;
37617}
37618#else
37619__ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
37620  uint8x8_t __ret;
37621  __ret = (uint8x8_t)(__p0);
37622  return __ret;
37623}
37624#endif
37625
37626#ifdef __LITTLE_ENDIAN__
37627__ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
37628  uint8x8_t __ret;
37629  __ret = (uint8x8_t)(__p0);
37630  return __ret;
37631}
37632#else
37633__ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
37634  uint8x8_t __ret;
37635  __ret = (uint8x8_t)(__p0);
37636  return __ret;
37637}
37638#endif
37639
37640#ifdef __LITTLE_ENDIAN__
37641__ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
37642  uint8x8_t __ret;
37643  __ret = (uint8x8_t)(__p0);
37644  return __ret;
37645}
37646#else
37647__ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
37648  uint8x8_t __ret;
37649  __ret = (uint8x8_t)(__p0);
37650  return __ret;
37651}
37652#endif
37653
37654#ifdef __LITTLE_ENDIAN__
37655__ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
37656  uint8x8_t __ret;
37657  __ret = (uint8x8_t)(__p0);
37658  return __ret;
37659}
37660#else
37661__ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
37662  uint8x8_t __ret;
37663  __ret = (uint8x8_t)(__p0);
37664  return __ret;
37665}
37666#endif
37667
37668#ifdef __LITTLE_ENDIAN__
37669__ai uint8x8_t vreinterpret_u8_f64(float64x1_t __p0) {
37670  uint8x8_t __ret;
37671  __ret = (uint8x8_t)(__p0);
37672  return __ret;
37673}
37674#else
37675__ai uint8x8_t vreinterpret_u8_f64(float64x1_t __p0) {
37676  uint8x8_t __ret;
37677  __ret = (uint8x8_t)(__p0);
37678  return __ret;
37679}
37680#endif
37681
37682#ifdef __LITTLE_ENDIAN__
37683__ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
37684  uint8x8_t __ret;
37685  __ret = (uint8x8_t)(__p0);
37686  return __ret;
37687}
37688#else
37689__ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
37690  uint8x8_t __ret;
37691  __ret = (uint8x8_t)(__p0);
37692  return __ret;
37693}
37694#endif
37695
37696#ifdef __LITTLE_ENDIAN__
37697__ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
37698  uint8x8_t __ret;
37699  __ret = (uint8x8_t)(__p0);
37700  return __ret;
37701}
37702#else
37703__ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
37704  uint8x8_t __ret;
37705  __ret = (uint8x8_t)(__p0);
37706  return __ret;
37707}
37708#endif
37709
37710#ifdef __LITTLE_ENDIAN__
37711__ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
37712  uint8x8_t __ret;
37713  __ret = (uint8x8_t)(__p0);
37714  return __ret;
37715}
37716#else
37717__ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
37718  uint8x8_t __ret;
37719  __ret = (uint8x8_t)(__p0);
37720  return __ret;
37721}
37722#endif
37723
37724#ifdef __LITTLE_ENDIAN__
37725__ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
37726  uint8x8_t __ret;
37727  __ret = (uint8x8_t)(__p0);
37728  return __ret;
37729}
37730#else
37731__ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
37732  uint8x8_t __ret;
37733  __ret = (uint8x8_t)(__p0);
37734  return __ret;
37735}
37736#endif
37737
37738#ifdef __LITTLE_ENDIAN__
37739__ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
37740  uint8x8_t __ret;
37741  __ret = (uint8x8_t)(__p0);
37742  return __ret;
37743}
37744#else
37745__ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
37746  uint8x8_t __ret;
37747  __ret = (uint8x8_t)(__p0);
37748  return __ret;
37749}
37750#endif
37751
37752#ifdef __LITTLE_ENDIAN__
37753__ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
37754  uint32x2_t __ret;
37755  __ret = (uint32x2_t)(__p0);
37756  return __ret;
37757}
37758#else
37759__ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
37760  uint32x2_t __ret;
37761  __ret = (uint32x2_t)(__p0);
37762  return __ret;
37763}
37764#endif
37765
37766#ifdef __LITTLE_ENDIAN__
37767__ai uint32x2_t vreinterpret_u32_p64(poly64x1_t __p0) {
37768  uint32x2_t __ret;
37769  __ret = (uint32x2_t)(__p0);
37770  return __ret;
37771}
37772#else
37773__ai uint32x2_t vreinterpret_u32_p64(poly64x1_t __p0) {
37774  uint32x2_t __ret;
37775  __ret = (uint32x2_t)(__p0);
37776  return __ret;
37777}
37778#endif
37779
37780#ifdef __LITTLE_ENDIAN__
37781__ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
37782  uint32x2_t __ret;
37783  __ret = (uint32x2_t)(__p0);
37784  return __ret;
37785}
37786#else
37787__ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
37788  uint32x2_t __ret;
37789  __ret = (uint32x2_t)(__p0);
37790  return __ret;
37791}
37792#endif
37793
37794#ifdef __LITTLE_ENDIAN__
37795__ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
37796  uint32x2_t __ret;
37797  __ret = (uint32x2_t)(__p0);
37798  return __ret;
37799}
37800#else
37801__ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
37802  uint32x2_t __ret;
37803  __ret = (uint32x2_t)(__p0);
37804  return __ret;
37805}
37806#endif
37807
37808#ifdef __LITTLE_ENDIAN__
37809__ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
37810  uint32x2_t __ret;
37811  __ret = (uint32x2_t)(__p0);
37812  return __ret;
37813}
37814#else
37815__ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
37816  uint32x2_t __ret;
37817  __ret = (uint32x2_t)(__p0);
37818  return __ret;
37819}
37820#endif
37821
37822#ifdef __LITTLE_ENDIAN__
37823__ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
37824  uint32x2_t __ret;
37825  __ret = (uint32x2_t)(__p0);
37826  return __ret;
37827}
37828#else
37829__ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
37830  uint32x2_t __ret;
37831  __ret = (uint32x2_t)(__p0);
37832  return __ret;
37833}
37834#endif
37835
37836#ifdef __LITTLE_ENDIAN__
37837__ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
37838  uint32x2_t __ret;
37839  __ret = (uint32x2_t)(__p0);
37840  return __ret;
37841}
37842#else
37843__ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
37844  uint32x2_t __ret;
37845  __ret = (uint32x2_t)(__p0);
37846  return __ret;
37847}
37848#endif
37849
37850#ifdef __LITTLE_ENDIAN__
37851__ai uint32x2_t vreinterpret_u32_f64(float64x1_t __p0) {
37852  uint32x2_t __ret;
37853  __ret = (uint32x2_t)(__p0);
37854  return __ret;
37855}
37856#else
37857__ai uint32x2_t vreinterpret_u32_f64(float64x1_t __p0) {
37858  uint32x2_t __ret;
37859  __ret = (uint32x2_t)(__p0);
37860  return __ret;
37861}
37862#endif
37863
37864#ifdef __LITTLE_ENDIAN__
37865__ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
37866  uint32x2_t __ret;
37867  __ret = (uint32x2_t)(__p0);
37868  return __ret;
37869}
37870#else
37871__ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
37872  uint32x2_t __ret;
37873  __ret = (uint32x2_t)(__p0);
37874  return __ret;
37875}
37876#endif
37877
37878#ifdef __LITTLE_ENDIAN__
37879__ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
37880  uint32x2_t __ret;
37881  __ret = (uint32x2_t)(__p0);
37882  return __ret;
37883}
37884#else
37885__ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
37886  uint32x2_t __ret;
37887  __ret = (uint32x2_t)(__p0);
37888  return __ret;
37889}
37890#endif
37891
37892#ifdef __LITTLE_ENDIAN__
37893__ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
37894  uint32x2_t __ret;
37895  __ret = (uint32x2_t)(__p0);
37896  return __ret;
37897}
37898#else
37899__ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
37900  uint32x2_t __ret;
37901  __ret = (uint32x2_t)(__p0);
37902  return __ret;
37903}
37904#endif
37905
37906#ifdef __LITTLE_ENDIAN__
37907__ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
37908  uint32x2_t __ret;
37909  __ret = (uint32x2_t)(__p0);
37910  return __ret;
37911}
37912#else
37913__ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
37914  uint32x2_t __ret;
37915  __ret = (uint32x2_t)(__p0);
37916  return __ret;
37917}
37918#endif
37919
37920#ifdef __LITTLE_ENDIAN__
37921__ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
37922  uint32x2_t __ret;
37923  __ret = (uint32x2_t)(__p0);
37924  return __ret;
37925}
37926#else
37927__ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
37928  uint32x2_t __ret;
37929  __ret = (uint32x2_t)(__p0);
37930  return __ret;
37931}
37932#endif
37933
37934#ifdef __LITTLE_ENDIAN__
37935__ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
37936  uint64x1_t __ret;
37937  __ret = (uint64x1_t)(__p0);
37938  return __ret;
37939}
37940#else
37941__ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
37942  uint64x1_t __ret;
37943  __ret = (uint64x1_t)(__p0);
37944  return __ret;
37945}
37946#endif
37947
37948#ifdef __LITTLE_ENDIAN__
37949__ai uint64x1_t vreinterpret_u64_p64(poly64x1_t __p0) {
37950  uint64x1_t __ret;
37951  __ret = (uint64x1_t)(__p0);
37952  return __ret;
37953}
37954#else
37955__ai uint64x1_t vreinterpret_u64_p64(poly64x1_t __p0) {
37956  uint64x1_t __ret;
37957  __ret = (uint64x1_t)(__p0);
37958  return __ret;
37959}
37960#endif
37961
37962#ifdef __LITTLE_ENDIAN__
37963__ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
37964  uint64x1_t __ret;
37965  __ret = (uint64x1_t)(__p0);
37966  return __ret;
37967}
37968#else
37969__ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
37970  uint64x1_t __ret;
37971  __ret = (uint64x1_t)(__p0);
37972  return __ret;
37973}
37974#endif
37975
37976#ifdef __LITTLE_ENDIAN__
37977__ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
37978  uint64x1_t __ret;
37979  __ret = (uint64x1_t)(__p0);
37980  return __ret;
37981}
37982#else
37983__ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
37984  uint64x1_t __ret;
37985  __ret = (uint64x1_t)(__p0);
37986  return __ret;
37987}
37988#endif
37989
37990#ifdef __LITTLE_ENDIAN__
37991__ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
37992  uint64x1_t __ret;
37993  __ret = (uint64x1_t)(__p0);
37994  return __ret;
37995}
37996#else
37997__ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
37998  uint64x1_t __ret;
37999  __ret = (uint64x1_t)(__p0);
38000  return __ret;
38001}
38002#endif
38003
38004#ifdef __LITTLE_ENDIAN__
38005__ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
38006  uint64x1_t __ret;
38007  __ret = (uint64x1_t)(__p0);
38008  return __ret;
38009}
38010#else
38011__ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
38012  uint64x1_t __ret;
38013  __ret = (uint64x1_t)(__p0);
38014  return __ret;
38015}
38016#endif
38017
38018#ifdef __LITTLE_ENDIAN__
38019__ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
38020  uint64x1_t __ret;
38021  __ret = (uint64x1_t)(__p0);
38022  return __ret;
38023}
38024#else
38025__ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
38026  uint64x1_t __ret;
38027  __ret = (uint64x1_t)(__p0);
38028  return __ret;
38029}
38030#endif
38031
38032#ifdef __LITTLE_ENDIAN__
38033__ai uint64x1_t vreinterpret_u64_f64(float64x1_t __p0) {
38034  uint64x1_t __ret;
38035  __ret = (uint64x1_t)(__p0);
38036  return __ret;
38037}
38038#else
38039__ai uint64x1_t vreinterpret_u64_f64(float64x1_t __p0) {
38040  uint64x1_t __ret;
38041  __ret = (uint64x1_t)(__p0);
38042  return __ret;
38043}
38044#endif
38045
38046#ifdef __LITTLE_ENDIAN__
38047__ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
38048  uint64x1_t __ret;
38049  __ret = (uint64x1_t)(__p0);
38050  return __ret;
38051}
38052#else
38053__ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
38054  uint64x1_t __ret;
38055  __ret = (uint64x1_t)(__p0);
38056  return __ret;
38057}
38058#endif
38059
38060#ifdef __LITTLE_ENDIAN__
38061__ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
38062  uint64x1_t __ret;
38063  __ret = (uint64x1_t)(__p0);
38064  return __ret;
38065}
38066#else
38067__ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
38068  uint64x1_t __ret;
38069  __ret = (uint64x1_t)(__p0);
38070  return __ret;
38071}
38072#endif
38073
38074#ifdef __LITTLE_ENDIAN__
38075__ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
38076  uint64x1_t __ret;
38077  __ret = (uint64x1_t)(__p0);
38078  return __ret;
38079}
38080#else
38081__ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
38082  uint64x1_t __ret;
38083  __ret = (uint64x1_t)(__p0);
38084  return __ret;
38085}
38086#endif
38087
38088#ifdef __LITTLE_ENDIAN__
38089__ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
38090  uint64x1_t __ret;
38091  __ret = (uint64x1_t)(__p0);
38092  return __ret;
38093}
38094#else
38095__ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
38096  uint64x1_t __ret;
38097  __ret = (uint64x1_t)(__p0);
38098  return __ret;
38099}
38100#endif
38101
38102#ifdef __LITTLE_ENDIAN__
38103__ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
38104  uint64x1_t __ret;
38105  __ret = (uint64x1_t)(__p0);
38106  return __ret;
38107}
38108#else
38109__ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
38110  uint64x1_t __ret;
38111  __ret = (uint64x1_t)(__p0);
38112  return __ret;
38113}
38114#endif
38115
38116#ifdef __LITTLE_ENDIAN__
38117__ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
38118  uint16x4_t __ret;
38119  __ret = (uint16x4_t)(__p0);
38120  return __ret;
38121}
38122#else
38123__ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
38124  uint16x4_t __ret;
38125  __ret = (uint16x4_t)(__p0);
38126  return __ret;
38127}
38128#endif
38129
38130#ifdef __LITTLE_ENDIAN__
38131__ai uint16x4_t vreinterpret_u16_p64(poly64x1_t __p0) {
38132  uint16x4_t __ret;
38133  __ret = (uint16x4_t)(__p0);
38134  return __ret;
38135}
38136#else
38137__ai uint16x4_t vreinterpret_u16_p64(poly64x1_t __p0) {
38138  uint16x4_t __ret;
38139  __ret = (uint16x4_t)(__p0);
38140  return __ret;
38141}
38142#endif
38143
38144#ifdef __LITTLE_ENDIAN__
38145__ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
38146  uint16x4_t __ret;
38147  __ret = (uint16x4_t)(__p0);
38148  return __ret;
38149}
38150#else
38151__ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
38152  uint16x4_t __ret;
38153  __ret = (uint16x4_t)(__p0);
38154  return __ret;
38155}
38156#endif
38157
38158#ifdef __LITTLE_ENDIAN__
38159__ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
38160  uint16x4_t __ret;
38161  __ret = (uint16x4_t)(__p0);
38162  return __ret;
38163}
38164#else
38165__ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
38166  uint16x4_t __ret;
38167  __ret = (uint16x4_t)(__p0);
38168  return __ret;
38169}
38170#endif
38171
38172#ifdef __LITTLE_ENDIAN__
38173__ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
38174  uint16x4_t __ret;
38175  __ret = (uint16x4_t)(__p0);
38176  return __ret;
38177}
38178#else
38179__ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
38180  uint16x4_t __ret;
38181  __ret = (uint16x4_t)(__p0);
38182  return __ret;
38183}
38184#endif
38185
38186#ifdef __LITTLE_ENDIAN__
38187__ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
38188  uint16x4_t __ret;
38189  __ret = (uint16x4_t)(__p0);
38190  return __ret;
38191}
38192#else
38193__ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
38194  uint16x4_t __ret;
38195  __ret = (uint16x4_t)(__p0);
38196  return __ret;
38197}
38198#endif
38199
38200#ifdef __LITTLE_ENDIAN__
38201__ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
38202  uint16x4_t __ret;
38203  __ret = (uint16x4_t)(__p0);
38204  return __ret;
38205}
38206#else
38207__ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
38208  uint16x4_t __ret;
38209  __ret = (uint16x4_t)(__p0);
38210  return __ret;
38211}
38212#endif
38213
38214#ifdef __LITTLE_ENDIAN__
38215__ai uint16x4_t vreinterpret_u16_f64(float64x1_t __p0) {
38216  uint16x4_t __ret;
38217  __ret = (uint16x4_t)(__p0);
38218  return __ret;
38219}
38220#else
38221__ai uint16x4_t vreinterpret_u16_f64(float64x1_t __p0) {
38222  uint16x4_t __ret;
38223  __ret = (uint16x4_t)(__p0);
38224  return __ret;
38225}
38226#endif
38227
38228#ifdef __LITTLE_ENDIAN__
38229__ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
38230  uint16x4_t __ret;
38231  __ret = (uint16x4_t)(__p0);
38232  return __ret;
38233}
38234#else
38235__ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
38236  uint16x4_t __ret;
38237  __ret = (uint16x4_t)(__p0);
38238  return __ret;
38239}
38240#endif
38241
38242#ifdef __LITTLE_ENDIAN__
38243__ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
38244  uint16x4_t __ret;
38245  __ret = (uint16x4_t)(__p0);
38246  return __ret;
38247}
38248#else
38249__ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
38250  uint16x4_t __ret;
38251  __ret = (uint16x4_t)(__p0);
38252  return __ret;
38253}
38254#endif
38255
38256#ifdef __LITTLE_ENDIAN__
38257__ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
38258  uint16x4_t __ret;
38259  __ret = (uint16x4_t)(__p0);
38260  return __ret;
38261}
38262#else
38263__ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
38264  uint16x4_t __ret;
38265  __ret = (uint16x4_t)(__p0);
38266  return __ret;
38267}
38268#endif
38269
38270#ifdef __LITTLE_ENDIAN__
38271__ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
38272  uint16x4_t __ret;
38273  __ret = (uint16x4_t)(__p0);
38274  return __ret;
38275}
38276#else
38277__ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
38278  uint16x4_t __ret;
38279  __ret = (uint16x4_t)(__p0);
38280  return __ret;
38281}
38282#endif
38283
38284#ifdef __LITTLE_ENDIAN__
38285__ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
38286  uint16x4_t __ret;
38287  __ret = (uint16x4_t)(__p0);
38288  return __ret;
38289}
38290#else
38291__ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
38292  uint16x4_t __ret;
38293  __ret = (uint16x4_t)(__p0);
38294  return __ret;
38295}
38296#endif
38297
38298#ifdef __LITTLE_ENDIAN__
38299__ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
38300  int8x8_t __ret;
38301  __ret = (int8x8_t)(__p0);
38302  return __ret;
38303}
38304#else
38305__ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
38306  int8x8_t __ret;
38307  __ret = (int8x8_t)(__p0);
38308  return __ret;
38309}
38310#endif
38311
38312#ifdef __LITTLE_ENDIAN__
38313__ai int8x8_t vreinterpret_s8_p64(poly64x1_t __p0) {
38314  int8x8_t __ret;
38315  __ret = (int8x8_t)(__p0);
38316  return __ret;
38317}
38318#else
38319__ai int8x8_t vreinterpret_s8_p64(poly64x1_t __p0) {
38320  int8x8_t __ret;
38321  __ret = (int8x8_t)(__p0);
38322  return __ret;
38323}
38324#endif
38325
38326#ifdef __LITTLE_ENDIAN__
38327__ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
38328  int8x8_t __ret;
38329  __ret = (int8x8_t)(__p0);
38330  return __ret;
38331}
38332#else
38333__ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
38334  int8x8_t __ret;
38335  __ret = (int8x8_t)(__p0);
38336  return __ret;
38337}
38338#endif
38339
38340#ifdef __LITTLE_ENDIAN__
38341__ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
38342  int8x8_t __ret;
38343  __ret = (int8x8_t)(__p0);
38344  return __ret;
38345}
38346#else
38347__ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
38348  int8x8_t __ret;
38349  __ret = (int8x8_t)(__p0);
38350  return __ret;
38351}
38352#endif
38353
38354#ifdef __LITTLE_ENDIAN__
38355__ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
38356  int8x8_t __ret;
38357  __ret = (int8x8_t)(__p0);
38358  return __ret;
38359}
38360#else
38361__ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
38362  int8x8_t __ret;
38363  __ret = (int8x8_t)(__p0);
38364  return __ret;
38365}
38366#endif
38367
38368#ifdef __LITTLE_ENDIAN__
38369__ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
38370  int8x8_t __ret;
38371  __ret = (int8x8_t)(__p0);
38372  return __ret;
38373}
38374#else
38375__ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
38376  int8x8_t __ret;
38377  __ret = (int8x8_t)(__p0);
38378  return __ret;
38379}
38380#endif
38381
38382#ifdef __LITTLE_ENDIAN__
38383__ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
38384  int8x8_t __ret;
38385  __ret = (int8x8_t)(__p0);
38386  return __ret;
38387}
38388#else
38389__ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
38390  int8x8_t __ret;
38391  __ret = (int8x8_t)(__p0);
38392  return __ret;
38393}
38394#endif
38395
38396#ifdef __LITTLE_ENDIAN__
38397__ai int8x8_t vreinterpret_s8_f64(float64x1_t __p0) {
38398  int8x8_t __ret;
38399  __ret = (int8x8_t)(__p0);
38400  return __ret;
38401}
38402#else
38403__ai int8x8_t vreinterpret_s8_f64(float64x1_t __p0) {
38404  int8x8_t __ret;
38405  __ret = (int8x8_t)(__p0);
38406  return __ret;
38407}
38408#endif
38409
38410#ifdef __LITTLE_ENDIAN__
38411__ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
38412  int8x8_t __ret;
38413  __ret = (int8x8_t)(__p0);
38414  return __ret;
38415}
38416#else
38417__ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
38418  int8x8_t __ret;
38419  __ret = (int8x8_t)(__p0);
38420  return __ret;
38421}
38422#endif
38423
38424#ifdef __LITTLE_ENDIAN__
38425__ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
38426  int8x8_t __ret;
38427  __ret = (int8x8_t)(__p0);
38428  return __ret;
38429}
38430#else
38431__ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
38432  int8x8_t __ret;
38433  __ret = (int8x8_t)(__p0);
38434  return __ret;
38435}
38436#endif
38437
38438#ifdef __LITTLE_ENDIAN__
38439__ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
38440  int8x8_t __ret;
38441  __ret = (int8x8_t)(__p0);
38442  return __ret;
38443}
38444#else
38445__ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
38446  int8x8_t __ret;
38447  __ret = (int8x8_t)(__p0);
38448  return __ret;
38449}
38450#endif
38451
38452#ifdef __LITTLE_ENDIAN__
38453__ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
38454  int8x8_t __ret;
38455  __ret = (int8x8_t)(__p0);
38456  return __ret;
38457}
38458#else
38459__ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
38460  int8x8_t __ret;
38461  __ret = (int8x8_t)(__p0);
38462  return __ret;
38463}
38464#endif
38465
38466#ifdef __LITTLE_ENDIAN__
38467__ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
38468  int8x8_t __ret;
38469  __ret = (int8x8_t)(__p0);
38470  return __ret;
38471}
38472#else
38473__ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
38474  int8x8_t __ret;
38475  __ret = (int8x8_t)(__p0);
38476  return __ret;
38477}
38478#endif
38479
38480#ifdef __LITTLE_ENDIAN__
38481__ai float64x1_t vreinterpret_f64_p8(poly8x8_t __p0) {
38482  float64x1_t __ret;
38483  __ret = (float64x1_t)(__p0);
38484  return __ret;
38485}
38486#else
38487__ai float64x1_t vreinterpret_f64_p8(poly8x8_t __p0) {
38488  float64x1_t __ret;
38489  __ret = (float64x1_t)(__p0);
38490  return __ret;
38491}
38492#endif
38493
38494#ifdef __LITTLE_ENDIAN__
38495__ai float64x1_t vreinterpret_f64_p64(poly64x1_t __p0) {
38496  float64x1_t __ret;
38497  __ret = (float64x1_t)(__p0);
38498  return __ret;
38499}
38500#else
38501__ai float64x1_t vreinterpret_f64_p64(poly64x1_t __p0) {
38502  float64x1_t __ret;
38503  __ret = (float64x1_t)(__p0);
38504  return __ret;
38505}
38506#endif
38507
38508#ifdef __LITTLE_ENDIAN__
38509__ai float64x1_t vreinterpret_f64_p16(poly16x4_t __p0) {
38510  float64x1_t __ret;
38511  __ret = (float64x1_t)(__p0);
38512  return __ret;
38513}
38514#else
38515__ai float64x1_t vreinterpret_f64_p16(poly16x4_t __p0) {
38516  float64x1_t __ret;
38517  __ret = (float64x1_t)(__p0);
38518  return __ret;
38519}
38520#endif
38521
38522#ifdef __LITTLE_ENDIAN__
38523__ai float64x1_t vreinterpret_f64_u8(uint8x8_t __p0) {
38524  float64x1_t __ret;
38525  __ret = (float64x1_t)(__p0);
38526  return __ret;
38527}
38528#else
38529__ai float64x1_t vreinterpret_f64_u8(uint8x8_t __p0) {
38530  float64x1_t __ret;
38531  __ret = (float64x1_t)(__p0);
38532  return __ret;
38533}
38534#endif
38535
38536#ifdef __LITTLE_ENDIAN__
38537__ai float64x1_t vreinterpret_f64_u32(uint32x2_t __p0) {
38538  float64x1_t __ret;
38539  __ret = (float64x1_t)(__p0);
38540  return __ret;
38541}
38542#else
38543__ai float64x1_t vreinterpret_f64_u32(uint32x2_t __p0) {
38544  float64x1_t __ret;
38545  __ret = (float64x1_t)(__p0);
38546  return __ret;
38547}
38548#endif
38549
38550#ifdef __LITTLE_ENDIAN__
38551__ai float64x1_t vreinterpret_f64_u64(uint64x1_t __p0) {
38552  float64x1_t __ret;
38553  __ret = (float64x1_t)(__p0);
38554  return __ret;
38555}
38556#else
38557__ai float64x1_t vreinterpret_f64_u64(uint64x1_t __p0) {
38558  float64x1_t __ret;
38559  __ret = (float64x1_t)(__p0);
38560  return __ret;
38561}
38562#endif
38563
38564#ifdef __LITTLE_ENDIAN__
38565__ai float64x1_t vreinterpret_f64_u16(uint16x4_t __p0) {
38566  float64x1_t __ret;
38567  __ret = (float64x1_t)(__p0);
38568  return __ret;
38569}
38570#else
38571__ai float64x1_t vreinterpret_f64_u16(uint16x4_t __p0) {
38572  float64x1_t __ret;
38573  __ret = (float64x1_t)(__p0);
38574  return __ret;
38575}
38576#endif
38577
38578#ifdef __LITTLE_ENDIAN__
38579__ai float64x1_t vreinterpret_f64_s8(int8x8_t __p0) {
38580  float64x1_t __ret;
38581  __ret = (float64x1_t)(__p0);
38582  return __ret;
38583}
38584#else
38585__ai float64x1_t vreinterpret_f64_s8(int8x8_t __p0) {
38586  float64x1_t __ret;
38587  __ret = (float64x1_t)(__p0);
38588  return __ret;
38589}
38590#endif
38591
38592#ifdef __LITTLE_ENDIAN__
38593__ai float64x1_t vreinterpret_f64_f32(float32x2_t __p0) {
38594  float64x1_t __ret;
38595  __ret = (float64x1_t)(__p0);
38596  return __ret;
38597}
38598#else
38599__ai float64x1_t vreinterpret_f64_f32(float32x2_t __p0) {
38600  float64x1_t __ret;
38601  __ret = (float64x1_t)(__p0);
38602  return __ret;
38603}
38604#endif
38605
38606#ifdef __LITTLE_ENDIAN__
38607__ai float64x1_t vreinterpret_f64_f16(float16x4_t __p0) {
38608  float64x1_t __ret;
38609  __ret = (float64x1_t)(__p0);
38610  return __ret;
38611}
38612#else
38613__ai float64x1_t vreinterpret_f64_f16(float16x4_t __p0) {
38614  float64x1_t __ret;
38615  __ret = (float64x1_t)(__p0);
38616  return __ret;
38617}
38618#endif
38619
38620#ifdef __LITTLE_ENDIAN__
38621__ai float64x1_t vreinterpret_f64_s32(int32x2_t __p0) {
38622  float64x1_t __ret;
38623  __ret = (float64x1_t)(__p0);
38624  return __ret;
38625}
38626#else
38627__ai float64x1_t vreinterpret_f64_s32(int32x2_t __p0) {
38628  float64x1_t __ret;
38629  __ret = (float64x1_t)(__p0);
38630  return __ret;
38631}
38632#endif
38633
38634#ifdef __LITTLE_ENDIAN__
38635__ai float64x1_t vreinterpret_f64_s64(int64x1_t __p0) {
38636  float64x1_t __ret;
38637  __ret = (float64x1_t)(__p0);
38638  return __ret;
38639}
38640#else
38641__ai float64x1_t vreinterpret_f64_s64(int64x1_t __p0) {
38642  float64x1_t __ret;
38643  __ret = (float64x1_t)(__p0);
38644  return __ret;
38645}
38646#endif
38647
38648#ifdef __LITTLE_ENDIAN__
38649__ai float64x1_t vreinterpret_f64_s16(int16x4_t __p0) {
38650  float64x1_t __ret;
38651  __ret = (float64x1_t)(__p0);
38652  return __ret;
38653}
38654#else
38655__ai float64x1_t vreinterpret_f64_s16(int16x4_t __p0) {
38656  float64x1_t __ret;
38657  __ret = (float64x1_t)(__p0);
38658  return __ret;
38659}
38660#endif
38661
38662#ifdef __LITTLE_ENDIAN__
38663__ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
38664  float32x2_t __ret;
38665  __ret = (float32x2_t)(__p0);
38666  return __ret;
38667}
38668#else
38669__ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
38670  float32x2_t __ret;
38671  __ret = (float32x2_t)(__p0);
38672  return __ret;
38673}
38674#endif
38675
38676#ifdef __LITTLE_ENDIAN__
38677__ai float32x2_t vreinterpret_f32_p64(poly64x1_t __p0) {
38678  float32x2_t __ret;
38679  __ret = (float32x2_t)(__p0);
38680  return __ret;
38681}
38682#else
38683__ai float32x2_t vreinterpret_f32_p64(poly64x1_t __p0) {
38684  float32x2_t __ret;
38685  __ret = (float32x2_t)(__p0);
38686  return __ret;
38687}
38688#endif
38689
38690#ifdef __LITTLE_ENDIAN__
38691__ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
38692  float32x2_t __ret;
38693  __ret = (float32x2_t)(__p0);
38694  return __ret;
38695}
38696#else
38697__ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
38698  float32x2_t __ret;
38699  __ret = (float32x2_t)(__p0);
38700  return __ret;
38701}
38702#endif
38703
38704#ifdef __LITTLE_ENDIAN__
38705__ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
38706  float32x2_t __ret;
38707  __ret = (float32x2_t)(__p0);
38708  return __ret;
38709}
38710#else
38711__ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
38712  float32x2_t __ret;
38713  __ret = (float32x2_t)(__p0);
38714  return __ret;
38715}
38716#endif
38717
38718#ifdef __LITTLE_ENDIAN__
38719__ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
38720  float32x2_t __ret;
38721  __ret = (float32x2_t)(__p0);
38722  return __ret;
38723}
38724#else
38725__ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
38726  float32x2_t __ret;
38727  __ret = (float32x2_t)(__p0);
38728  return __ret;
38729}
38730#endif
38731
38732#ifdef __LITTLE_ENDIAN__
38733__ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
38734  float32x2_t __ret;
38735  __ret = (float32x2_t)(__p0);
38736  return __ret;
38737}
38738#else
38739__ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
38740  float32x2_t __ret;
38741  __ret = (float32x2_t)(__p0);
38742  return __ret;
38743}
38744#endif
38745
38746#ifdef __LITTLE_ENDIAN__
38747__ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
38748  float32x2_t __ret;
38749  __ret = (float32x2_t)(__p0);
38750  return __ret;
38751}
38752#else
38753__ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
38754  float32x2_t __ret;
38755  __ret = (float32x2_t)(__p0);
38756  return __ret;
38757}
38758#endif
38759
38760#ifdef __LITTLE_ENDIAN__
38761__ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
38762  float32x2_t __ret;
38763  __ret = (float32x2_t)(__p0);
38764  return __ret;
38765}
38766#else
38767__ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
38768  float32x2_t __ret;
38769  __ret = (float32x2_t)(__p0);
38770  return __ret;
38771}
38772#endif
38773
38774#ifdef __LITTLE_ENDIAN__
38775__ai float32x2_t vreinterpret_f32_f64(float64x1_t __p0) {
38776  float32x2_t __ret;
38777  __ret = (float32x2_t)(__p0);
38778  return __ret;
38779}
38780#else
38781__ai float32x2_t vreinterpret_f32_f64(float64x1_t __p0) {
38782  float32x2_t __ret;
38783  __ret = (float32x2_t)(__p0);
38784  return __ret;
38785}
38786#endif
38787
38788#ifdef __LITTLE_ENDIAN__
38789__ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
38790  float32x2_t __ret;
38791  __ret = (float32x2_t)(__p0);
38792  return __ret;
38793}
38794#else
38795__ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
38796  float32x2_t __ret;
38797  __ret = (float32x2_t)(__p0);
38798  return __ret;
38799}
38800#endif
38801
38802#ifdef __LITTLE_ENDIAN__
38803__ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
38804  float32x2_t __ret;
38805  __ret = (float32x2_t)(__p0);
38806  return __ret;
38807}
38808#else
38809__ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
38810  float32x2_t __ret;
38811  __ret = (float32x2_t)(__p0);
38812  return __ret;
38813}
38814#endif
38815
38816#ifdef __LITTLE_ENDIAN__
38817__ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
38818  float32x2_t __ret;
38819  __ret = (float32x2_t)(__p0);
38820  return __ret;
38821}
38822#else
38823__ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
38824  float32x2_t __ret;
38825  __ret = (float32x2_t)(__p0);
38826  return __ret;
38827}
38828#endif
38829
38830#ifdef __LITTLE_ENDIAN__
38831__ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
38832  float32x2_t __ret;
38833  __ret = (float32x2_t)(__p0);
38834  return __ret;
38835}
38836#else
38837__ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
38838  float32x2_t __ret;
38839  __ret = (float32x2_t)(__p0);
38840  return __ret;
38841}
38842#endif
38843
38844#ifdef __LITTLE_ENDIAN__
38845__ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
38846  float16x4_t __ret;
38847  __ret = (float16x4_t)(__p0);
38848  return __ret;
38849}
38850#else
38851__ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
38852  float16x4_t __ret;
38853  __ret = (float16x4_t)(__p0);
38854  return __ret;
38855}
38856#endif
38857
38858#ifdef __LITTLE_ENDIAN__
38859__ai float16x4_t vreinterpret_f16_p64(poly64x1_t __p0) {
38860  float16x4_t __ret;
38861  __ret = (float16x4_t)(__p0);
38862  return __ret;
38863}
38864#else
38865__ai float16x4_t vreinterpret_f16_p64(poly64x1_t __p0) {
38866  float16x4_t __ret;
38867  __ret = (float16x4_t)(__p0);
38868  return __ret;
38869}
38870#endif
38871
38872#ifdef __LITTLE_ENDIAN__
38873__ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
38874  float16x4_t __ret;
38875  __ret = (float16x4_t)(__p0);
38876  return __ret;
38877}
38878#else
38879__ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
38880  float16x4_t __ret;
38881  __ret = (float16x4_t)(__p0);
38882  return __ret;
38883}
38884#endif
38885
38886#ifdef __LITTLE_ENDIAN__
38887__ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
38888  float16x4_t __ret;
38889  __ret = (float16x4_t)(__p0);
38890  return __ret;
38891}
38892#else
38893__ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
38894  float16x4_t __ret;
38895  __ret = (float16x4_t)(__p0);
38896  return __ret;
38897}
38898#endif
38899
38900#ifdef __LITTLE_ENDIAN__
38901__ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
38902  float16x4_t __ret;
38903  __ret = (float16x4_t)(__p0);
38904  return __ret;
38905}
38906#else
38907__ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
38908  float16x4_t __ret;
38909  __ret = (float16x4_t)(__p0);
38910  return __ret;
38911}
38912#endif
38913
38914#ifdef __LITTLE_ENDIAN__
38915__ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
38916  float16x4_t __ret;
38917  __ret = (float16x4_t)(__p0);
38918  return __ret;
38919}
38920#else
38921__ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
38922  float16x4_t __ret;
38923  __ret = (float16x4_t)(__p0);
38924  return __ret;
38925}
38926#endif
38927
38928#ifdef __LITTLE_ENDIAN__
38929__ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
38930  float16x4_t __ret;
38931  __ret = (float16x4_t)(__p0);
38932  return __ret;
38933}
38934#else
38935__ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
38936  float16x4_t __ret;
38937  __ret = (float16x4_t)(__p0);
38938  return __ret;
38939}
38940#endif
38941
38942#ifdef __LITTLE_ENDIAN__
38943__ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
38944  float16x4_t __ret;
38945  __ret = (float16x4_t)(__p0);
38946  return __ret;
38947}
38948#else
38949__ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
38950  float16x4_t __ret;
38951  __ret = (float16x4_t)(__p0);
38952  return __ret;
38953}
38954#endif
38955
38956#ifdef __LITTLE_ENDIAN__
38957__ai float16x4_t vreinterpret_f16_f64(float64x1_t __p0) {
38958  float16x4_t __ret;
38959  __ret = (float16x4_t)(__p0);
38960  return __ret;
38961}
38962#else
38963__ai float16x4_t vreinterpret_f16_f64(float64x1_t __p0) {
38964  float16x4_t __ret;
38965  __ret = (float16x4_t)(__p0);
38966  return __ret;
38967}
38968#endif
38969
38970#ifdef __LITTLE_ENDIAN__
38971__ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
38972  float16x4_t __ret;
38973  __ret = (float16x4_t)(__p0);
38974  return __ret;
38975}
38976#else
38977__ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
38978  float16x4_t __ret;
38979  __ret = (float16x4_t)(__p0);
38980  return __ret;
38981}
38982#endif
38983
38984#ifdef __LITTLE_ENDIAN__
38985__ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
38986  float16x4_t __ret;
38987  __ret = (float16x4_t)(__p0);
38988  return __ret;
38989}
38990#else
38991__ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
38992  float16x4_t __ret;
38993  __ret = (float16x4_t)(__p0);
38994  return __ret;
38995}
38996#endif
38997
38998#ifdef __LITTLE_ENDIAN__
38999__ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
39000  float16x4_t __ret;
39001  __ret = (float16x4_t)(__p0);
39002  return __ret;
39003}
39004#else
39005__ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
39006  float16x4_t __ret;
39007  __ret = (float16x4_t)(__p0);
39008  return __ret;
39009}
39010#endif
39011
39012#ifdef __LITTLE_ENDIAN__
39013__ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
39014  float16x4_t __ret;
39015  __ret = (float16x4_t)(__p0);
39016  return __ret;
39017}
39018#else
39019__ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
39020  float16x4_t __ret;
39021  __ret = (float16x4_t)(__p0);
39022  return __ret;
39023}
39024#endif
39025
39026#ifdef __LITTLE_ENDIAN__
39027__ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
39028  int32x2_t __ret;
39029  __ret = (int32x2_t)(__p0);
39030  return __ret;
39031}
39032#else
39033__ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
39034  int32x2_t __ret;
39035  __ret = (int32x2_t)(__p0);
39036  return __ret;
39037}
39038#endif
39039
39040#ifdef __LITTLE_ENDIAN__
39041__ai int32x2_t vreinterpret_s32_p64(poly64x1_t __p0) {
39042  int32x2_t __ret;
39043  __ret = (int32x2_t)(__p0);
39044  return __ret;
39045}
39046#else
39047__ai int32x2_t vreinterpret_s32_p64(poly64x1_t __p0) {
39048  int32x2_t __ret;
39049  __ret = (int32x2_t)(__p0);
39050  return __ret;
39051}
39052#endif
39053
39054#ifdef __LITTLE_ENDIAN__
39055__ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
39056  int32x2_t __ret;
39057  __ret = (int32x2_t)(__p0);
39058  return __ret;
39059}
39060#else
39061__ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
39062  int32x2_t __ret;
39063  __ret = (int32x2_t)(__p0);
39064  return __ret;
39065}
39066#endif
39067
39068#ifdef __LITTLE_ENDIAN__
39069__ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
39070  int32x2_t __ret;
39071  __ret = (int32x2_t)(__p0);
39072  return __ret;
39073}
39074#else
39075__ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
39076  int32x2_t __ret;
39077  __ret = (int32x2_t)(__p0);
39078  return __ret;
39079}
39080#endif
39081
39082#ifdef __LITTLE_ENDIAN__
39083__ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
39084  int32x2_t __ret;
39085  __ret = (int32x2_t)(__p0);
39086  return __ret;
39087}
39088#else
39089__ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
39090  int32x2_t __ret;
39091  __ret = (int32x2_t)(__p0);
39092  return __ret;
39093}
39094#endif
39095
39096#ifdef __LITTLE_ENDIAN__
39097__ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
39098  int32x2_t __ret;
39099  __ret = (int32x2_t)(__p0);
39100  return __ret;
39101}
39102#else
39103__ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
39104  int32x2_t __ret;
39105  __ret = (int32x2_t)(__p0);
39106  return __ret;
39107}
39108#endif
39109
39110#ifdef __LITTLE_ENDIAN__
39111__ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
39112  int32x2_t __ret;
39113  __ret = (int32x2_t)(__p0);
39114  return __ret;
39115}
39116#else
39117__ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
39118  int32x2_t __ret;
39119  __ret = (int32x2_t)(__p0);
39120  return __ret;
39121}
39122#endif
39123
39124#ifdef __LITTLE_ENDIAN__
39125__ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
39126  int32x2_t __ret;
39127  __ret = (int32x2_t)(__p0);
39128  return __ret;
39129}
39130#else
39131__ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
39132  int32x2_t __ret;
39133  __ret = (int32x2_t)(__p0);
39134  return __ret;
39135}
39136#endif
39137
39138#ifdef __LITTLE_ENDIAN__
39139__ai int32x2_t vreinterpret_s32_f64(float64x1_t __p0) {
39140  int32x2_t __ret;
39141  __ret = (int32x2_t)(__p0);
39142  return __ret;
39143}
39144#else
39145__ai int32x2_t vreinterpret_s32_f64(float64x1_t __p0) {
39146  int32x2_t __ret;
39147  __ret = (int32x2_t)(__p0);
39148  return __ret;
39149}
39150#endif
39151
39152#ifdef __LITTLE_ENDIAN__
39153__ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
39154  int32x2_t __ret;
39155  __ret = (int32x2_t)(__p0);
39156  return __ret;
39157}
39158#else
39159__ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
39160  int32x2_t __ret;
39161  __ret = (int32x2_t)(__p0);
39162  return __ret;
39163}
39164#endif
39165
39166#ifdef __LITTLE_ENDIAN__
39167__ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
39168  int32x2_t __ret;
39169  __ret = (int32x2_t)(__p0);
39170  return __ret;
39171}
39172#else
39173__ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
39174  int32x2_t __ret;
39175  __ret = (int32x2_t)(__p0);
39176  return __ret;
39177}
39178#endif
39179
39180#ifdef __LITTLE_ENDIAN__
39181__ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
39182  int32x2_t __ret;
39183  __ret = (int32x2_t)(__p0);
39184  return __ret;
39185}
39186#else
39187__ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
39188  int32x2_t __ret;
39189  __ret = (int32x2_t)(__p0);
39190  return __ret;
39191}
39192#endif
39193
39194#ifdef __LITTLE_ENDIAN__
39195__ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
39196  int32x2_t __ret;
39197  __ret = (int32x2_t)(__p0);
39198  return __ret;
39199}
39200#else
39201__ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
39202  int32x2_t __ret;
39203  __ret = (int32x2_t)(__p0);
39204  return __ret;
39205}
39206#endif
39207
39208#ifdef __LITTLE_ENDIAN__
39209__ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
39210  int64x1_t __ret;
39211  __ret = (int64x1_t)(__p0);
39212  return __ret;
39213}
39214#else
39215__ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
39216  int64x1_t __ret;
39217  __ret = (int64x1_t)(__p0);
39218  return __ret;
39219}
39220#endif
39221
39222#ifdef __LITTLE_ENDIAN__
39223__ai int64x1_t vreinterpret_s64_p64(poly64x1_t __p0) {
39224  int64x1_t __ret;
39225  __ret = (int64x1_t)(__p0);
39226  return __ret;
39227}
39228#else
39229__ai int64x1_t vreinterpret_s64_p64(poly64x1_t __p0) {
39230  int64x1_t __ret;
39231  __ret = (int64x1_t)(__p0);
39232  return __ret;
39233}
39234#endif
39235
39236#ifdef __LITTLE_ENDIAN__
39237__ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
39238  int64x1_t __ret;
39239  __ret = (int64x1_t)(__p0);
39240  return __ret;
39241}
39242#else
39243__ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
39244  int64x1_t __ret;
39245  __ret = (int64x1_t)(__p0);
39246  return __ret;
39247}
39248#endif
39249
39250#ifdef __LITTLE_ENDIAN__
39251__ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
39252  int64x1_t __ret;
39253  __ret = (int64x1_t)(__p0);
39254  return __ret;
39255}
39256#else
39257__ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
39258  int64x1_t __ret;
39259  __ret = (int64x1_t)(__p0);
39260  return __ret;
39261}
39262#endif
39263
39264#ifdef __LITTLE_ENDIAN__
39265__ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
39266  int64x1_t __ret;
39267  __ret = (int64x1_t)(__p0);
39268  return __ret;
39269}
39270#else
39271__ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
39272  int64x1_t __ret;
39273  __ret = (int64x1_t)(__p0);
39274  return __ret;
39275}
39276#endif
39277
39278#ifdef __LITTLE_ENDIAN__
39279__ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
39280  int64x1_t __ret;
39281  __ret = (int64x1_t)(__p0);
39282  return __ret;
39283}
39284#else
39285__ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
39286  int64x1_t __ret;
39287  __ret = (int64x1_t)(__p0);
39288  return __ret;
39289}
39290#endif
39291
39292#ifdef __LITTLE_ENDIAN__
39293__ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
39294  int64x1_t __ret;
39295  __ret = (int64x1_t)(__p0);
39296  return __ret;
39297}
39298#else
39299__ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
39300  int64x1_t __ret;
39301  __ret = (int64x1_t)(__p0);
39302  return __ret;
39303}
39304#endif
39305
39306#ifdef __LITTLE_ENDIAN__
39307__ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
39308  int64x1_t __ret;
39309  __ret = (int64x1_t)(__p0);
39310  return __ret;
39311}
39312#else
39313__ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
39314  int64x1_t __ret;
39315  __ret = (int64x1_t)(__p0);
39316  return __ret;
39317}
39318#endif
39319
39320#ifdef __LITTLE_ENDIAN__
39321__ai int64x1_t vreinterpret_s64_f64(float64x1_t __p0) {
39322  int64x1_t __ret;
39323  __ret = (int64x1_t)(__p0);
39324  return __ret;
39325}
39326#else
39327__ai int64x1_t vreinterpret_s64_f64(float64x1_t __p0) {
39328  int64x1_t __ret;
39329  __ret = (int64x1_t)(__p0);
39330  return __ret;
39331}
39332#endif
39333
39334#ifdef __LITTLE_ENDIAN__
39335__ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
39336  int64x1_t __ret;
39337  __ret = (int64x1_t)(__p0);
39338  return __ret;
39339}
39340#else
39341__ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
39342  int64x1_t __ret;
39343  __ret = (int64x1_t)(__p0);
39344  return __ret;
39345}
39346#endif
39347
39348#ifdef __LITTLE_ENDIAN__
39349__ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
39350  int64x1_t __ret;
39351  __ret = (int64x1_t)(__p0);
39352  return __ret;
39353}
39354#else
39355__ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
39356  int64x1_t __ret;
39357  __ret = (int64x1_t)(__p0);
39358  return __ret;
39359}
39360#endif
39361
39362#ifdef __LITTLE_ENDIAN__
39363__ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
39364  int64x1_t __ret;
39365  __ret = (int64x1_t)(__p0);
39366  return __ret;
39367}
39368#else
39369__ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
39370  int64x1_t __ret;
39371  __ret = (int64x1_t)(__p0);
39372  return __ret;
39373}
39374#endif
39375
39376#ifdef __LITTLE_ENDIAN__
39377__ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
39378  int64x1_t __ret;
39379  __ret = (int64x1_t)(__p0);
39380  return __ret;
39381}
39382#else
39383__ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
39384  int64x1_t __ret;
39385  __ret = (int64x1_t)(__p0);
39386  return __ret;
39387}
39388#endif
39389
39390#ifdef __LITTLE_ENDIAN__
39391__ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
39392  int16x4_t __ret;
39393  __ret = (int16x4_t)(__p0);
39394  return __ret;
39395}
39396#else
39397__ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
39398  int16x4_t __ret;
39399  __ret = (int16x4_t)(__p0);
39400  return __ret;
39401}
39402#endif
39403
39404#ifdef __LITTLE_ENDIAN__
39405__ai int16x4_t vreinterpret_s16_p64(poly64x1_t __p0) {
39406  int16x4_t __ret;
39407  __ret = (int16x4_t)(__p0);
39408  return __ret;
39409}
39410#else
39411__ai int16x4_t vreinterpret_s16_p64(poly64x1_t __p0) {
39412  int16x4_t __ret;
39413  __ret = (int16x4_t)(__p0);
39414  return __ret;
39415}
39416#endif
39417
39418#ifdef __LITTLE_ENDIAN__
39419__ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
39420  int16x4_t __ret;
39421  __ret = (int16x4_t)(__p0);
39422  return __ret;
39423}
39424#else
39425__ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
39426  int16x4_t __ret;
39427  __ret = (int16x4_t)(__p0);
39428  return __ret;
39429}
39430#endif
39431
39432#ifdef __LITTLE_ENDIAN__
39433__ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
39434  int16x4_t __ret;
39435  __ret = (int16x4_t)(__p0);
39436  return __ret;
39437}
39438#else
39439__ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
39440  int16x4_t __ret;
39441  __ret = (int16x4_t)(__p0);
39442  return __ret;
39443}
39444#endif
39445
39446#ifdef __LITTLE_ENDIAN__
39447__ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
39448  int16x4_t __ret;
39449  __ret = (int16x4_t)(__p0);
39450  return __ret;
39451}
39452#else
39453__ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
39454  int16x4_t __ret;
39455  __ret = (int16x4_t)(__p0);
39456  return __ret;
39457}
39458#endif
39459
39460#ifdef __LITTLE_ENDIAN__
39461__ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
39462  int16x4_t __ret;
39463  __ret = (int16x4_t)(__p0);
39464  return __ret;
39465}
39466#else
39467__ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
39468  int16x4_t __ret;
39469  __ret = (int16x4_t)(__p0);
39470  return __ret;
39471}
39472#endif
39473
39474#ifdef __LITTLE_ENDIAN__
39475__ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
39476  int16x4_t __ret;
39477  __ret = (int16x4_t)(__p0);
39478  return __ret;
39479}
39480#else
39481__ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
39482  int16x4_t __ret;
39483  __ret = (int16x4_t)(__p0);
39484  return __ret;
39485}
39486#endif
39487
39488#ifdef __LITTLE_ENDIAN__
39489__ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
39490  int16x4_t __ret;
39491  __ret = (int16x4_t)(__p0);
39492  return __ret;
39493}
39494#else
39495__ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
39496  int16x4_t __ret;
39497  __ret = (int16x4_t)(__p0);
39498  return __ret;
39499}
39500#endif
39501
39502#ifdef __LITTLE_ENDIAN__
39503__ai int16x4_t vreinterpret_s16_f64(float64x1_t __p0) {
39504  int16x4_t __ret;
39505  __ret = (int16x4_t)(__p0);
39506  return __ret;
39507}
39508#else
39509__ai int16x4_t vreinterpret_s16_f64(float64x1_t __p0) {
39510  int16x4_t __ret;
39511  __ret = (int16x4_t)(__p0);
39512  return __ret;
39513}
39514#endif
39515
39516#ifdef __LITTLE_ENDIAN__
39517__ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
39518  int16x4_t __ret;
39519  __ret = (int16x4_t)(__p0);
39520  return __ret;
39521}
39522#else
39523__ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
39524  int16x4_t __ret;
39525  __ret = (int16x4_t)(__p0);
39526  return __ret;
39527}
39528#endif
39529
39530#ifdef __LITTLE_ENDIAN__
39531__ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
39532  int16x4_t __ret;
39533  __ret = (int16x4_t)(__p0);
39534  return __ret;
39535}
39536#else
39537__ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
39538  int16x4_t __ret;
39539  __ret = (int16x4_t)(__p0);
39540  return __ret;
39541}
39542#endif
39543
39544#ifdef __LITTLE_ENDIAN__
39545__ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
39546  int16x4_t __ret;
39547  __ret = (int16x4_t)(__p0);
39548  return __ret;
39549}
39550#else
39551__ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
39552  int16x4_t __ret;
39553  __ret = (int16x4_t)(__p0);
39554  return __ret;
39555}
39556#endif
39557
39558#ifdef __LITTLE_ENDIAN__
39559__ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
39560  int16x4_t __ret;
39561  __ret = (int16x4_t)(__p0);
39562  return __ret;
39563}
39564#else
39565__ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
39566  int16x4_t __ret;
39567  __ret = (int16x4_t)(__p0);
39568  return __ret;
39569}
39570#endif
39571
39572#endif
39573#if __ARM_FEATURE_CRYPTO
39574#ifdef __LITTLE_ENDIAN__
39575__ai uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
39576  uint8x16_t __ret;
39577  __ret = (uint8x16_t) __builtin_neon_vaesdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
39578  return __ret;
39579}
39580#else
39581__ai uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
39582  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
39583  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
39584  uint8x16_t __ret;
39585  __ret = (uint8x16_t) __builtin_neon_vaesdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
39586  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
39587  return __ret;
39588}
39589#endif
39590
39591#ifdef __LITTLE_ENDIAN__
39592__ai uint8x16_t vaeseq_u8(uint8x16_t __p0, uint8x16_t __p1) {
39593  uint8x16_t __ret;
39594  __ret = (uint8x16_t) __builtin_neon_vaeseq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
39595  return __ret;
39596}
39597#else
39598__ai uint8x16_t vaeseq_u8(uint8x16_t __p0, uint8x16_t __p1) {
39599  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
39600  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
39601  uint8x16_t __ret;
39602  __ret = (uint8x16_t) __builtin_neon_vaeseq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
39603  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
39604  return __ret;
39605}
39606#endif
39607
39608#ifdef __LITTLE_ENDIAN__
39609__ai uint8x16_t vaesimcq_u8(uint8x16_t __p0) {
39610  uint8x16_t __ret;
39611  __ret = (uint8x16_t) __builtin_neon_vaesimcq_v((int8x16_t)__p0, 48);
39612  return __ret;
39613}
39614#else
39615__ai uint8x16_t vaesimcq_u8(uint8x16_t __p0) {
39616  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
39617  uint8x16_t __ret;
39618  __ret = (uint8x16_t) __builtin_neon_vaesimcq_v((int8x16_t)__rev0, 48);
39619  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
39620  return __ret;
39621}
39622#endif
39623
39624#ifdef __LITTLE_ENDIAN__
39625__ai uint8x16_t vaesmcq_u8(uint8x16_t __p0) {
39626  uint8x16_t __ret;
39627  __ret = (uint8x16_t) __builtin_neon_vaesmcq_v((int8x16_t)__p0, 48);
39628  return __ret;
39629}
39630#else
39631__ai uint8x16_t vaesmcq_u8(uint8x16_t __p0) {
39632  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
39633  uint8x16_t __ret;
39634  __ret = (uint8x16_t) __builtin_neon_vaesmcq_v((int8x16_t)__rev0, 48);
39635  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
39636  return __ret;
39637}
39638#endif
39639
39640#ifdef __LITTLE_ENDIAN__
39641__ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
39642  uint32x4_t __ret;
39643  __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
39644  return __ret;
39645}
39646#else
39647__ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
39648  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39649  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
39650  uint32x4_t __ret;
39651  __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
39652  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39653  return __ret;
39654}
39655#endif
39656
39657#ifdef __LITTLE_ENDIAN__
39658__ai uint32_t vsha1h_u32(uint32_t __p0) {
39659  uint32_t __ret;
39660  __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0);
39661  return __ret;
39662}
39663#else
39664__ai uint32_t vsha1h_u32(uint32_t __p0) {
39665  uint32_t __ret;
39666  __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0);
39667  return __ret;
39668}
39669#endif
39670
39671#ifdef __LITTLE_ENDIAN__
39672__ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
39673  uint32x4_t __ret;
39674  __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
39675  return __ret;
39676}
39677#else
39678__ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
39679  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39680  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
39681  uint32x4_t __ret;
39682  __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
39683  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39684  return __ret;
39685}
39686#endif
39687
39688#ifdef __LITTLE_ENDIAN__
39689__ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
39690  uint32x4_t __ret;
39691  __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
39692  return __ret;
39693}
39694#else
39695__ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
39696  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39697  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
39698  uint32x4_t __ret;
39699  __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
39700  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39701  return __ret;
39702}
39703#endif
39704
39705#ifdef __LITTLE_ENDIAN__
39706__ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
39707  uint32x4_t __ret;
39708  __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
39709  return __ret;
39710}
39711#else
39712__ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
39713  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39714  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
39715  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
39716  uint32x4_t __ret;
39717  __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
39718  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39719  return __ret;
39720}
39721#endif
39722
39723#ifdef __LITTLE_ENDIAN__
39724__ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
39725  uint32x4_t __ret;
39726  __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
39727  return __ret;
39728}
39729#else
39730__ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
39731  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39732  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
39733  uint32x4_t __ret;
39734  __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
39735  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39736  return __ret;
39737}
39738#endif
39739
39740#ifdef __LITTLE_ENDIAN__
39741__ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
39742  uint32x4_t __ret;
39743  __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
39744  return __ret;
39745}
39746#else
39747__ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
39748  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39749  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
39750  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
39751  uint32x4_t __ret;
39752  __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
39753  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39754  return __ret;
39755}
39756#endif
39757
39758#ifdef __LITTLE_ENDIAN__
39759__ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
39760  uint32x4_t __ret;
39761  __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
39762  return __ret;
39763}
39764#else
39765__ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
39766  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39767  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
39768  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
39769  uint32x4_t __ret;
39770  __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
39771  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39772  return __ret;
39773}
39774#endif
39775
39776#ifdef __LITTLE_ENDIAN__
39777__ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) {
39778  uint32x4_t __ret;
39779  __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
39780  return __ret;
39781}
39782#else
39783__ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) {
39784  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39785  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
39786  uint32x4_t __ret;
39787  __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
39788  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39789  return __ret;
39790}
39791#endif
39792
39793#ifdef __LITTLE_ENDIAN__
39794__ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
39795  uint32x4_t __ret;
39796  __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
39797  return __ret;
39798}
39799#else
39800__ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
39801  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39802  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
39803  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
39804  uint32x4_t __ret;
39805  __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
39806  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39807  return __ret;
39808}
39809#endif
39810
39811#endif
39812#if defined(__aarch64__)
39813#ifdef __LITTLE_ENDIAN__
39814__ai float64x2_t vabdq_f64(float64x2_t __p0, float64x2_t __p1) {
39815  float64x2_t __ret;
39816  __ret = (float64x2_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
39817  return __ret;
39818}
39819#else
39820__ai float64x2_t vabdq_f64(float64x2_t __p0, float64x2_t __p1) {
39821  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39822  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
39823  float64x2_t __ret;
39824  __ret = (float64x2_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
39825  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39826  return __ret;
39827}
39828#endif
39829
39830#ifdef __LITTLE_ENDIAN__
39831__ai float64x1_t vabd_f64(float64x1_t __p0, float64x1_t __p1) {
39832  float64x1_t __ret;
39833  __ret = (float64x1_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
39834  return __ret;
39835}
39836#else
39837__ai float64x1_t vabd_f64(float64x1_t __p0, float64x1_t __p1) {
39838  float64x1_t __ret;
39839  __ret = (float64x1_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
39840  return __ret;
39841}
39842#endif
39843
39844#ifdef __LITTLE_ENDIAN__
39845__ai float64_t vabdd_f64(float64_t __p0, float64_t __p1) {
39846  float64_t __ret;
39847  __ret = (float64_t) __builtin_neon_vabdd_f64(__p0, __p1);
39848  return __ret;
39849}
39850#else
39851__ai float64_t vabdd_f64(float64_t __p0, float64_t __p1) {
39852  float64_t __ret;
39853  __ret = (float64_t) __builtin_neon_vabdd_f64(__p0, __p1);
39854  return __ret;
39855}
39856#endif
39857
39858#ifdef __LITTLE_ENDIAN__
39859__ai float32_t vabds_f32(float32_t __p0, float32_t __p1) {
39860  float32_t __ret;
39861  __ret = (float32_t) __builtin_neon_vabds_f32(__p0, __p1);
39862  return __ret;
39863}
39864#else
39865__ai float32_t vabds_f32(float32_t __p0, float32_t __p1) {
39866  float32_t __ret;
39867  __ret = (float32_t) __builtin_neon_vabds_f32(__p0, __p1);
39868  return __ret;
39869}
39870#endif
39871
39872#ifdef __LITTLE_ENDIAN__
39873__ai float64x2_t vabsq_f64(float64x2_t __p0) {
39874  float64x2_t __ret;
39875  __ret = (float64x2_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 42);
39876  return __ret;
39877}
39878#else
39879__ai float64x2_t vabsq_f64(float64x2_t __p0) {
39880  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39881  float64x2_t __ret;
39882  __ret = (float64x2_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 42);
39883  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39884  return __ret;
39885}
39886#endif
39887
39888#ifdef __LITTLE_ENDIAN__
39889__ai int64x2_t vabsq_s64(int64x2_t __p0) {
39890  int64x2_t __ret;
39891  __ret = (int64x2_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 35);
39892  return __ret;
39893}
39894#else
39895__ai int64x2_t vabsq_s64(int64x2_t __p0) {
39896  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39897  int64x2_t __ret;
39898  __ret = (int64x2_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 35);
39899  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39900  return __ret;
39901}
39902#endif
39903
39904#ifdef __LITTLE_ENDIAN__
39905__ai float64x1_t vabs_f64(float64x1_t __p0) {
39906  float64x1_t __ret;
39907  __ret = (float64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 10);
39908  return __ret;
39909}
39910#else
39911__ai float64x1_t vabs_f64(float64x1_t __p0) {
39912  float64x1_t __ret;
39913  __ret = (float64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 10);
39914  return __ret;
39915}
39916#endif
39917
39918#ifdef __LITTLE_ENDIAN__
39919__ai int64x1_t vabs_s64(int64x1_t __p0) {
39920  int64x1_t __ret;
39921  __ret = (int64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 3);
39922  return __ret;
39923}
39924#else
39925__ai int64x1_t vabs_s64(int64x1_t __p0) {
39926  int64x1_t __ret;
39927  __ret = (int64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 3);
39928  return __ret;
39929}
39930#endif
39931
39932#ifdef __LITTLE_ENDIAN__
39933__ai int64_t vabsd_s64(int64_t __p0) {
39934  int64_t __ret;
39935  __ret = (int64_t) __builtin_neon_vabsd_s64(__p0);
39936  return __ret;
39937}
39938#else
39939__ai int64_t vabsd_s64(int64_t __p0) {
39940  int64_t __ret;
39941  __ret = (int64_t) __builtin_neon_vabsd_s64(__p0);
39942  return __ret;
39943}
39944#endif
39945
39946#ifdef __LITTLE_ENDIAN__
39947__ai float64x2_t vaddq_f64(float64x2_t __p0, float64x2_t __p1) {
39948  float64x2_t __ret;
39949  __ret = __p0 + __p1;
39950  return __ret;
39951}
39952#else
39953__ai float64x2_t vaddq_f64(float64x2_t __p0, float64x2_t __p1) {
39954  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39955  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
39956  float64x2_t __ret;
39957  __ret = __rev0 + __rev1;
39958  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39959  return __ret;
39960}
39961#endif
39962
39963#ifdef __LITTLE_ENDIAN__
39964__ai float64x1_t vadd_f64(float64x1_t __p0, float64x1_t __p1) {
39965  float64x1_t __ret;
39966  __ret = __p0 + __p1;
39967  return __ret;
39968}
39969#else
39970__ai float64x1_t vadd_f64(float64x1_t __p0, float64x1_t __p1) {
39971  float64x1_t __ret;
39972  __ret = __p0 + __p1;
39973  return __ret;
39974}
39975#endif
39976
39977#ifdef __LITTLE_ENDIAN__
39978__ai uint64_t vaddd_u64(uint64_t __p0, uint64_t __p1) {
39979  uint64_t __ret;
39980  __ret = (uint64_t) __builtin_neon_vaddd_u64(__p0, __p1);
39981  return __ret;
39982}
39983#else
39984__ai uint64_t vaddd_u64(uint64_t __p0, uint64_t __p1) {
39985  uint64_t __ret;
39986  __ret = (uint64_t) __builtin_neon_vaddd_u64(__p0, __p1);
39987  return __ret;
39988}
39989#endif
39990
39991#ifdef __LITTLE_ENDIAN__
39992__ai int64_t vaddd_s64(int64_t __p0, int64_t __p1) {
39993  int64_t __ret;
39994  __ret = (int64_t) __builtin_neon_vaddd_s64(__p0, __p1);
39995  return __ret;
39996}
39997#else
39998__ai int64_t vaddd_s64(int64_t __p0, int64_t __p1) {
39999  int64_t __ret;
40000  __ret = (int64_t) __builtin_neon_vaddd_s64(__p0, __p1);
40001  return __ret;
40002}
40003#endif
40004
40005#ifdef __LITTLE_ENDIAN__
40006__ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40007  uint16x8_t __ret;
40008  __ret = vcombine_u16(__p0, vaddhn_u32(__p1, __p2));
40009  return __ret;
40010}
40011#else
40012__ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40013  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40014  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40015  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40016  uint16x8_t __ret;
40017  __ret = __noswap_vcombine_u16(__rev0, __noswap_vaddhn_u32(__rev1, __rev2));
40018  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40019  return __ret;
40020}
40021#endif
40022
40023#ifdef __LITTLE_ENDIAN__
40024__ai uint32x4_t vaddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
40025  uint32x4_t __ret;
40026  __ret = vcombine_u32(__p0, vaddhn_u64(__p1, __p2));
40027  return __ret;
40028}
40029#else
40030__ai uint32x4_t vaddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
40031  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40032  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40033  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
40034  uint32x4_t __ret;
40035  __ret = __noswap_vcombine_u32(__rev0, __noswap_vaddhn_u64(__rev1, __rev2));
40036  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40037  return __ret;
40038}
40039#endif
40040
40041#ifdef __LITTLE_ENDIAN__
40042__ai uint8x16_t vaddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
40043  uint8x16_t __ret;
40044  __ret = vcombine_u8(__p0, vaddhn_u16(__p1, __p2));
40045  return __ret;
40046}
40047#else
40048__ai uint8x16_t vaddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
40049  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40050  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
40051  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
40052  uint8x16_t __ret;
40053  __ret = __noswap_vcombine_u8(__rev0, __noswap_vaddhn_u16(__rev1, __rev2));
40054  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40055  return __ret;
40056}
40057#endif
40058
40059#ifdef __LITTLE_ENDIAN__
40060__ai int16x8_t vaddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
40061  int16x8_t __ret;
40062  __ret = vcombine_s16(__p0, vaddhn_s32(__p1, __p2));
40063  return __ret;
40064}
40065#else
40066__ai int16x8_t vaddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
40067  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40068  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40069  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40070  int16x8_t __ret;
40071  __ret = __noswap_vcombine_s16(__rev0, __noswap_vaddhn_s32(__rev1, __rev2));
40072  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40073  return __ret;
40074}
40075#endif
40076
40077#ifdef __LITTLE_ENDIAN__
40078__ai int32x4_t vaddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
40079  int32x4_t __ret;
40080  __ret = vcombine_s32(__p0, vaddhn_s64(__p1, __p2));
40081  return __ret;
40082}
40083#else
40084__ai int32x4_t vaddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
40085  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40086  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40087  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
40088  int32x4_t __ret;
40089  __ret = __noswap_vcombine_s32(__rev0, __noswap_vaddhn_s64(__rev1, __rev2));
40090  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40091  return __ret;
40092}
40093#endif
40094
40095#ifdef __LITTLE_ENDIAN__
40096__ai int8x16_t vaddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
40097  int8x16_t __ret;
40098  __ret = vcombine_s8(__p0, vaddhn_s16(__p1, __p2));
40099  return __ret;
40100}
40101#else
40102__ai int8x16_t vaddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
40103  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40104  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
40105  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
40106  int8x16_t __ret;
40107  __ret = __noswap_vcombine_s8(__rev0, __noswap_vaddhn_s16(__rev1, __rev2));
40108  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40109  return __ret;
40110}
40111#endif
40112
40113#ifdef __LITTLE_ENDIAN__
40114__ai uint16_t vaddlvq_u8(uint8x16_t __p0) {
40115  uint16_t __ret;
40116  __ret = (uint16_t) __builtin_neon_vaddlvq_u8((int8x16_t)__p0);
40117  return __ret;
40118}
40119#else
40120__ai uint16_t vaddlvq_u8(uint8x16_t __p0) {
40121  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40122  uint16_t __ret;
40123  __ret = (uint16_t) __builtin_neon_vaddlvq_u8((int8x16_t)__rev0);
40124  return __ret;
40125}
40126#endif
40127
40128#ifdef __LITTLE_ENDIAN__
40129__ai uint64_t vaddlvq_u32(uint32x4_t __p0) {
40130  uint64_t __ret;
40131  __ret = (uint64_t) __builtin_neon_vaddlvq_u32((int8x16_t)__p0);
40132  return __ret;
40133}
40134#else
40135__ai uint64_t vaddlvq_u32(uint32x4_t __p0) {
40136  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40137  uint64_t __ret;
40138  __ret = (uint64_t) __builtin_neon_vaddlvq_u32((int8x16_t)__rev0);
40139  return __ret;
40140}
40141#endif
40142
40143#ifdef __LITTLE_ENDIAN__
40144__ai uint32_t vaddlvq_u16(uint16x8_t __p0) {
40145  uint32_t __ret;
40146  __ret = (uint32_t) __builtin_neon_vaddlvq_u16((int8x16_t)__p0);
40147  return __ret;
40148}
40149#else
40150__ai uint32_t vaddlvq_u16(uint16x8_t __p0) {
40151  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40152  uint32_t __ret;
40153  __ret = (uint32_t) __builtin_neon_vaddlvq_u16((int8x16_t)__rev0);
40154  return __ret;
40155}
40156#endif
40157
40158#ifdef __LITTLE_ENDIAN__
40159__ai int16_t vaddlvq_s8(int8x16_t __p0) {
40160  int16_t __ret;
40161  __ret = (int16_t) __builtin_neon_vaddlvq_s8((int8x16_t)__p0);
40162  return __ret;
40163}
40164#else
40165__ai int16_t vaddlvq_s8(int8x16_t __p0) {
40166  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40167  int16_t __ret;
40168  __ret = (int16_t) __builtin_neon_vaddlvq_s8((int8x16_t)__rev0);
40169  return __ret;
40170}
40171#endif
40172
40173#ifdef __LITTLE_ENDIAN__
40174__ai int64_t vaddlvq_s32(int32x4_t __p0) {
40175  int64_t __ret;
40176  __ret = (int64_t) __builtin_neon_vaddlvq_s32((int8x16_t)__p0);
40177  return __ret;
40178}
40179#else
40180__ai int64_t vaddlvq_s32(int32x4_t __p0) {
40181  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40182  int64_t __ret;
40183  __ret = (int64_t) __builtin_neon_vaddlvq_s32((int8x16_t)__rev0);
40184  return __ret;
40185}
40186#endif
40187
40188#ifdef __LITTLE_ENDIAN__
40189__ai int32_t vaddlvq_s16(int16x8_t __p0) {
40190  int32_t __ret;
40191  __ret = (int32_t) __builtin_neon_vaddlvq_s16((int8x16_t)__p0);
40192  return __ret;
40193}
40194#else
40195__ai int32_t vaddlvq_s16(int16x8_t __p0) {
40196  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40197  int32_t __ret;
40198  __ret = (int32_t) __builtin_neon_vaddlvq_s16((int8x16_t)__rev0);
40199  return __ret;
40200}
40201#endif
40202
40203#ifdef __LITTLE_ENDIAN__
40204__ai uint16_t vaddlv_u8(uint8x8_t __p0) {
40205  uint16_t __ret;
40206  __ret = (uint16_t) __builtin_neon_vaddlv_u8((int8x8_t)__p0);
40207  return __ret;
40208}
40209#else
40210__ai uint16_t vaddlv_u8(uint8x8_t __p0) {
40211  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40212  uint16_t __ret;
40213  __ret = (uint16_t) __builtin_neon_vaddlv_u8((int8x8_t)__rev0);
40214  return __ret;
40215}
40216#endif
40217
40218#ifdef __LITTLE_ENDIAN__
40219__ai uint64_t vaddlv_u32(uint32x2_t __p0) {
40220  uint64_t __ret;
40221  __ret = (uint64_t) __builtin_neon_vaddlv_u32((int8x8_t)__p0);
40222  return __ret;
40223}
40224#else
40225__ai uint64_t vaddlv_u32(uint32x2_t __p0) {
40226  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40227  uint64_t __ret;
40228  __ret = (uint64_t) __builtin_neon_vaddlv_u32((int8x8_t)__rev0);
40229  return __ret;
40230}
40231#endif
40232
40233#ifdef __LITTLE_ENDIAN__
40234__ai uint32_t vaddlv_u16(uint16x4_t __p0) {
40235  uint32_t __ret;
40236  __ret = (uint32_t) __builtin_neon_vaddlv_u16((int8x8_t)__p0);
40237  return __ret;
40238}
40239#else
40240__ai uint32_t vaddlv_u16(uint16x4_t __p0) {
40241  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40242  uint32_t __ret;
40243  __ret = (uint32_t) __builtin_neon_vaddlv_u16((int8x8_t)__rev0);
40244  return __ret;
40245}
40246#endif
40247
40248#ifdef __LITTLE_ENDIAN__
40249__ai int16_t vaddlv_s8(int8x8_t __p0) {
40250  int16_t __ret;
40251  __ret = (int16_t) __builtin_neon_vaddlv_s8((int8x8_t)__p0);
40252  return __ret;
40253}
40254#else
40255__ai int16_t vaddlv_s8(int8x8_t __p0) {
40256  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40257  int16_t __ret;
40258  __ret = (int16_t) __builtin_neon_vaddlv_s8((int8x8_t)__rev0);
40259  return __ret;
40260}
40261#endif
40262
40263#ifdef __LITTLE_ENDIAN__
40264__ai int64_t vaddlv_s32(int32x2_t __p0) {
40265  int64_t __ret;
40266  __ret = (int64_t) __builtin_neon_vaddlv_s32((int8x8_t)__p0);
40267  return __ret;
40268}
40269#else
40270__ai int64_t vaddlv_s32(int32x2_t __p0) {
40271  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40272  int64_t __ret;
40273  __ret = (int64_t) __builtin_neon_vaddlv_s32((int8x8_t)__rev0);
40274  return __ret;
40275}
40276#endif
40277
40278#ifdef __LITTLE_ENDIAN__
40279__ai int32_t vaddlv_s16(int16x4_t __p0) {
40280  int32_t __ret;
40281  __ret = (int32_t) __builtin_neon_vaddlv_s16((int8x8_t)__p0);
40282  return __ret;
40283}
40284#else
40285__ai int32_t vaddlv_s16(int16x4_t __p0) {
40286  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40287  int32_t __ret;
40288  __ret = (int32_t) __builtin_neon_vaddlv_s16((int8x8_t)__rev0);
40289  return __ret;
40290}
40291#endif
40292
40293#ifdef __LITTLE_ENDIAN__
40294__ai uint8_t vaddvq_u8(uint8x16_t __p0) {
40295  uint8_t __ret;
40296  __ret = (uint8_t) __builtin_neon_vaddvq_u8((int8x16_t)__p0);
40297  return __ret;
40298}
40299#else
40300__ai uint8_t vaddvq_u8(uint8x16_t __p0) {
40301  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40302  uint8_t __ret;
40303  __ret = (uint8_t) __builtin_neon_vaddvq_u8((int8x16_t)__rev0);
40304  return __ret;
40305}
40306#endif
40307
40308#ifdef __LITTLE_ENDIAN__
40309__ai uint32_t vaddvq_u32(uint32x4_t __p0) {
40310  uint32_t __ret;
40311  __ret = (uint32_t) __builtin_neon_vaddvq_u32((int8x16_t)__p0);
40312  return __ret;
40313}
40314#else
40315__ai uint32_t vaddvq_u32(uint32x4_t __p0) {
40316  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40317  uint32_t __ret;
40318  __ret = (uint32_t) __builtin_neon_vaddvq_u32((int8x16_t)__rev0);
40319  return __ret;
40320}
40321#endif
40322
40323#ifdef __LITTLE_ENDIAN__
40324__ai uint64_t vaddvq_u64(uint64x2_t __p0) {
40325  uint64_t __ret;
40326  __ret = (uint64_t) __builtin_neon_vaddvq_u64((int8x16_t)__p0);
40327  return __ret;
40328}
40329#else
40330__ai uint64_t vaddvq_u64(uint64x2_t __p0) {
40331  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40332  uint64_t __ret;
40333  __ret = (uint64_t) __builtin_neon_vaddvq_u64((int8x16_t)__rev0);
40334  return __ret;
40335}
40336#endif
40337
40338#ifdef __LITTLE_ENDIAN__
40339__ai uint16_t vaddvq_u16(uint16x8_t __p0) {
40340  uint16_t __ret;
40341  __ret = (uint16_t) __builtin_neon_vaddvq_u16((int8x16_t)__p0);
40342  return __ret;
40343}
40344#else
40345__ai uint16_t vaddvq_u16(uint16x8_t __p0) {
40346  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40347  uint16_t __ret;
40348  __ret = (uint16_t) __builtin_neon_vaddvq_u16((int8x16_t)__rev0);
40349  return __ret;
40350}
40351#endif
40352
40353#ifdef __LITTLE_ENDIAN__
40354__ai int8_t vaddvq_s8(int8x16_t __p0) {
40355  int8_t __ret;
40356  __ret = (int8_t) __builtin_neon_vaddvq_s8((int8x16_t)__p0);
40357  return __ret;
40358}
40359#else
40360__ai int8_t vaddvq_s8(int8x16_t __p0) {
40361  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40362  int8_t __ret;
40363  __ret = (int8_t) __builtin_neon_vaddvq_s8((int8x16_t)__rev0);
40364  return __ret;
40365}
40366#endif
40367
40368#ifdef __LITTLE_ENDIAN__
40369__ai float64_t vaddvq_f64(float64x2_t __p0) {
40370  float64_t __ret;
40371  __ret = (float64_t) __builtin_neon_vaddvq_f64((int8x16_t)__p0);
40372  return __ret;
40373}
40374#else
40375__ai float64_t vaddvq_f64(float64x2_t __p0) {
40376  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40377  float64_t __ret;
40378  __ret = (float64_t) __builtin_neon_vaddvq_f64((int8x16_t)__rev0);
40379  return __ret;
40380}
40381#endif
40382
40383#ifdef __LITTLE_ENDIAN__
40384__ai float32_t vaddvq_f32(float32x4_t __p0) {
40385  float32_t __ret;
40386  __ret = (float32_t) __builtin_neon_vaddvq_f32((int8x16_t)__p0);
40387  return __ret;
40388}
40389#else
40390__ai float32_t vaddvq_f32(float32x4_t __p0) {
40391  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40392  float32_t __ret;
40393  __ret = (float32_t) __builtin_neon_vaddvq_f32((int8x16_t)__rev0);
40394  return __ret;
40395}
40396#endif
40397
40398#ifdef __LITTLE_ENDIAN__
40399__ai int32_t vaddvq_s32(int32x4_t __p0) {
40400  int32_t __ret;
40401  __ret = (int32_t) __builtin_neon_vaddvq_s32((int8x16_t)__p0);
40402  return __ret;
40403}
40404#else
40405__ai int32_t vaddvq_s32(int32x4_t __p0) {
40406  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40407  int32_t __ret;
40408  __ret = (int32_t) __builtin_neon_vaddvq_s32((int8x16_t)__rev0);
40409  return __ret;
40410}
40411#endif
40412
40413#ifdef __LITTLE_ENDIAN__
40414__ai int64_t vaddvq_s64(int64x2_t __p0) {
40415  int64_t __ret;
40416  __ret = (int64_t) __builtin_neon_vaddvq_s64((int8x16_t)__p0);
40417  return __ret;
40418}
40419#else
40420__ai int64_t vaddvq_s64(int64x2_t __p0) {
40421  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40422  int64_t __ret;
40423  __ret = (int64_t) __builtin_neon_vaddvq_s64((int8x16_t)__rev0);
40424  return __ret;
40425}
40426#endif
40427
40428#ifdef __LITTLE_ENDIAN__
40429__ai int16_t vaddvq_s16(int16x8_t __p0) {
40430  int16_t __ret;
40431  __ret = (int16_t) __builtin_neon_vaddvq_s16((int8x16_t)__p0);
40432  return __ret;
40433}
40434#else
40435__ai int16_t vaddvq_s16(int16x8_t __p0) {
40436  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40437  int16_t __ret;
40438  __ret = (int16_t) __builtin_neon_vaddvq_s16((int8x16_t)__rev0);
40439  return __ret;
40440}
40441#endif
40442
40443#ifdef __LITTLE_ENDIAN__
40444__ai uint8_t vaddv_u8(uint8x8_t __p0) {
40445  uint8_t __ret;
40446  __ret = (uint8_t) __builtin_neon_vaddv_u8((int8x8_t)__p0);
40447  return __ret;
40448}
40449#else
40450__ai uint8_t vaddv_u8(uint8x8_t __p0) {
40451  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40452  uint8_t __ret;
40453  __ret = (uint8_t) __builtin_neon_vaddv_u8((int8x8_t)__rev0);
40454  return __ret;
40455}
40456#endif
40457
40458#ifdef __LITTLE_ENDIAN__
40459__ai uint32_t vaddv_u32(uint32x2_t __p0) {
40460  uint32_t __ret;
40461  __ret = (uint32_t) __builtin_neon_vaddv_u32((int8x8_t)__p0);
40462  return __ret;
40463}
40464#else
40465__ai uint32_t vaddv_u32(uint32x2_t __p0) {
40466  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40467  uint32_t __ret;
40468  __ret = (uint32_t) __builtin_neon_vaddv_u32((int8x8_t)__rev0);
40469  return __ret;
40470}
40471#endif
40472
40473#ifdef __LITTLE_ENDIAN__
40474__ai uint16_t vaddv_u16(uint16x4_t __p0) {
40475  uint16_t __ret;
40476  __ret = (uint16_t) __builtin_neon_vaddv_u16((int8x8_t)__p0);
40477  return __ret;
40478}
40479#else
40480__ai uint16_t vaddv_u16(uint16x4_t __p0) {
40481  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40482  uint16_t __ret;
40483  __ret = (uint16_t) __builtin_neon_vaddv_u16((int8x8_t)__rev0);
40484  return __ret;
40485}
40486#endif
40487
40488#ifdef __LITTLE_ENDIAN__
40489__ai int8_t vaddv_s8(int8x8_t __p0) {
40490  int8_t __ret;
40491  __ret = (int8_t) __builtin_neon_vaddv_s8((int8x8_t)__p0);
40492  return __ret;
40493}
40494#else
40495__ai int8_t vaddv_s8(int8x8_t __p0) {
40496  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40497  int8_t __ret;
40498  __ret = (int8_t) __builtin_neon_vaddv_s8((int8x8_t)__rev0);
40499  return __ret;
40500}
40501#endif
40502
40503#ifdef __LITTLE_ENDIAN__
40504__ai float32_t vaddv_f32(float32x2_t __p0) {
40505  float32_t __ret;
40506  __ret = (float32_t) __builtin_neon_vaddv_f32((int8x8_t)__p0);
40507  return __ret;
40508}
40509#else
40510__ai float32_t vaddv_f32(float32x2_t __p0) {
40511  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40512  float32_t __ret;
40513  __ret = (float32_t) __builtin_neon_vaddv_f32((int8x8_t)__rev0);
40514  return __ret;
40515}
40516#endif
40517
40518#ifdef __LITTLE_ENDIAN__
40519__ai int32_t vaddv_s32(int32x2_t __p0) {
40520  int32_t __ret;
40521  __ret = (int32_t) __builtin_neon_vaddv_s32((int8x8_t)__p0);
40522  return __ret;
40523}
40524#else
40525__ai int32_t vaddv_s32(int32x2_t __p0) {
40526  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40527  int32_t __ret;
40528  __ret = (int32_t) __builtin_neon_vaddv_s32((int8x8_t)__rev0);
40529  return __ret;
40530}
40531#endif
40532
40533#ifdef __LITTLE_ENDIAN__
40534__ai int16_t vaddv_s16(int16x4_t __p0) {
40535  int16_t __ret;
40536  __ret = (int16_t) __builtin_neon_vaddv_s16((int8x8_t)__p0);
40537  return __ret;
40538}
40539#else
40540__ai int16_t vaddv_s16(int16x4_t __p0) {
40541  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40542  int16_t __ret;
40543  __ret = (int16_t) __builtin_neon_vaddv_s16((int8x8_t)__rev0);
40544  return __ret;
40545}
40546#endif
40547
40548#ifdef __LITTLE_ENDIAN__
40549__ai poly64x1_t vbsl_p64(uint64x1_t __p0, poly64x1_t __p1, poly64x1_t __p2) {
40550  poly64x1_t __ret;
40551  __ret = (poly64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 6);
40552  return __ret;
40553}
40554#else
40555__ai poly64x1_t vbsl_p64(uint64x1_t __p0, poly64x1_t __p1, poly64x1_t __p2) {
40556  poly64x1_t __ret;
40557  __ret = (poly64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 6);
40558  return __ret;
40559}
40560#endif
40561
40562#ifdef __LITTLE_ENDIAN__
40563__ai poly64x2_t vbslq_p64(uint64x2_t __p0, poly64x2_t __p1, poly64x2_t __p2) {
40564  poly64x2_t __ret;
40565  __ret = (poly64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 38);
40566  return __ret;
40567}
40568#else
40569__ai poly64x2_t vbslq_p64(uint64x2_t __p0, poly64x2_t __p1, poly64x2_t __p2) {
40570  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40571  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40572  poly64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
40573  poly64x2_t __ret;
40574  __ret = (poly64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 38);
40575  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40576  return __ret;
40577}
40578#endif
40579
40580#ifdef __LITTLE_ENDIAN__
40581__ai float64x2_t vbslq_f64(uint64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
40582  float64x2_t __ret;
40583  __ret = (float64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
40584  return __ret;
40585}
40586#else
40587__ai float64x2_t vbslq_f64(uint64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
40588  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40589  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40590  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
40591  float64x2_t __ret;
40592  __ret = (float64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
40593  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40594  return __ret;
40595}
40596#endif
40597
40598#ifdef __LITTLE_ENDIAN__
40599__ai float64x1_t vbsl_f64(uint64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
40600  float64x1_t __ret;
40601  __ret = (float64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
40602  return __ret;
40603}
40604#else
40605__ai float64x1_t vbsl_f64(uint64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
40606  float64x1_t __ret;
40607  __ret = (float64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
40608  return __ret;
40609}
40610#endif
40611
40612#ifdef __LITTLE_ENDIAN__
40613__ai uint64x2_t vcageq_f64(float64x2_t __p0, float64x2_t __p1) {
40614  uint64x2_t __ret;
40615  __ret = (uint64x2_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
40616  return __ret;
40617}
40618#else
40619__ai uint64x2_t vcageq_f64(float64x2_t __p0, float64x2_t __p1) {
40620  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40621  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40622  uint64x2_t __ret;
40623  __ret = (uint64x2_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
40624  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40625  return __ret;
40626}
40627#endif
40628
40629#ifdef __LITTLE_ENDIAN__
40630__ai uint64x1_t vcage_f64(float64x1_t __p0, float64x1_t __p1) {
40631  uint64x1_t __ret;
40632  __ret = (uint64x1_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
40633  return __ret;
40634}
40635#else
40636__ai uint64x1_t vcage_f64(float64x1_t __p0, float64x1_t __p1) {
40637  uint64x1_t __ret;
40638  __ret = (uint64x1_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
40639  return __ret;
40640}
40641#endif
40642
40643#ifdef __LITTLE_ENDIAN__
40644__ai uint64_t vcaged_f64(float64_t __p0, float64_t __p1) {
40645  uint64_t __ret;
40646  __ret = (uint64_t) __builtin_neon_vcaged_f64(__p0, __p1);
40647  return __ret;
40648}
40649#else
40650__ai uint64_t vcaged_f64(float64_t __p0, float64_t __p1) {
40651  uint64_t __ret;
40652  __ret = (uint64_t) __builtin_neon_vcaged_f64(__p0, __p1);
40653  return __ret;
40654}
40655#endif
40656
40657#ifdef __LITTLE_ENDIAN__
40658__ai uint32_t vcages_f32(float32_t __p0, float32_t __p1) {
40659  uint32_t __ret;
40660  __ret = (uint32_t) __builtin_neon_vcages_f32(__p0, __p1);
40661  return __ret;
40662}
40663#else
40664__ai uint32_t vcages_f32(float32_t __p0, float32_t __p1) {
40665  uint32_t __ret;
40666  __ret = (uint32_t) __builtin_neon_vcages_f32(__p0, __p1);
40667  return __ret;
40668}
40669#endif
40670
40671#ifdef __LITTLE_ENDIAN__
40672__ai uint64x2_t vcagtq_f64(float64x2_t __p0, float64x2_t __p1) {
40673  uint64x2_t __ret;
40674  __ret = (uint64x2_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
40675  return __ret;
40676}
40677#else
40678__ai uint64x2_t vcagtq_f64(float64x2_t __p0, float64x2_t __p1) {
40679  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40680  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40681  uint64x2_t __ret;
40682  __ret = (uint64x2_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
40683  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40684  return __ret;
40685}
40686#endif
40687
40688#ifdef __LITTLE_ENDIAN__
40689__ai uint64x1_t vcagt_f64(float64x1_t __p0, float64x1_t __p1) {
40690  uint64x1_t __ret;
40691  __ret = (uint64x1_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
40692  return __ret;
40693}
40694#else
40695__ai uint64x1_t vcagt_f64(float64x1_t __p0, float64x1_t __p1) {
40696  uint64x1_t __ret;
40697  __ret = (uint64x1_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
40698  return __ret;
40699}
40700#endif
40701
40702#ifdef __LITTLE_ENDIAN__
40703__ai uint64_t vcagtd_f64(float64_t __p0, float64_t __p1) {
40704  uint64_t __ret;
40705  __ret = (uint64_t) __builtin_neon_vcagtd_f64(__p0, __p1);
40706  return __ret;
40707}
40708#else
40709__ai uint64_t vcagtd_f64(float64_t __p0, float64_t __p1) {
40710  uint64_t __ret;
40711  __ret = (uint64_t) __builtin_neon_vcagtd_f64(__p0, __p1);
40712  return __ret;
40713}
40714#endif
40715
40716#ifdef __LITTLE_ENDIAN__
40717__ai uint32_t vcagts_f32(float32_t __p0, float32_t __p1) {
40718  uint32_t __ret;
40719  __ret = (uint32_t) __builtin_neon_vcagts_f32(__p0, __p1);
40720  return __ret;
40721}
40722#else
40723__ai uint32_t vcagts_f32(float32_t __p0, float32_t __p1) {
40724  uint32_t __ret;
40725  __ret = (uint32_t) __builtin_neon_vcagts_f32(__p0, __p1);
40726  return __ret;
40727}
40728#endif
40729
40730#ifdef __LITTLE_ENDIAN__
40731__ai uint64x2_t vcaleq_f64(float64x2_t __p0, float64x2_t __p1) {
40732  uint64x2_t __ret;
40733  __ret = (uint64x2_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
40734  return __ret;
40735}
40736#else
40737__ai uint64x2_t vcaleq_f64(float64x2_t __p0, float64x2_t __p1) {
40738  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40739  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40740  uint64x2_t __ret;
40741  __ret = (uint64x2_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
40742  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40743  return __ret;
40744}
40745#endif
40746
40747#ifdef __LITTLE_ENDIAN__
40748__ai uint64x1_t vcale_f64(float64x1_t __p0, float64x1_t __p1) {
40749  uint64x1_t __ret;
40750  __ret = (uint64x1_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
40751  return __ret;
40752}
40753#else
40754__ai uint64x1_t vcale_f64(float64x1_t __p0, float64x1_t __p1) {
40755  uint64x1_t __ret;
40756  __ret = (uint64x1_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
40757  return __ret;
40758}
40759#endif
40760
40761#ifdef __LITTLE_ENDIAN__
40762__ai uint64_t vcaled_f64(float64_t __p0, float64_t __p1) {
40763  uint64_t __ret;
40764  __ret = (uint64_t) __builtin_neon_vcaled_f64(__p0, __p1);
40765  return __ret;
40766}
40767#else
40768__ai uint64_t vcaled_f64(float64_t __p0, float64_t __p1) {
40769  uint64_t __ret;
40770  __ret = (uint64_t) __builtin_neon_vcaled_f64(__p0, __p1);
40771  return __ret;
40772}
40773#endif
40774
40775#ifdef __LITTLE_ENDIAN__
40776__ai uint32_t vcales_f32(float32_t __p0, float32_t __p1) {
40777  uint32_t __ret;
40778  __ret = (uint32_t) __builtin_neon_vcales_f32(__p0, __p1);
40779  return __ret;
40780}
40781#else
40782__ai uint32_t vcales_f32(float32_t __p0, float32_t __p1) {
40783  uint32_t __ret;
40784  __ret = (uint32_t) __builtin_neon_vcales_f32(__p0, __p1);
40785  return __ret;
40786}
40787#endif
40788
40789#ifdef __LITTLE_ENDIAN__
40790__ai uint64x2_t vcaltq_f64(float64x2_t __p0, float64x2_t __p1) {
40791  uint64x2_t __ret;
40792  __ret = (uint64x2_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
40793  return __ret;
40794}
40795#else
40796__ai uint64x2_t vcaltq_f64(float64x2_t __p0, float64x2_t __p1) {
40797  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40798  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40799  uint64x2_t __ret;
40800  __ret = (uint64x2_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
40801  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40802  return __ret;
40803}
40804#endif
40805
40806#ifdef __LITTLE_ENDIAN__
40807__ai uint64x1_t vcalt_f64(float64x1_t __p0, float64x1_t __p1) {
40808  uint64x1_t __ret;
40809  __ret = (uint64x1_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
40810  return __ret;
40811}
40812#else
40813__ai uint64x1_t vcalt_f64(float64x1_t __p0, float64x1_t __p1) {
40814  uint64x1_t __ret;
40815  __ret = (uint64x1_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
40816  return __ret;
40817}
40818#endif
40819
40820#ifdef __LITTLE_ENDIAN__
40821__ai uint64_t vcaltd_f64(float64_t __p0, float64_t __p1) {
40822  uint64_t __ret;
40823  __ret = (uint64_t) __builtin_neon_vcaltd_f64(__p0, __p1);
40824  return __ret;
40825}
40826#else
40827__ai uint64_t vcaltd_f64(float64_t __p0, float64_t __p1) {
40828  uint64_t __ret;
40829  __ret = (uint64_t) __builtin_neon_vcaltd_f64(__p0, __p1);
40830  return __ret;
40831}
40832#endif
40833
40834#ifdef __LITTLE_ENDIAN__
40835__ai uint32_t vcalts_f32(float32_t __p0, float32_t __p1) {
40836  uint32_t __ret;
40837  __ret = (uint32_t) __builtin_neon_vcalts_f32(__p0, __p1);
40838  return __ret;
40839}
40840#else
40841__ai uint32_t vcalts_f32(float32_t __p0, float32_t __p1) {
40842  uint32_t __ret;
40843  __ret = (uint32_t) __builtin_neon_vcalts_f32(__p0, __p1);
40844  return __ret;
40845}
40846#endif
40847
40848#ifdef __LITTLE_ENDIAN__
40849__ai uint64x1_t vceq_p64(poly64x1_t __p0, poly64x1_t __p1) {
40850  uint64x1_t __ret;
40851  __ret = (uint64x1_t)(__p0 == __p1);
40852  return __ret;
40853}
40854#else
40855__ai uint64x1_t vceq_p64(poly64x1_t __p0, poly64x1_t __p1) {
40856  uint64x1_t __ret;
40857  __ret = (uint64x1_t)(__p0 == __p1);
40858  return __ret;
40859}
40860#endif
40861
40862#ifdef __LITTLE_ENDIAN__
40863__ai uint64x2_t vceqq_p64(poly64x2_t __p0, poly64x2_t __p1) {
40864  uint64x2_t __ret;
40865  __ret = (uint64x2_t)(__p0 == __p1);
40866  return __ret;
40867}
40868#else
40869__ai uint64x2_t vceqq_p64(poly64x2_t __p0, poly64x2_t __p1) {
40870  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40871  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40872  uint64x2_t __ret;
40873  __ret = (uint64x2_t)(__rev0 == __rev1);
40874  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40875  return __ret;
40876}
40877#endif
40878
40879#ifdef __LITTLE_ENDIAN__
40880__ai uint64x2_t vceqq_u64(uint64x2_t __p0, uint64x2_t __p1) {
40881  uint64x2_t __ret;
40882  __ret = (uint64x2_t)(__p0 == __p1);
40883  return __ret;
40884}
40885#else
40886__ai uint64x2_t vceqq_u64(uint64x2_t __p0, uint64x2_t __p1) {
40887  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40888  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40889  uint64x2_t __ret;
40890  __ret = (uint64x2_t)(__rev0 == __rev1);
40891  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40892  return __ret;
40893}
40894#endif
40895
40896#ifdef __LITTLE_ENDIAN__
40897__ai uint64x2_t vceqq_f64(float64x2_t __p0, float64x2_t __p1) {
40898  uint64x2_t __ret;
40899  __ret = (uint64x2_t)(__p0 == __p1);
40900  return __ret;
40901}
40902#else
40903__ai uint64x2_t vceqq_f64(float64x2_t __p0, float64x2_t __p1) {
40904  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40905  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40906  uint64x2_t __ret;
40907  __ret = (uint64x2_t)(__rev0 == __rev1);
40908  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40909  return __ret;
40910}
40911#endif
40912
40913#ifdef __LITTLE_ENDIAN__
40914__ai uint64x2_t vceqq_s64(int64x2_t __p0, int64x2_t __p1) {
40915  uint64x2_t __ret;
40916  __ret = (uint64x2_t)(__p0 == __p1);
40917  return __ret;
40918}
40919#else
40920__ai uint64x2_t vceqq_s64(int64x2_t __p0, int64x2_t __p1) {
40921  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40922  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40923  uint64x2_t __ret;
40924  __ret = (uint64x2_t)(__rev0 == __rev1);
40925  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40926  return __ret;
40927}
40928#endif
40929
40930#ifdef __LITTLE_ENDIAN__
40931__ai uint64x1_t vceq_u64(uint64x1_t __p0, uint64x1_t __p1) {
40932  uint64x1_t __ret;
40933  __ret = (uint64x1_t)(__p0 == __p1);
40934  return __ret;
40935}
40936#else
40937__ai uint64x1_t vceq_u64(uint64x1_t __p0, uint64x1_t __p1) {
40938  uint64x1_t __ret;
40939  __ret = (uint64x1_t)(__p0 == __p1);
40940  return __ret;
40941}
40942#endif
40943
40944#ifdef __LITTLE_ENDIAN__
40945__ai uint64x1_t vceq_f64(float64x1_t __p0, float64x1_t __p1) {
40946  uint64x1_t __ret;
40947  __ret = (uint64x1_t)(__p0 == __p1);
40948  return __ret;
40949}
40950#else
40951__ai uint64x1_t vceq_f64(float64x1_t __p0, float64x1_t __p1) {
40952  uint64x1_t __ret;
40953  __ret = (uint64x1_t)(__p0 == __p1);
40954  return __ret;
40955}
40956#endif
40957
40958#ifdef __LITTLE_ENDIAN__
40959__ai uint64x1_t vceq_s64(int64x1_t __p0, int64x1_t __p1) {
40960  uint64x1_t __ret;
40961  __ret = (uint64x1_t)(__p0 == __p1);
40962  return __ret;
40963}
40964#else
40965__ai uint64x1_t vceq_s64(int64x1_t __p0, int64x1_t __p1) {
40966  uint64x1_t __ret;
40967  __ret = (uint64x1_t)(__p0 == __p1);
40968  return __ret;
40969}
40970#endif
40971
40972#ifdef __LITTLE_ENDIAN__
40973__ai uint64_t vceqd_u64(uint64_t __p0, uint64_t __p1) {
40974  uint64_t __ret;
40975  __ret = (uint64_t) __builtin_neon_vceqd_u64(__p0, __p1);
40976  return __ret;
40977}
40978#else
40979__ai uint64_t vceqd_u64(uint64_t __p0, uint64_t __p1) {
40980  uint64_t __ret;
40981  __ret = (uint64_t) __builtin_neon_vceqd_u64(__p0, __p1);
40982  return __ret;
40983}
40984#endif
40985
40986#ifdef __LITTLE_ENDIAN__
40987__ai int64_t vceqd_s64(int64_t __p0, int64_t __p1) {
40988  int64_t __ret;
40989  __ret = (int64_t) __builtin_neon_vceqd_s64(__p0, __p1);
40990  return __ret;
40991}
40992#else
40993__ai int64_t vceqd_s64(int64_t __p0, int64_t __p1) {
40994  int64_t __ret;
40995  __ret = (int64_t) __builtin_neon_vceqd_s64(__p0, __p1);
40996  return __ret;
40997}
40998#endif
40999
41000#ifdef __LITTLE_ENDIAN__
41001__ai uint64_t vceqd_f64(float64_t __p0, float64_t __p1) {
41002  uint64_t __ret;
41003  __ret = (uint64_t) __builtin_neon_vceqd_f64(__p0, __p1);
41004  return __ret;
41005}
41006#else
41007__ai uint64_t vceqd_f64(float64_t __p0, float64_t __p1) {
41008  uint64_t __ret;
41009  __ret = (uint64_t) __builtin_neon_vceqd_f64(__p0, __p1);
41010  return __ret;
41011}
41012#endif
41013
41014#ifdef __LITTLE_ENDIAN__
41015__ai uint32_t vceqs_f32(float32_t __p0, float32_t __p1) {
41016  uint32_t __ret;
41017  __ret = (uint32_t) __builtin_neon_vceqs_f32(__p0, __p1);
41018  return __ret;
41019}
41020#else
41021__ai uint32_t vceqs_f32(float32_t __p0, float32_t __p1) {
41022  uint32_t __ret;
41023  __ret = (uint32_t) __builtin_neon_vceqs_f32(__p0, __p1);
41024  return __ret;
41025}
41026#endif
41027
41028#ifdef __LITTLE_ENDIAN__
41029__ai uint8x8_t vceqz_p8(poly8x8_t __p0) {
41030  uint8x8_t __ret;
41031  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
41032  return __ret;
41033}
41034#else
41035__ai uint8x8_t vceqz_p8(poly8x8_t __p0) {
41036  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41037  uint8x8_t __ret;
41038  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
41039  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41040  return __ret;
41041}
41042#endif
41043
41044#ifdef __LITTLE_ENDIAN__
41045__ai uint64x1_t vceqz_p64(poly64x1_t __p0) {
41046  uint64x1_t __ret;
41047  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
41048  return __ret;
41049}
41050#else
41051__ai uint64x1_t vceqz_p64(poly64x1_t __p0) {
41052  uint64x1_t __ret;
41053  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
41054  return __ret;
41055}
41056#endif
41057
41058#ifdef __LITTLE_ENDIAN__
41059__ai uint16x4_t vceqz_p16(poly16x4_t __p0) {
41060  uint16x4_t __ret;
41061  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
41062  return __ret;
41063}
41064#else
41065__ai uint16x4_t vceqz_p16(poly16x4_t __p0) {
41066  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41067  uint16x4_t __ret;
41068  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
41069  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41070  return __ret;
41071}
41072#endif
41073
41074#ifdef __LITTLE_ENDIAN__
41075__ai uint8x16_t vceqzq_p8(poly8x16_t __p0) {
41076  uint8x16_t __ret;
41077  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
41078  return __ret;
41079}
41080#else
41081__ai uint8x16_t vceqzq_p8(poly8x16_t __p0) {
41082  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41083  uint8x16_t __ret;
41084  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
41085  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41086  return __ret;
41087}
41088#endif
41089
41090#ifdef __LITTLE_ENDIAN__
41091__ai uint64x2_t vceqzq_p64(poly64x2_t __p0) {
41092  uint64x2_t __ret;
41093  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
41094  return __ret;
41095}
41096#else
41097__ai uint64x2_t vceqzq_p64(poly64x2_t __p0) {
41098  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41099  uint64x2_t __ret;
41100  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
41101  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41102  return __ret;
41103}
41104#endif
41105
41106#ifdef __LITTLE_ENDIAN__
41107__ai uint16x8_t vceqzq_p16(poly16x8_t __p0) {
41108  uint16x8_t __ret;
41109  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
41110  return __ret;
41111}
41112#else
41113__ai uint16x8_t vceqzq_p16(poly16x8_t __p0) {
41114  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41115  uint16x8_t __ret;
41116  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
41117  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41118  return __ret;
41119}
41120#endif
41121
41122#ifdef __LITTLE_ENDIAN__
41123__ai uint8x16_t vceqzq_u8(uint8x16_t __p0) {
41124  uint8x16_t __ret;
41125  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
41126  return __ret;
41127}
41128#else
41129__ai uint8x16_t vceqzq_u8(uint8x16_t __p0) {
41130  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41131  uint8x16_t __ret;
41132  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
41133  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41134  return __ret;
41135}
41136#endif
41137
41138#ifdef __LITTLE_ENDIAN__
41139__ai uint32x4_t vceqzq_u32(uint32x4_t __p0) {
41140  uint32x4_t __ret;
41141  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
41142  return __ret;
41143}
41144#else
41145__ai uint32x4_t vceqzq_u32(uint32x4_t __p0) {
41146  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41147  uint32x4_t __ret;
41148  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
41149  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41150  return __ret;
41151}
41152#endif
41153
41154#ifdef __LITTLE_ENDIAN__
41155__ai uint64x2_t vceqzq_u64(uint64x2_t __p0) {
41156  uint64x2_t __ret;
41157  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
41158  return __ret;
41159}
41160#else
41161__ai uint64x2_t vceqzq_u64(uint64x2_t __p0) {
41162  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41163  uint64x2_t __ret;
41164  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
41165  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41166  return __ret;
41167}
41168#endif
41169
41170#ifdef __LITTLE_ENDIAN__
41171__ai uint16x8_t vceqzq_u16(uint16x8_t __p0) {
41172  uint16x8_t __ret;
41173  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
41174  return __ret;
41175}
41176#else
41177__ai uint16x8_t vceqzq_u16(uint16x8_t __p0) {
41178  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41179  uint16x8_t __ret;
41180  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
41181  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41182  return __ret;
41183}
41184#endif
41185
41186#ifdef __LITTLE_ENDIAN__
41187__ai uint8x16_t vceqzq_s8(int8x16_t __p0) {
41188  uint8x16_t __ret;
41189  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
41190  return __ret;
41191}
41192#else
41193__ai uint8x16_t vceqzq_s8(int8x16_t __p0) {
41194  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41195  uint8x16_t __ret;
41196  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
41197  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41198  return __ret;
41199}
41200#endif
41201
41202#ifdef __LITTLE_ENDIAN__
41203__ai uint64x2_t vceqzq_f64(float64x2_t __p0) {
41204  uint64x2_t __ret;
41205  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
41206  return __ret;
41207}
41208#else
41209__ai uint64x2_t vceqzq_f64(float64x2_t __p0) {
41210  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41211  uint64x2_t __ret;
41212  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
41213  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41214  return __ret;
41215}
41216#endif
41217
41218#ifdef __LITTLE_ENDIAN__
41219__ai uint32x4_t vceqzq_f32(float32x4_t __p0) {
41220  uint32x4_t __ret;
41221  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
41222  return __ret;
41223}
41224#else
41225__ai uint32x4_t vceqzq_f32(float32x4_t __p0) {
41226  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41227  uint32x4_t __ret;
41228  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
41229  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41230  return __ret;
41231}
41232#endif
41233
41234#ifdef __LITTLE_ENDIAN__
41235__ai uint32x4_t vceqzq_s32(int32x4_t __p0) {
41236  uint32x4_t __ret;
41237  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
41238  return __ret;
41239}
41240#else
41241__ai uint32x4_t vceqzq_s32(int32x4_t __p0) {
41242  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41243  uint32x4_t __ret;
41244  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
41245  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41246  return __ret;
41247}
41248#endif
41249
41250#ifdef __LITTLE_ENDIAN__
41251__ai uint64x2_t vceqzq_s64(int64x2_t __p0) {
41252  uint64x2_t __ret;
41253  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
41254  return __ret;
41255}
41256#else
41257__ai uint64x2_t vceqzq_s64(int64x2_t __p0) {
41258  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41259  uint64x2_t __ret;
41260  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
41261  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41262  return __ret;
41263}
41264#endif
41265
41266#ifdef __LITTLE_ENDIAN__
41267__ai uint16x8_t vceqzq_s16(int16x8_t __p0) {
41268  uint16x8_t __ret;
41269  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
41270  return __ret;
41271}
41272#else
41273__ai uint16x8_t vceqzq_s16(int16x8_t __p0) {
41274  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41275  uint16x8_t __ret;
41276  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
41277  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41278  return __ret;
41279}
41280#endif
41281
41282#ifdef __LITTLE_ENDIAN__
41283__ai uint8x8_t vceqz_u8(uint8x8_t __p0) {
41284  uint8x8_t __ret;
41285  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
41286  return __ret;
41287}
41288#else
41289__ai uint8x8_t vceqz_u8(uint8x8_t __p0) {
41290  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41291  uint8x8_t __ret;
41292  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
41293  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41294  return __ret;
41295}
41296#endif
41297
41298#ifdef __LITTLE_ENDIAN__
41299__ai uint32x2_t vceqz_u32(uint32x2_t __p0) {
41300  uint32x2_t __ret;
41301  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
41302  return __ret;
41303}
41304#else
41305__ai uint32x2_t vceqz_u32(uint32x2_t __p0) {
41306  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41307  uint32x2_t __ret;
41308  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
41309  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41310  return __ret;
41311}
41312#endif
41313
41314#ifdef __LITTLE_ENDIAN__
41315__ai uint64x1_t vceqz_u64(uint64x1_t __p0) {
41316  uint64x1_t __ret;
41317  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
41318  return __ret;
41319}
41320#else
41321__ai uint64x1_t vceqz_u64(uint64x1_t __p0) {
41322  uint64x1_t __ret;
41323  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
41324  return __ret;
41325}
41326#endif
41327
41328#ifdef __LITTLE_ENDIAN__
41329__ai uint16x4_t vceqz_u16(uint16x4_t __p0) {
41330  uint16x4_t __ret;
41331  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
41332  return __ret;
41333}
41334#else
41335__ai uint16x4_t vceqz_u16(uint16x4_t __p0) {
41336  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41337  uint16x4_t __ret;
41338  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
41339  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41340  return __ret;
41341}
41342#endif
41343
41344#ifdef __LITTLE_ENDIAN__
41345__ai uint8x8_t vceqz_s8(int8x8_t __p0) {
41346  uint8x8_t __ret;
41347  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
41348  return __ret;
41349}
41350#else
41351__ai uint8x8_t vceqz_s8(int8x8_t __p0) {
41352  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41353  uint8x8_t __ret;
41354  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
41355  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41356  return __ret;
41357}
41358#endif
41359
41360#ifdef __LITTLE_ENDIAN__
41361__ai uint64x1_t vceqz_f64(float64x1_t __p0) {
41362  uint64x1_t __ret;
41363  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
41364  return __ret;
41365}
41366#else
41367__ai uint64x1_t vceqz_f64(float64x1_t __p0) {
41368  uint64x1_t __ret;
41369  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
41370  return __ret;
41371}
41372#endif
41373
41374#ifdef __LITTLE_ENDIAN__
41375__ai uint32x2_t vceqz_f32(float32x2_t __p0) {
41376  uint32x2_t __ret;
41377  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
41378  return __ret;
41379}
41380#else
41381__ai uint32x2_t vceqz_f32(float32x2_t __p0) {
41382  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41383  uint32x2_t __ret;
41384  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
41385  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41386  return __ret;
41387}
41388#endif
41389
41390#ifdef __LITTLE_ENDIAN__
41391__ai uint32x2_t vceqz_s32(int32x2_t __p0) {
41392  uint32x2_t __ret;
41393  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
41394  return __ret;
41395}
41396#else
41397__ai uint32x2_t vceqz_s32(int32x2_t __p0) {
41398  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41399  uint32x2_t __ret;
41400  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
41401  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41402  return __ret;
41403}
41404#endif
41405
41406#ifdef __LITTLE_ENDIAN__
41407__ai uint64x1_t vceqz_s64(int64x1_t __p0) {
41408  uint64x1_t __ret;
41409  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
41410  return __ret;
41411}
41412#else
41413__ai uint64x1_t vceqz_s64(int64x1_t __p0) {
41414  uint64x1_t __ret;
41415  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
41416  return __ret;
41417}
41418#endif
41419
41420#ifdef __LITTLE_ENDIAN__
41421__ai uint16x4_t vceqz_s16(int16x4_t __p0) {
41422  uint16x4_t __ret;
41423  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
41424  return __ret;
41425}
41426#else
41427__ai uint16x4_t vceqz_s16(int16x4_t __p0) {
41428  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41429  uint16x4_t __ret;
41430  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
41431  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41432  return __ret;
41433}
41434#endif
41435
41436#ifdef __LITTLE_ENDIAN__
41437__ai uint64_t vceqzd_u64(uint64_t __p0) {
41438  uint64_t __ret;
41439  __ret = (uint64_t) __builtin_neon_vceqzd_u64(__p0);
41440  return __ret;
41441}
41442#else
41443__ai uint64_t vceqzd_u64(uint64_t __p0) {
41444  uint64_t __ret;
41445  __ret = (uint64_t) __builtin_neon_vceqzd_u64(__p0);
41446  return __ret;
41447}
41448#endif
41449
41450#ifdef __LITTLE_ENDIAN__
41451__ai int64_t vceqzd_s64(int64_t __p0) {
41452  int64_t __ret;
41453  __ret = (int64_t) __builtin_neon_vceqzd_s64(__p0);
41454  return __ret;
41455}
41456#else
41457__ai int64_t vceqzd_s64(int64_t __p0) {
41458  int64_t __ret;
41459  __ret = (int64_t) __builtin_neon_vceqzd_s64(__p0);
41460  return __ret;
41461}
41462#endif
41463
41464#ifdef __LITTLE_ENDIAN__
41465__ai uint64_t vceqzd_f64(float64_t __p0) {
41466  uint64_t __ret;
41467  __ret = (uint64_t) __builtin_neon_vceqzd_f64(__p0);
41468  return __ret;
41469}
41470#else
41471__ai uint64_t vceqzd_f64(float64_t __p0) {
41472  uint64_t __ret;
41473  __ret = (uint64_t) __builtin_neon_vceqzd_f64(__p0);
41474  return __ret;
41475}
41476#endif
41477
41478#ifdef __LITTLE_ENDIAN__
41479__ai uint32_t vceqzs_f32(float32_t __p0) {
41480  uint32_t __ret;
41481  __ret = (uint32_t) __builtin_neon_vceqzs_f32(__p0);
41482  return __ret;
41483}
41484#else
41485__ai uint32_t vceqzs_f32(float32_t __p0) {
41486  uint32_t __ret;
41487  __ret = (uint32_t) __builtin_neon_vceqzs_f32(__p0);
41488  return __ret;
41489}
41490#endif
41491
41492#ifdef __LITTLE_ENDIAN__
41493__ai uint64x2_t vcgeq_u64(uint64x2_t __p0, uint64x2_t __p1) {
41494  uint64x2_t __ret;
41495  __ret = (uint64x2_t)(__p0 >= __p1);
41496  return __ret;
41497}
41498#else
41499__ai uint64x2_t vcgeq_u64(uint64x2_t __p0, uint64x2_t __p1) {
41500  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41501  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41502  uint64x2_t __ret;
41503  __ret = (uint64x2_t)(__rev0 >= __rev1);
41504  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41505  return __ret;
41506}
41507#endif
41508
41509#ifdef __LITTLE_ENDIAN__
41510__ai uint64x2_t vcgeq_f64(float64x2_t __p0, float64x2_t __p1) {
41511  uint64x2_t __ret;
41512  __ret = (uint64x2_t)(__p0 >= __p1);
41513  return __ret;
41514}
41515#else
41516__ai uint64x2_t vcgeq_f64(float64x2_t __p0, float64x2_t __p1) {
41517  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41518  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41519  uint64x2_t __ret;
41520  __ret = (uint64x2_t)(__rev0 >= __rev1);
41521  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41522  return __ret;
41523}
41524#endif
41525
41526#ifdef __LITTLE_ENDIAN__
41527__ai uint64x2_t vcgeq_s64(int64x2_t __p0, int64x2_t __p1) {
41528  uint64x2_t __ret;
41529  __ret = (uint64x2_t)(__p0 >= __p1);
41530  return __ret;
41531}
41532#else
41533__ai uint64x2_t vcgeq_s64(int64x2_t __p0, int64x2_t __p1) {
41534  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41535  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41536  uint64x2_t __ret;
41537  __ret = (uint64x2_t)(__rev0 >= __rev1);
41538  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41539  return __ret;
41540}
41541#endif
41542
41543#ifdef __LITTLE_ENDIAN__
41544__ai uint64x1_t vcge_u64(uint64x1_t __p0, uint64x1_t __p1) {
41545  uint64x1_t __ret;
41546  __ret = (uint64x1_t)(__p0 >= __p1);
41547  return __ret;
41548}
41549#else
41550__ai uint64x1_t vcge_u64(uint64x1_t __p0, uint64x1_t __p1) {
41551  uint64x1_t __ret;
41552  __ret = (uint64x1_t)(__p0 >= __p1);
41553  return __ret;
41554}
41555#endif
41556
41557#ifdef __LITTLE_ENDIAN__
41558__ai uint64x1_t vcge_f64(float64x1_t __p0, float64x1_t __p1) {
41559  uint64x1_t __ret;
41560  __ret = (uint64x1_t)(__p0 >= __p1);
41561  return __ret;
41562}
41563#else
41564__ai uint64x1_t vcge_f64(float64x1_t __p0, float64x1_t __p1) {
41565  uint64x1_t __ret;
41566  __ret = (uint64x1_t)(__p0 >= __p1);
41567  return __ret;
41568}
41569#endif
41570
41571#ifdef __LITTLE_ENDIAN__
41572__ai uint64x1_t vcge_s64(int64x1_t __p0, int64x1_t __p1) {
41573  uint64x1_t __ret;
41574  __ret = (uint64x1_t)(__p0 >= __p1);
41575  return __ret;
41576}
41577#else
41578__ai uint64x1_t vcge_s64(int64x1_t __p0, int64x1_t __p1) {
41579  uint64x1_t __ret;
41580  __ret = (uint64x1_t)(__p0 >= __p1);
41581  return __ret;
41582}
41583#endif
41584
41585#ifdef __LITTLE_ENDIAN__
41586__ai int64_t vcged_s64(int64_t __p0, int64_t __p1) {
41587  int64_t __ret;
41588  __ret = (int64_t) __builtin_neon_vcged_s64(__p0, __p1);
41589  return __ret;
41590}
41591#else
41592__ai int64_t vcged_s64(int64_t __p0, int64_t __p1) {
41593  int64_t __ret;
41594  __ret = (int64_t) __builtin_neon_vcged_s64(__p0, __p1);
41595  return __ret;
41596}
41597#endif
41598
41599#ifdef __LITTLE_ENDIAN__
41600__ai uint64_t vcged_u64(uint64_t __p0, uint64_t __p1) {
41601  uint64_t __ret;
41602  __ret = (uint64_t) __builtin_neon_vcged_u64(__p0, __p1);
41603  return __ret;
41604}
41605#else
41606__ai uint64_t vcged_u64(uint64_t __p0, uint64_t __p1) {
41607  uint64_t __ret;
41608  __ret = (uint64_t) __builtin_neon_vcged_u64(__p0, __p1);
41609  return __ret;
41610}
41611#endif
41612
41613#ifdef __LITTLE_ENDIAN__
41614__ai uint64_t vcged_f64(float64_t __p0, float64_t __p1) {
41615  uint64_t __ret;
41616  __ret = (uint64_t) __builtin_neon_vcged_f64(__p0, __p1);
41617  return __ret;
41618}
41619#else
41620__ai uint64_t vcged_f64(float64_t __p0, float64_t __p1) {
41621  uint64_t __ret;
41622  __ret = (uint64_t) __builtin_neon_vcged_f64(__p0, __p1);
41623  return __ret;
41624}
41625#endif
41626
41627#ifdef __LITTLE_ENDIAN__
41628__ai uint32_t vcges_f32(float32_t __p0, float32_t __p1) {
41629  uint32_t __ret;
41630  __ret = (uint32_t) __builtin_neon_vcges_f32(__p0, __p1);
41631  return __ret;
41632}
41633#else
41634__ai uint32_t vcges_f32(float32_t __p0, float32_t __p1) {
41635  uint32_t __ret;
41636  __ret = (uint32_t) __builtin_neon_vcges_f32(__p0, __p1);
41637  return __ret;
41638}
41639#endif
41640
41641#ifdef __LITTLE_ENDIAN__
41642__ai uint8x16_t vcgezq_s8(int8x16_t __p0) {
41643  uint8x16_t __ret;
41644  __ret = (uint8x16_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 48);
41645  return __ret;
41646}
41647#else
41648__ai uint8x16_t vcgezq_s8(int8x16_t __p0) {
41649  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41650  uint8x16_t __ret;
41651  __ret = (uint8x16_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 48);
41652  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41653  return __ret;
41654}
41655#endif
41656
41657#ifdef __LITTLE_ENDIAN__
41658__ai uint64x2_t vcgezq_f64(float64x2_t __p0) {
41659  uint64x2_t __ret;
41660  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 51);
41661  return __ret;
41662}
41663#else
41664__ai uint64x2_t vcgezq_f64(float64x2_t __p0) {
41665  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41666  uint64x2_t __ret;
41667  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 51);
41668  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41669  return __ret;
41670}
41671#endif
41672
41673#ifdef __LITTLE_ENDIAN__
41674__ai uint32x4_t vcgezq_f32(float32x4_t __p0) {
41675  uint32x4_t __ret;
41676  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 50);
41677  return __ret;
41678}
41679#else
41680__ai uint32x4_t vcgezq_f32(float32x4_t __p0) {
41681  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41682  uint32x4_t __ret;
41683  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 50);
41684  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41685  return __ret;
41686}
41687#endif
41688
41689#ifdef __LITTLE_ENDIAN__
41690__ai uint32x4_t vcgezq_s32(int32x4_t __p0) {
41691  uint32x4_t __ret;
41692  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 50);
41693  return __ret;
41694}
41695#else
41696__ai uint32x4_t vcgezq_s32(int32x4_t __p0) {
41697  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41698  uint32x4_t __ret;
41699  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 50);
41700  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41701  return __ret;
41702}
41703#endif
41704
41705#ifdef __LITTLE_ENDIAN__
41706__ai uint64x2_t vcgezq_s64(int64x2_t __p0) {
41707  uint64x2_t __ret;
41708  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 51);
41709  return __ret;
41710}
41711#else
41712__ai uint64x2_t vcgezq_s64(int64x2_t __p0) {
41713  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41714  uint64x2_t __ret;
41715  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 51);
41716  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41717  return __ret;
41718}
41719#endif
41720
41721#ifdef __LITTLE_ENDIAN__
41722__ai uint16x8_t vcgezq_s16(int16x8_t __p0) {
41723  uint16x8_t __ret;
41724  __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 49);
41725  return __ret;
41726}
41727#else
41728__ai uint16x8_t vcgezq_s16(int16x8_t __p0) {
41729  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41730  uint16x8_t __ret;
41731  __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 49);
41732  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41733  return __ret;
41734}
41735#endif
41736
41737#ifdef __LITTLE_ENDIAN__
41738__ai uint8x8_t vcgez_s8(int8x8_t __p0) {
41739  uint8x8_t __ret;
41740  __ret = (uint8x8_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 16);
41741  return __ret;
41742}
41743#else
41744__ai uint8x8_t vcgez_s8(int8x8_t __p0) {
41745  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41746  uint8x8_t __ret;
41747  __ret = (uint8x8_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 16);
41748  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41749  return __ret;
41750}
41751#endif
41752
41753#ifdef __LITTLE_ENDIAN__
41754__ai uint64x1_t vcgez_f64(float64x1_t __p0) {
41755  uint64x1_t __ret;
41756  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
41757  return __ret;
41758}
41759#else
41760__ai uint64x1_t vcgez_f64(float64x1_t __p0) {
41761  uint64x1_t __ret;
41762  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
41763  return __ret;
41764}
41765#endif
41766
41767#ifdef __LITTLE_ENDIAN__
41768__ai uint32x2_t vcgez_f32(float32x2_t __p0) {
41769  uint32x2_t __ret;
41770  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 18);
41771  return __ret;
41772}
41773#else
41774__ai uint32x2_t vcgez_f32(float32x2_t __p0) {
41775  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41776  uint32x2_t __ret;
41777  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 18);
41778  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41779  return __ret;
41780}
41781#endif
41782
41783#ifdef __LITTLE_ENDIAN__
41784__ai uint32x2_t vcgez_s32(int32x2_t __p0) {
41785  uint32x2_t __ret;
41786  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 18);
41787  return __ret;
41788}
41789#else
41790__ai uint32x2_t vcgez_s32(int32x2_t __p0) {
41791  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41792  uint32x2_t __ret;
41793  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 18);
41794  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41795  return __ret;
41796}
41797#endif
41798
41799#ifdef __LITTLE_ENDIAN__
41800__ai uint64x1_t vcgez_s64(int64x1_t __p0) {
41801  uint64x1_t __ret;
41802  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
41803  return __ret;
41804}
41805#else
41806__ai uint64x1_t vcgez_s64(int64x1_t __p0) {
41807  uint64x1_t __ret;
41808  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
41809  return __ret;
41810}
41811#endif
41812
41813#ifdef __LITTLE_ENDIAN__
41814__ai uint16x4_t vcgez_s16(int16x4_t __p0) {
41815  uint16x4_t __ret;
41816  __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 17);
41817  return __ret;
41818}
41819#else
41820__ai uint16x4_t vcgez_s16(int16x4_t __p0) {
41821  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41822  uint16x4_t __ret;
41823  __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 17);
41824  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41825  return __ret;
41826}
41827#endif
41828
41829#ifdef __LITTLE_ENDIAN__
41830__ai int64_t vcgezd_s64(int64_t __p0) {
41831  int64_t __ret;
41832  __ret = (int64_t) __builtin_neon_vcgezd_s64(__p0);
41833  return __ret;
41834}
41835#else
41836__ai int64_t vcgezd_s64(int64_t __p0) {
41837  int64_t __ret;
41838  __ret = (int64_t) __builtin_neon_vcgezd_s64(__p0);
41839  return __ret;
41840}
41841#endif
41842
41843#ifdef __LITTLE_ENDIAN__
41844__ai uint64_t vcgezd_f64(float64_t __p0) {
41845  uint64_t __ret;
41846  __ret = (uint64_t) __builtin_neon_vcgezd_f64(__p0);
41847  return __ret;
41848}
41849#else
41850__ai uint64_t vcgezd_f64(float64_t __p0) {
41851  uint64_t __ret;
41852  __ret = (uint64_t) __builtin_neon_vcgezd_f64(__p0);
41853  return __ret;
41854}
41855#endif
41856
41857#ifdef __LITTLE_ENDIAN__
41858__ai uint32_t vcgezs_f32(float32_t __p0) {
41859  uint32_t __ret;
41860  __ret = (uint32_t) __builtin_neon_vcgezs_f32(__p0);
41861  return __ret;
41862}
41863#else
41864__ai uint32_t vcgezs_f32(float32_t __p0) {
41865  uint32_t __ret;
41866  __ret = (uint32_t) __builtin_neon_vcgezs_f32(__p0);
41867  return __ret;
41868}
41869#endif
41870
41871#ifdef __LITTLE_ENDIAN__
41872__ai uint64x2_t vcgtq_u64(uint64x2_t __p0, uint64x2_t __p1) {
41873  uint64x2_t __ret;
41874  __ret = (uint64x2_t)(__p0 > __p1);
41875  return __ret;
41876}
41877#else
41878__ai uint64x2_t vcgtq_u64(uint64x2_t __p0, uint64x2_t __p1) {
41879  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41880  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41881  uint64x2_t __ret;
41882  __ret = (uint64x2_t)(__rev0 > __rev1);
41883  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41884  return __ret;
41885}
41886#endif
41887
41888#ifdef __LITTLE_ENDIAN__
41889__ai uint64x2_t vcgtq_f64(float64x2_t __p0, float64x2_t __p1) {
41890  uint64x2_t __ret;
41891  __ret = (uint64x2_t)(__p0 > __p1);
41892  return __ret;
41893}
41894#else
41895__ai uint64x2_t vcgtq_f64(float64x2_t __p0, float64x2_t __p1) {
41896  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41897  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41898  uint64x2_t __ret;
41899  __ret = (uint64x2_t)(__rev0 > __rev1);
41900  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41901  return __ret;
41902}
41903#endif
41904
41905#ifdef __LITTLE_ENDIAN__
41906__ai uint64x2_t vcgtq_s64(int64x2_t __p0, int64x2_t __p1) {
41907  uint64x2_t __ret;
41908  __ret = (uint64x2_t)(__p0 > __p1);
41909  return __ret;
41910}
41911#else
41912__ai uint64x2_t vcgtq_s64(int64x2_t __p0, int64x2_t __p1) {
41913  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41914  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41915  uint64x2_t __ret;
41916  __ret = (uint64x2_t)(__rev0 > __rev1);
41917  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41918  return __ret;
41919}
41920#endif
41921
41922#ifdef __LITTLE_ENDIAN__
41923__ai uint64x1_t vcgt_u64(uint64x1_t __p0, uint64x1_t __p1) {
41924  uint64x1_t __ret;
41925  __ret = (uint64x1_t)(__p0 > __p1);
41926  return __ret;
41927}
41928#else
41929__ai uint64x1_t vcgt_u64(uint64x1_t __p0, uint64x1_t __p1) {
41930  uint64x1_t __ret;
41931  __ret = (uint64x1_t)(__p0 > __p1);
41932  return __ret;
41933}
41934#endif
41935
41936#ifdef __LITTLE_ENDIAN__
41937__ai uint64x1_t vcgt_f64(float64x1_t __p0, float64x1_t __p1) {
41938  uint64x1_t __ret;
41939  __ret = (uint64x1_t)(__p0 > __p1);
41940  return __ret;
41941}
41942#else
41943__ai uint64x1_t vcgt_f64(float64x1_t __p0, float64x1_t __p1) {
41944  uint64x1_t __ret;
41945  __ret = (uint64x1_t)(__p0 > __p1);
41946  return __ret;
41947}
41948#endif
41949
41950#ifdef __LITTLE_ENDIAN__
41951__ai uint64x1_t vcgt_s64(int64x1_t __p0, int64x1_t __p1) {
41952  uint64x1_t __ret;
41953  __ret = (uint64x1_t)(__p0 > __p1);
41954  return __ret;
41955}
41956#else
41957__ai uint64x1_t vcgt_s64(int64x1_t __p0, int64x1_t __p1) {
41958  uint64x1_t __ret;
41959  __ret = (uint64x1_t)(__p0 > __p1);
41960  return __ret;
41961}
41962#endif
41963
41964#ifdef __LITTLE_ENDIAN__
41965__ai int64_t vcgtd_s64(int64_t __p0, int64_t __p1) {
41966  int64_t __ret;
41967  __ret = (int64_t) __builtin_neon_vcgtd_s64(__p0, __p1);
41968  return __ret;
41969}
41970#else
41971__ai int64_t vcgtd_s64(int64_t __p0, int64_t __p1) {
41972  int64_t __ret;
41973  __ret = (int64_t) __builtin_neon_vcgtd_s64(__p0, __p1);
41974  return __ret;
41975}
41976#endif
41977
41978#ifdef __LITTLE_ENDIAN__
41979__ai uint64_t vcgtd_u64(uint64_t __p0, uint64_t __p1) {
41980  uint64_t __ret;
41981  __ret = (uint64_t) __builtin_neon_vcgtd_u64(__p0, __p1);
41982  return __ret;
41983}
41984#else
41985__ai uint64_t vcgtd_u64(uint64_t __p0, uint64_t __p1) {
41986  uint64_t __ret;
41987  __ret = (uint64_t) __builtin_neon_vcgtd_u64(__p0, __p1);
41988  return __ret;
41989}
41990#endif
41991
41992#ifdef __LITTLE_ENDIAN__
41993__ai uint64_t vcgtd_f64(float64_t __p0, float64_t __p1) {
41994  uint64_t __ret;
41995  __ret = (uint64_t) __builtin_neon_vcgtd_f64(__p0, __p1);
41996  return __ret;
41997}
41998#else
41999__ai uint64_t vcgtd_f64(float64_t __p0, float64_t __p1) {
42000  uint64_t __ret;
42001  __ret = (uint64_t) __builtin_neon_vcgtd_f64(__p0, __p1);
42002  return __ret;
42003}
42004#endif
42005
42006#ifdef __LITTLE_ENDIAN__
42007__ai uint32_t vcgts_f32(float32_t __p0, float32_t __p1) {
42008  uint32_t __ret;
42009  __ret = (uint32_t) __builtin_neon_vcgts_f32(__p0, __p1);
42010  return __ret;
42011}
42012#else
42013__ai uint32_t vcgts_f32(float32_t __p0, float32_t __p1) {
42014  uint32_t __ret;
42015  __ret = (uint32_t) __builtin_neon_vcgts_f32(__p0, __p1);
42016  return __ret;
42017}
42018#endif
42019
42020#ifdef __LITTLE_ENDIAN__
42021__ai uint8x16_t vcgtzq_s8(int8x16_t __p0) {
42022  uint8x16_t __ret;
42023  __ret = (uint8x16_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 48);
42024  return __ret;
42025}
42026#else
42027__ai uint8x16_t vcgtzq_s8(int8x16_t __p0) {
42028  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42029  uint8x16_t __ret;
42030  __ret = (uint8x16_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 48);
42031  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42032  return __ret;
42033}
42034#endif
42035
42036#ifdef __LITTLE_ENDIAN__
42037__ai uint64x2_t vcgtzq_f64(float64x2_t __p0) {
42038  uint64x2_t __ret;
42039  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 51);
42040  return __ret;
42041}
42042#else
42043__ai uint64x2_t vcgtzq_f64(float64x2_t __p0) {
42044  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42045  uint64x2_t __ret;
42046  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 51);
42047  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42048  return __ret;
42049}
42050#endif
42051
42052#ifdef __LITTLE_ENDIAN__
42053__ai uint32x4_t vcgtzq_f32(float32x4_t __p0) {
42054  uint32x4_t __ret;
42055  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 50);
42056  return __ret;
42057}
42058#else
42059__ai uint32x4_t vcgtzq_f32(float32x4_t __p0) {
42060  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42061  uint32x4_t __ret;
42062  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 50);
42063  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42064  return __ret;
42065}
42066#endif
42067
42068#ifdef __LITTLE_ENDIAN__
42069__ai uint32x4_t vcgtzq_s32(int32x4_t __p0) {
42070  uint32x4_t __ret;
42071  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 50);
42072  return __ret;
42073}
42074#else
42075__ai uint32x4_t vcgtzq_s32(int32x4_t __p0) {
42076  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42077  uint32x4_t __ret;
42078  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 50);
42079  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42080  return __ret;
42081}
42082#endif
42083
42084#ifdef __LITTLE_ENDIAN__
42085__ai uint64x2_t vcgtzq_s64(int64x2_t __p0) {
42086  uint64x2_t __ret;
42087  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 51);
42088  return __ret;
42089}
42090#else
42091__ai uint64x2_t vcgtzq_s64(int64x2_t __p0) {
42092  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42093  uint64x2_t __ret;
42094  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 51);
42095  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42096  return __ret;
42097}
42098#endif
42099
42100#ifdef __LITTLE_ENDIAN__
42101__ai uint16x8_t vcgtzq_s16(int16x8_t __p0) {
42102  uint16x8_t __ret;
42103  __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 49);
42104  return __ret;
42105}
42106#else
42107__ai uint16x8_t vcgtzq_s16(int16x8_t __p0) {
42108  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42109  uint16x8_t __ret;
42110  __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 49);
42111  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42112  return __ret;
42113}
42114#endif
42115
42116#ifdef __LITTLE_ENDIAN__
42117__ai uint8x8_t vcgtz_s8(int8x8_t __p0) {
42118  uint8x8_t __ret;
42119  __ret = (uint8x8_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 16);
42120  return __ret;
42121}
42122#else
42123__ai uint8x8_t vcgtz_s8(int8x8_t __p0) {
42124  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42125  uint8x8_t __ret;
42126  __ret = (uint8x8_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 16);
42127  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42128  return __ret;
42129}
42130#endif
42131
42132#ifdef __LITTLE_ENDIAN__
42133__ai uint64x1_t vcgtz_f64(float64x1_t __p0) {
42134  uint64x1_t __ret;
42135  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
42136  return __ret;
42137}
42138#else
42139__ai uint64x1_t vcgtz_f64(float64x1_t __p0) {
42140  uint64x1_t __ret;
42141  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
42142  return __ret;
42143}
42144#endif
42145
42146#ifdef __LITTLE_ENDIAN__
42147__ai uint32x2_t vcgtz_f32(float32x2_t __p0) {
42148  uint32x2_t __ret;
42149  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 18);
42150  return __ret;
42151}
42152#else
42153__ai uint32x2_t vcgtz_f32(float32x2_t __p0) {
42154  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42155  uint32x2_t __ret;
42156  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 18);
42157  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42158  return __ret;
42159}
42160#endif
42161
42162#ifdef __LITTLE_ENDIAN__
42163__ai uint32x2_t vcgtz_s32(int32x2_t __p0) {
42164  uint32x2_t __ret;
42165  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 18);
42166  return __ret;
42167}
42168#else
42169__ai uint32x2_t vcgtz_s32(int32x2_t __p0) {
42170  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42171  uint32x2_t __ret;
42172  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 18);
42173  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42174  return __ret;
42175}
42176#endif
42177
42178#ifdef __LITTLE_ENDIAN__
42179__ai uint64x1_t vcgtz_s64(int64x1_t __p0) {
42180  uint64x1_t __ret;
42181  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
42182  return __ret;
42183}
42184#else
42185__ai uint64x1_t vcgtz_s64(int64x1_t __p0) {
42186  uint64x1_t __ret;
42187  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
42188  return __ret;
42189}
42190#endif
42191
42192#ifdef __LITTLE_ENDIAN__
42193__ai uint16x4_t vcgtz_s16(int16x4_t __p0) {
42194  uint16x4_t __ret;
42195  __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 17);
42196  return __ret;
42197}
42198#else
42199__ai uint16x4_t vcgtz_s16(int16x4_t __p0) {
42200  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42201  uint16x4_t __ret;
42202  __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 17);
42203  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42204  return __ret;
42205}
42206#endif
42207
42208#ifdef __LITTLE_ENDIAN__
42209__ai int64_t vcgtzd_s64(int64_t __p0) {
42210  int64_t __ret;
42211  __ret = (int64_t) __builtin_neon_vcgtzd_s64(__p0);
42212  return __ret;
42213}
42214#else
42215__ai int64_t vcgtzd_s64(int64_t __p0) {
42216  int64_t __ret;
42217  __ret = (int64_t) __builtin_neon_vcgtzd_s64(__p0);
42218  return __ret;
42219}
42220#endif
42221
42222#ifdef __LITTLE_ENDIAN__
42223__ai uint64_t vcgtzd_f64(float64_t __p0) {
42224  uint64_t __ret;
42225  __ret = (uint64_t) __builtin_neon_vcgtzd_f64(__p0);
42226  return __ret;
42227}
42228#else
42229__ai uint64_t vcgtzd_f64(float64_t __p0) {
42230  uint64_t __ret;
42231  __ret = (uint64_t) __builtin_neon_vcgtzd_f64(__p0);
42232  return __ret;
42233}
42234#endif
42235
42236#ifdef __LITTLE_ENDIAN__
42237__ai uint32_t vcgtzs_f32(float32_t __p0) {
42238  uint32_t __ret;
42239  __ret = (uint32_t) __builtin_neon_vcgtzs_f32(__p0);
42240  return __ret;
42241}
42242#else
42243__ai uint32_t vcgtzs_f32(float32_t __p0) {
42244  uint32_t __ret;
42245  __ret = (uint32_t) __builtin_neon_vcgtzs_f32(__p0);
42246  return __ret;
42247}
42248#endif
42249
42250#ifdef __LITTLE_ENDIAN__
42251__ai uint64x2_t vcleq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42252  uint64x2_t __ret;
42253  __ret = (uint64x2_t)(__p0 <= __p1);
42254  return __ret;
42255}
42256#else
42257__ai uint64x2_t vcleq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42258  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42259  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42260  uint64x2_t __ret;
42261  __ret = (uint64x2_t)(__rev0 <= __rev1);
42262  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42263  return __ret;
42264}
42265#endif
42266
42267#ifdef __LITTLE_ENDIAN__
42268__ai uint64x2_t vcleq_f64(float64x2_t __p0, float64x2_t __p1) {
42269  uint64x2_t __ret;
42270  __ret = (uint64x2_t)(__p0 <= __p1);
42271  return __ret;
42272}
42273#else
42274__ai uint64x2_t vcleq_f64(float64x2_t __p0, float64x2_t __p1) {
42275  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42276  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42277  uint64x2_t __ret;
42278  __ret = (uint64x2_t)(__rev0 <= __rev1);
42279  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42280  return __ret;
42281}
42282#endif
42283
42284#ifdef __LITTLE_ENDIAN__
42285__ai uint64x2_t vcleq_s64(int64x2_t __p0, int64x2_t __p1) {
42286  uint64x2_t __ret;
42287  __ret = (uint64x2_t)(__p0 <= __p1);
42288  return __ret;
42289}
42290#else
42291__ai uint64x2_t vcleq_s64(int64x2_t __p0, int64x2_t __p1) {
42292  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42293  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42294  uint64x2_t __ret;
42295  __ret = (uint64x2_t)(__rev0 <= __rev1);
42296  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42297  return __ret;
42298}
42299#endif
42300
42301#ifdef __LITTLE_ENDIAN__
42302__ai uint64x1_t vcle_u64(uint64x1_t __p0, uint64x1_t __p1) {
42303  uint64x1_t __ret;
42304  __ret = (uint64x1_t)(__p0 <= __p1);
42305  return __ret;
42306}
42307#else
42308__ai uint64x1_t vcle_u64(uint64x1_t __p0, uint64x1_t __p1) {
42309  uint64x1_t __ret;
42310  __ret = (uint64x1_t)(__p0 <= __p1);
42311  return __ret;
42312}
42313#endif
42314
42315#ifdef __LITTLE_ENDIAN__
42316__ai uint64x1_t vcle_f64(float64x1_t __p0, float64x1_t __p1) {
42317  uint64x1_t __ret;
42318  __ret = (uint64x1_t)(__p0 <= __p1);
42319  return __ret;
42320}
42321#else
42322__ai uint64x1_t vcle_f64(float64x1_t __p0, float64x1_t __p1) {
42323  uint64x1_t __ret;
42324  __ret = (uint64x1_t)(__p0 <= __p1);
42325  return __ret;
42326}
42327#endif
42328
42329#ifdef __LITTLE_ENDIAN__
42330__ai uint64x1_t vcle_s64(int64x1_t __p0, int64x1_t __p1) {
42331  uint64x1_t __ret;
42332  __ret = (uint64x1_t)(__p0 <= __p1);
42333  return __ret;
42334}
42335#else
42336__ai uint64x1_t vcle_s64(int64x1_t __p0, int64x1_t __p1) {
42337  uint64x1_t __ret;
42338  __ret = (uint64x1_t)(__p0 <= __p1);
42339  return __ret;
42340}
42341#endif
42342
42343#ifdef __LITTLE_ENDIAN__
42344__ai uint64_t vcled_u64(uint64_t __p0, uint64_t __p1) {
42345  uint64_t __ret;
42346  __ret = (uint64_t) __builtin_neon_vcled_u64(__p0, __p1);
42347  return __ret;
42348}
42349#else
42350__ai uint64_t vcled_u64(uint64_t __p0, uint64_t __p1) {
42351  uint64_t __ret;
42352  __ret = (uint64_t) __builtin_neon_vcled_u64(__p0, __p1);
42353  return __ret;
42354}
42355#endif
42356
42357#ifdef __LITTLE_ENDIAN__
42358__ai int64_t vcled_s64(int64_t __p0, int64_t __p1) {
42359  int64_t __ret;
42360  __ret = (int64_t) __builtin_neon_vcled_s64(__p0, __p1);
42361  return __ret;
42362}
42363#else
42364__ai int64_t vcled_s64(int64_t __p0, int64_t __p1) {
42365  int64_t __ret;
42366  __ret = (int64_t) __builtin_neon_vcled_s64(__p0, __p1);
42367  return __ret;
42368}
42369#endif
42370
42371#ifdef __LITTLE_ENDIAN__
42372__ai uint64_t vcled_f64(float64_t __p0, float64_t __p1) {
42373  uint64_t __ret;
42374  __ret = (uint64_t) __builtin_neon_vcled_f64(__p0, __p1);
42375  return __ret;
42376}
42377#else
42378__ai uint64_t vcled_f64(float64_t __p0, float64_t __p1) {
42379  uint64_t __ret;
42380  __ret = (uint64_t) __builtin_neon_vcled_f64(__p0, __p1);
42381  return __ret;
42382}
42383#endif
42384
42385#ifdef __LITTLE_ENDIAN__
42386__ai uint32_t vcles_f32(float32_t __p0, float32_t __p1) {
42387  uint32_t __ret;
42388  __ret = (uint32_t) __builtin_neon_vcles_f32(__p0, __p1);
42389  return __ret;
42390}
42391#else
42392__ai uint32_t vcles_f32(float32_t __p0, float32_t __p1) {
42393  uint32_t __ret;
42394  __ret = (uint32_t) __builtin_neon_vcles_f32(__p0, __p1);
42395  return __ret;
42396}
42397#endif
42398
42399#ifdef __LITTLE_ENDIAN__
42400__ai uint8x16_t vclezq_s8(int8x16_t __p0) {
42401  uint8x16_t __ret;
42402  __ret = (uint8x16_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 48);
42403  return __ret;
42404}
42405#else
42406__ai uint8x16_t vclezq_s8(int8x16_t __p0) {
42407  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42408  uint8x16_t __ret;
42409  __ret = (uint8x16_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 48);
42410  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42411  return __ret;
42412}
42413#endif
42414
42415#ifdef __LITTLE_ENDIAN__
42416__ai uint64x2_t vclezq_f64(float64x2_t __p0) {
42417  uint64x2_t __ret;
42418  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 51);
42419  return __ret;
42420}
42421#else
42422__ai uint64x2_t vclezq_f64(float64x2_t __p0) {
42423  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42424  uint64x2_t __ret;
42425  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 51);
42426  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42427  return __ret;
42428}
42429#endif
42430
42431#ifdef __LITTLE_ENDIAN__
42432__ai uint32x4_t vclezq_f32(float32x4_t __p0) {
42433  uint32x4_t __ret;
42434  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 50);
42435  return __ret;
42436}
42437#else
42438__ai uint32x4_t vclezq_f32(float32x4_t __p0) {
42439  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42440  uint32x4_t __ret;
42441  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 50);
42442  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42443  return __ret;
42444}
42445#endif
42446
42447#ifdef __LITTLE_ENDIAN__
42448__ai uint32x4_t vclezq_s32(int32x4_t __p0) {
42449  uint32x4_t __ret;
42450  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 50);
42451  return __ret;
42452}
42453#else
42454__ai uint32x4_t vclezq_s32(int32x4_t __p0) {
42455  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42456  uint32x4_t __ret;
42457  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 50);
42458  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42459  return __ret;
42460}
42461#endif
42462
42463#ifdef __LITTLE_ENDIAN__
42464__ai uint64x2_t vclezq_s64(int64x2_t __p0) {
42465  uint64x2_t __ret;
42466  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 51);
42467  return __ret;
42468}
42469#else
42470__ai uint64x2_t vclezq_s64(int64x2_t __p0) {
42471  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42472  uint64x2_t __ret;
42473  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 51);
42474  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42475  return __ret;
42476}
42477#endif
42478
42479#ifdef __LITTLE_ENDIAN__
42480__ai uint16x8_t vclezq_s16(int16x8_t __p0) {
42481  uint16x8_t __ret;
42482  __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 49);
42483  return __ret;
42484}
42485#else
42486__ai uint16x8_t vclezq_s16(int16x8_t __p0) {
42487  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42488  uint16x8_t __ret;
42489  __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 49);
42490  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42491  return __ret;
42492}
42493#endif
42494
42495#ifdef __LITTLE_ENDIAN__
42496__ai uint8x8_t vclez_s8(int8x8_t __p0) {
42497  uint8x8_t __ret;
42498  __ret = (uint8x8_t) __builtin_neon_vclez_v((int8x8_t)__p0, 16);
42499  return __ret;
42500}
42501#else
42502__ai uint8x8_t vclez_s8(int8x8_t __p0) {
42503  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42504  uint8x8_t __ret;
42505  __ret = (uint8x8_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 16);
42506  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42507  return __ret;
42508}
42509#endif
42510
42511#ifdef __LITTLE_ENDIAN__
42512__ai uint64x1_t vclez_f64(float64x1_t __p0) {
42513  uint64x1_t __ret;
42514  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
42515  return __ret;
42516}
42517#else
42518__ai uint64x1_t vclez_f64(float64x1_t __p0) {
42519  uint64x1_t __ret;
42520  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
42521  return __ret;
42522}
42523#endif
42524
42525#ifdef __LITTLE_ENDIAN__
42526__ai uint32x2_t vclez_f32(float32x2_t __p0) {
42527  uint32x2_t __ret;
42528  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__p0, 18);
42529  return __ret;
42530}
42531#else
42532__ai uint32x2_t vclez_f32(float32x2_t __p0) {
42533  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42534  uint32x2_t __ret;
42535  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 18);
42536  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42537  return __ret;
42538}
42539#endif
42540
42541#ifdef __LITTLE_ENDIAN__
42542__ai uint32x2_t vclez_s32(int32x2_t __p0) {
42543  uint32x2_t __ret;
42544  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__p0, 18);
42545  return __ret;
42546}
42547#else
42548__ai uint32x2_t vclez_s32(int32x2_t __p0) {
42549  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42550  uint32x2_t __ret;
42551  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 18);
42552  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42553  return __ret;
42554}
42555#endif
42556
42557#ifdef __LITTLE_ENDIAN__
42558__ai uint64x1_t vclez_s64(int64x1_t __p0) {
42559  uint64x1_t __ret;
42560  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
42561  return __ret;
42562}
42563#else
42564__ai uint64x1_t vclez_s64(int64x1_t __p0) {
42565  uint64x1_t __ret;
42566  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
42567  return __ret;
42568}
42569#endif
42570
42571#ifdef __LITTLE_ENDIAN__
42572__ai uint16x4_t vclez_s16(int16x4_t __p0) {
42573  uint16x4_t __ret;
42574  __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__p0, 17);
42575  return __ret;
42576}
42577#else
42578__ai uint16x4_t vclez_s16(int16x4_t __p0) {
42579  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42580  uint16x4_t __ret;
42581  __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 17);
42582  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42583  return __ret;
42584}
42585#endif
42586
42587#ifdef __LITTLE_ENDIAN__
42588__ai int64_t vclezd_s64(int64_t __p0) {
42589  int64_t __ret;
42590  __ret = (int64_t) __builtin_neon_vclezd_s64(__p0);
42591  return __ret;
42592}
42593#else
42594__ai int64_t vclezd_s64(int64_t __p0) {
42595  int64_t __ret;
42596  __ret = (int64_t) __builtin_neon_vclezd_s64(__p0);
42597  return __ret;
42598}
42599#endif
42600
42601#ifdef __LITTLE_ENDIAN__
42602__ai uint64_t vclezd_f64(float64_t __p0) {
42603  uint64_t __ret;
42604  __ret = (uint64_t) __builtin_neon_vclezd_f64(__p0);
42605  return __ret;
42606}
42607#else
42608__ai uint64_t vclezd_f64(float64_t __p0) {
42609  uint64_t __ret;
42610  __ret = (uint64_t) __builtin_neon_vclezd_f64(__p0);
42611  return __ret;
42612}
42613#endif
42614
42615#ifdef __LITTLE_ENDIAN__
42616__ai uint32_t vclezs_f32(float32_t __p0) {
42617  uint32_t __ret;
42618  __ret = (uint32_t) __builtin_neon_vclezs_f32(__p0);
42619  return __ret;
42620}
42621#else
42622__ai uint32_t vclezs_f32(float32_t __p0) {
42623  uint32_t __ret;
42624  __ret = (uint32_t) __builtin_neon_vclezs_f32(__p0);
42625  return __ret;
42626}
42627#endif
42628
42629#ifdef __LITTLE_ENDIAN__
42630__ai uint64x2_t vcltq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42631  uint64x2_t __ret;
42632  __ret = (uint64x2_t)(__p0 < __p1);
42633  return __ret;
42634}
42635#else
42636__ai uint64x2_t vcltq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42637  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42638  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42639  uint64x2_t __ret;
42640  __ret = (uint64x2_t)(__rev0 < __rev1);
42641  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42642  return __ret;
42643}
42644#endif
42645
42646#ifdef __LITTLE_ENDIAN__
42647__ai uint64x2_t vcltq_f64(float64x2_t __p0, float64x2_t __p1) {
42648  uint64x2_t __ret;
42649  __ret = (uint64x2_t)(__p0 < __p1);
42650  return __ret;
42651}
42652#else
42653__ai uint64x2_t vcltq_f64(float64x2_t __p0, float64x2_t __p1) {
42654  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42655  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42656  uint64x2_t __ret;
42657  __ret = (uint64x2_t)(__rev0 < __rev1);
42658  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42659  return __ret;
42660}
42661#endif
42662
42663#ifdef __LITTLE_ENDIAN__
42664__ai uint64x2_t vcltq_s64(int64x2_t __p0, int64x2_t __p1) {
42665  uint64x2_t __ret;
42666  __ret = (uint64x2_t)(__p0 < __p1);
42667  return __ret;
42668}
42669#else
42670__ai uint64x2_t vcltq_s64(int64x2_t __p0, int64x2_t __p1) {
42671  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42672  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42673  uint64x2_t __ret;
42674  __ret = (uint64x2_t)(__rev0 < __rev1);
42675  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42676  return __ret;
42677}
42678#endif
42679
42680#ifdef __LITTLE_ENDIAN__
42681__ai uint64x1_t vclt_u64(uint64x1_t __p0, uint64x1_t __p1) {
42682  uint64x1_t __ret;
42683  __ret = (uint64x1_t)(__p0 < __p1);
42684  return __ret;
42685}
42686#else
42687__ai uint64x1_t vclt_u64(uint64x1_t __p0, uint64x1_t __p1) {
42688  uint64x1_t __ret;
42689  __ret = (uint64x1_t)(__p0 < __p1);
42690  return __ret;
42691}
42692#endif
42693
42694#ifdef __LITTLE_ENDIAN__
42695__ai uint64x1_t vclt_f64(float64x1_t __p0, float64x1_t __p1) {
42696  uint64x1_t __ret;
42697  __ret = (uint64x1_t)(__p0 < __p1);
42698  return __ret;
42699}
42700#else
42701__ai uint64x1_t vclt_f64(float64x1_t __p0, float64x1_t __p1) {
42702  uint64x1_t __ret;
42703  __ret = (uint64x1_t)(__p0 < __p1);
42704  return __ret;
42705}
42706#endif
42707
42708#ifdef __LITTLE_ENDIAN__
42709__ai uint64x1_t vclt_s64(int64x1_t __p0, int64x1_t __p1) {
42710  uint64x1_t __ret;
42711  __ret = (uint64x1_t)(__p0 < __p1);
42712  return __ret;
42713}
42714#else
42715__ai uint64x1_t vclt_s64(int64x1_t __p0, int64x1_t __p1) {
42716  uint64x1_t __ret;
42717  __ret = (uint64x1_t)(__p0 < __p1);
42718  return __ret;
42719}
42720#endif
42721
42722#ifdef __LITTLE_ENDIAN__
42723__ai uint64_t vcltd_u64(uint64_t __p0, uint64_t __p1) {
42724  uint64_t __ret;
42725  __ret = (uint64_t) __builtin_neon_vcltd_u64(__p0, __p1);
42726  return __ret;
42727}
42728#else
42729__ai uint64_t vcltd_u64(uint64_t __p0, uint64_t __p1) {
42730  uint64_t __ret;
42731  __ret = (uint64_t) __builtin_neon_vcltd_u64(__p0, __p1);
42732  return __ret;
42733}
42734#endif
42735
42736#ifdef __LITTLE_ENDIAN__
42737__ai int64_t vcltd_s64(int64_t __p0, int64_t __p1) {
42738  int64_t __ret;
42739  __ret = (int64_t) __builtin_neon_vcltd_s64(__p0, __p1);
42740  return __ret;
42741}
42742#else
42743__ai int64_t vcltd_s64(int64_t __p0, int64_t __p1) {
42744  int64_t __ret;
42745  __ret = (int64_t) __builtin_neon_vcltd_s64(__p0, __p1);
42746  return __ret;
42747}
42748#endif
42749
42750#ifdef __LITTLE_ENDIAN__
42751__ai uint64_t vcltd_f64(float64_t __p0, float64_t __p1) {
42752  uint64_t __ret;
42753  __ret = (uint64_t) __builtin_neon_vcltd_f64(__p0, __p1);
42754  return __ret;
42755}
42756#else
42757__ai uint64_t vcltd_f64(float64_t __p0, float64_t __p1) {
42758  uint64_t __ret;
42759  __ret = (uint64_t) __builtin_neon_vcltd_f64(__p0, __p1);
42760  return __ret;
42761}
42762#endif
42763
42764#ifdef __LITTLE_ENDIAN__
42765__ai uint32_t vclts_f32(float32_t __p0, float32_t __p1) {
42766  uint32_t __ret;
42767  __ret = (uint32_t) __builtin_neon_vclts_f32(__p0, __p1);
42768  return __ret;
42769}
42770#else
42771__ai uint32_t vclts_f32(float32_t __p0, float32_t __p1) {
42772  uint32_t __ret;
42773  __ret = (uint32_t) __builtin_neon_vclts_f32(__p0, __p1);
42774  return __ret;
42775}
42776#endif
42777
42778#ifdef __LITTLE_ENDIAN__
42779__ai uint8x16_t vcltzq_s8(int8x16_t __p0) {
42780  uint8x16_t __ret;
42781  __ret = (uint8x16_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 48);
42782  return __ret;
42783}
42784#else
42785__ai uint8x16_t vcltzq_s8(int8x16_t __p0) {
42786  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42787  uint8x16_t __ret;
42788  __ret = (uint8x16_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 48);
42789  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42790  return __ret;
42791}
42792#endif
42793
42794#ifdef __LITTLE_ENDIAN__
42795__ai uint64x2_t vcltzq_f64(float64x2_t __p0) {
42796  uint64x2_t __ret;
42797  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 51);
42798  return __ret;
42799}
42800#else
42801__ai uint64x2_t vcltzq_f64(float64x2_t __p0) {
42802  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42803  uint64x2_t __ret;
42804  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 51);
42805  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42806  return __ret;
42807}
42808#endif
42809
42810#ifdef __LITTLE_ENDIAN__
42811__ai uint32x4_t vcltzq_f32(float32x4_t __p0) {
42812  uint32x4_t __ret;
42813  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 50);
42814  return __ret;
42815}
42816#else
42817__ai uint32x4_t vcltzq_f32(float32x4_t __p0) {
42818  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42819  uint32x4_t __ret;
42820  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 50);
42821  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42822  return __ret;
42823}
42824#endif
42825
42826#ifdef __LITTLE_ENDIAN__
42827__ai uint32x4_t vcltzq_s32(int32x4_t __p0) {
42828  uint32x4_t __ret;
42829  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 50);
42830  return __ret;
42831}
42832#else
42833__ai uint32x4_t vcltzq_s32(int32x4_t __p0) {
42834  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42835  uint32x4_t __ret;
42836  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 50);
42837  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42838  return __ret;
42839}
42840#endif
42841
42842#ifdef __LITTLE_ENDIAN__
42843__ai uint64x2_t vcltzq_s64(int64x2_t __p0) {
42844  uint64x2_t __ret;
42845  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 51);
42846  return __ret;
42847}
42848#else
42849__ai uint64x2_t vcltzq_s64(int64x2_t __p0) {
42850  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42851  uint64x2_t __ret;
42852  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 51);
42853  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42854  return __ret;
42855}
42856#endif
42857
42858#ifdef __LITTLE_ENDIAN__
42859__ai uint16x8_t vcltzq_s16(int16x8_t __p0) {
42860  uint16x8_t __ret;
42861  __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 49);
42862  return __ret;
42863}
42864#else
42865__ai uint16x8_t vcltzq_s16(int16x8_t __p0) {
42866  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42867  uint16x8_t __ret;
42868  __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 49);
42869  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42870  return __ret;
42871}
42872#endif
42873
42874#ifdef __LITTLE_ENDIAN__
42875__ai uint8x8_t vcltz_s8(int8x8_t __p0) {
42876  uint8x8_t __ret;
42877  __ret = (uint8x8_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 16);
42878  return __ret;
42879}
42880#else
42881__ai uint8x8_t vcltz_s8(int8x8_t __p0) {
42882  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42883  uint8x8_t __ret;
42884  __ret = (uint8x8_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 16);
42885  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42886  return __ret;
42887}
42888#endif
42889
42890#ifdef __LITTLE_ENDIAN__
42891__ai uint64x1_t vcltz_f64(float64x1_t __p0) {
42892  uint64x1_t __ret;
42893  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
42894  return __ret;
42895}
42896#else
42897__ai uint64x1_t vcltz_f64(float64x1_t __p0) {
42898  uint64x1_t __ret;
42899  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
42900  return __ret;
42901}
42902#endif
42903
42904#ifdef __LITTLE_ENDIAN__
42905__ai uint32x2_t vcltz_f32(float32x2_t __p0) {
42906  uint32x2_t __ret;
42907  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 18);
42908  return __ret;
42909}
42910#else
42911__ai uint32x2_t vcltz_f32(float32x2_t __p0) {
42912  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42913  uint32x2_t __ret;
42914  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 18);
42915  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42916  return __ret;
42917}
42918#endif
42919
42920#ifdef __LITTLE_ENDIAN__
42921__ai uint32x2_t vcltz_s32(int32x2_t __p0) {
42922  uint32x2_t __ret;
42923  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 18);
42924  return __ret;
42925}
42926#else
42927__ai uint32x2_t vcltz_s32(int32x2_t __p0) {
42928  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42929  uint32x2_t __ret;
42930  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 18);
42931  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42932  return __ret;
42933}
42934#endif
42935
42936#ifdef __LITTLE_ENDIAN__
42937__ai uint64x1_t vcltz_s64(int64x1_t __p0) {
42938  uint64x1_t __ret;
42939  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
42940  return __ret;
42941}
42942#else
42943__ai uint64x1_t vcltz_s64(int64x1_t __p0) {
42944  uint64x1_t __ret;
42945  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
42946  return __ret;
42947}
42948#endif
42949
42950#ifdef __LITTLE_ENDIAN__
42951__ai uint16x4_t vcltz_s16(int16x4_t __p0) {
42952  uint16x4_t __ret;
42953  __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 17);
42954  return __ret;
42955}
42956#else
42957__ai uint16x4_t vcltz_s16(int16x4_t __p0) {
42958  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42959  uint16x4_t __ret;
42960  __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 17);
42961  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42962  return __ret;
42963}
42964#endif
42965
42966#ifdef __LITTLE_ENDIAN__
42967__ai int64_t vcltzd_s64(int64_t __p0) {
42968  int64_t __ret;
42969  __ret = (int64_t) __builtin_neon_vcltzd_s64(__p0);
42970  return __ret;
42971}
42972#else
42973__ai int64_t vcltzd_s64(int64_t __p0) {
42974  int64_t __ret;
42975  __ret = (int64_t) __builtin_neon_vcltzd_s64(__p0);
42976  return __ret;
42977}
42978#endif
42979
42980#ifdef __LITTLE_ENDIAN__
42981__ai uint64_t vcltzd_f64(float64_t __p0) {
42982  uint64_t __ret;
42983  __ret = (uint64_t) __builtin_neon_vcltzd_f64(__p0);
42984  return __ret;
42985}
42986#else
42987__ai uint64_t vcltzd_f64(float64_t __p0) {
42988  uint64_t __ret;
42989  __ret = (uint64_t) __builtin_neon_vcltzd_f64(__p0);
42990  return __ret;
42991}
42992#endif
42993
42994#ifdef __LITTLE_ENDIAN__
42995__ai uint32_t vcltzs_f32(float32_t __p0) {
42996  uint32_t __ret;
42997  __ret = (uint32_t) __builtin_neon_vcltzs_f32(__p0);
42998  return __ret;
42999}
43000#else
43001__ai uint32_t vcltzs_f32(float32_t __p0) {
43002  uint32_t __ret;
43003  __ret = (uint32_t) __builtin_neon_vcltzs_f32(__p0);
43004  return __ret;
43005}
43006#endif
43007
43008#ifdef __LITTLE_ENDIAN__
43009__ai poly64x2_t vcombine_p64(poly64x1_t __p0, poly64x1_t __p1) {
43010  poly64x2_t __ret;
43011  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
43012  return __ret;
43013}
43014#else
43015__ai poly64x2_t vcombine_p64(poly64x1_t __p0, poly64x1_t __p1) {
43016  poly64x2_t __ret;
43017  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
43018  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43019  return __ret;
43020}
43021#endif
43022
43023#ifdef __LITTLE_ENDIAN__
43024__ai float64x2_t vcombine_f64(float64x1_t __p0, float64x1_t __p1) {
43025  float64x2_t __ret;
43026  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
43027  return __ret;
43028}
43029#else
43030__ai float64x2_t vcombine_f64(float64x1_t __p0, float64x1_t __p1) {
43031  float64x2_t __ret;
43032  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
43033  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43034  return __ret;
43035}
43036#endif
43037
43038#ifdef __LITTLE_ENDIAN__
43039#define vcopyq_lane_p8(__p0_0, __p1_0, __p2_0, __p3_0) __extension__ ({ \
43040  poly8x16_t __s0_0 = __p0_0; \
43041  poly8x8_t __s2_0 = __p2_0; \
43042  poly8x16_t __ret_0; \
43043  __ret_0 = vsetq_lane_p8(vget_lane_p8(__s2_0, __p3_0), __s0_0, __p1_0); \
43044  __ret_0; \
43045})
43046#else
43047#define vcopyq_lane_p8(__p0_1, __p1_1, __p2_1, __p3_1) __extension__ ({ \
43048  poly8x16_t __s0_1 = __p0_1; \
43049  poly8x8_t __s2_1 = __p2_1; \
43050  poly8x16_t __rev0_1;  __rev0_1 = __builtin_shufflevector(__s0_1, __s0_1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43051  poly8x8_t __rev2_1;  __rev2_1 = __builtin_shufflevector(__s2_1, __s2_1, 7, 6, 5, 4, 3, 2, 1, 0); \
43052  poly8x16_t __ret_1; \
43053  __ret_1 = __noswap_vsetq_lane_p8(__noswap_vget_lane_p8(__rev2_1, __p3_1), __rev0_1, __p1_1); \
43054  __ret_1 = __builtin_shufflevector(__ret_1, __ret_1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43055  __ret_1; \
43056})
43057#endif
43058
43059#ifdef __LITTLE_ENDIAN__
43060#define vcopyq_lane_p16(__p0_2, __p1_2, __p2_2, __p3_2) __extension__ ({ \
43061  poly16x8_t __s0_2 = __p0_2; \
43062  poly16x4_t __s2_2 = __p2_2; \
43063  poly16x8_t __ret_2; \
43064  __ret_2 = vsetq_lane_p16(vget_lane_p16(__s2_2, __p3_2), __s0_2, __p1_2); \
43065  __ret_2; \
43066})
43067#else
43068#define vcopyq_lane_p16(__p0_3, __p1_3, __p2_3, __p3_3) __extension__ ({ \
43069  poly16x8_t __s0_3 = __p0_3; \
43070  poly16x4_t __s2_3 = __p2_3; \
43071  poly16x8_t __rev0_3;  __rev0_3 = __builtin_shufflevector(__s0_3, __s0_3, 7, 6, 5, 4, 3, 2, 1, 0); \
43072  poly16x4_t __rev2_3;  __rev2_3 = __builtin_shufflevector(__s2_3, __s2_3, 3, 2, 1, 0); \
43073  poly16x8_t __ret_3; \
43074  __ret_3 = __noswap_vsetq_lane_p16(__noswap_vget_lane_p16(__rev2_3, __p3_3), __rev0_3, __p1_3); \
43075  __ret_3 = __builtin_shufflevector(__ret_3, __ret_3, 7, 6, 5, 4, 3, 2, 1, 0); \
43076  __ret_3; \
43077})
43078#endif
43079
43080#ifdef __LITTLE_ENDIAN__
43081#define vcopyq_lane_u8(__p0_4, __p1_4, __p2_4, __p3_4) __extension__ ({ \
43082  uint8x16_t __s0_4 = __p0_4; \
43083  uint8x8_t __s2_4 = __p2_4; \
43084  uint8x16_t __ret_4; \
43085  __ret_4 = vsetq_lane_u8(vget_lane_u8(__s2_4, __p3_4), __s0_4, __p1_4); \
43086  __ret_4; \
43087})
43088#else
43089#define vcopyq_lane_u8(__p0_5, __p1_5, __p2_5, __p3_5) __extension__ ({ \
43090  uint8x16_t __s0_5 = __p0_5; \
43091  uint8x8_t __s2_5 = __p2_5; \
43092  uint8x16_t __rev0_5;  __rev0_5 = __builtin_shufflevector(__s0_5, __s0_5, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43093  uint8x8_t __rev2_5;  __rev2_5 = __builtin_shufflevector(__s2_5, __s2_5, 7, 6, 5, 4, 3, 2, 1, 0); \
43094  uint8x16_t __ret_5; \
43095  __ret_5 = __noswap_vsetq_lane_u8(__noswap_vget_lane_u8(__rev2_5, __p3_5), __rev0_5, __p1_5); \
43096  __ret_5 = __builtin_shufflevector(__ret_5, __ret_5, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43097  __ret_5; \
43098})
43099#endif
43100
43101#ifdef __LITTLE_ENDIAN__
43102#define vcopyq_lane_u32(__p0_6, __p1_6, __p2_6, __p3_6) __extension__ ({ \
43103  uint32x4_t __s0_6 = __p0_6; \
43104  uint32x2_t __s2_6 = __p2_6; \
43105  uint32x4_t __ret_6; \
43106  __ret_6 = vsetq_lane_u32(vget_lane_u32(__s2_6, __p3_6), __s0_6, __p1_6); \
43107  __ret_6; \
43108})
43109#else
43110#define vcopyq_lane_u32(__p0_7, __p1_7, __p2_7, __p3_7) __extension__ ({ \
43111  uint32x4_t __s0_7 = __p0_7; \
43112  uint32x2_t __s2_7 = __p2_7; \
43113  uint32x4_t __rev0_7;  __rev0_7 = __builtin_shufflevector(__s0_7, __s0_7, 3, 2, 1, 0); \
43114  uint32x2_t __rev2_7;  __rev2_7 = __builtin_shufflevector(__s2_7, __s2_7, 1, 0); \
43115  uint32x4_t __ret_7; \
43116  __ret_7 = __noswap_vsetq_lane_u32(__noswap_vget_lane_u32(__rev2_7, __p3_7), __rev0_7, __p1_7); \
43117  __ret_7 = __builtin_shufflevector(__ret_7, __ret_7, 3, 2, 1, 0); \
43118  __ret_7; \
43119})
43120#endif
43121
43122#ifdef __LITTLE_ENDIAN__
43123#define vcopyq_lane_u64(__p0_8, __p1_8, __p2_8, __p3_8) __extension__ ({ \
43124  uint64x2_t __s0_8 = __p0_8; \
43125  uint64x1_t __s2_8 = __p2_8; \
43126  uint64x2_t __ret_8; \
43127  __ret_8 = vsetq_lane_u64(vget_lane_u64(__s2_8, __p3_8), __s0_8, __p1_8); \
43128  __ret_8; \
43129})
43130#else
43131#define vcopyq_lane_u64(__p0_9, __p1_9, __p2_9, __p3_9) __extension__ ({ \
43132  uint64x2_t __s0_9 = __p0_9; \
43133  uint64x1_t __s2_9 = __p2_9; \
43134  uint64x2_t __rev0_9;  __rev0_9 = __builtin_shufflevector(__s0_9, __s0_9, 1, 0); \
43135  uint64x2_t __ret_9; \
43136  __ret_9 = __noswap_vsetq_lane_u64(__noswap_vget_lane_u64(__s2_9, __p3_9), __rev0_9, __p1_9); \
43137  __ret_9 = __builtin_shufflevector(__ret_9, __ret_9, 1, 0); \
43138  __ret_9; \
43139})
43140#endif
43141
43142#ifdef __LITTLE_ENDIAN__
43143#define vcopyq_lane_u16(__p0_10, __p1_10, __p2_10, __p3_10) __extension__ ({ \
43144  uint16x8_t __s0_10 = __p0_10; \
43145  uint16x4_t __s2_10 = __p2_10; \
43146  uint16x8_t __ret_10; \
43147  __ret_10 = vsetq_lane_u16(vget_lane_u16(__s2_10, __p3_10), __s0_10, __p1_10); \
43148  __ret_10; \
43149})
43150#else
43151#define vcopyq_lane_u16(__p0_11, __p1_11, __p2_11, __p3_11) __extension__ ({ \
43152  uint16x8_t __s0_11 = __p0_11; \
43153  uint16x4_t __s2_11 = __p2_11; \
43154  uint16x8_t __rev0_11;  __rev0_11 = __builtin_shufflevector(__s0_11, __s0_11, 7, 6, 5, 4, 3, 2, 1, 0); \
43155  uint16x4_t __rev2_11;  __rev2_11 = __builtin_shufflevector(__s2_11, __s2_11, 3, 2, 1, 0); \
43156  uint16x8_t __ret_11; \
43157  __ret_11 = __noswap_vsetq_lane_u16(__noswap_vget_lane_u16(__rev2_11, __p3_11), __rev0_11, __p1_11); \
43158  __ret_11 = __builtin_shufflevector(__ret_11, __ret_11, 7, 6, 5, 4, 3, 2, 1, 0); \
43159  __ret_11; \
43160})
43161#endif
43162
43163#ifdef __LITTLE_ENDIAN__
43164#define vcopyq_lane_s8(__p0_12, __p1_12, __p2_12, __p3_12) __extension__ ({ \
43165  int8x16_t __s0_12 = __p0_12; \
43166  int8x8_t __s2_12 = __p2_12; \
43167  int8x16_t __ret_12; \
43168  __ret_12 = vsetq_lane_s8(vget_lane_s8(__s2_12, __p3_12), __s0_12, __p1_12); \
43169  __ret_12; \
43170})
43171#else
43172#define vcopyq_lane_s8(__p0_13, __p1_13, __p2_13, __p3_13) __extension__ ({ \
43173  int8x16_t __s0_13 = __p0_13; \
43174  int8x8_t __s2_13 = __p2_13; \
43175  int8x16_t __rev0_13;  __rev0_13 = __builtin_shufflevector(__s0_13, __s0_13, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43176  int8x8_t __rev2_13;  __rev2_13 = __builtin_shufflevector(__s2_13, __s2_13, 7, 6, 5, 4, 3, 2, 1, 0); \
43177  int8x16_t __ret_13; \
43178  __ret_13 = __noswap_vsetq_lane_s8(__noswap_vget_lane_s8(__rev2_13, __p3_13), __rev0_13, __p1_13); \
43179  __ret_13 = __builtin_shufflevector(__ret_13, __ret_13, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43180  __ret_13; \
43181})
43182#endif
43183
43184#ifdef __LITTLE_ENDIAN__
43185#define vcopyq_lane_f32(__p0_14, __p1_14, __p2_14, __p3_14) __extension__ ({ \
43186  float32x4_t __s0_14 = __p0_14; \
43187  float32x2_t __s2_14 = __p2_14; \
43188  float32x4_t __ret_14; \
43189  __ret_14 = vsetq_lane_f32(vget_lane_f32(__s2_14, __p3_14), __s0_14, __p1_14); \
43190  __ret_14; \
43191})
43192#else
43193#define vcopyq_lane_f32(__p0_15, __p1_15, __p2_15, __p3_15) __extension__ ({ \
43194  float32x4_t __s0_15 = __p0_15; \
43195  float32x2_t __s2_15 = __p2_15; \
43196  float32x4_t __rev0_15;  __rev0_15 = __builtin_shufflevector(__s0_15, __s0_15, 3, 2, 1, 0); \
43197  float32x2_t __rev2_15;  __rev2_15 = __builtin_shufflevector(__s2_15, __s2_15, 1, 0); \
43198  float32x4_t __ret_15; \
43199  __ret_15 = __noswap_vsetq_lane_f32(__noswap_vget_lane_f32(__rev2_15, __p3_15), __rev0_15, __p1_15); \
43200  __ret_15 = __builtin_shufflevector(__ret_15, __ret_15, 3, 2, 1, 0); \
43201  __ret_15; \
43202})
43203#endif
43204
43205#ifdef __LITTLE_ENDIAN__
43206#define vcopyq_lane_s32(__p0_16, __p1_16, __p2_16, __p3_16) __extension__ ({ \
43207  int32x4_t __s0_16 = __p0_16; \
43208  int32x2_t __s2_16 = __p2_16; \
43209  int32x4_t __ret_16; \
43210  __ret_16 = vsetq_lane_s32(vget_lane_s32(__s2_16, __p3_16), __s0_16, __p1_16); \
43211  __ret_16; \
43212})
43213#else
43214#define vcopyq_lane_s32(__p0_17, __p1_17, __p2_17, __p3_17) __extension__ ({ \
43215  int32x4_t __s0_17 = __p0_17; \
43216  int32x2_t __s2_17 = __p2_17; \
43217  int32x4_t __rev0_17;  __rev0_17 = __builtin_shufflevector(__s0_17, __s0_17, 3, 2, 1, 0); \
43218  int32x2_t __rev2_17;  __rev2_17 = __builtin_shufflevector(__s2_17, __s2_17, 1, 0); \
43219  int32x4_t __ret_17; \
43220  __ret_17 = __noswap_vsetq_lane_s32(__noswap_vget_lane_s32(__rev2_17, __p3_17), __rev0_17, __p1_17); \
43221  __ret_17 = __builtin_shufflevector(__ret_17, __ret_17, 3, 2, 1, 0); \
43222  __ret_17; \
43223})
43224#endif
43225
43226#ifdef __LITTLE_ENDIAN__
43227#define vcopyq_lane_s64(__p0_18, __p1_18, __p2_18, __p3_18) __extension__ ({ \
43228  int64x2_t __s0_18 = __p0_18; \
43229  int64x1_t __s2_18 = __p2_18; \
43230  int64x2_t __ret_18; \
43231  __ret_18 = vsetq_lane_s64(vget_lane_s64(__s2_18, __p3_18), __s0_18, __p1_18); \
43232  __ret_18; \
43233})
43234#else
43235#define vcopyq_lane_s64(__p0_19, __p1_19, __p2_19, __p3_19) __extension__ ({ \
43236  int64x2_t __s0_19 = __p0_19; \
43237  int64x1_t __s2_19 = __p2_19; \
43238  int64x2_t __rev0_19;  __rev0_19 = __builtin_shufflevector(__s0_19, __s0_19, 1, 0); \
43239  int64x2_t __ret_19; \
43240  __ret_19 = __noswap_vsetq_lane_s64(__noswap_vget_lane_s64(__s2_19, __p3_19), __rev0_19, __p1_19); \
43241  __ret_19 = __builtin_shufflevector(__ret_19, __ret_19, 1, 0); \
43242  __ret_19; \
43243})
43244#endif
43245
43246#ifdef __LITTLE_ENDIAN__
43247#define vcopyq_lane_s16(__p0_20, __p1_20, __p2_20, __p3_20) __extension__ ({ \
43248  int16x8_t __s0_20 = __p0_20; \
43249  int16x4_t __s2_20 = __p2_20; \
43250  int16x8_t __ret_20; \
43251  __ret_20 = vsetq_lane_s16(vget_lane_s16(__s2_20, __p3_20), __s0_20, __p1_20); \
43252  __ret_20; \
43253})
43254#else
43255#define vcopyq_lane_s16(__p0_21, __p1_21, __p2_21, __p3_21) __extension__ ({ \
43256  int16x8_t __s0_21 = __p0_21; \
43257  int16x4_t __s2_21 = __p2_21; \
43258  int16x8_t __rev0_21;  __rev0_21 = __builtin_shufflevector(__s0_21, __s0_21, 7, 6, 5, 4, 3, 2, 1, 0); \
43259  int16x4_t __rev2_21;  __rev2_21 = __builtin_shufflevector(__s2_21, __s2_21, 3, 2, 1, 0); \
43260  int16x8_t __ret_21; \
43261  __ret_21 = __noswap_vsetq_lane_s16(__noswap_vget_lane_s16(__rev2_21, __p3_21), __rev0_21, __p1_21); \
43262  __ret_21 = __builtin_shufflevector(__ret_21, __ret_21, 7, 6, 5, 4, 3, 2, 1, 0); \
43263  __ret_21; \
43264})
43265#endif
43266
43267#ifdef __LITTLE_ENDIAN__
43268#define vcopy_lane_p8(__p0_22, __p1_22, __p2_22, __p3_22) __extension__ ({ \
43269  poly8x8_t __s0_22 = __p0_22; \
43270  poly8x8_t __s2_22 = __p2_22; \
43271  poly8x8_t __ret_22; \
43272  __ret_22 = vset_lane_p8(vget_lane_p8(__s2_22, __p3_22), __s0_22, __p1_22); \
43273  __ret_22; \
43274})
43275#else
43276#define vcopy_lane_p8(__p0_23, __p1_23, __p2_23, __p3_23) __extension__ ({ \
43277  poly8x8_t __s0_23 = __p0_23; \
43278  poly8x8_t __s2_23 = __p2_23; \
43279  poly8x8_t __rev0_23;  __rev0_23 = __builtin_shufflevector(__s0_23, __s0_23, 7, 6, 5, 4, 3, 2, 1, 0); \
43280  poly8x8_t __rev2_23;  __rev2_23 = __builtin_shufflevector(__s2_23, __s2_23, 7, 6, 5, 4, 3, 2, 1, 0); \
43281  poly8x8_t __ret_23; \
43282  __ret_23 = __noswap_vset_lane_p8(__noswap_vget_lane_p8(__rev2_23, __p3_23), __rev0_23, __p1_23); \
43283  __ret_23 = __builtin_shufflevector(__ret_23, __ret_23, 7, 6, 5, 4, 3, 2, 1, 0); \
43284  __ret_23; \
43285})
43286#endif
43287
43288#ifdef __LITTLE_ENDIAN__
43289#define vcopy_lane_p16(__p0_24, __p1_24, __p2_24, __p3_24) __extension__ ({ \
43290  poly16x4_t __s0_24 = __p0_24; \
43291  poly16x4_t __s2_24 = __p2_24; \
43292  poly16x4_t __ret_24; \
43293  __ret_24 = vset_lane_p16(vget_lane_p16(__s2_24, __p3_24), __s0_24, __p1_24); \
43294  __ret_24; \
43295})
43296#else
43297#define vcopy_lane_p16(__p0_25, __p1_25, __p2_25, __p3_25) __extension__ ({ \
43298  poly16x4_t __s0_25 = __p0_25; \
43299  poly16x4_t __s2_25 = __p2_25; \
43300  poly16x4_t __rev0_25;  __rev0_25 = __builtin_shufflevector(__s0_25, __s0_25, 3, 2, 1, 0); \
43301  poly16x4_t __rev2_25;  __rev2_25 = __builtin_shufflevector(__s2_25, __s2_25, 3, 2, 1, 0); \
43302  poly16x4_t __ret_25; \
43303  __ret_25 = __noswap_vset_lane_p16(__noswap_vget_lane_p16(__rev2_25, __p3_25), __rev0_25, __p1_25); \
43304  __ret_25 = __builtin_shufflevector(__ret_25, __ret_25, 3, 2, 1, 0); \
43305  __ret_25; \
43306})
43307#endif
43308
43309#ifdef __LITTLE_ENDIAN__
43310#define vcopy_lane_u8(__p0_26, __p1_26, __p2_26, __p3_26) __extension__ ({ \
43311  uint8x8_t __s0_26 = __p0_26; \
43312  uint8x8_t __s2_26 = __p2_26; \
43313  uint8x8_t __ret_26; \
43314  __ret_26 = vset_lane_u8(vget_lane_u8(__s2_26, __p3_26), __s0_26, __p1_26); \
43315  __ret_26; \
43316})
43317#else
43318#define vcopy_lane_u8(__p0_27, __p1_27, __p2_27, __p3_27) __extension__ ({ \
43319  uint8x8_t __s0_27 = __p0_27; \
43320  uint8x8_t __s2_27 = __p2_27; \
43321  uint8x8_t __rev0_27;  __rev0_27 = __builtin_shufflevector(__s0_27, __s0_27, 7, 6, 5, 4, 3, 2, 1, 0); \
43322  uint8x8_t __rev2_27;  __rev2_27 = __builtin_shufflevector(__s2_27, __s2_27, 7, 6, 5, 4, 3, 2, 1, 0); \
43323  uint8x8_t __ret_27; \
43324  __ret_27 = __noswap_vset_lane_u8(__noswap_vget_lane_u8(__rev2_27, __p3_27), __rev0_27, __p1_27); \
43325  __ret_27 = __builtin_shufflevector(__ret_27, __ret_27, 7, 6, 5, 4, 3, 2, 1, 0); \
43326  __ret_27; \
43327})
43328#endif
43329
43330#ifdef __LITTLE_ENDIAN__
43331#define vcopy_lane_u32(__p0_28, __p1_28, __p2_28, __p3_28) __extension__ ({ \
43332  uint32x2_t __s0_28 = __p0_28; \
43333  uint32x2_t __s2_28 = __p2_28; \
43334  uint32x2_t __ret_28; \
43335  __ret_28 = vset_lane_u32(vget_lane_u32(__s2_28, __p3_28), __s0_28, __p1_28); \
43336  __ret_28; \
43337})
43338#else
43339#define vcopy_lane_u32(__p0_29, __p1_29, __p2_29, __p3_29) __extension__ ({ \
43340  uint32x2_t __s0_29 = __p0_29; \
43341  uint32x2_t __s2_29 = __p2_29; \
43342  uint32x2_t __rev0_29;  __rev0_29 = __builtin_shufflevector(__s0_29, __s0_29, 1, 0); \
43343  uint32x2_t __rev2_29;  __rev2_29 = __builtin_shufflevector(__s2_29, __s2_29, 1, 0); \
43344  uint32x2_t __ret_29; \
43345  __ret_29 = __noswap_vset_lane_u32(__noswap_vget_lane_u32(__rev2_29, __p3_29), __rev0_29, __p1_29); \
43346  __ret_29 = __builtin_shufflevector(__ret_29, __ret_29, 1, 0); \
43347  __ret_29; \
43348})
43349#endif
43350
43351#ifdef __LITTLE_ENDIAN__
43352#define vcopy_lane_u64(__p0_30, __p1_30, __p2_30, __p3_30) __extension__ ({ \
43353  uint64x1_t __s0_30 = __p0_30; \
43354  uint64x1_t __s2_30 = __p2_30; \
43355  uint64x1_t __ret_30; \
43356  __ret_30 = vset_lane_u64(vget_lane_u64(__s2_30, __p3_30), __s0_30, __p1_30); \
43357  __ret_30; \
43358})
43359#else
43360#define vcopy_lane_u64(__p0_31, __p1_31, __p2_31, __p3_31) __extension__ ({ \
43361  uint64x1_t __s0_31 = __p0_31; \
43362  uint64x1_t __s2_31 = __p2_31; \
43363  uint64x1_t __ret_31; \
43364  __ret_31 = __noswap_vset_lane_u64(__noswap_vget_lane_u64(__s2_31, __p3_31), __s0_31, __p1_31); \
43365  __ret_31; \
43366})
43367#endif
43368
43369#ifdef __LITTLE_ENDIAN__
43370#define vcopy_lane_u16(__p0_32, __p1_32, __p2_32, __p3_32) __extension__ ({ \
43371  uint16x4_t __s0_32 = __p0_32; \
43372  uint16x4_t __s2_32 = __p2_32; \
43373  uint16x4_t __ret_32; \
43374  __ret_32 = vset_lane_u16(vget_lane_u16(__s2_32, __p3_32), __s0_32, __p1_32); \
43375  __ret_32; \
43376})
43377#else
43378#define vcopy_lane_u16(__p0_33, __p1_33, __p2_33, __p3_33) __extension__ ({ \
43379  uint16x4_t __s0_33 = __p0_33; \
43380  uint16x4_t __s2_33 = __p2_33; \
43381  uint16x4_t __rev0_33;  __rev0_33 = __builtin_shufflevector(__s0_33, __s0_33, 3, 2, 1, 0); \
43382  uint16x4_t __rev2_33;  __rev2_33 = __builtin_shufflevector(__s2_33, __s2_33, 3, 2, 1, 0); \
43383  uint16x4_t __ret_33; \
43384  __ret_33 = __noswap_vset_lane_u16(__noswap_vget_lane_u16(__rev2_33, __p3_33), __rev0_33, __p1_33); \
43385  __ret_33 = __builtin_shufflevector(__ret_33, __ret_33, 3, 2, 1, 0); \
43386  __ret_33; \
43387})
43388#endif
43389
43390#ifdef __LITTLE_ENDIAN__
43391#define vcopy_lane_s8(__p0_34, __p1_34, __p2_34, __p3_34) __extension__ ({ \
43392  int8x8_t __s0_34 = __p0_34; \
43393  int8x8_t __s2_34 = __p2_34; \
43394  int8x8_t __ret_34; \
43395  __ret_34 = vset_lane_s8(vget_lane_s8(__s2_34, __p3_34), __s0_34, __p1_34); \
43396  __ret_34; \
43397})
43398#else
43399#define vcopy_lane_s8(__p0_35, __p1_35, __p2_35, __p3_35) __extension__ ({ \
43400  int8x8_t __s0_35 = __p0_35; \
43401  int8x8_t __s2_35 = __p2_35; \
43402  int8x8_t __rev0_35;  __rev0_35 = __builtin_shufflevector(__s0_35, __s0_35, 7, 6, 5, 4, 3, 2, 1, 0); \
43403  int8x8_t __rev2_35;  __rev2_35 = __builtin_shufflevector(__s2_35, __s2_35, 7, 6, 5, 4, 3, 2, 1, 0); \
43404  int8x8_t __ret_35; \
43405  __ret_35 = __noswap_vset_lane_s8(__noswap_vget_lane_s8(__rev2_35, __p3_35), __rev0_35, __p1_35); \
43406  __ret_35 = __builtin_shufflevector(__ret_35, __ret_35, 7, 6, 5, 4, 3, 2, 1, 0); \
43407  __ret_35; \
43408})
43409#endif
43410
43411#ifdef __LITTLE_ENDIAN__
43412#define vcopy_lane_f32(__p0_36, __p1_36, __p2_36, __p3_36) __extension__ ({ \
43413  float32x2_t __s0_36 = __p0_36; \
43414  float32x2_t __s2_36 = __p2_36; \
43415  float32x2_t __ret_36; \
43416  __ret_36 = vset_lane_f32(vget_lane_f32(__s2_36, __p3_36), __s0_36, __p1_36); \
43417  __ret_36; \
43418})
43419#else
43420#define vcopy_lane_f32(__p0_37, __p1_37, __p2_37, __p3_37) __extension__ ({ \
43421  float32x2_t __s0_37 = __p0_37; \
43422  float32x2_t __s2_37 = __p2_37; \
43423  float32x2_t __rev0_37;  __rev0_37 = __builtin_shufflevector(__s0_37, __s0_37, 1, 0); \
43424  float32x2_t __rev2_37;  __rev2_37 = __builtin_shufflevector(__s2_37, __s2_37, 1, 0); \
43425  float32x2_t __ret_37; \
43426  __ret_37 = __noswap_vset_lane_f32(__noswap_vget_lane_f32(__rev2_37, __p3_37), __rev0_37, __p1_37); \
43427  __ret_37 = __builtin_shufflevector(__ret_37, __ret_37, 1, 0); \
43428  __ret_37; \
43429})
43430#endif
43431
43432#ifdef __LITTLE_ENDIAN__
43433#define vcopy_lane_s32(__p0_38, __p1_38, __p2_38, __p3_38) __extension__ ({ \
43434  int32x2_t __s0_38 = __p0_38; \
43435  int32x2_t __s2_38 = __p2_38; \
43436  int32x2_t __ret_38; \
43437  __ret_38 = vset_lane_s32(vget_lane_s32(__s2_38, __p3_38), __s0_38, __p1_38); \
43438  __ret_38; \
43439})
43440#else
43441#define vcopy_lane_s32(__p0_39, __p1_39, __p2_39, __p3_39) __extension__ ({ \
43442  int32x2_t __s0_39 = __p0_39; \
43443  int32x2_t __s2_39 = __p2_39; \
43444  int32x2_t __rev0_39;  __rev0_39 = __builtin_shufflevector(__s0_39, __s0_39, 1, 0); \
43445  int32x2_t __rev2_39;  __rev2_39 = __builtin_shufflevector(__s2_39, __s2_39, 1, 0); \
43446  int32x2_t __ret_39; \
43447  __ret_39 = __noswap_vset_lane_s32(__noswap_vget_lane_s32(__rev2_39, __p3_39), __rev0_39, __p1_39); \
43448  __ret_39 = __builtin_shufflevector(__ret_39, __ret_39, 1, 0); \
43449  __ret_39; \
43450})
43451#endif
43452
43453#ifdef __LITTLE_ENDIAN__
43454#define vcopy_lane_s64(__p0_40, __p1_40, __p2_40, __p3_40) __extension__ ({ \
43455  int64x1_t __s0_40 = __p0_40; \
43456  int64x1_t __s2_40 = __p2_40; \
43457  int64x1_t __ret_40; \
43458  __ret_40 = vset_lane_s64(vget_lane_s64(__s2_40, __p3_40), __s0_40, __p1_40); \
43459  __ret_40; \
43460})
43461#else
43462#define vcopy_lane_s64(__p0_41, __p1_41, __p2_41, __p3_41) __extension__ ({ \
43463  int64x1_t __s0_41 = __p0_41; \
43464  int64x1_t __s2_41 = __p2_41; \
43465  int64x1_t __ret_41; \
43466  __ret_41 = __noswap_vset_lane_s64(__noswap_vget_lane_s64(__s2_41, __p3_41), __s0_41, __p1_41); \
43467  __ret_41; \
43468})
43469#endif
43470
43471#ifdef __LITTLE_ENDIAN__
43472#define vcopy_lane_s16(__p0_42, __p1_42, __p2_42, __p3_42) __extension__ ({ \
43473  int16x4_t __s0_42 = __p0_42; \
43474  int16x4_t __s2_42 = __p2_42; \
43475  int16x4_t __ret_42; \
43476  __ret_42 = vset_lane_s16(vget_lane_s16(__s2_42, __p3_42), __s0_42, __p1_42); \
43477  __ret_42; \
43478})
43479#else
43480#define vcopy_lane_s16(__p0_43, __p1_43, __p2_43, __p3_43) __extension__ ({ \
43481  int16x4_t __s0_43 = __p0_43; \
43482  int16x4_t __s2_43 = __p2_43; \
43483  int16x4_t __rev0_43;  __rev0_43 = __builtin_shufflevector(__s0_43, __s0_43, 3, 2, 1, 0); \
43484  int16x4_t __rev2_43;  __rev2_43 = __builtin_shufflevector(__s2_43, __s2_43, 3, 2, 1, 0); \
43485  int16x4_t __ret_43; \
43486  __ret_43 = __noswap_vset_lane_s16(__noswap_vget_lane_s16(__rev2_43, __p3_43), __rev0_43, __p1_43); \
43487  __ret_43 = __builtin_shufflevector(__ret_43, __ret_43, 3, 2, 1, 0); \
43488  __ret_43; \
43489})
43490#endif
43491
43492#ifdef __LITTLE_ENDIAN__
43493#define vcopyq_laneq_p8(__p0_44, __p1_44, __p2_44, __p3_44) __extension__ ({ \
43494  poly8x16_t __s0_44 = __p0_44; \
43495  poly8x16_t __s2_44 = __p2_44; \
43496  poly8x16_t __ret_44; \
43497  __ret_44 = vsetq_lane_p8(vgetq_lane_p8(__s2_44, __p3_44), __s0_44, __p1_44); \
43498  __ret_44; \
43499})
43500#else
43501#define vcopyq_laneq_p8(__p0_45, __p1_45, __p2_45, __p3_45) __extension__ ({ \
43502  poly8x16_t __s0_45 = __p0_45; \
43503  poly8x16_t __s2_45 = __p2_45; \
43504  poly8x16_t __rev0_45;  __rev0_45 = __builtin_shufflevector(__s0_45, __s0_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43505  poly8x16_t __rev2_45;  __rev2_45 = __builtin_shufflevector(__s2_45, __s2_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43506  poly8x16_t __ret_45; \
43507  __ret_45 = __noswap_vsetq_lane_p8(__noswap_vgetq_lane_p8(__rev2_45, __p3_45), __rev0_45, __p1_45); \
43508  __ret_45 = __builtin_shufflevector(__ret_45, __ret_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43509  __ret_45; \
43510})
43511#endif
43512
43513#ifdef __LITTLE_ENDIAN__
43514#define vcopyq_laneq_p16(__p0_46, __p1_46, __p2_46, __p3_46) __extension__ ({ \
43515  poly16x8_t __s0_46 = __p0_46; \
43516  poly16x8_t __s2_46 = __p2_46; \
43517  poly16x8_t __ret_46; \
43518  __ret_46 = vsetq_lane_p16(vgetq_lane_p16(__s2_46, __p3_46), __s0_46, __p1_46); \
43519  __ret_46; \
43520})
43521#else
43522#define vcopyq_laneq_p16(__p0_47, __p1_47, __p2_47, __p3_47) __extension__ ({ \
43523  poly16x8_t __s0_47 = __p0_47; \
43524  poly16x8_t __s2_47 = __p2_47; \
43525  poly16x8_t __rev0_47;  __rev0_47 = __builtin_shufflevector(__s0_47, __s0_47, 7, 6, 5, 4, 3, 2, 1, 0); \
43526  poly16x8_t __rev2_47;  __rev2_47 = __builtin_shufflevector(__s2_47, __s2_47, 7, 6, 5, 4, 3, 2, 1, 0); \
43527  poly16x8_t __ret_47; \
43528  __ret_47 = __noswap_vsetq_lane_p16(__noswap_vgetq_lane_p16(__rev2_47, __p3_47), __rev0_47, __p1_47); \
43529  __ret_47 = __builtin_shufflevector(__ret_47, __ret_47, 7, 6, 5, 4, 3, 2, 1, 0); \
43530  __ret_47; \
43531})
43532#endif
43533
43534#ifdef __LITTLE_ENDIAN__
43535#define vcopyq_laneq_u8(__p0_48, __p1_48, __p2_48, __p3_48) __extension__ ({ \
43536  uint8x16_t __s0_48 = __p0_48; \
43537  uint8x16_t __s2_48 = __p2_48; \
43538  uint8x16_t __ret_48; \
43539  __ret_48 = vsetq_lane_u8(vgetq_lane_u8(__s2_48, __p3_48), __s0_48, __p1_48); \
43540  __ret_48; \
43541})
43542#else
43543#define vcopyq_laneq_u8(__p0_49, __p1_49, __p2_49, __p3_49) __extension__ ({ \
43544  uint8x16_t __s0_49 = __p0_49; \
43545  uint8x16_t __s2_49 = __p2_49; \
43546  uint8x16_t __rev0_49;  __rev0_49 = __builtin_shufflevector(__s0_49, __s0_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43547  uint8x16_t __rev2_49;  __rev2_49 = __builtin_shufflevector(__s2_49, __s2_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43548  uint8x16_t __ret_49; \
43549  __ret_49 = __noswap_vsetq_lane_u8(__noswap_vgetq_lane_u8(__rev2_49, __p3_49), __rev0_49, __p1_49); \
43550  __ret_49 = __builtin_shufflevector(__ret_49, __ret_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43551  __ret_49; \
43552})
43553#endif
43554
43555#ifdef __LITTLE_ENDIAN__
43556#define vcopyq_laneq_u32(__p0_50, __p1_50, __p2_50, __p3_50) __extension__ ({ \
43557  uint32x4_t __s0_50 = __p0_50; \
43558  uint32x4_t __s2_50 = __p2_50; \
43559  uint32x4_t __ret_50; \
43560  __ret_50 = vsetq_lane_u32(vgetq_lane_u32(__s2_50, __p3_50), __s0_50, __p1_50); \
43561  __ret_50; \
43562})
43563#else
43564#define vcopyq_laneq_u32(__p0_51, __p1_51, __p2_51, __p3_51) __extension__ ({ \
43565  uint32x4_t __s0_51 = __p0_51; \
43566  uint32x4_t __s2_51 = __p2_51; \
43567  uint32x4_t __rev0_51;  __rev0_51 = __builtin_shufflevector(__s0_51, __s0_51, 3, 2, 1, 0); \
43568  uint32x4_t __rev2_51;  __rev2_51 = __builtin_shufflevector(__s2_51, __s2_51, 3, 2, 1, 0); \
43569  uint32x4_t __ret_51; \
43570  __ret_51 = __noswap_vsetq_lane_u32(__noswap_vgetq_lane_u32(__rev2_51, __p3_51), __rev0_51, __p1_51); \
43571  __ret_51 = __builtin_shufflevector(__ret_51, __ret_51, 3, 2, 1, 0); \
43572  __ret_51; \
43573})
43574#endif
43575
43576#ifdef __LITTLE_ENDIAN__
43577#define vcopyq_laneq_u64(__p0_52, __p1_52, __p2_52, __p3_52) __extension__ ({ \
43578  uint64x2_t __s0_52 = __p0_52; \
43579  uint64x2_t __s2_52 = __p2_52; \
43580  uint64x2_t __ret_52; \
43581  __ret_52 = vsetq_lane_u64(vgetq_lane_u64(__s2_52, __p3_52), __s0_52, __p1_52); \
43582  __ret_52; \
43583})
43584#else
43585#define vcopyq_laneq_u64(__p0_53, __p1_53, __p2_53, __p3_53) __extension__ ({ \
43586  uint64x2_t __s0_53 = __p0_53; \
43587  uint64x2_t __s2_53 = __p2_53; \
43588  uint64x2_t __rev0_53;  __rev0_53 = __builtin_shufflevector(__s0_53, __s0_53, 1, 0); \
43589  uint64x2_t __rev2_53;  __rev2_53 = __builtin_shufflevector(__s2_53, __s2_53, 1, 0); \
43590  uint64x2_t __ret_53; \
43591  __ret_53 = __noswap_vsetq_lane_u64(__noswap_vgetq_lane_u64(__rev2_53, __p3_53), __rev0_53, __p1_53); \
43592  __ret_53 = __builtin_shufflevector(__ret_53, __ret_53, 1, 0); \
43593  __ret_53; \
43594})
43595#endif
43596
43597#ifdef __LITTLE_ENDIAN__
43598#define vcopyq_laneq_u16(__p0_54, __p1_54, __p2_54, __p3_54) __extension__ ({ \
43599  uint16x8_t __s0_54 = __p0_54; \
43600  uint16x8_t __s2_54 = __p2_54; \
43601  uint16x8_t __ret_54; \
43602  __ret_54 = vsetq_lane_u16(vgetq_lane_u16(__s2_54, __p3_54), __s0_54, __p1_54); \
43603  __ret_54; \
43604})
43605#else
43606#define vcopyq_laneq_u16(__p0_55, __p1_55, __p2_55, __p3_55) __extension__ ({ \
43607  uint16x8_t __s0_55 = __p0_55; \
43608  uint16x8_t __s2_55 = __p2_55; \
43609  uint16x8_t __rev0_55;  __rev0_55 = __builtin_shufflevector(__s0_55, __s0_55, 7, 6, 5, 4, 3, 2, 1, 0); \
43610  uint16x8_t __rev2_55;  __rev2_55 = __builtin_shufflevector(__s2_55, __s2_55, 7, 6, 5, 4, 3, 2, 1, 0); \
43611  uint16x8_t __ret_55; \
43612  __ret_55 = __noswap_vsetq_lane_u16(__noswap_vgetq_lane_u16(__rev2_55, __p3_55), __rev0_55, __p1_55); \
43613  __ret_55 = __builtin_shufflevector(__ret_55, __ret_55, 7, 6, 5, 4, 3, 2, 1, 0); \
43614  __ret_55; \
43615})
43616#endif
43617
43618#ifdef __LITTLE_ENDIAN__
43619#define vcopyq_laneq_s8(__p0_56, __p1_56, __p2_56, __p3_56) __extension__ ({ \
43620  int8x16_t __s0_56 = __p0_56; \
43621  int8x16_t __s2_56 = __p2_56; \
43622  int8x16_t __ret_56; \
43623  __ret_56 = vsetq_lane_s8(vgetq_lane_s8(__s2_56, __p3_56), __s0_56, __p1_56); \
43624  __ret_56; \
43625})
43626#else
43627#define vcopyq_laneq_s8(__p0_57, __p1_57, __p2_57, __p3_57) __extension__ ({ \
43628  int8x16_t __s0_57 = __p0_57; \
43629  int8x16_t __s2_57 = __p2_57; \
43630  int8x16_t __rev0_57;  __rev0_57 = __builtin_shufflevector(__s0_57, __s0_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43631  int8x16_t __rev2_57;  __rev2_57 = __builtin_shufflevector(__s2_57, __s2_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43632  int8x16_t __ret_57; \
43633  __ret_57 = __noswap_vsetq_lane_s8(__noswap_vgetq_lane_s8(__rev2_57, __p3_57), __rev0_57, __p1_57); \
43634  __ret_57 = __builtin_shufflevector(__ret_57, __ret_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43635  __ret_57; \
43636})
43637#endif
43638
43639#ifdef __LITTLE_ENDIAN__
43640#define vcopyq_laneq_f32(__p0_58, __p1_58, __p2_58, __p3_58) __extension__ ({ \
43641  float32x4_t __s0_58 = __p0_58; \
43642  float32x4_t __s2_58 = __p2_58; \
43643  float32x4_t __ret_58; \
43644  __ret_58 = vsetq_lane_f32(vgetq_lane_f32(__s2_58, __p3_58), __s0_58, __p1_58); \
43645  __ret_58; \
43646})
43647#else
43648#define vcopyq_laneq_f32(__p0_59, __p1_59, __p2_59, __p3_59) __extension__ ({ \
43649  float32x4_t __s0_59 = __p0_59; \
43650  float32x4_t __s2_59 = __p2_59; \
43651  float32x4_t __rev0_59;  __rev0_59 = __builtin_shufflevector(__s0_59, __s0_59, 3, 2, 1, 0); \
43652  float32x4_t __rev2_59;  __rev2_59 = __builtin_shufflevector(__s2_59, __s2_59, 3, 2, 1, 0); \
43653  float32x4_t __ret_59; \
43654  __ret_59 = __noswap_vsetq_lane_f32(__noswap_vgetq_lane_f32(__rev2_59, __p3_59), __rev0_59, __p1_59); \
43655  __ret_59 = __builtin_shufflevector(__ret_59, __ret_59, 3, 2, 1, 0); \
43656  __ret_59; \
43657})
43658#endif
43659
43660#ifdef __LITTLE_ENDIAN__
43661#define vcopyq_laneq_s32(__p0_60, __p1_60, __p2_60, __p3_60) __extension__ ({ \
43662  int32x4_t __s0_60 = __p0_60; \
43663  int32x4_t __s2_60 = __p2_60; \
43664  int32x4_t __ret_60; \
43665  __ret_60 = vsetq_lane_s32(vgetq_lane_s32(__s2_60, __p3_60), __s0_60, __p1_60); \
43666  __ret_60; \
43667})
43668#else
43669#define vcopyq_laneq_s32(__p0_61, __p1_61, __p2_61, __p3_61) __extension__ ({ \
43670  int32x4_t __s0_61 = __p0_61; \
43671  int32x4_t __s2_61 = __p2_61; \
43672  int32x4_t __rev0_61;  __rev0_61 = __builtin_shufflevector(__s0_61, __s0_61, 3, 2, 1, 0); \
43673  int32x4_t __rev2_61;  __rev2_61 = __builtin_shufflevector(__s2_61, __s2_61, 3, 2, 1, 0); \
43674  int32x4_t __ret_61; \
43675  __ret_61 = __noswap_vsetq_lane_s32(__noswap_vgetq_lane_s32(__rev2_61, __p3_61), __rev0_61, __p1_61); \
43676  __ret_61 = __builtin_shufflevector(__ret_61, __ret_61, 3, 2, 1, 0); \
43677  __ret_61; \
43678})
43679#endif
43680
43681#ifdef __LITTLE_ENDIAN__
43682#define vcopyq_laneq_s64(__p0_62, __p1_62, __p2_62, __p3_62) __extension__ ({ \
43683  int64x2_t __s0_62 = __p0_62; \
43684  int64x2_t __s2_62 = __p2_62; \
43685  int64x2_t __ret_62; \
43686  __ret_62 = vsetq_lane_s64(vgetq_lane_s64(__s2_62, __p3_62), __s0_62, __p1_62); \
43687  __ret_62; \
43688})
43689#else
43690#define vcopyq_laneq_s64(__p0_63, __p1_63, __p2_63, __p3_63) __extension__ ({ \
43691  int64x2_t __s0_63 = __p0_63; \
43692  int64x2_t __s2_63 = __p2_63; \
43693  int64x2_t __rev0_63;  __rev0_63 = __builtin_shufflevector(__s0_63, __s0_63, 1, 0); \
43694  int64x2_t __rev2_63;  __rev2_63 = __builtin_shufflevector(__s2_63, __s2_63, 1, 0); \
43695  int64x2_t __ret_63; \
43696  __ret_63 = __noswap_vsetq_lane_s64(__noswap_vgetq_lane_s64(__rev2_63, __p3_63), __rev0_63, __p1_63); \
43697  __ret_63 = __builtin_shufflevector(__ret_63, __ret_63, 1, 0); \
43698  __ret_63; \
43699})
43700#endif
43701
43702#ifdef __LITTLE_ENDIAN__
43703#define vcopyq_laneq_s16(__p0_64, __p1_64, __p2_64, __p3_64) __extension__ ({ \
43704  int16x8_t __s0_64 = __p0_64; \
43705  int16x8_t __s2_64 = __p2_64; \
43706  int16x8_t __ret_64; \
43707  __ret_64 = vsetq_lane_s16(vgetq_lane_s16(__s2_64, __p3_64), __s0_64, __p1_64); \
43708  __ret_64; \
43709})
43710#else
43711#define vcopyq_laneq_s16(__p0_65, __p1_65, __p2_65, __p3_65) __extension__ ({ \
43712  int16x8_t __s0_65 = __p0_65; \
43713  int16x8_t __s2_65 = __p2_65; \
43714  int16x8_t __rev0_65;  __rev0_65 = __builtin_shufflevector(__s0_65, __s0_65, 7, 6, 5, 4, 3, 2, 1, 0); \
43715  int16x8_t __rev2_65;  __rev2_65 = __builtin_shufflevector(__s2_65, __s2_65, 7, 6, 5, 4, 3, 2, 1, 0); \
43716  int16x8_t __ret_65; \
43717  __ret_65 = __noswap_vsetq_lane_s16(__noswap_vgetq_lane_s16(__rev2_65, __p3_65), __rev0_65, __p1_65); \
43718  __ret_65 = __builtin_shufflevector(__ret_65, __ret_65, 7, 6, 5, 4, 3, 2, 1, 0); \
43719  __ret_65; \
43720})
43721#endif
43722
43723#ifdef __LITTLE_ENDIAN__
43724#define vcopy_laneq_p8(__p0_66, __p1_66, __p2_66, __p3_66) __extension__ ({ \
43725  poly8x8_t __s0_66 = __p0_66; \
43726  poly8x16_t __s2_66 = __p2_66; \
43727  poly8x8_t __ret_66; \
43728  __ret_66 = vset_lane_p8(vgetq_lane_p8(__s2_66, __p3_66), __s0_66, __p1_66); \
43729  __ret_66; \
43730})
43731#else
43732#define vcopy_laneq_p8(__p0_67, __p1_67, __p2_67, __p3_67) __extension__ ({ \
43733  poly8x8_t __s0_67 = __p0_67; \
43734  poly8x16_t __s2_67 = __p2_67; \
43735  poly8x8_t __rev0_67;  __rev0_67 = __builtin_shufflevector(__s0_67, __s0_67, 7, 6, 5, 4, 3, 2, 1, 0); \
43736  poly8x16_t __rev2_67;  __rev2_67 = __builtin_shufflevector(__s2_67, __s2_67, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43737  poly8x8_t __ret_67; \
43738  __ret_67 = __noswap_vset_lane_p8(__noswap_vgetq_lane_p8(__rev2_67, __p3_67), __rev0_67, __p1_67); \
43739  __ret_67 = __builtin_shufflevector(__ret_67, __ret_67, 7, 6, 5, 4, 3, 2, 1, 0); \
43740  __ret_67; \
43741})
43742#endif
43743
43744#ifdef __LITTLE_ENDIAN__
43745#define vcopy_laneq_p16(__p0_68, __p1_68, __p2_68, __p3_68) __extension__ ({ \
43746  poly16x4_t __s0_68 = __p0_68; \
43747  poly16x8_t __s2_68 = __p2_68; \
43748  poly16x4_t __ret_68; \
43749  __ret_68 = vset_lane_p16(vgetq_lane_p16(__s2_68, __p3_68), __s0_68, __p1_68); \
43750  __ret_68; \
43751})
43752#else
43753#define vcopy_laneq_p16(__p0_69, __p1_69, __p2_69, __p3_69) __extension__ ({ \
43754  poly16x4_t __s0_69 = __p0_69; \
43755  poly16x8_t __s2_69 = __p2_69; \
43756  poly16x4_t __rev0_69;  __rev0_69 = __builtin_shufflevector(__s0_69, __s0_69, 3, 2, 1, 0); \
43757  poly16x8_t __rev2_69;  __rev2_69 = __builtin_shufflevector(__s2_69, __s2_69, 7, 6, 5, 4, 3, 2, 1, 0); \
43758  poly16x4_t __ret_69; \
43759  __ret_69 = __noswap_vset_lane_p16(__noswap_vgetq_lane_p16(__rev2_69, __p3_69), __rev0_69, __p1_69); \
43760  __ret_69 = __builtin_shufflevector(__ret_69, __ret_69, 3, 2, 1, 0); \
43761  __ret_69; \
43762})
43763#endif
43764
43765#ifdef __LITTLE_ENDIAN__
43766#define vcopy_laneq_u8(__p0_70, __p1_70, __p2_70, __p3_70) __extension__ ({ \
43767  uint8x8_t __s0_70 = __p0_70; \
43768  uint8x16_t __s2_70 = __p2_70; \
43769  uint8x8_t __ret_70; \
43770  __ret_70 = vset_lane_u8(vgetq_lane_u8(__s2_70, __p3_70), __s0_70, __p1_70); \
43771  __ret_70; \
43772})
43773#else
43774#define vcopy_laneq_u8(__p0_71, __p1_71, __p2_71, __p3_71) __extension__ ({ \
43775  uint8x8_t __s0_71 = __p0_71; \
43776  uint8x16_t __s2_71 = __p2_71; \
43777  uint8x8_t __rev0_71;  __rev0_71 = __builtin_shufflevector(__s0_71, __s0_71, 7, 6, 5, 4, 3, 2, 1, 0); \
43778  uint8x16_t __rev2_71;  __rev2_71 = __builtin_shufflevector(__s2_71, __s2_71, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43779  uint8x8_t __ret_71; \
43780  __ret_71 = __noswap_vset_lane_u8(__noswap_vgetq_lane_u8(__rev2_71, __p3_71), __rev0_71, __p1_71); \
43781  __ret_71 = __builtin_shufflevector(__ret_71, __ret_71, 7, 6, 5, 4, 3, 2, 1, 0); \
43782  __ret_71; \
43783})
43784#endif
43785
43786#ifdef __LITTLE_ENDIAN__
43787#define vcopy_laneq_u32(__p0_72, __p1_72, __p2_72, __p3_72) __extension__ ({ \
43788  uint32x2_t __s0_72 = __p0_72; \
43789  uint32x4_t __s2_72 = __p2_72; \
43790  uint32x2_t __ret_72; \
43791  __ret_72 = vset_lane_u32(vgetq_lane_u32(__s2_72, __p3_72), __s0_72, __p1_72); \
43792  __ret_72; \
43793})
43794#else
43795#define vcopy_laneq_u32(__p0_73, __p1_73, __p2_73, __p3_73) __extension__ ({ \
43796  uint32x2_t __s0_73 = __p0_73; \
43797  uint32x4_t __s2_73 = __p2_73; \
43798  uint32x2_t __rev0_73;  __rev0_73 = __builtin_shufflevector(__s0_73, __s0_73, 1, 0); \
43799  uint32x4_t __rev2_73;  __rev2_73 = __builtin_shufflevector(__s2_73, __s2_73, 3, 2, 1, 0); \
43800  uint32x2_t __ret_73; \
43801  __ret_73 = __noswap_vset_lane_u32(__noswap_vgetq_lane_u32(__rev2_73, __p3_73), __rev0_73, __p1_73); \
43802  __ret_73 = __builtin_shufflevector(__ret_73, __ret_73, 1, 0); \
43803  __ret_73; \
43804})
43805#endif
43806
43807#ifdef __LITTLE_ENDIAN__
43808#define vcopy_laneq_u64(__p0_74, __p1_74, __p2_74, __p3_74) __extension__ ({ \
43809  uint64x1_t __s0_74 = __p0_74; \
43810  uint64x2_t __s2_74 = __p2_74; \
43811  uint64x1_t __ret_74; \
43812  __ret_74 = vset_lane_u64(vgetq_lane_u64(__s2_74, __p3_74), __s0_74, __p1_74); \
43813  __ret_74; \
43814})
43815#else
43816#define vcopy_laneq_u64(__p0_75, __p1_75, __p2_75, __p3_75) __extension__ ({ \
43817  uint64x1_t __s0_75 = __p0_75; \
43818  uint64x2_t __s2_75 = __p2_75; \
43819  uint64x2_t __rev2_75;  __rev2_75 = __builtin_shufflevector(__s2_75, __s2_75, 1, 0); \
43820  uint64x1_t __ret_75; \
43821  __ret_75 = __noswap_vset_lane_u64(__noswap_vgetq_lane_u64(__rev2_75, __p3_75), __s0_75, __p1_75); \
43822  __ret_75; \
43823})
43824#endif
43825
43826#ifdef __LITTLE_ENDIAN__
43827#define vcopy_laneq_u16(__p0_76, __p1_76, __p2_76, __p3_76) __extension__ ({ \
43828  uint16x4_t __s0_76 = __p0_76; \
43829  uint16x8_t __s2_76 = __p2_76; \
43830  uint16x4_t __ret_76; \
43831  __ret_76 = vset_lane_u16(vgetq_lane_u16(__s2_76, __p3_76), __s0_76, __p1_76); \
43832  __ret_76; \
43833})
43834#else
43835#define vcopy_laneq_u16(__p0_77, __p1_77, __p2_77, __p3_77) __extension__ ({ \
43836  uint16x4_t __s0_77 = __p0_77; \
43837  uint16x8_t __s2_77 = __p2_77; \
43838  uint16x4_t __rev0_77;  __rev0_77 = __builtin_shufflevector(__s0_77, __s0_77, 3, 2, 1, 0); \
43839  uint16x8_t __rev2_77;  __rev2_77 = __builtin_shufflevector(__s2_77, __s2_77, 7, 6, 5, 4, 3, 2, 1, 0); \
43840  uint16x4_t __ret_77; \
43841  __ret_77 = __noswap_vset_lane_u16(__noswap_vgetq_lane_u16(__rev2_77, __p3_77), __rev0_77, __p1_77); \
43842  __ret_77 = __builtin_shufflevector(__ret_77, __ret_77, 3, 2, 1, 0); \
43843  __ret_77; \
43844})
43845#endif
43846
43847#ifdef __LITTLE_ENDIAN__
43848#define vcopy_laneq_s8(__p0_78, __p1_78, __p2_78, __p3_78) __extension__ ({ \
43849  int8x8_t __s0_78 = __p0_78; \
43850  int8x16_t __s2_78 = __p2_78; \
43851  int8x8_t __ret_78; \
43852  __ret_78 = vset_lane_s8(vgetq_lane_s8(__s2_78, __p3_78), __s0_78, __p1_78); \
43853  __ret_78; \
43854})
43855#else
43856#define vcopy_laneq_s8(__p0_79, __p1_79, __p2_79, __p3_79) __extension__ ({ \
43857  int8x8_t __s0_79 = __p0_79; \
43858  int8x16_t __s2_79 = __p2_79; \
43859  int8x8_t __rev0_79;  __rev0_79 = __builtin_shufflevector(__s0_79, __s0_79, 7, 6, 5, 4, 3, 2, 1, 0); \
43860  int8x16_t __rev2_79;  __rev2_79 = __builtin_shufflevector(__s2_79, __s2_79, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
43861  int8x8_t __ret_79; \
43862  __ret_79 = __noswap_vset_lane_s8(__noswap_vgetq_lane_s8(__rev2_79, __p3_79), __rev0_79, __p1_79); \
43863  __ret_79 = __builtin_shufflevector(__ret_79, __ret_79, 7, 6, 5, 4, 3, 2, 1, 0); \
43864  __ret_79; \
43865})
43866#endif
43867
43868#ifdef __LITTLE_ENDIAN__
43869#define vcopy_laneq_f32(__p0_80, __p1_80, __p2_80, __p3_80) __extension__ ({ \
43870  float32x2_t __s0_80 = __p0_80; \
43871  float32x4_t __s2_80 = __p2_80; \
43872  float32x2_t __ret_80; \
43873  __ret_80 = vset_lane_f32(vgetq_lane_f32(__s2_80, __p3_80), __s0_80, __p1_80); \
43874  __ret_80; \
43875})
43876#else
43877#define vcopy_laneq_f32(__p0_81, __p1_81, __p2_81, __p3_81) __extension__ ({ \
43878  float32x2_t __s0_81 = __p0_81; \
43879  float32x4_t __s2_81 = __p2_81; \
43880  float32x2_t __rev0_81;  __rev0_81 = __builtin_shufflevector(__s0_81, __s0_81, 1, 0); \
43881  float32x4_t __rev2_81;  __rev2_81 = __builtin_shufflevector(__s2_81, __s2_81, 3, 2, 1, 0); \
43882  float32x2_t __ret_81; \
43883  __ret_81 = __noswap_vset_lane_f32(__noswap_vgetq_lane_f32(__rev2_81, __p3_81), __rev0_81, __p1_81); \
43884  __ret_81 = __builtin_shufflevector(__ret_81, __ret_81, 1, 0); \
43885  __ret_81; \
43886})
43887#endif
43888
43889#ifdef __LITTLE_ENDIAN__
43890#define vcopy_laneq_s32(__p0_82, __p1_82, __p2_82, __p3_82) __extension__ ({ \
43891  int32x2_t __s0_82 = __p0_82; \
43892  int32x4_t __s2_82 = __p2_82; \
43893  int32x2_t __ret_82; \
43894  __ret_82 = vset_lane_s32(vgetq_lane_s32(__s2_82, __p3_82), __s0_82, __p1_82); \
43895  __ret_82; \
43896})
43897#else
43898#define vcopy_laneq_s32(__p0_83, __p1_83, __p2_83, __p3_83) __extension__ ({ \
43899  int32x2_t __s0_83 = __p0_83; \
43900  int32x4_t __s2_83 = __p2_83; \
43901  int32x2_t __rev0_83;  __rev0_83 = __builtin_shufflevector(__s0_83, __s0_83, 1, 0); \
43902  int32x4_t __rev2_83;  __rev2_83 = __builtin_shufflevector(__s2_83, __s2_83, 3, 2, 1, 0); \
43903  int32x2_t __ret_83; \
43904  __ret_83 = __noswap_vset_lane_s32(__noswap_vgetq_lane_s32(__rev2_83, __p3_83), __rev0_83, __p1_83); \
43905  __ret_83 = __builtin_shufflevector(__ret_83, __ret_83, 1, 0); \
43906  __ret_83; \
43907})
43908#endif
43909
43910#ifdef __LITTLE_ENDIAN__
43911#define vcopy_laneq_s64(__p0_84, __p1_84, __p2_84, __p3_84) __extension__ ({ \
43912  int64x1_t __s0_84 = __p0_84; \
43913  int64x2_t __s2_84 = __p2_84; \
43914  int64x1_t __ret_84; \
43915  __ret_84 = vset_lane_s64(vgetq_lane_s64(__s2_84, __p3_84), __s0_84, __p1_84); \
43916  __ret_84; \
43917})
43918#else
43919#define vcopy_laneq_s64(__p0_85, __p1_85, __p2_85, __p3_85) __extension__ ({ \
43920  int64x1_t __s0_85 = __p0_85; \
43921  int64x2_t __s2_85 = __p2_85; \
43922  int64x2_t __rev2_85;  __rev2_85 = __builtin_shufflevector(__s2_85, __s2_85, 1, 0); \
43923  int64x1_t __ret_85; \
43924  __ret_85 = __noswap_vset_lane_s64(__noswap_vgetq_lane_s64(__rev2_85, __p3_85), __s0_85, __p1_85); \
43925  __ret_85; \
43926})
43927#endif
43928
43929#ifdef __LITTLE_ENDIAN__
43930#define vcopy_laneq_s16(__p0_86, __p1_86, __p2_86, __p3_86) __extension__ ({ \
43931  int16x4_t __s0_86 = __p0_86; \
43932  int16x8_t __s2_86 = __p2_86; \
43933  int16x4_t __ret_86; \
43934  __ret_86 = vset_lane_s16(vgetq_lane_s16(__s2_86, __p3_86), __s0_86, __p1_86); \
43935  __ret_86; \
43936})
43937#else
43938#define vcopy_laneq_s16(__p0_87, __p1_87, __p2_87, __p3_87) __extension__ ({ \
43939  int16x4_t __s0_87 = __p0_87; \
43940  int16x8_t __s2_87 = __p2_87; \
43941  int16x4_t __rev0_87;  __rev0_87 = __builtin_shufflevector(__s0_87, __s0_87, 3, 2, 1, 0); \
43942  int16x8_t __rev2_87;  __rev2_87 = __builtin_shufflevector(__s2_87, __s2_87, 7, 6, 5, 4, 3, 2, 1, 0); \
43943  int16x4_t __ret_87; \
43944  __ret_87 = __noswap_vset_lane_s16(__noswap_vgetq_lane_s16(__rev2_87, __p3_87), __rev0_87, __p1_87); \
43945  __ret_87 = __builtin_shufflevector(__ret_87, __ret_87, 3, 2, 1, 0); \
43946  __ret_87; \
43947})
43948#endif
43949
43950#ifdef __LITTLE_ENDIAN__
43951__ai poly64x1_t vcreate_p64(uint64_t __p0) {
43952  poly64x1_t __ret;
43953  __ret = (poly64x1_t)(__p0);
43954  return __ret;
43955}
43956#else
43957__ai poly64x1_t vcreate_p64(uint64_t __p0) {
43958  poly64x1_t __ret;
43959  __ret = (poly64x1_t)(__p0);
43960  return __ret;
43961}
43962#endif
43963
43964#ifdef __LITTLE_ENDIAN__
43965__ai float64x1_t vcreate_f64(uint64_t __p0) {
43966  float64x1_t __ret;
43967  __ret = (float64x1_t)(__p0);
43968  return __ret;
43969}
43970#else
43971__ai float64x1_t vcreate_f64(uint64_t __p0) {
43972  float64x1_t __ret;
43973  __ret = (float64x1_t)(__p0);
43974  return __ret;
43975}
43976#endif
43977
43978#ifdef __LITTLE_ENDIAN__
43979__ai float32_t vcvts_f32_s32(int32_t __p0) {
43980  float32_t __ret;
43981  __ret = (float32_t) __builtin_neon_vcvts_f32_s32(__p0);
43982  return __ret;
43983}
43984#else
43985__ai float32_t vcvts_f32_s32(int32_t __p0) {
43986  float32_t __ret;
43987  __ret = (float32_t) __builtin_neon_vcvts_f32_s32(__p0);
43988  return __ret;
43989}
43990#endif
43991
43992#ifdef __LITTLE_ENDIAN__
43993__ai float32_t vcvts_f32_u32(uint32_t __p0) {
43994  float32_t __ret;
43995  __ret = (float32_t) __builtin_neon_vcvts_f32_u32(__p0);
43996  return __ret;
43997}
43998#else
43999__ai float32_t vcvts_f32_u32(uint32_t __p0) {
44000  float32_t __ret;
44001  __ret = (float32_t) __builtin_neon_vcvts_f32_u32(__p0);
44002  return __ret;
44003}
44004#endif
44005
44006#ifdef __LITTLE_ENDIAN__
44007__ai float32x2_t vcvt_f32_f64(float64x2_t __p0) {
44008  float32x2_t __ret;
44009  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__p0, 9);
44010  return __ret;
44011}
44012#else
44013__ai float32x2_t vcvt_f32_f64(float64x2_t __p0) {
44014  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44015  float32x2_t __ret;
44016  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__rev0, 9);
44017  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44018  return __ret;
44019}
44020__ai float32x2_t __noswap_vcvt_f32_f64(float64x2_t __p0) {
44021  float32x2_t __ret;
44022  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__p0, 9);
44023  return __ret;
44024}
44025#endif
44026
44027#ifdef __LITTLE_ENDIAN__
44028__ai float64_t vcvtd_f64_s64(int64_t __p0) {
44029  float64_t __ret;
44030  __ret = (float64_t) __builtin_neon_vcvtd_f64_s64(__p0);
44031  return __ret;
44032}
44033#else
44034__ai float64_t vcvtd_f64_s64(int64_t __p0) {
44035  float64_t __ret;
44036  __ret = (float64_t) __builtin_neon_vcvtd_f64_s64(__p0);
44037  return __ret;
44038}
44039#endif
44040
44041#ifdef __LITTLE_ENDIAN__
44042__ai float64_t vcvtd_f64_u64(uint64_t __p0) {
44043  float64_t __ret;
44044  __ret = (float64_t) __builtin_neon_vcvtd_f64_u64(__p0);
44045  return __ret;
44046}
44047#else
44048__ai float64_t vcvtd_f64_u64(uint64_t __p0) {
44049  float64_t __ret;
44050  __ret = (float64_t) __builtin_neon_vcvtd_f64_u64(__p0);
44051  return __ret;
44052}
44053#endif
44054
44055#ifdef __LITTLE_ENDIAN__
44056__ai float64x2_t vcvtq_f64_u64(uint64x2_t __p0) {
44057  float64x2_t __ret;
44058  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__p0, 51);
44059  return __ret;
44060}
44061#else
44062__ai float64x2_t vcvtq_f64_u64(uint64x2_t __p0) {
44063  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44064  float64x2_t __ret;
44065  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__rev0, 51);
44066  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44067  return __ret;
44068}
44069#endif
44070
44071#ifdef __LITTLE_ENDIAN__
44072__ai float64x2_t vcvtq_f64_s64(int64x2_t __p0) {
44073  float64x2_t __ret;
44074  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__p0, 35);
44075  return __ret;
44076}
44077#else
44078__ai float64x2_t vcvtq_f64_s64(int64x2_t __p0) {
44079  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44080  float64x2_t __ret;
44081  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__rev0, 35);
44082  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44083  return __ret;
44084}
44085#endif
44086
44087#ifdef __LITTLE_ENDIAN__
44088__ai float64x1_t vcvt_f64_u64(uint64x1_t __p0) {
44089  float64x1_t __ret;
44090  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 19);
44091  return __ret;
44092}
44093#else
44094__ai float64x1_t vcvt_f64_u64(uint64x1_t __p0) {
44095  float64x1_t __ret;
44096  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 19);
44097  return __ret;
44098}
44099#endif
44100
44101#ifdef __LITTLE_ENDIAN__
44102__ai float64x1_t vcvt_f64_s64(int64x1_t __p0) {
44103  float64x1_t __ret;
44104  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 3);
44105  return __ret;
44106}
44107#else
44108__ai float64x1_t vcvt_f64_s64(int64x1_t __p0) {
44109  float64x1_t __ret;
44110  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 3);
44111  return __ret;
44112}
44113#endif
44114
44115#ifdef __LITTLE_ENDIAN__
44116__ai float64x2_t vcvt_f64_f32(float32x2_t __p0) {
44117  float64x2_t __ret;
44118  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__p0, 42);
44119  return __ret;
44120}
44121#else
44122__ai float64x2_t vcvt_f64_f32(float32x2_t __p0) {
44123  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44124  float64x2_t __ret;
44125  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__rev0, 42);
44126  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44127  return __ret;
44128}
44129__ai float64x2_t __noswap_vcvt_f64_f32(float32x2_t __p0) {
44130  float64x2_t __ret;
44131  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__p0, 42);
44132  return __ret;
44133}
44134#endif
44135
44136#ifdef __LITTLE_ENDIAN__
44137__ai float16x8_t vcvt_high_f16_f32(float16x4_t __p0, float32x4_t __p1) {
44138  float16x8_t __ret;
44139  __ret = vcombine_f16(__p0, vcvt_f16_f32(__p1));
44140  return __ret;
44141}
44142#else
44143__ai float16x8_t vcvt_high_f16_f32(float16x4_t __p0, float32x4_t __p1) {
44144  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44145  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
44146  float16x8_t __ret;
44147  __ret = __noswap_vcombine_f16(__rev0, __noswap_vcvt_f16_f32(__rev1));
44148  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44149  return __ret;
44150}
44151#endif
44152
44153#ifdef __LITTLE_ENDIAN__
44154__ai float32x4_t vcvt_high_f32_f16(float16x8_t __p0) {
44155  float32x4_t __ret;
44156  __ret = vcvt_f32_f16(vget_high_f16(__p0));
44157  return __ret;
44158}
44159#else
44160__ai float32x4_t vcvt_high_f32_f16(float16x8_t __p0) {
44161  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44162  float32x4_t __ret;
44163  __ret = __noswap_vcvt_f32_f16(__noswap_vget_high_f16(__rev0));
44164  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44165  return __ret;
44166}
44167#endif
44168
44169#ifdef __LITTLE_ENDIAN__
44170__ai float32x4_t vcvt_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
44171  float32x4_t __ret;
44172  __ret = vcombine_f32(__p0, vcvt_f32_f64(__p1));
44173  return __ret;
44174}
44175#else
44176__ai float32x4_t vcvt_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
44177  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44178  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
44179  float32x4_t __ret;
44180  __ret = __noswap_vcombine_f32(__rev0, __noswap_vcvt_f32_f64(__rev1));
44181  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44182  return __ret;
44183}
44184#endif
44185
44186#ifdef __LITTLE_ENDIAN__
44187__ai float64x2_t vcvt_high_f64_f32(float32x4_t __p0) {
44188  float64x2_t __ret;
44189  __ret = vcvt_f64_f32(vget_high_f32(__p0));
44190  return __ret;
44191}
44192#else
44193__ai float64x2_t vcvt_high_f64_f32(float32x4_t __p0) {
44194  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44195  float64x2_t __ret;
44196  __ret = __noswap_vcvt_f64_f32(__noswap_vget_high_f32(__rev0));
44197  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44198  return __ret;
44199}
44200#endif
44201
44202#ifdef __LITTLE_ENDIAN__
44203#define vcvts_n_f32_u32(__p0, __p1) __extension__ ({ \
44204  uint32_t __s0 = __p0; \
44205  float32_t __ret; \
44206  __ret = (float32_t) __builtin_neon_vcvts_n_f32_u32(__s0, __p1); \
44207  __ret; \
44208})
44209#else
44210#define vcvts_n_f32_u32(__p0, __p1) __extension__ ({ \
44211  uint32_t __s0 = __p0; \
44212  float32_t __ret; \
44213  __ret = (float32_t) __builtin_neon_vcvts_n_f32_u32(__s0, __p1); \
44214  __ret; \
44215})
44216#endif
44217
44218#ifdef __LITTLE_ENDIAN__
44219#define vcvts_n_f32_s32(__p0, __p1) __extension__ ({ \
44220  int32_t __s0 = __p0; \
44221  float32_t __ret; \
44222  __ret = (float32_t) __builtin_neon_vcvts_n_f32_s32(__s0, __p1); \
44223  __ret; \
44224})
44225#else
44226#define vcvts_n_f32_s32(__p0, __p1) __extension__ ({ \
44227  int32_t __s0 = __p0; \
44228  float32_t __ret; \
44229  __ret = (float32_t) __builtin_neon_vcvts_n_f32_s32(__s0, __p1); \
44230  __ret; \
44231})
44232#endif
44233
44234#ifdef __LITTLE_ENDIAN__
44235#define vcvtq_n_f64_u64(__p0, __p1) __extension__ ({ \
44236  uint64x2_t __s0 = __p0; \
44237  float64x2_t __ret; \
44238  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__s0, __p1, 51); \
44239  __ret; \
44240})
44241#else
44242#define vcvtq_n_f64_u64(__p0, __p1) __extension__ ({ \
44243  uint64x2_t __s0 = __p0; \
44244  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
44245  float64x2_t __ret; \
44246  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__rev0, __p1, 51); \
44247  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
44248  __ret; \
44249})
44250#endif
44251
44252#ifdef __LITTLE_ENDIAN__
44253#define vcvtq_n_f64_s64(__p0, __p1) __extension__ ({ \
44254  int64x2_t __s0 = __p0; \
44255  float64x2_t __ret; \
44256  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__s0, __p1, 35); \
44257  __ret; \
44258})
44259#else
44260#define vcvtq_n_f64_s64(__p0, __p1) __extension__ ({ \
44261  int64x2_t __s0 = __p0; \
44262  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
44263  float64x2_t __ret; \
44264  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__rev0, __p1, 35); \
44265  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
44266  __ret; \
44267})
44268#endif
44269
44270#ifdef __LITTLE_ENDIAN__
44271#define vcvt_n_f64_u64(__p0, __p1) __extension__ ({ \
44272  uint64x1_t __s0 = __p0; \
44273  float64x1_t __ret; \
44274  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 19); \
44275  __ret; \
44276})
44277#else
44278#define vcvt_n_f64_u64(__p0, __p1) __extension__ ({ \
44279  uint64x1_t __s0 = __p0; \
44280  float64x1_t __ret; \
44281  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 19); \
44282  __ret; \
44283})
44284#endif
44285
44286#ifdef __LITTLE_ENDIAN__
44287#define vcvt_n_f64_s64(__p0, __p1) __extension__ ({ \
44288  int64x1_t __s0 = __p0; \
44289  float64x1_t __ret; \
44290  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 3); \
44291  __ret; \
44292})
44293#else
44294#define vcvt_n_f64_s64(__p0, __p1) __extension__ ({ \
44295  int64x1_t __s0 = __p0; \
44296  float64x1_t __ret; \
44297  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 3); \
44298  __ret; \
44299})
44300#endif
44301
44302#ifdef __LITTLE_ENDIAN__
44303#define vcvtd_n_f64_u64(__p0, __p1) __extension__ ({ \
44304  uint64_t __s0 = __p0; \
44305  float64_t __ret; \
44306  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_u64(__s0, __p1); \
44307  __ret; \
44308})
44309#else
44310#define vcvtd_n_f64_u64(__p0, __p1) __extension__ ({ \
44311  uint64_t __s0 = __p0; \
44312  float64_t __ret; \
44313  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_u64(__s0, __p1); \
44314  __ret; \
44315})
44316#endif
44317
44318#ifdef __LITTLE_ENDIAN__
44319#define vcvtd_n_f64_s64(__p0, __p1) __extension__ ({ \
44320  int64_t __s0 = __p0; \
44321  float64_t __ret; \
44322  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_s64(__s0, __p1); \
44323  __ret; \
44324})
44325#else
44326#define vcvtd_n_f64_s64(__p0, __p1) __extension__ ({ \
44327  int64_t __s0 = __p0; \
44328  float64_t __ret; \
44329  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_s64(__s0, __p1); \
44330  __ret; \
44331})
44332#endif
44333
44334#ifdef __LITTLE_ENDIAN__
44335#define vcvts_n_s32_f32(__p0, __p1) __extension__ ({ \
44336  float32_t __s0 = __p0; \
44337  int32_t __ret; \
44338  __ret = (int32_t) __builtin_neon_vcvts_n_s32_f32(__s0, __p1); \
44339  __ret; \
44340})
44341#else
44342#define vcvts_n_s32_f32(__p0, __p1) __extension__ ({ \
44343  float32_t __s0 = __p0; \
44344  int32_t __ret; \
44345  __ret = (int32_t) __builtin_neon_vcvts_n_s32_f32(__s0, __p1); \
44346  __ret; \
44347})
44348#endif
44349
44350#ifdef __LITTLE_ENDIAN__
44351#define vcvtq_n_s64_f64(__p0, __p1) __extension__ ({ \
44352  float64x2_t __s0 = __p0; \
44353  int64x2_t __ret; \
44354  __ret = (int64x2_t) __builtin_neon_vcvtq_n_s64_v((int8x16_t)__s0, __p1, 35); \
44355  __ret; \
44356})
44357#else
44358#define vcvtq_n_s64_f64(__p0, __p1) __extension__ ({ \
44359  float64x2_t __s0 = __p0; \
44360  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
44361  int64x2_t __ret; \
44362  __ret = (int64x2_t) __builtin_neon_vcvtq_n_s64_v((int8x16_t)__rev0, __p1, 35); \
44363  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
44364  __ret; \
44365})
44366#endif
44367
44368#ifdef __LITTLE_ENDIAN__
44369#define vcvt_n_s64_f64(__p0, __p1) __extension__ ({ \
44370  float64x1_t __s0 = __p0; \
44371  int64x1_t __ret; \
44372  __ret = (int64x1_t) __builtin_neon_vcvt_n_s64_v((int8x8_t)__s0, __p1, 3); \
44373  __ret; \
44374})
44375#else
44376#define vcvt_n_s64_f64(__p0, __p1) __extension__ ({ \
44377  float64x1_t __s0 = __p0; \
44378  int64x1_t __ret; \
44379  __ret = (int64x1_t) __builtin_neon_vcvt_n_s64_v((int8x8_t)__s0, __p1, 3); \
44380  __ret; \
44381})
44382#endif
44383
44384#ifdef __LITTLE_ENDIAN__
44385#define vcvtd_n_s64_f64(__p0, __p1) __extension__ ({ \
44386  float64_t __s0 = __p0; \
44387  int64_t __ret; \
44388  __ret = (int64_t) __builtin_neon_vcvtd_n_s64_f64(__s0, __p1); \
44389  __ret; \
44390})
44391#else
44392#define vcvtd_n_s64_f64(__p0, __p1) __extension__ ({ \
44393  float64_t __s0 = __p0; \
44394  int64_t __ret; \
44395  __ret = (int64_t) __builtin_neon_vcvtd_n_s64_f64(__s0, __p1); \
44396  __ret; \
44397})
44398#endif
44399
44400#ifdef __LITTLE_ENDIAN__
44401#define vcvts_n_u32_f32(__p0, __p1) __extension__ ({ \
44402  float32_t __s0 = __p0; \
44403  uint32_t __ret; \
44404  __ret = (uint32_t) __builtin_neon_vcvts_n_u32_f32(__s0, __p1); \
44405  __ret; \
44406})
44407#else
44408#define vcvts_n_u32_f32(__p0, __p1) __extension__ ({ \
44409  float32_t __s0 = __p0; \
44410  uint32_t __ret; \
44411  __ret = (uint32_t) __builtin_neon_vcvts_n_u32_f32(__s0, __p1); \
44412  __ret; \
44413})
44414#endif
44415
44416#ifdef __LITTLE_ENDIAN__
44417#define vcvtq_n_u64_f64(__p0, __p1) __extension__ ({ \
44418  float64x2_t __s0 = __p0; \
44419  uint64x2_t __ret; \
44420  __ret = (uint64x2_t) __builtin_neon_vcvtq_n_u64_v((int8x16_t)__s0, __p1, 51); \
44421  __ret; \
44422})
44423#else
44424#define vcvtq_n_u64_f64(__p0, __p1) __extension__ ({ \
44425  float64x2_t __s0 = __p0; \
44426  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
44427  uint64x2_t __ret; \
44428  __ret = (uint64x2_t) __builtin_neon_vcvtq_n_u64_v((int8x16_t)__rev0, __p1, 51); \
44429  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
44430  __ret; \
44431})
44432#endif
44433
44434#ifdef __LITTLE_ENDIAN__
44435#define vcvt_n_u64_f64(__p0, __p1) __extension__ ({ \
44436  float64x1_t __s0 = __p0; \
44437  uint64x1_t __ret; \
44438  __ret = (uint64x1_t) __builtin_neon_vcvt_n_u64_v((int8x8_t)__s0, __p1, 19); \
44439  __ret; \
44440})
44441#else
44442#define vcvt_n_u64_f64(__p0, __p1) __extension__ ({ \
44443  float64x1_t __s0 = __p0; \
44444  uint64x1_t __ret; \
44445  __ret = (uint64x1_t) __builtin_neon_vcvt_n_u64_v((int8x8_t)__s0, __p1, 19); \
44446  __ret; \
44447})
44448#endif
44449
44450#ifdef __LITTLE_ENDIAN__
44451#define vcvtd_n_u64_f64(__p0, __p1) __extension__ ({ \
44452  float64_t __s0 = __p0; \
44453  uint64_t __ret; \
44454  __ret = (uint64_t) __builtin_neon_vcvtd_n_u64_f64(__s0, __p1); \
44455  __ret; \
44456})
44457#else
44458#define vcvtd_n_u64_f64(__p0, __p1) __extension__ ({ \
44459  float64_t __s0 = __p0; \
44460  uint64_t __ret; \
44461  __ret = (uint64_t) __builtin_neon_vcvtd_n_u64_f64(__s0, __p1); \
44462  __ret; \
44463})
44464#endif
44465
44466#ifdef __LITTLE_ENDIAN__
44467__ai int32_t vcvts_s32_f32(float32_t __p0) {
44468  int32_t __ret;
44469  __ret = (int32_t) __builtin_neon_vcvts_s32_f32(__p0);
44470  return __ret;
44471}
44472#else
44473__ai int32_t vcvts_s32_f32(float32_t __p0) {
44474  int32_t __ret;
44475  __ret = (int32_t) __builtin_neon_vcvts_s32_f32(__p0);
44476  return __ret;
44477}
44478#endif
44479
44480#ifdef __LITTLE_ENDIAN__
44481__ai int64_t vcvtd_s64_f64(float64_t __p0) {
44482  int64_t __ret;
44483  __ret = (int64_t) __builtin_neon_vcvtd_s64_f64(__p0);
44484  return __ret;
44485}
44486#else
44487__ai int64_t vcvtd_s64_f64(float64_t __p0) {
44488  int64_t __ret;
44489  __ret = (int64_t) __builtin_neon_vcvtd_s64_f64(__p0);
44490  return __ret;
44491}
44492#endif
44493
44494#ifdef __LITTLE_ENDIAN__
44495__ai int64x2_t vcvtq_s64_f64(float64x2_t __p0) {
44496  int64x2_t __ret;
44497  __ret = (int64x2_t) __builtin_neon_vcvtq_s64_v((int8x16_t)__p0, 35);
44498  return __ret;
44499}
44500#else
44501__ai int64x2_t vcvtq_s64_f64(float64x2_t __p0) {
44502  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44503  int64x2_t __ret;
44504  __ret = (int64x2_t) __builtin_neon_vcvtq_s64_v((int8x16_t)__rev0, 35);
44505  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44506  return __ret;
44507}
44508#endif
44509
44510#ifdef __LITTLE_ENDIAN__
44511__ai int64x1_t vcvt_s64_f64(float64x1_t __p0) {
44512  int64x1_t __ret;
44513  __ret = (int64x1_t) __builtin_neon_vcvt_s64_v((int8x8_t)__p0, 3);
44514  return __ret;
44515}
44516#else
44517__ai int64x1_t vcvt_s64_f64(float64x1_t __p0) {
44518  int64x1_t __ret;
44519  __ret = (int64x1_t) __builtin_neon_vcvt_s64_v((int8x8_t)__p0, 3);
44520  return __ret;
44521}
44522#endif
44523
44524#ifdef __LITTLE_ENDIAN__
44525__ai uint32_t vcvts_u32_f32(float32_t __p0) {
44526  uint32_t __ret;
44527  __ret = (uint32_t) __builtin_neon_vcvts_u32_f32(__p0);
44528  return __ret;
44529}
44530#else
44531__ai uint32_t vcvts_u32_f32(float32_t __p0) {
44532  uint32_t __ret;
44533  __ret = (uint32_t) __builtin_neon_vcvts_u32_f32(__p0);
44534  return __ret;
44535}
44536#endif
44537
44538#ifdef __LITTLE_ENDIAN__
44539__ai uint64_t vcvtd_u64_f64(float64_t __p0) {
44540  uint64_t __ret;
44541  __ret = (uint64_t) __builtin_neon_vcvtd_u64_f64(__p0);
44542  return __ret;
44543}
44544#else
44545__ai uint64_t vcvtd_u64_f64(float64_t __p0) {
44546  uint64_t __ret;
44547  __ret = (uint64_t) __builtin_neon_vcvtd_u64_f64(__p0);
44548  return __ret;
44549}
44550#endif
44551
44552#ifdef __LITTLE_ENDIAN__
44553__ai uint64x2_t vcvtq_u64_f64(float64x2_t __p0) {
44554  uint64x2_t __ret;
44555  __ret = (uint64x2_t) __builtin_neon_vcvtq_u64_v((int8x16_t)__p0, 51);
44556  return __ret;
44557}
44558#else
44559__ai uint64x2_t vcvtq_u64_f64(float64x2_t __p0) {
44560  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44561  uint64x2_t __ret;
44562  __ret = (uint64x2_t) __builtin_neon_vcvtq_u64_v((int8x16_t)__rev0, 51);
44563  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44564  return __ret;
44565}
44566#endif
44567
44568#ifdef __LITTLE_ENDIAN__
44569__ai uint64x1_t vcvt_u64_f64(float64x1_t __p0) {
44570  uint64x1_t __ret;
44571  __ret = (uint64x1_t) __builtin_neon_vcvt_u64_v((int8x8_t)__p0, 19);
44572  return __ret;
44573}
44574#else
44575__ai uint64x1_t vcvt_u64_f64(float64x1_t __p0) {
44576  uint64x1_t __ret;
44577  __ret = (uint64x1_t) __builtin_neon_vcvt_u64_v((int8x8_t)__p0, 19);
44578  return __ret;
44579}
44580#endif
44581
44582#ifdef __LITTLE_ENDIAN__
44583__ai int32_t vcvtas_s32_f32(float32_t __p0) {
44584  int32_t __ret;
44585  __ret = (int32_t) __builtin_neon_vcvtas_s32_f32(__p0);
44586  return __ret;
44587}
44588#else
44589__ai int32_t vcvtas_s32_f32(float32_t __p0) {
44590  int32_t __ret;
44591  __ret = (int32_t) __builtin_neon_vcvtas_s32_f32(__p0);
44592  return __ret;
44593}
44594#endif
44595
44596#ifdef __LITTLE_ENDIAN__
44597__ai int64_t vcvtad_s64_f64(float64_t __p0) {
44598  int64_t __ret;
44599  __ret = (int64_t) __builtin_neon_vcvtad_s64_f64(__p0);
44600  return __ret;
44601}
44602#else
44603__ai int64_t vcvtad_s64_f64(float64_t __p0) {
44604  int64_t __ret;
44605  __ret = (int64_t) __builtin_neon_vcvtad_s64_f64(__p0);
44606  return __ret;
44607}
44608#endif
44609
44610#ifdef __LITTLE_ENDIAN__
44611__ai uint32_t vcvtas_u32_f32(float32_t __p0) {
44612  uint32_t __ret;
44613  __ret = (uint32_t) __builtin_neon_vcvtas_u32_f32(__p0);
44614  return __ret;
44615}
44616#else
44617__ai uint32_t vcvtas_u32_f32(float32_t __p0) {
44618  uint32_t __ret;
44619  __ret = (uint32_t) __builtin_neon_vcvtas_u32_f32(__p0);
44620  return __ret;
44621}
44622#endif
44623
44624#ifdef __LITTLE_ENDIAN__
44625__ai uint64_t vcvtad_u64_f64(float64_t __p0) {
44626  uint64_t __ret;
44627  __ret = (uint64_t) __builtin_neon_vcvtad_u64_f64(__p0);
44628  return __ret;
44629}
44630#else
44631__ai uint64_t vcvtad_u64_f64(float64_t __p0) {
44632  uint64_t __ret;
44633  __ret = (uint64_t) __builtin_neon_vcvtad_u64_f64(__p0);
44634  return __ret;
44635}
44636#endif
44637
44638#ifdef __LITTLE_ENDIAN__
44639__ai int32_t vcvtms_s32_f32(float32_t __p0) {
44640  int32_t __ret;
44641  __ret = (int32_t) __builtin_neon_vcvtms_s32_f32(__p0);
44642  return __ret;
44643}
44644#else
44645__ai int32_t vcvtms_s32_f32(float32_t __p0) {
44646  int32_t __ret;
44647  __ret = (int32_t) __builtin_neon_vcvtms_s32_f32(__p0);
44648  return __ret;
44649}
44650#endif
44651
44652#ifdef __LITTLE_ENDIAN__
44653__ai int64_t vcvtmd_s64_f64(float64_t __p0) {
44654  int64_t __ret;
44655  __ret = (int64_t) __builtin_neon_vcvtmd_s64_f64(__p0);
44656  return __ret;
44657}
44658#else
44659__ai int64_t vcvtmd_s64_f64(float64_t __p0) {
44660  int64_t __ret;
44661  __ret = (int64_t) __builtin_neon_vcvtmd_s64_f64(__p0);
44662  return __ret;
44663}
44664#endif
44665
44666#ifdef __LITTLE_ENDIAN__
44667__ai uint32_t vcvtms_u32_f32(float32_t __p0) {
44668  uint32_t __ret;
44669  __ret = (uint32_t) __builtin_neon_vcvtms_u32_f32(__p0);
44670  return __ret;
44671}
44672#else
44673__ai uint32_t vcvtms_u32_f32(float32_t __p0) {
44674  uint32_t __ret;
44675  __ret = (uint32_t) __builtin_neon_vcvtms_u32_f32(__p0);
44676  return __ret;
44677}
44678#endif
44679
44680#ifdef __LITTLE_ENDIAN__
44681__ai uint64_t vcvtmd_u64_f64(float64_t __p0) {
44682  uint64_t __ret;
44683  __ret = (uint64_t) __builtin_neon_vcvtmd_u64_f64(__p0);
44684  return __ret;
44685}
44686#else
44687__ai uint64_t vcvtmd_u64_f64(float64_t __p0) {
44688  uint64_t __ret;
44689  __ret = (uint64_t) __builtin_neon_vcvtmd_u64_f64(__p0);
44690  return __ret;
44691}
44692#endif
44693
44694#ifdef __LITTLE_ENDIAN__
44695__ai int32_t vcvtns_s32_f32(float32_t __p0) {
44696  int32_t __ret;
44697  __ret = (int32_t) __builtin_neon_vcvtns_s32_f32(__p0);
44698  return __ret;
44699}
44700#else
44701__ai int32_t vcvtns_s32_f32(float32_t __p0) {
44702  int32_t __ret;
44703  __ret = (int32_t) __builtin_neon_vcvtns_s32_f32(__p0);
44704  return __ret;
44705}
44706#endif
44707
44708#ifdef __LITTLE_ENDIAN__
44709__ai int64_t vcvtnd_s64_f64(float64_t __p0) {
44710  int64_t __ret;
44711  __ret = (int64_t) __builtin_neon_vcvtnd_s64_f64(__p0);
44712  return __ret;
44713}
44714#else
44715__ai int64_t vcvtnd_s64_f64(float64_t __p0) {
44716  int64_t __ret;
44717  __ret = (int64_t) __builtin_neon_vcvtnd_s64_f64(__p0);
44718  return __ret;
44719}
44720#endif
44721
44722#ifdef __LITTLE_ENDIAN__
44723__ai uint32_t vcvtns_u32_f32(float32_t __p0) {
44724  uint32_t __ret;
44725  __ret = (uint32_t) __builtin_neon_vcvtns_u32_f32(__p0);
44726  return __ret;
44727}
44728#else
44729__ai uint32_t vcvtns_u32_f32(float32_t __p0) {
44730  uint32_t __ret;
44731  __ret = (uint32_t) __builtin_neon_vcvtns_u32_f32(__p0);
44732  return __ret;
44733}
44734#endif
44735
44736#ifdef __LITTLE_ENDIAN__
44737__ai uint64_t vcvtnd_u64_f64(float64_t __p0) {
44738  uint64_t __ret;
44739  __ret = (uint64_t) __builtin_neon_vcvtnd_u64_f64(__p0);
44740  return __ret;
44741}
44742#else
44743__ai uint64_t vcvtnd_u64_f64(float64_t __p0) {
44744  uint64_t __ret;
44745  __ret = (uint64_t) __builtin_neon_vcvtnd_u64_f64(__p0);
44746  return __ret;
44747}
44748#endif
44749
44750#ifdef __LITTLE_ENDIAN__
44751__ai int32_t vcvtps_s32_f32(float32_t __p0) {
44752  int32_t __ret;
44753  __ret = (int32_t) __builtin_neon_vcvtps_s32_f32(__p0);
44754  return __ret;
44755}
44756#else
44757__ai int32_t vcvtps_s32_f32(float32_t __p0) {
44758  int32_t __ret;
44759  __ret = (int32_t) __builtin_neon_vcvtps_s32_f32(__p0);
44760  return __ret;
44761}
44762#endif
44763
44764#ifdef __LITTLE_ENDIAN__
44765__ai int64_t vcvtpd_s64_f64(float64_t __p0) {
44766  int64_t __ret;
44767  __ret = (int64_t) __builtin_neon_vcvtpd_s64_f64(__p0);
44768  return __ret;
44769}
44770#else
44771__ai int64_t vcvtpd_s64_f64(float64_t __p0) {
44772  int64_t __ret;
44773  __ret = (int64_t) __builtin_neon_vcvtpd_s64_f64(__p0);
44774  return __ret;
44775}
44776#endif
44777
44778#ifdef __LITTLE_ENDIAN__
44779__ai uint32_t vcvtps_u32_f32(float32_t __p0) {
44780  uint32_t __ret;
44781  __ret = (uint32_t) __builtin_neon_vcvtps_u32_f32(__p0);
44782  return __ret;
44783}
44784#else
44785__ai uint32_t vcvtps_u32_f32(float32_t __p0) {
44786  uint32_t __ret;
44787  __ret = (uint32_t) __builtin_neon_vcvtps_u32_f32(__p0);
44788  return __ret;
44789}
44790#endif
44791
44792#ifdef __LITTLE_ENDIAN__
44793__ai uint64_t vcvtpd_u64_f64(float64_t __p0) {
44794  uint64_t __ret;
44795  __ret = (uint64_t) __builtin_neon_vcvtpd_u64_f64(__p0);
44796  return __ret;
44797}
44798#else
44799__ai uint64_t vcvtpd_u64_f64(float64_t __p0) {
44800  uint64_t __ret;
44801  __ret = (uint64_t) __builtin_neon_vcvtpd_u64_f64(__p0);
44802  return __ret;
44803}
44804#endif
44805
44806#ifdef __LITTLE_ENDIAN__
44807__ai float32_t vcvtxd_f32_f64(float64_t __p0) {
44808  float32_t __ret;
44809  __ret = (float32_t) __builtin_neon_vcvtxd_f32_f64(__p0);
44810  return __ret;
44811}
44812#else
44813__ai float32_t vcvtxd_f32_f64(float64_t __p0) {
44814  float32_t __ret;
44815  __ret = (float32_t) __builtin_neon_vcvtxd_f32_f64(__p0);
44816  return __ret;
44817}
44818#endif
44819
44820#ifdef __LITTLE_ENDIAN__
44821__ai float32x2_t vcvtx_f32_f64(float64x2_t __p0) {
44822  float32x2_t __ret;
44823  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__p0, 42);
44824  return __ret;
44825}
44826#else
44827__ai float32x2_t vcvtx_f32_f64(float64x2_t __p0) {
44828  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44829  float32x2_t __ret;
44830  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__rev0, 42);
44831  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44832  return __ret;
44833}
44834__ai float32x2_t __noswap_vcvtx_f32_f64(float64x2_t __p0) {
44835  float32x2_t __ret;
44836  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__p0, 42);
44837  return __ret;
44838}
44839#endif
44840
44841#ifdef __LITTLE_ENDIAN__
44842__ai float32x4_t vcvtx_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
44843  float32x4_t __ret;
44844  __ret = vcombine_f32(__p0, vcvtx_f32_f64(__p1));
44845  return __ret;
44846}
44847#else
44848__ai float32x4_t vcvtx_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
44849  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44850  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
44851  float32x4_t __ret;
44852  __ret = __noswap_vcombine_f32(__rev0, __noswap_vcvtx_f32_f64(__rev1));
44853  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44854  return __ret;
44855}
44856#endif
44857
44858#ifdef __LITTLE_ENDIAN__
44859__ai float64x2_t vdivq_f64(float64x2_t __p0, float64x2_t __p1) {
44860  float64x2_t __ret;
44861  __ret = __p0 / __p1;
44862  return __ret;
44863}
44864#else
44865__ai float64x2_t vdivq_f64(float64x2_t __p0, float64x2_t __p1) {
44866  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44867  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
44868  float64x2_t __ret;
44869  __ret = __rev0 / __rev1;
44870  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44871  return __ret;
44872}
44873#endif
44874
44875#ifdef __LITTLE_ENDIAN__
44876__ai float32x4_t vdivq_f32(float32x4_t __p0, float32x4_t __p1) {
44877  float32x4_t __ret;
44878  __ret = __p0 / __p1;
44879  return __ret;
44880}
44881#else
44882__ai float32x4_t vdivq_f32(float32x4_t __p0, float32x4_t __p1) {
44883  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44884  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
44885  float32x4_t __ret;
44886  __ret = __rev0 / __rev1;
44887  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44888  return __ret;
44889}
44890#endif
44891
44892#ifdef __LITTLE_ENDIAN__
44893__ai float64x1_t vdiv_f64(float64x1_t __p0, float64x1_t __p1) {
44894  float64x1_t __ret;
44895  __ret = __p0 / __p1;
44896  return __ret;
44897}
44898#else
44899__ai float64x1_t vdiv_f64(float64x1_t __p0, float64x1_t __p1) {
44900  float64x1_t __ret;
44901  __ret = __p0 / __p1;
44902  return __ret;
44903}
44904#endif
44905
44906#ifdef __LITTLE_ENDIAN__
44907__ai float32x2_t vdiv_f32(float32x2_t __p0, float32x2_t __p1) {
44908  float32x2_t __ret;
44909  __ret = __p0 / __p1;
44910  return __ret;
44911}
44912#else
44913__ai float32x2_t vdiv_f32(float32x2_t __p0, float32x2_t __p1) {
44914  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44915  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
44916  float32x2_t __ret;
44917  __ret = __rev0 / __rev1;
44918  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44919  return __ret;
44920}
44921#endif
44922
44923#ifdef __LITTLE_ENDIAN__
44924#define vdupb_lane_p8(__p0, __p1) __extension__ ({ \
44925  poly8x8_t __s0 = __p0; \
44926  poly8_t __ret; \
44927  __ret = (poly8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
44928  __ret; \
44929})
44930#else
44931#define vdupb_lane_p8(__p0, __p1) __extension__ ({ \
44932  poly8x8_t __s0 = __p0; \
44933  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
44934  poly8_t __ret; \
44935  __ret = (poly8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
44936  __ret; \
44937})
44938#endif
44939
44940#ifdef __LITTLE_ENDIAN__
44941#define vduph_lane_p16(__p0, __p1) __extension__ ({ \
44942  poly16x4_t __s0 = __p0; \
44943  poly16_t __ret; \
44944  __ret = (poly16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
44945  __ret; \
44946})
44947#else
44948#define vduph_lane_p16(__p0, __p1) __extension__ ({ \
44949  poly16x4_t __s0 = __p0; \
44950  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
44951  poly16_t __ret; \
44952  __ret = (poly16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
44953  __ret; \
44954})
44955#endif
44956
44957#ifdef __LITTLE_ENDIAN__
44958#define vdupb_lane_u8(__p0, __p1) __extension__ ({ \
44959  uint8x8_t __s0 = __p0; \
44960  uint8_t __ret; \
44961  __ret = (uint8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
44962  __ret; \
44963})
44964#else
44965#define vdupb_lane_u8(__p0, __p1) __extension__ ({ \
44966  uint8x8_t __s0 = __p0; \
44967  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
44968  uint8_t __ret; \
44969  __ret = (uint8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
44970  __ret; \
44971})
44972#endif
44973
44974#ifdef __LITTLE_ENDIAN__
44975#define vdups_lane_u32(__p0, __p1) __extension__ ({ \
44976  uint32x2_t __s0 = __p0; \
44977  uint32_t __ret; \
44978  __ret = (uint32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__s0, __p1); \
44979  __ret; \
44980})
44981#else
44982#define vdups_lane_u32(__p0, __p1) __extension__ ({ \
44983  uint32x2_t __s0 = __p0; \
44984  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
44985  uint32_t __ret; \
44986  __ret = (uint32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__rev0, __p1); \
44987  __ret; \
44988})
44989#endif
44990
44991#ifdef __LITTLE_ENDIAN__
44992#define vdupd_lane_u64(__p0, __p1) __extension__ ({ \
44993  uint64x1_t __s0 = __p0; \
44994  uint64_t __ret; \
44995  __ret = (uint64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
44996  __ret; \
44997})
44998#else
44999#define vdupd_lane_u64(__p0, __p1) __extension__ ({ \
45000  uint64x1_t __s0 = __p0; \
45001  uint64_t __ret; \
45002  __ret = (uint64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
45003  __ret; \
45004})
45005#endif
45006
45007#ifdef __LITTLE_ENDIAN__
45008#define vduph_lane_u16(__p0, __p1) __extension__ ({ \
45009  uint16x4_t __s0 = __p0; \
45010  uint16_t __ret; \
45011  __ret = (uint16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
45012  __ret; \
45013})
45014#else
45015#define vduph_lane_u16(__p0, __p1) __extension__ ({ \
45016  uint16x4_t __s0 = __p0; \
45017  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45018  uint16_t __ret; \
45019  __ret = (uint16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
45020  __ret; \
45021})
45022#endif
45023
45024#ifdef __LITTLE_ENDIAN__
45025#define vdupb_lane_s8(__p0, __p1) __extension__ ({ \
45026  int8x8_t __s0 = __p0; \
45027  int8_t __ret; \
45028  __ret = (int8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
45029  __ret; \
45030})
45031#else
45032#define vdupb_lane_s8(__p0, __p1) __extension__ ({ \
45033  int8x8_t __s0 = __p0; \
45034  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45035  int8_t __ret; \
45036  __ret = (int8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
45037  __ret; \
45038})
45039#endif
45040
45041#ifdef __LITTLE_ENDIAN__
45042#define vdupd_lane_f64(__p0, __p1) __extension__ ({ \
45043  float64x1_t __s0 = __p0; \
45044  float64_t __ret; \
45045  __ret = (float64_t) __builtin_neon_vdupd_lane_f64((int8x8_t)__s0, __p1); \
45046  __ret; \
45047})
45048#else
45049#define vdupd_lane_f64(__p0, __p1) __extension__ ({ \
45050  float64x1_t __s0 = __p0; \
45051  float64_t __ret; \
45052  __ret = (float64_t) __builtin_neon_vdupd_lane_f64((int8x8_t)__s0, __p1); \
45053  __ret; \
45054})
45055#endif
45056
45057#ifdef __LITTLE_ENDIAN__
45058#define vdups_lane_f32(__p0, __p1) __extension__ ({ \
45059  float32x2_t __s0 = __p0; \
45060  float32_t __ret; \
45061  __ret = (float32_t) __builtin_neon_vdups_lane_f32((int8x8_t)__s0, __p1); \
45062  __ret; \
45063})
45064#else
45065#define vdups_lane_f32(__p0, __p1) __extension__ ({ \
45066  float32x2_t __s0 = __p0; \
45067  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45068  float32_t __ret; \
45069  __ret = (float32_t) __builtin_neon_vdups_lane_f32((int8x8_t)__rev0, __p1); \
45070  __ret; \
45071})
45072#endif
45073
45074#ifdef __LITTLE_ENDIAN__
45075#define vdups_lane_s32(__p0, __p1) __extension__ ({ \
45076  int32x2_t __s0 = __p0; \
45077  int32_t __ret; \
45078  __ret = (int32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__s0, __p1); \
45079  __ret; \
45080})
45081#else
45082#define vdups_lane_s32(__p0, __p1) __extension__ ({ \
45083  int32x2_t __s0 = __p0; \
45084  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45085  int32_t __ret; \
45086  __ret = (int32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__rev0, __p1); \
45087  __ret; \
45088})
45089#endif
45090
45091#ifdef __LITTLE_ENDIAN__
45092#define vdupd_lane_s64(__p0, __p1) __extension__ ({ \
45093  int64x1_t __s0 = __p0; \
45094  int64_t __ret; \
45095  __ret = (int64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
45096  __ret; \
45097})
45098#else
45099#define vdupd_lane_s64(__p0, __p1) __extension__ ({ \
45100  int64x1_t __s0 = __p0; \
45101  int64_t __ret; \
45102  __ret = (int64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
45103  __ret; \
45104})
45105#endif
45106
45107#ifdef __LITTLE_ENDIAN__
45108#define vduph_lane_s16(__p0, __p1) __extension__ ({ \
45109  int16x4_t __s0 = __p0; \
45110  int16_t __ret; \
45111  __ret = (int16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
45112  __ret; \
45113})
45114#else
45115#define vduph_lane_s16(__p0, __p1) __extension__ ({ \
45116  int16x4_t __s0 = __p0; \
45117  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45118  int16_t __ret; \
45119  __ret = (int16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
45120  __ret; \
45121})
45122#endif
45123
45124#ifdef __LITTLE_ENDIAN__
45125#define vdup_lane_p64(__p0, __p1) __extension__ ({ \
45126  poly64x1_t __s0 = __p0; \
45127  poly64x1_t __ret; \
45128  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
45129  __ret; \
45130})
45131#else
45132#define vdup_lane_p64(__p0, __p1) __extension__ ({ \
45133  poly64x1_t __s0 = __p0; \
45134  poly64x1_t __ret; \
45135  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
45136  __ret; \
45137})
45138#endif
45139
45140#ifdef __LITTLE_ENDIAN__
45141#define vdupq_lane_p64(__p0, __p1) __extension__ ({ \
45142  poly64x1_t __s0 = __p0; \
45143  poly64x2_t __ret; \
45144  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
45145  __ret; \
45146})
45147#else
45148#define vdupq_lane_p64(__p0, __p1) __extension__ ({ \
45149  poly64x1_t __s0 = __p0; \
45150  poly64x2_t __ret; \
45151  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
45152  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45153  __ret; \
45154})
45155#endif
45156
45157#ifdef __LITTLE_ENDIAN__
45158#define vdupq_lane_f64(__p0, __p1) __extension__ ({ \
45159  float64x1_t __s0 = __p0; \
45160  float64x2_t __ret; \
45161  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
45162  __ret; \
45163})
45164#else
45165#define vdupq_lane_f64(__p0, __p1) __extension__ ({ \
45166  float64x1_t __s0 = __p0; \
45167  float64x2_t __ret; \
45168  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
45169  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45170  __ret; \
45171})
45172#endif
45173
45174#ifdef __LITTLE_ENDIAN__
45175#define vdupq_lane_f16(__p0, __p1) __extension__ ({ \
45176  float16x4_t __s0 = __p0; \
45177  float16x8_t __ret; \
45178  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45179  __ret; \
45180})
45181#else
45182#define vdupq_lane_f16(__p0, __p1) __extension__ ({ \
45183  float16x4_t __s0 = __p0; \
45184  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45185  float16x8_t __ret; \
45186  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45187  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
45188  __ret; \
45189})
45190#endif
45191
45192#ifdef __LITTLE_ENDIAN__
45193#define vdup_lane_f64(__p0, __p1) __extension__ ({ \
45194  float64x1_t __s0 = __p0; \
45195  float64x1_t __ret; \
45196  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
45197  __ret; \
45198})
45199#else
45200#define vdup_lane_f64(__p0, __p1) __extension__ ({ \
45201  float64x1_t __s0 = __p0; \
45202  float64x1_t __ret; \
45203  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
45204  __ret; \
45205})
45206#endif
45207
45208#ifdef __LITTLE_ENDIAN__
45209#define vdup_lane_f16(__p0, __p1) __extension__ ({ \
45210  float16x4_t __s0 = __p0; \
45211  float16x4_t __ret; \
45212  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
45213  __ret; \
45214})
45215#else
45216#define vdup_lane_f16(__p0, __p1) __extension__ ({ \
45217  float16x4_t __s0 = __p0; \
45218  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45219  float16x4_t __ret; \
45220  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
45221  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
45222  __ret; \
45223})
45224#endif
45225
45226#ifdef __LITTLE_ENDIAN__
45227#define vdupb_laneq_p8(__p0, __p1) __extension__ ({ \
45228  poly8x16_t __s0 = __p0; \
45229  poly8_t __ret; \
45230  __ret = (poly8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
45231  __ret; \
45232})
45233#else
45234#define vdupb_laneq_p8(__p0, __p1) __extension__ ({ \
45235  poly8x16_t __s0 = __p0; \
45236  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45237  poly8_t __ret; \
45238  __ret = (poly8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
45239  __ret; \
45240})
45241#endif
45242
45243#ifdef __LITTLE_ENDIAN__
45244#define vduph_laneq_p16(__p0, __p1) __extension__ ({ \
45245  poly16x8_t __s0 = __p0; \
45246  poly16_t __ret; \
45247  __ret = (poly16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
45248  __ret; \
45249})
45250#else
45251#define vduph_laneq_p16(__p0, __p1) __extension__ ({ \
45252  poly16x8_t __s0 = __p0; \
45253  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45254  poly16_t __ret; \
45255  __ret = (poly16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
45256  __ret; \
45257})
45258#endif
45259
45260#ifdef __LITTLE_ENDIAN__
45261#define vdupb_laneq_u8(__p0, __p1) __extension__ ({ \
45262  uint8x16_t __s0 = __p0; \
45263  uint8_t __ret; \
45264  __ret = (uint8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
45265  __ret; \
45266})
45267#else
45268#define vdupb_laneq_u8(__p0, __p1) __extension__ ({ \
45269  uint8x16_t __s0 = __p0; \
45270  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45271  uint8_t __ret; \
45272  __ret = (uint8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
45273  __ret; \
45274})
45275#endif
45276
45277#ifdef __LITTLE_ENDIAN__
45278#define vdups_laneq_u32(__p0, __p1) __extension__ ({ \
45279  uint32x4_t __s0 = __p0; \
45280  uint32_t __ret; \
45281  __ret = (uint32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__s0, __p1); \
45282  __ret; \
45283})
45284#else
45285#define vdups_laneq_u32(__p0, __p1) __extension__ ({ \
45286  uint32x4_t __s0 = __p0; \
45287  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45288  uint32_t __ret; \
45289  __ret = (uint32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__rev0, __p1); \
45290  __ret; \
45291})
45292#endif
45293
45294#ifdef __LITTLE_ENDIAN__
45295#define vdupd_laneq_u64(__p0, __p1) __extension__ ({ \
45296  uint64x2_t __s0 = __p0; \
45297  uint64_t __ret; \
45298  __ret = (uint64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__s0, __p1); \
45299  __ret; \
45300})
45301#else
45302#define vdupd_laneq_u64(__p0, __p1) __extension__ ({ \
45303  uint64x2_t __s0 = __p0; \
45304  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45305  uint64_t __ret; \
45306  __ret = (uint64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__rev0, __p1); \
45307  __ret; \
45308})
45309#endif
45310
45311#ifdef __LITTLE_ENDIAN__
45312#define vduph_laneq_u16(__p0, __p1) __extension__ ({ \
45313  uint16x8_t __s0 = __p0; \
45314  uint16_t __ret; \
45315  __ret = (uint16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
45316  __ret; \
45317})
45318#else
45319#define vduph_laneq_u16(__p0, __p1) __extension__ ({ \
45320  uint16x8_t __s0 = __p0; \
45321  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45322  uint16_t __ret; \
45323  __ret = (uint16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
45324  __ret; \
45325})
45326#endif
45327
45328#ifdef __LITTLE_ENDIAN__
45329#define vdupb_laneq_s8(__p0, __p1) __extension__ ({ \
45330  int8x16_t __s0 = __p0; \
45331  int8_t __ret; \
45332  __ret = (int8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
45333  __ret; \
45334})
45335#else
45336#define vdupb_laneq_s8(__p0, __p1) __extension__ ({ \
45337  int8x16_t __s0 = __p0; \
45338  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45339  int8_t __ret; \
45340  __ret = (int8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
45341  __ret; \
45342})
45343#endif
45344
45345#ifdef __LITTLE_ENDIAN__
45346#define vdupd_laneq_f64(__p0, __p1) __extension__ ({ \
45347  float64x2_t __s0 = __p0; \
45348  float64_t __ret; \
45349  __ret = (float64_t) __builtin_neon_vdupd_laneq_f64((int8x16_t)__s0, __p1); \
45350  __ret; \
45351})
45352#else
45353#define vdupd_laneq_f64(__p0, __p1) __extension__ ({ \
45354  float64x2_t __s0 = __p0; \
45355  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45356  float64_t __ret; \
45357  __ret = (float64_t) __builtin_neon_vdupd_laneq_f64((int8x16_t)__rev0, __p1); \
45358  __ret; \
45359})
45360#endif
45361
45362#ifdef __LITTLE_ENDIAN__
45363#define vdups_laneq_f32(__p0, __p1) __extension__ ({ \
45364  float32x4_t __s0 = __p0; \
45365  float32_t __ret; \
45366  __ret = (float32_t) __builtin_neon_vdups_laneq_f32((int8x16_t)__s0, __p1); \
45367  __ret; \
45368})
45369#else
45370#define vdups_laneq_f32(__p0, __p1) __extension__ ({ \
45371  float32x4_t __s0 = __p0; \
45372  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45373  float32_t __ret; \
45374  __ret = (float32_t) __builtin_neon_vdups_laneq_f32((int8x16_t)__rev0, __p1); \
45375  __ret; \
45376})
45377#endif
45378
45379#ifdef __LITTLE_ENDIAN__
45380#define vdups_laneq_s32(__p0, __p1) __extension__ ({ \
45381  int32x4_t __s0 = __p0; \
45382  int32_t __ret; \
45383  __ret = (int32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__s0, __p1); \
45384  __ret; \
45385})
45386#else
45387#define vdups_laneq_s32(__p0, __p1) __extension__ ({ \
45388  int32x4_t __s0 = __p0; \
45389  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45390  int32_t __ret; \
45391  __ret = (int32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__rev0, __p1); \
45392  __ret; \
45393})
45394#endif
45395
45396#ifdef __LITTLE_ENDIAN__
45397#define vdupd_laneq_s64(__p0, __p1) __extension__ ({ \
45398  int64x2_t __s0 = __p0; \
45399  int64_t __ret; \
45400  __ret = (int64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__s0, __p1); \
45401  __ret; \
45402})
45403#else
45404#define vdupd_laneq_s64(__p0, __p1) __extension__ ({ \
45405  int64x2_t __s0 = __p0; \
45406  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45407  int64_t __ret; \
45408  __ret = (int64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__rev0, __p1); \
45409  __ret; \
45410})
45411#endif
45412
45413#ifdef __LITTLE_ENDIAN__
45414#define vduph_laneq_s16(__p0, __p1) __extension__ ({ \
45415  int16x8_t __s0 = __p0; \
45416  int16_t __ret; \
45417  __ret = (int16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
45418  __ret; \
45419})
45420#else
45421#define vduph_laneq_s16(__p0, __p1) __extension__ ({ \
45422  int16x8_t __s0 = __p0; \
45423  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45424  int16_t __ret; \
45425  __ret = (int16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
45426  __ret; \
45427})
45428#endif
45429
45430#ifdef __LITTLE_ENDIAN__
45431#define vdup_laneq_p8(__p0, __p1) __extension__ ({ \
45432  poly8x16_t __s0 = __p0; \
45433  poly8x8_t __ret; \
45434  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45435  __ret; \
45436})
45437#else
45438#define vdup_laneq_p8(__p0, __p1) __extension__ ({ \
45439  poly8x16_t __s0 = __p0; \
45440  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45441  poly8x8_t __ret; \
45442  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45443  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
45444  __ret; \
45445})
45446#endif
45447
45448#ifdef __LITTLE_ENDIAN__
45449#define vdup_laneq_p64(__p0, __p1) __extension__ ({ \
45450  poly64x2_t __s0 = __p0; \
45451  poly64x1_t __ret; \
45452  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
45453  __ret; \
45454})
45455#else
45456#define vdup_laneq_p64(__p0, __p1) __extension__ ({ \
45457  poly64x2_t __s0 = __p0; \
45458  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45459  poly64x1_t __ret; \
45460  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
45461  __ret; \
45462})
45463#endif
45464
45465#ifdef __LITTLE_ENDIAN__
45466#define vdup_laneq_p16(__p0, __p1) __extension__ ({ \
45467  poly16x8_t __s0 = __p0; \
45468  poly16x4_t __ret; \
45469  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
45470  __ret; \
45471})
45472#else
45473#define vdup_laneq_p16(__p0, __p1) __extension__ ({ \
45474  poly16x8_t __s0 = __p0; \
45475  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45476  poly16x4_t __ret; \
45477  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
45478  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
45479  __ret; \
45480})
45481#endif
45482
45483#ifdef __LITTLE_ENDIAN__
45484#define vdupq_laneq_p8(__p0, __p1) __extension__ ({ \
45485  poly8x16_t __s0 = __p0; \
45486  poly8x16_t __ret; \
45487  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45488  __ret; \
45489})
45490#else
45491#define vdupq_laneq_p8(__p0, __p1) __extension__ ({ \
45492  poly8x16_t __s0 = __p0; \
45493  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45494  poly8x16_t __ret; \
45495  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45496  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45497  __ret; \
45498})
45499#endif
45500
45501#ifdef __LITTLE_ENDIAN__
45502#define vdupq_laneq_p64(__p0, __p1) __extension__ ({ \
45503  poly64x2_t __s0 = __p0; \
45504  poly64x2_t __ret; \
45505  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
45506  __ret; \
45507})
45508#else
45509#define vdupq_laneq_p64(__p0, __p1) __extension__ ({ \
45510  poly64x2_t __s0 = __p0; \
45511  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45512  poly64x2_t __ret; \
45513  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
45514  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45515  __ret; \
45516})
45517#endif
45518
45519#ifdef __LITTLE_ENDIAN__
45520#define vdupq_laneq_p16(__p0, __p1) __extension__ ({ \
45521  poly16x8_t __s0 = __p0; \
45522  poly16x8_t __ret; \
45523  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45524  __ret; \
45525})
45526#else
45527#define vdupq_laneq_p16(__p0, __p1) __extension__ ({ \
45528  poly16x8_t __s0 = __p0; \
45529  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45530  poly16x8_t __ret; \
45531  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45532  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
45533  __ret; \
45534})
45535#endif
45536
45537#ifdef __LITTLE_ENDIAN__
45538#define vdupq_laneq_u8(__p0, __p1) __extension__ ({ \
45539  uint8x16_t __s0 = __p0; \
45540  uint8x16_t __ret; \
45541  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45542  __ret; \
45543})
45544#else
45545#define vdupq_laneq_u8(__p0, __p1) __extension__ ({ \
45546  uint8x16_t __s0 = __p0; \
45547  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45548  uint8x16_t __ret; \
45549  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45550  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45551  __ret; \
45552})
45553#endif
45554
45555#ifdef __LITTLE_ENDIAN__
45556#define vdupq_laneq_u32(__p0, __p1) __extension__ ({ \
45557  uint32x4_t __s0 = __p0; \
45558  uint32x4_t __ret; \
45559  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
45560  __ret; \
45561})
45562#else
45563#define vdupq_laneq_u32(__p0, __p1) __extension__ ({ \
45564  uint32x4_t __s0 = __p0; \
45565  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45566  uint32x4_t __ret; \
45567  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
45568  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
45569  __ret; \
45570})
45571#endif
45572
45573#ifdef __LITTLE_ENDIAN__
45574#define vdupq_laneq_u64(__p0, __p1) __extension__ ({ \
45575  uint64x2_t __s0 = __p0; \
45576  uint64x2_t __ret; \
45577  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
45578  __ret; \
45579})
45580#else
45581#define vdupq_laneq_u64(__p0, __p1) __extension__ ({ \
45582  uint64x2_t __s0 = __p0; \
45583  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45584  uint64x2_t __ret; \
45585  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
45586  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45587  __ret; \
45588})
45589#endif
45590
45591#ifdef __LITTLE_ENDIAN__
45592#define vdupq_laneq_u16(__p0, __p1) __extension__ ({ \
45593  uint16x8_t __s0 = __p0; \
45594  uint16x8_t __ret; \
45595  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45596  __ret; \
45597})
45598#else
45599#define vdupq_laneq_u16(__p0, __p1) __extension__ ({ \
45600  uint16x8_t __s0 = __p0; \
45601  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45602  uint16x8_t __ret; \
45603  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45604  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
45605  __ret; \
45606})
45607#endif
45608
45609#ifdef __LITTLE_ENDIAN__
45610#define vdupq_laneq_s8(__p0, __p1) __extension__ ({ \
45611  int8x16_t __s0 = __p0; \
45612  int8x16_t __ret; \
45613  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45614  __ret; \
45615})
45616#else
45617#define vdupq_laneq_s8(__p0, __p1) __extension__ ({ \
45618  int8x16_t __s0 = __p0; \
45619  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45620  int8x16_t __ret; \
45621  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45622  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45623  __ret; \
45624})
45625#endif
45626
45627#ifdef __LITTLE_ENDIAN__
45628#define vdupq_laneq_f64(__p0, __p1) __extension__ ({ \
45629  float64x2_t __s0 = __p0; \
45630  float64x2_t __ret; \
45631  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
45632  __ret; \
45633})
45634#else
45635#define vdupq_laneq_f64(__p0, __p1) __extension__ ({ \
45636  float64x2_t __s0 = __p0; \
45637  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45638  float64x2_t __ret; \
45639  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
45640  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45641  __ret; \
45642})
45643#endif
45644
45645#ifdef __LITTLE_ENDIAN__
45646#define vdupq_laneq_f32(__p0, __p1) __extension__ ({ \
45647  float32x4_t __s0 = __p0; \
45648  float32x4_t __ret; \
45649  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
45650  __ret; \
45651})
45652#else
45653#define vdupq_laneq_f32(__p0, __p1) __extension__ ({ \
45654  float32x4_t __s0 = __p0; \
45655  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45656  float32x4_t __ret; \
45657  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
45658  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
45659  __ret; \
45660})
45661#endif
45662
45663#ifdef __LITTLE_ENDIAN__
45664#define vdupq_laneq_f16(__p0, __p1) __extension__ ({ \
45665  float16x8_t __s0 = __p0; \
45666  float16x8_t __ret; \
45667  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45668  __ret; \
45669})
45670#else
45671#define vdupq_laneq_f16(__p0, __p1) __extension__ ({ \
45672  float16x8_t __s0 = __p0; \
45673  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45674  float16x8_t __ret; \
45675  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45676  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
45677  __ret; \
45678})
45679#endif
45680
45681#ifdef __LITTLE_ENDIAN__
45682#define vdupq_laneq_s32(__p0, __p1) __extension__ ({ \
45683  int32x4_t __s0 = __p0; \
45684  int32x4_t __ret; \
45685  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
45686  __ret; \
45687})
45688#else
45689#define vdupq_laneq_s32(__p0, __p1) __extension__ ({ \
45690  int32x4_t __s0 = __p0; \
45691  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45692  int32x4_t __ret; \
45693  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
45694  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
45695  __ret; \
45696})
45697#endif
45698
45699#ifdef __LITTLE_ENDIAN__
45700#define vdupq_laneq_s64(__p0, __p1) __extension__ ({ \
45701  int64x2_t __s0 = __p0; \
45702  int64x2_t __ret; \
45703  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
45704  __ret; \
45705})
45706#else
45707#define vdupq_laneq_s64(__p0, __p1) __extension__ ({ \
45708  int64x2_t __s0 = __p0; \
45709  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45710  int64x2_t __ret; \
45711  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
45712  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45713  __ret; \
45714})
45715#endif
45716
45717#ifdef __LITTLE_ENDIAN__
45718#define vdupq_laneq_s16(__p0, __p1) __extension__ ({ \
45719  int16x8_t __s0 = __p0; \
45720  int16x8_t __ret; \
45721  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45722  __ret; \
45723})
45724#else
45725#define vdupq_laneq_s16(__p0, __p1) __extension__ ({ \
45726  int16x8_t __s0 = __p0; \
45727  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45728  int16x8_t __ret; \
45729  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45730  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
45731  __ret; \
45732})
45733#endif
45734
45735#ifdef __LITTLE_ENDIAN__
45736#define vdup_laneq_u8(__p0, __p1) __extension__ ({ \
45737  uint8x16_t __s0 = __p0; \
45738  uint8x8_t __ret; \
45739  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45740  __ret; \
45741})
45742#else
45743#define vdup_laneq_u8(__p0, __p1) __extension__ ({ \
45744  uint8x16_t __s0 = __p0; \
45745  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45746  uint8x8_t __ret; \
45747  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45748  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
45749  __ret; \
45750})
45751#endif
45752
45753#ifdef __LITTLE_ENDIAN__
45754#define vdup_laneq_u32(__p0, __p1) __extension__ ({ \
45755  uint32x4_t __s0 = __p0; \
45756  uint32x2_t __ret; \
45757  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
45758  __ret; \
45759})
45760#else
45761#define vdup_laneq_u32(__p0, __p1) __extension__ ({ \
45762  uint32x4_t __s0 = __p0; \
45763  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45764  uint32x2_t __ret; \
45765  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
45766  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45767  __ret; \
45768})
45769#endif
45770
45771#ifdef __LITTLE_ENDIAN__
45772#define vdup_laneq_u64(__p0, __p1) __extension__ ({ \
45773  uint64x2_t __s0 = __p0; \
45774  uint64x1_t __ret; \
45775  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
45776  __ret; \
45777})
45778#else
45779#define vdup_laneq_u64(__p0, __p1) __extension__ ({ \
45780  uint64x2_t __s0 = __p0; \
45781  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45782  uint64x1_t __ret; \
45783  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
45784  __ret; \
45785})
45786#endif
45787
45788#ifdef __LITTLE_ENDIAN__
45789#define vdup_laneq_u16(__p0, __p1) __extension__ ({ \
45790  uint16x8_t __s0 = __p0; \
45791  uint16x4_t __ret; \
45792  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
45793  __ret; \
45794})
45795#else
45796#define vdup_laneq_u16(__p0, __p1) __extension__ ({ \
45797  uint16x8_t __s0 = __p0; \
45798  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45799  uint16x4_t __ret; \
45800  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
45801  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
45802  __ret; \
45803})
45804#endif
45805
45806#ifdef __LITTLE_ENDIAN__
45807#define vdup_laneq_s8(__p0, __p1) __extension__ ({ \
45808  int8x16_t __s0 = __p0; \
45809  int8x8_t __ret; \
45810  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45811  __ret; \
45812})
45813#else
45814#define vdup_laneq_s8(__p0, __p1) __extension__ ({ \
45815  int8x16_t __s0 = __p0; \
45816  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45817  int8x8_t __ret; \
45818  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
45819  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
45820  __ret; \
45821})
45822#endif
45823
45824#ifdef __LITTLE_ENDIAN__
45825#define vdup_laneq_f64(__p0, __p1) __extension__ ({ \
45826  float64x2_t __s0 = __p0; \
45827  float64x1_t __ret; \
45828  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
45829  __ret; \
45830})
45831#else
45832#define vdup_laneq_f64(__p0, __p1) __extension__ ({ \
45833  float64x2_t __s0 = __p0; \
45834  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45835  float64x1_t __ret; \
45836  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
45837  __ret; \
45838})
45839#endif
45840
45841#ifdef __LITTLE_ENDIAN__
45842#define vdup_laneq_f32(__p0, __p1) __extension__ ({ \
45843  float32x4_t __s0 = __p0; \
45844  float32x2_t __ret; \
45845  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
45846  __ret; \
45847})
45848#else
45849#define vdup_laneq_f32(__p0, __p1) __extension__ ({ \
45850  float32x4_t __s0 = __p0; \
45851  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45852  float32x2_t __ret; \
45853  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
45854  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45855  __ret; \
45856})
45857#endif
45858
45859#ifdef __LITTLE_ENDIAN__
45860#define vdup_laneq_f16(__p0, __p1) __extension__ ({ \
45861  float16x8_t __s0 = __p0; \
45862  float16x4_t __ret; \
45863  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
45864  __ret; \
45865})
45866#else
45867#define vdup_laneq_f16(__p0, __p1) __extension__ ({ \
45868  float16x8_t __s0 = __p0; \
45869  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45870  float16x4_t __ret; \
45871  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
45872  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
45873  __ret; \
45874})
45875#endif
45876
45877#ifdef __LITTLE_ENDIAN__
45878#define vdup_laneq_s32(__p0, __p1) __extension__ ({ \
45879  int32x4_t __s0 = __p0; \
45880  int32x2_t __ret; \
45881  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
45882  __ret; \
45883})
45884#else
45885#define vdup_laneq_s32(__p0, __p1) __extension__ ({ \
45886  int32x4_t __s0 = __p0; \
45887  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
45888  int32x2_t __ret; \
45889  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
45890  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45891  __ret; \
45892})
45893#endif
45894
45895#ifdef __LITTLE_ENDIAN__
45896#define vdup_laneq_s64(__p0, __p1) __extension__ ({ \
45897  int64x2_t __s0 = __p0; \
45898  int64x1_t __ret; \
45899  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
45900  __ret; \
45901})
45902#else
45903#define vdup_laneq_s64(__p0, __p1) __extension__ ({ \
45904  int64x2_t __s0 = __p0; \
45905  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45906  int64x1_t __ret; \
45907  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
45908  __ret; \
45909})
45910#endif
45911
45912#ifdef __LITTLE_ENDIAN__
45913#define vdup_laneq_s16(__p0, __p1) __extension__ ({ \
45914  int16x8_t __s0 = __p0; \
45915  int16x4_t __ret; \
45916  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
45917  __ret; \
45918})
45919#else
45920#define vdup_laneq_s16(__p0, __p1) __extension__ ({ \
45921  int16x8_t __s0 = __p0; \
45922  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
45923  int16x4_t __ret; \
45924  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
45925  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
45926  __ret; \
45927})
45928#endif
45929
45930#ifdef __LITTLE_ENDIAN__
45931__ai poly64x1_t vdup_n_p64(poly64_t __p0) {
45932  poly64x1_t __ret;
45933  __ret = (poly64x1_t) {__p0};
45934  return __ret;
45935}
45936#else
45937__ai poly64x1_t vdup_n_p64(poly64_t __p0) {
45938  poly64x1_t __ret;
45939  __ret = (poly64x1_t) {__p0};
45940  return __ret;
45941}
45942#endif
45943
45944#ifdef __LITTLE_ENDIAN__
45945__ai poly64x2_t vdupq_n_p64(poly64_t __p0) {
45946  poly64x2_t __ret;
45947  __ret = (poly64x2_t) {__p0, __p0};
45948  return __ret;
45949}
45950#else
45951__ai poly64x2_t vdupq_n_p64(poly64_t __p0) {
45952  poly64x2_t __ret;
45953  __ret = (poly64x2_t) {__p0, __p0};
45954  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45955  return __ret;
45956}
45957#endif
45958
45959#ifdef __LITTLE_ENDIAN__
45960__ai float64x2_t vdupq_n_f64(float64_t __p0) {
45961  float64x2_t __ret;
45962  __ret = (float64x2_t) {__p0, __p0};
45963  return __ret;
45964}
45965#else
45966__ai float64x2_t vdupq_n_f64(float64_t __p0) {
45967  float64x2_t __ret;
45968  __ret = (float64x2_t) {__p0, __p0};
45969  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45970  return __ret;
45971}
45972#endif
45973
45974#ifdef __LITTLE_ENDIAN__
45975__ai float64x1_t vdup_n_f64(float64_t __p0) {
45976  float64x1_t __ret;
45977  __ret = (float64x1_t) {__p0};
45978  return __ret;
45979}
45980#else
45981__ai float64x1_t vdup_n_f64(float64_t __p0) {
45982  float64x1_t __ret;
45983  __ret = (float64x1_t) {__p0};
45984  return __ret;
45985}
45986#endif
45987
45988#ifdef __LITTLE_ENDIAN__
45989#define vext_p64(__p0, __p1, __p2) __extension__ ({ \
45990  poly64x1_t __s0 = __p0; \
45991  poly64x1_t __s1 = __p1; \
45992  poly64x1_t __ret; \
45993  __ret = (poly64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
45994  __ret; \
45995})
45996#else
45997#define vext_p64(__p0, __p1, __p2) __extension__ ({ \
45998  poly64x1_t __s0 = __p0; \
45999  poly64x1_t __s1 = __p1; \
46000  poly64x1_t __ret; \
46001  __ret = (poly64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
46002  __ret; \
46003})
46004#endif
46005
46006#ifdef __LITTLE_ENDIAN__
46007#define vextq_p64(__p0, __p1, __p2) __extension__ ({ \
46008  poly64x2_t __s0 = __p0; \
46009  poly64x2_t __s1 = __p1; \
46010  poly64x2_t __ret; \
46011  __ret = (poly64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
46012  __ret; \
46013})
46014#else
46015#define vextq_p64(__p0, __p1, __p2) __extension__ ({ \
46016  poly64x2_t __s0 = __p0; \
46017  poly64x2_t __s1 = __p1; \
46018  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46019  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
46020  poly64x2_t __ret; \
46021  __ret = (poly64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
46022  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46023  __ret; \
46024})
46025#endif
46026
46027#ifdef __LITTLE_ENDIAN__
46028#define vextq_f64(__p0, __p1, __p2) __extension__ ({ \
46029  float64x2_t __s0 = __p0; \
46030  float64x2_t __s1 = __p1; \
46031  float64x2_t __ret; \
46032  __ret = (float64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 42); \
46033  __ret; \
46034})
46035#else
46036#define vextq_f64(__p0, __p1, __p2) __extension__ ({ \
46037  float64x2_t __s0 = __p0; \
46038  float64x2_t __s1 = __p1; \
46039  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46040  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
46041  float64x2_t __ret; \
46042  __ret = (float64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 42); \
46043  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46044  __ret; \
46045})
46046#endif
46047
46048#ifdef __LITTLE_ENDIAN__
46049#define vext_f64(__p0, __p1, __p2) __extension__ ({ \
46050  float64x1_t __s0 = __p0; \
46051  float64x1_t __s1 = __p1; \
46052  float64x1_t __ret; \
46053  __ret = (float64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
46054  __ret; \
46055})
46056#else
46057#define vext_f64(__p0, __p1, __p2) __extension__ ({ \
46058  float64x1_t __s0 = __p0; \
46059  float64x1_t __s1 = __p1; \
46060  float64x1_t __ret; \
46061  __ret = (float64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
46062  __ret; \
46063})
46064#endif
46065
46066#ifdef __LITTLE_ENDIAN__
46067__ai float64x2_t vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
46068  float64x2_t __ret;
46069  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
46070  return __ret;
46071}
46072#else
46073__ai float64x2_t vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
46074  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46075  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46076  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
46077  float64x2_t __ret;
46078  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
46079  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46080  return __ret;
46081}
46082__ai float64x2_t __noswap_vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
46083  float64x2_t __ret;
46084  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
46085  return __ret;
46086}
46087#endif
46088
46089#ifdef __LITTLE_ENDIAN__
46090__ai float64x1_t vfma_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
46091  float64x1_t __ret;
46092  __ret = (float64x1_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
46093  return __ret;
46094}
46095#else
46096__ai float64x1_t vfma_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
46097  float64x1_t __ret;
46098  __ret = (float64x1_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
46099  return __ret;
46100}
46101#endif
46102
46103#ifdef __LITTLE_ENDIAN__
46104#define vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46105  float64_t __s0 = __p0; \
46106  float64_t __s1 = __p1; \
46107  float64x1_t __s2 = __p2; \
46108  float64_t __ret; \
46109  __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
46110  __ret; \
46111})
46112#else
46113#define vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46114  float64_t __s0 = __p0; \
46115  float64_t __s1 = __p1; \
46116  float64x1_t __s2 = __p2; \
46117  float64_t __ret; \
46118  __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
46119  __ret; \
46120})
46121#define __noswap_vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46122  float64_t __s0 = __p0; \
46123  float64_t __s1 = __p1; \
46124  float64x1_t __s2 = __p2; \
46125  float64_t __ret; \
46126  __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
46127  __ret; \
46128})
46129#endif
46130
46131#ifdef __LITTLE_ENDIAN__
46132#define vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46133  float32_t __s0 = __p0; \
46134  float32_t __s1 = __p1; \
46135  float32x2_t __s2 = __p2; \
46136  float32_t __ret; \
46137  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__s2, __p3); \
46138  __ret; \
46139})
46140#else
46141#define vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46142  float32_t __s0 = __p0; \
46143  float32_t __s1 = __p1; \
46144  float32x2_t __s2 = __p2; \
46145  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
46146  float32_t __ret; \
46147  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__rev2, __p3); \
46148  __ret; \
46149})
46150#define __noswap_vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46151  float32_t __s0 = __p0; \
46152  float32_t __s1 = __p1; \
46153  float32x2_t __s2 = __p2; \
46154  float32_t __ret; \
46155  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__s2, __p3); \
46156  __ret; \
46157})
46158#endif
46159
46160#ifdef __LITTLE_ENDIAN__
46161#define vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46162  float64x2_t __s0 = __p0; \
46163  float64x2_t __s1 = __p1; \
46164  float64x1_t __s2 = __p2; \
46165  float64x2_t __ret; \
46166  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 42); \
46167  __ret; \
46168})
46169#else
46170#define vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46171  float64x2_t __s0 = __p0; \
46172  float64x2_t __s1 = __p1; \
46173  float64x1_t __s2 = __p2; \
46174  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46175  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
46176  float64x2_t __ret; \
46177  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__s2, __p3, 42); \
46178  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46179  __ret; \
46180})
46181#define __noswap_vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46182  float64x2_t __s0 = __p0; \
46183  float64x2_t __s1 = __p1; \
46184  float64x1_t __s2 = __p2; \
46185  float64x2_t __ret; \
46186  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 42); \
46187  __ret; \
46188})
46189#endif
46190
46191#ifdef __LITTLE_ENDIAN__
46192#define vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46193  float32x4_t __s0 = __p0; \
46194  float32x4_t __s1 = __p1; \
46195  float32x2_t __s2 = __p2; \
46196  float32x4_t __ret; \
46197  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 41); \
46198  __ret; \
46199})
46200#else
46201#define vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46202  float32x4_t __s0 = __p0; \
46203  float32x4_t __s1 = __p1; \
46204  float32x2_t __s2 = __p2; \
46205  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46206  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
46207  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
46208  float32x4_t __ret; \
46209  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, __p3, 41); \
46210  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46211  __ret; \
46212})
46213#define __noswap_vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46214  float32x4_t __s0 = __p0; \
46215  float32x4_t __s1 = __p1; \
46216  float32x2_t __s2 = __p2; \
46217  float32x4_t __ret; \
46218  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 41); \
46219  __ret; \
46220})
46221#endif
46222
46223#ifdef __LITTLE_ENDIAN__
46224#define vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46225  float64x1_t __s0 = __p0; \
46226  float64x1_t __s1 = __p1; \
46227  float64x1_t __s2 = __p2; \
46228  float64x1_t __ret; \
46229  __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
46230  __ret; \
46231})
46232#else
46233#define vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46234  float64x1_t __s0 = __p0; \
46235  float64x1_t __s1 = __p1; \
46236  float64x1_t __s2 = __p2; \
46237  float64x1_t __ret; \
46238  __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
46239  __ret; \
46240})
46241#define __noswap_vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46242  float64x1_t __s0 = __p0; \
46243  float64x1_t __s1 = __p1; \
46244  float64x1_t __s2 = __p2; \
46245  float64x1_t __ret; \
46246  __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
46247  __ret; \
46248})
46249#endif
46250
46251#ifdef __LITTLE_ENDIAN__
46252#define vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46253  float32x2_t __s0 = __p0; \
46254  float32x2_t __s1 = __p1; \
46255  float32x2_t __s2 = __p2; \
46256  float32x2_t __ret; \
46257  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 9); \
46258  __ret; \
46259})
46260#else
46261#define vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46262  float32x2_t __s0 = __p0; \
46263  float32x2_t __s1 = __p1; \
46264  float32x2_t __s2 = __p2; \
46265  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46266  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
46267  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
46268  float32x2_t __ret; \
46269  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, __p3, 9); \
46270  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46271  __ret; \
46272})
46273#define __noswap_vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46274  float32x2_t __s0 = __p0; \
46275  float32x2_t __s1 = __p1; \
46276  float32x2_t __s2 = __p2; \
46277  float32x2_t __ret; \
46278  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 9); \
46279  __ret; \
46280})
46281#endif
46282
46283#ifdef __LITTLE_ENDIAN__
46284#define vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46285  float64_t __s0 = __p0; \
46286  float64_t __s1 = __p1; \
46287  float64x2_t __s2 = __p2; \
46288  float64_t __ret; \
46289  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__s2, __p3); \
46290  __ret; \
46291})
46292#else
46293#define vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46294  float64_t __s0 = __p0; \
46295  float64_t __s1 = __p1; \
46296  float64x2_t __s2 = __p2; \
46297  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
46298  float64_t __ret; \
46299  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__rev2, __p3); \
46300  __ret; \
46301})
46302#define __noswap_vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46303  float64_t __s0 = __p0; \
46304  float64_t __s1 = __p1; \
46305  float64x2_t __s2 = __p2; \
46306  float64_t __ret; \
46307  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__s2, __p3); \
46308  __ret; \
46309})
46310#endif
46311
46312#ifdef __LITTLE_ENDIAN__
46313#define vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46314  float32_t __s0 = __p0; \
46315  float32_t __s1 = __p1; \
46316  float32x4_t __s2 = __p2; \
46317  float32_t __ret; \
46318  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__s2, __p3); \
46319  __ret; \
46320})
46321#else
46322#define vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46323  float32_t __s0 = __p0; \
46324  float32_t __s1 = __p1; \
46325  float32x4_t __s2 = __p2; \
46326  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
46327  float32_t __ret; \
46328  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__rev2, __p3); \
46329  __ret; \
46330})
46331#define __noswap_vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46332  float32_t __s0 = __p0; \
46333  float32_t __s1 = __p1; \
46334  float32x4_t __s2 = __p2; \
46335  float32_t __ret; \
46336  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__s2, __p3); \
46337  __ret; \
46338})
46339#endif
46340
46341#ifdef __LITTLE_ENDIAN__
46342#define vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46343  float64x2_t __s0 = __p0; \
46344  float64x2_t __s1 = __p1; \
46345  float64x2_t __s2 = __p2; \
46346  float64x2_t __ret; \
46347  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 42); \
46348  __ret; \
46349})
46350#else
46351#define vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46352  float64x2_t __s0 = __p0; \
46353  float64x2_t __s1 = __p1; \
46354  float64x2_t __s2 = __p2; \
46355  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46356  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
46357  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
46358  float64x2_t __ret; \
46359  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 42); \
46360  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46361  __ret; \
46362})
46363#define __noswap_vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46364  float64x2_t __s0 = __p0; \
46365  float64x2_t __s1 = __p1; \
46366  float64x2_t __s2 = __p2; \
46367  float64x2_t __ret; \
46368  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 42); \
46369  __ret; \
46370})
46371#endif
46372
46373#ifdef __LITTLE_ENDIAN__
46374#define vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46375  float32x4_t __s0 = __p0; \
46376  float32x4_t __s1 = __p1; \
46377  float32x4_t __s2 = __p2; \
46378  float32x4_t __ret; \
46379  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 41); \
46380  __ret; \
46381})
46382#else
46383#define vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46384  float32x4_t __s0 = __p0; \
46385  float32x4_t __s1 = __p1; \
46386  float32x4_t __s2 = __p2; \
46387  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46388  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
46389  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
46390  float32x4_t __ret; \
46391  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 41); \
46392  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46393  __ret; \
46394})
46395#define __noswap_vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46396  float32x4_t __s0 = __p0; \
46397  float32x4_t __s1 = __p1; \
46398  float32x4_t __s2 = __p2; \
46399  float32x4_t __ret; \
46400  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 41); \
46401  __ret; \
46402})
46403#endif
46404
46405#ifdef __LITTLE_ENDIAN__
46406#define vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46407  float64x1_t __s0 = __p0; \
46408  float64x1_t __s1 = __p1; \
46409  float64x2_t __s2 = __p2; \
46410  float64x1_t __ret; \
46411  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 10); \
46412  __ret; \
46413})
46414#else
46415#define vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46416  float64x1_t __s0 = __p0; \
46417  float64x1_t __s1 = __p1; \
46418  float64x2_t __s2 = __p2; \
46419  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
46420  float64x1_t __ret; \
46421  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__rev2, __p3, 10); \
46422  __ret; \
46423})
46424#define __noswap_vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
46425  float64x1_t __s0 = __p0; \
46426  float64x1_t __s1 = __p1; \
46427  float64x2_t __s2 = __p2; \
46428  float64x1_t __ret; \
46429  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 10); \
46430  __ret; \
46431})
46432#endif
46433
46434#ifdef __LITTLE_ENDIAN__
46435#define vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46436  float32x2_t __s0 = __p0; \
46437  float32x2_t __s1 = __p1; \
46438  float32x4_t __s2 = __p2; \
46439  float32x2_t __ret; \
46440  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 9); \
46441  __ret; \
46442})
46443#else
46444#define vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46445  float32x2_t __s0 = __p0; \
46446  float32x2_t __s1 = __p1; \
46447  float32x4_t __s2 = __p2; \
46448  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46449  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
46450  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
46451  float32x2_t __ret; \
46452  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x16_t)__rev2, __p3, 9); \
46453  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46454  __ret; \
46455})
46456#define __noswap_vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
46457  float32x2_t __s0 = __p0; \
46458  float32x2_t __s1 = __p1; \
46459  float32x4_t __s2 = __p2; \
46460  float32x2_t __ret; \
46461  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 9); \
46462  __ret; \
46463})
46464#endif
46465
46466#ifdef __LITTLE_ENDIAN__
46467__ai float64x2_t vfmaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
46468  float64x2_t __ret;
46469  __ret = vfmaq_f64(__p0, __p1, (float64x2_t) {__p2, __p2});
46470  return __ret;
46471}
46472#else
46473__ai float64x2_t vfmaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
46474  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46475  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46476  float64x2_t __ret;
46477  __ret = __noswap_vfmaq_f64(__rev0, __rev1, (float64x2_t) {__p2, __p2});
46478  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46479  return __ret;
46480}
46481#endif
46482
46483#ifdef __LITTLE_ENDIAN__
46484__ai float32x4_t vfmaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
46485  float32x4_t __ret;
46486  __ret = vfmaq_f32(__p0, __p1, (float32x4_t) {__p2, __p2, __p2, __p2});
46487  return __ret;
46488}
46489#else
46490__ai float32x4_t vfmaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
46491  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46492  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
46493  float32x4_t __ret;
46494  __ret = __noswap_vfmaq_f32(__rev0, __rev1, (float32x4_t) {__p2, __p2, __p2, __p2});
46495  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46496  return __ret;
46497}
46498#endif
46499
46500#ifdef __LITTLE_ENDIAN__
46501__ai float32x2_t vfma_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
46502  float32x2_t __ret;
46503  __ret = vfma_f32(__p0, __p1, (float32x2_t) {__p2, __p2});
46504  return __ret;
46505}
46506#else
46507__ai float32x2_t vfma_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
46508  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46509  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46510  float32x2_t __ret;
46511  __ret = __noswap_vfma_f32(__rev0, __rev1, (float32x2_t) {__p2, __p2});
46512  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46513  return __ret;
46514}
46515#endif
46516
46517#ifdef __LITTLE_ENDIAN__
46518__ai float64x2_t vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
46519  float64x2_t __ret;
46520  __ret = (float64x2_t) __builtin_neon_vfmsq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
46521  return __ret;
46522}
46523#else
46524__ai float64x2_t vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
46525  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46526  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46527  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
46528  float64x2_t __ret;
46529  __ret = (float64x2_t) __builtin_neon_vfmsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
46530  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46531  return __ret;
46532}
46533__ai float64x2_t __noswap_vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
46534  float64x2_t __ret;
46535  __ret = (float64x2_t) __builtin_neon_vfmsq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
46536  return __ret;
46537}
46538#endif
46539
46540#ifdef __LITTLE_ENDIAN__
46541__ai float32x4_t vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
46542  float32x4_t __ret;
46543  __ret = (float32x4_t) __builtin_neon_vfmsq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
46544  return __ret;
46545}
46546#else
46547__ai float32x4_t vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
46548  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46549  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
46550  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
46551  float32x4_t __ret;
46552  __ret = (float32x4_t) __builtin_neon_vfmsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
46553  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46554  return __ret;
46555}
46556__ai float32x4_t __noswap_vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
46557  float32x4_t __ret;
46558  __ret = (float32x4_t) __builtin_neon_vfmsq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
46559  return __ret;
46560}
46561#endif
46562
46563#ifdef __LITTLE_ENDIAN__
46564__ai float64x1_t vfms_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
46565  float64x1_t __ret;
46566  __ret = (float64x1_t) __builtin_neon_vfms_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
46567  return __ret;
46568}
46569#else
46570__ai float64x1_t vfms_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
46571  float64x1_t __ret;
46572  __ret = (float64x1_t) __builtin_neon_vfms_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
46573  return __ret;
46574}
46575#endif
46576
46577#ifdef __LITTLE_ENDIAN__
46578__ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
46579  float32x2_t __ret;
46580  __ret = (float32x2_t) __builtin_neon_vfms_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
46581  return __ret;
46582}
46583#else
46584__ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
46585  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46586  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46587  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
46588  float32x2_t __ret;
46589  __ret = (float32x2_t) __builtin_neon_vfms_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
46590  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46591  return __ret;
46592}
46593__ai float32x2_t __noswap_vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
46594  float32x2_t __ret;
46595  __ret = (float32x2_t) __builtin_neon_vfms_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
46596  return __ret;
46597}
46598#endif
46599
46600#ifdef __LITTLE_ENDIAN__
46601#define vfmsd_lane_f64(__p0_88, __p1_88, __p2_88, __p3_88) __extension__ ({ \
46602  float64_t __s0_88 = __p0_88; \
46603  float64_t __s1_88 = __p1_88; \
46604  float64x1_t __s2_88 = __p2_88; \
46605  float64_t __ret_88; \
46606  __ret_88 = vfmad_lane_f64(__s0_88, __s1_88, -__s2_88, __p3_88); \
46607  __ret_88; \
46608})
46609#else
46610#define vfmsd_lane_f64(__p0_89, __p1_89, __p2_89, __p3_89) __extension__ ({ \
46611  float64_t __s0_89 = __p0_89; \
46612  float64_t __s1_89 = __p1_89; \
46613  float64x1_t __s2_89 = __p2_89; \
46614  float64_t __ret_89; \
46615  __ret_89 = __noswap_vfmad_lane_f64(__s0_89, __s1_89, -__s2_89, __p3_89); \
46616  __ret_89; \
46617})
46618#endif
46619
46620#ifdef __LITTLE_ENDIAN__
46621#define vfmss_lane_f32(__p0_90, __p1_90, __p2_90, __p3_90) __extension__ ({ \
46622  float32_t __s0_90 = __p0_90; \
46623  float32_t __s1_90 = __p1_90; \
46624  float32x2_t __s2_90 = __p2_90; \
46625  float32_t __ret_90; \
46626  __ret_90 = vfmas_lane_f32(__s0_90, __s1_90, -__s2_90, __p3_90); \
46627  __ret_90; \
46628})
46629#else
46630#define vfmss_lane_f32(__p0_91, __p1_91, __p2_91, __p3_91) __extension__ ({ \
46631  float32_t __s0_91 = __p0_91; \
46632  float32_t __s1_91 = __p1_91; \
46633  float32x2_t __s2_91 = __p2_91; \
46634  float32x2_t __rev2_91;  __rev2_91 = __builtin_shufflevector(__s2_91, __s2_91, 1, 0); \
46635  float32_t __ret_91; \
46636  __ret_91 = __noswap_vfmas_lane_f32(__s0_91, __s1_91, -__rev2_91, __p3_91); \
46637  __ret_91; \
46638})
46639#endif
46640
46641#ifdef __LITTLE_ENDIAN__
46642#define vfmsq_lane_f64(__p0_92, __p1_92, __p2_92, __p3_92) __extension__ ({ \
46643  float64x2_t __s0_92 = __p0_92; \
46644  float64x2_t __s1_92 = __p1_92; \
46645  float64x1_t __s2_92 = __p2_92; \
46646  float64x2_t __ret_92; \
46647  __ret_92 = vfmaq_lane_f64(__s0_92, __s1_92, -__s2_92, __p3_92); \
46648  __ret_92; \
46649})
46650#else
46651#define vfmsq_lane_f64(__p0_93, __p1_93, __p2_93, __p3_93) __extension__ ({ \
46652  float64x2_t __s0_93 = __p0_93; \
46653  float64x2_t __s1_93 = __p1_93; \
46654  float64x1_t __s2_93 = __p2_93; \
46655  float64x2_t __rev0_93;  __rev0_93 = __builtin_shufflevector(__s0_93, __s0_93, 1, 0); \
46656  float64x2_t __rev1_93;  __rev1_93 = __builtin_shufflevector(__s1_93, __s1_93, 1, 0); \
46657  float64x2_t __ret_93; \
46658  __ret_93 = __noswap_vfmaq_lane_f64(__rev0_93, __rev1_93, -__s2_93, __p3_93); \
46659  __ret_93 = __builtin_shufflevector(__ret_93, __ret_93, 1, 0); \
46660  __ret_93; \
46661})
46662#endif
46663
46664#ifdef __LITTLE_ENDIAN__
46665#define vfmsq_lane_f32(__p0_94, __p1_94, __p2_94, __p3_94) __extension__ ({ \
46666  float32x4_t __s0_94 = __p0_94; \
46667  float32x4_t __s1_94 = __p1_94; \
46668  float32x2_t __s2_94 = __p2_94; \
46669  float32x4_t __ret_94; \
46670  __ret_94 = vfmaq_lane_f32(__s0_94, __s1_94, -__s2_94, __p3_94); \
46671  __ret_94; \
46672})
46673#else
46674#define vfmsq_lane_f32(__p0_95, __p1_95, __p2_95, __p3_95) __extension__ ({ \
46675  float32x4_t __s0_95 = __p0_95; \
46676  float32x4_t __s1_95 = __p1_95; \
46677  float32x2_t __s2_95 = __p2_95; \
46678  float32x4_t __rev0_95;  __rev0_95 = __builtin_shufflevector(__s0_95, __s0_95, 3, 2, 1, 0); \
46679  float32x4_t __rev1_95;  __rev1_95 = __builtin_shufflevector(__s1_95, __s1_95, 3, 2, 1, 0); \
46680  float32x2_t __rev2_95;  __rev2_95 = __builtin_shufflevector(__s2_95, __s2_95, 1, 0); \
46681  float32x4_t __ret_95; \
46682  __ret_95 = __noswap_vfmaq_lane_f32(__rev0_95, __rev1_95, -__rev2_95, __p3_95); \
46683  __ret_95 = __builtin_shufflevector(__ret_95, __ret_95, 3, 2, 1, 0); \
46684  __ret_95; \
46685})
46686#endif
46687
46688#ifdef __LITTLE_ENDIAN__
46689#define vfms_lane_f64(__p0_96, __p1_96, __p2_96, __p3_96) __extension__ ({ \
46690  float64x1_t __s0_96 = __p0_96; \
46691  float64x1_t __s1_96 = __p1_96; \
46692  float64x1_t __s2_96 = __p2_96; \
46693  float64x1_t __ret_96; \
46694  __ret_96 = vfma_lane_f64(__s0_96, __s1_96, -__s2_96, __p3_96); \
46695  __ret_96; \
46696})
46697#else
46698#define vfms_lane_f64(__p0_97, __p1_97, __p2_97, __p3_97) __extension__ ({ \
46699  float64x1_t __s0_97 = __p0_97; \
46700  float64x1_t __s1_97 = __p1_97; \
46701  float64x1_t __s2_97 = __p2_97; \
46702  float64x1_t __ret_97; \
46703  __ret_97 = __noswap_vfma_lane_f64(__s0_97, __s1_97, -__s2_97, __p3_97); \
46704  __ret_97; \
46705})
46706#endif
46707
46708#ifdef __LITTLE_ENDIAN__
46709#define vfms_lane_f32(__p0_98, __p1_98, __p2_98, __p3_98) __extension__ ({ \
46710  float32x2_t __s0_98 = __p0_98; \
46711  float32x2_t __s1_98 = __p1_98; \
46712  float32x2_t __s2_98 = __p2_98; \
46713  float32x2_t __ret_98; \
46714  __ret_98 = vfma_lane_f32(__s0_98, __s1_98, -__s2_98, __p3_98); \
46715  __ret_98; \
46716})
46717#else
46718#define vfms_lane_f32(__p0_99, __p1_99, __p2_99, __p3_99) __extension__ ({ \
46719  float32x2_t __s0_99 = __p0_99; \
46720  float32x2_t __s1_99 = __p1_99; \
46721  float32x2_t __s2_99 = __p2_99; \
46722  float32x2_t __rev0_99;  __rev0_99 = __builtin_shufflevector(__s0_99, __s0_99, 1, 0); \
46723  float32x2_t __rev1_99;  __rev1_99 = __builtin_shufflevector(__s1_99, __s1_99, 1, 0); \
46724  float32x2_t __rev2_99;  __rev2_99 = __builtin_shufflevector(__s2_99, __s2_99, 1, 0); \
46725  float32x2_t __ret_99; \
46726  __ret_99 = __noswap_vfma_lane_f32(__rev0_99, __rev1_99, -__rev2_99, __p3_99); \
46727  __ret_99 = __builtin_shufflevector(__ret_99, __ret_99, 1, 0); \
46728  __ret_99; \
46729})
46730#endif
46731
46732#ifdef __LITTLE_ENDIAN__
46733#define vfmsd_laneq_f64(__p0_100, __p1_100, __p2_100, __p3_100) __extension__ ({ \
46734  float64_t __s0_100 = __p0_100; \
46735  float64_t __s1_100 = __p1_100; \
46736  float64x2_t __s2_100 = __p2_100; \
46737  float64_t __ret_100; \
46738  __ret_100 = vfmad_laneq_f64(__s0_100, __s1_100, -__s2_100, __p3_100); \
46739  __ret_100; \
46740})
46741#else
46742#define vfmsd_laneq_f64(__p0_101, __p1_101, __p2_101, __p3_101) __extension__ ({ \
46743  float64_t __s0_101 = __p0_101; \
46744  float64_t __s1_101 = __p1_101; \
46745  float64x2_t __s2_101 = __p2_101; \
46746  float64x2_t __rev2_101;  __rev2_101 = __builtin_shufflevector(__s2_101, __s2_101, 1, 0); \
46747  float64_t __ret_101; \
46748  __ret_101 = __noswap_vfmad_laneq_f64(__s0_101, __s1_101, -__rev2_101, __p3_101); \
46749  __ret_101; \
46750})
46751#endif
46752
46753#ifdef __LITTLE_ENDIAN__
46754#define vfmss_laneq_f32(__p0_102, __p1_102, __p2_102, __p3_102) __extension__ ({ \
46755  float32_t __s0_102 = __p0_102; \
46756  float32_t __s1_102 = __p1_102; \
46757  float32x4_t __s2_102 = __p2_102; \
46758  float32_t __ret_102; \
46759  __ret_102 = vfmas_laneq_f32(__s0_102, __s1_102, -__s2_102, __p3_102); \
46760  __ret_102; \
46761})
46762#else
46763#define vfmss_laneq_f32(__p0_103, __p1_103, __p2_103, __p3_103) __extension__ ({ \
46764  float32_t __s0_103 = __p0_103; \
46765  float32_t __s1_103 = __p1_103; \
46766  float32x4_t __s2_103 = __p2_103; \
46767  float32x4_t __rev2_103;  __rev2_103 = __builtin_shufflevector(__s2_103, __s2_103, 3, 2, 1, 0); \
46768  float32_t __ret_103; \
46769  __ret_103 = __noswap_vfmas_laneq_f32(__s0_103, __s1_103, -__rev2_103, __p3_103); \
46770  __ret_103; \
46771})
46772#endif
46773
46774#ifdef __LITTLE_ENDIAN__
46775#define vfmsq_laneq_f64(__p0_104, __p1_104, __p2_104, __p3_104) __extension__ ({ \
46776  float64x2_t __s0_104 = __p0_104; \
46777  float64x2_t __s1_104 = __p1_104; \
46778  float64x2_t __s2_104 = __p2_104; \
46779  float64x2_t __ret_104; \
46780  __ret_104 = vfmaq_laneq_f64(__s0_104, __s1_104, -__s2_104, __p3_104); \
46781  __ret_104; \
46782})
46783#else
46784#define vfmsq_laneq_f64(__p0_105, __p1_105, __p2_105, __p3_105) __extension__ ({ \
46785  float64x2_t __s0_105 = __p0_105; \
46786  float64x2_t __s1_105 = __p1_105; \
46787  float64x2_t __s2_105 = __p2_105; \
46788  float64x2_t __rev0_105;  __rev0_105 = __builtin_shufflevector(__s0_105, __s0_105, 1, 0); \
46789  float64x2_t __rev1_105;  __rev1_105 = __builtin_shufflevector(__s1_105, __s1_105, 1, 0); \
46790  float64x2_t __rev2_105;  __rev2_105 = __builtin_shufflevector(__s2_105, __s2_105, 1, 0); \
46791  float64x2_t __ret_105; \
46792  __ret_105 = __noswap_vfmaq_laneq_f64(__rev0_105, __rev1_105, -__rev2_105, __p3_105); \
46793  __ret_105 = __builtin_shufflevector(__ret_105, __ret_105, 1, 0); \
46794  __ret_105; \
46795})
46796#endif
46797
46798#ifdef __LITTLE_ENDIAN__
46799#define vfmsq_laneq_f32(__p0_106, __p1_106, __p2_106, __p3_106) __extension__ ({ \
46800  float32x4_t __s0_106 = __p0_106; \
46801  float32x4_t __s1_106 = __p1_106; \
46802  float32x4_t __s2_106 = __p2_106; \
46803  float32x4_t __ret_106; \
46804  __ret_106 = vfmaq_laneq_f32(__s0_106, __s1_106, -__s2_106, __p3_106); \
46805  __ret_106; \
46806})
46807#else
46808#define vfmsq_laneq_f32(__p0_107, __p1_107, __p2_107, __p3_107) __extension__ ({ \
46809  float32x4_t __s0_107 = __p0_107; \
46810  float32x4_t __s1_107 = __p1_107; \
46811  float32x4_t __s2_107 = __p2_107; \
46812  float32x4_t __rev0_107;  __rev0_107 = __builtin_shufflevector(__s0_107, __s0_107, 3, 2, 1, 0); \
46813  float32x4_t __rev1_107;  __rev1_107 = __builtin_shufflevector(__s1_107, __s1_107, 3, 2, 1, 0); \
46814  float32x4_t __rev2_107;  __rev2_107 = __builtin_shufflevector(__s2_107, __s2_107, 3, 2, 1, 0); \
46815  float32x4_t __ret_107; \
46816  __ret_107 = __noswap_vfmaq_laneq_f32(__rev0_107, __rev1_107, -__rev2_107, __p3_107); \
46817  __ret_107 = __builtin_shufflevector(__ret_107, __ret_107, 3, 2, 1, 0); \
46818  __ret_107; \
46819})
46820#endif
46821
46822#ifdef __LITTLE_ENDIAN__
46823#define vfms_laneq_f64(__p0_108, __p1_108, __p2_108, __p3_108) __extension__ ({ \
46824  float64x1_t __s0_108 = __p0_108; \
46825  float64x1_t __s1_108 = __p1_108; \
46826  float64x2_t __s2_108 = __p2_108; \
46827  float64x1_t __ret_108; \
46828  __ret_108 = vfma_laneq_f64(__s0_108, __s1_108, -__s2_108, __p3_108); \
46829  __ret_108; \
46830})
46831#else
46832#define vfms_laneq_f64(__p0_109, __p1_109, __p2_109, __p3_109) __extension__ ({ \
46833  float64x1_t __s0_109 = __p0_109; \
46834  float64x1_t __s1_109 = __p1_109; \
46835  float64x2_t __s2_109 = __p2_109; \
46836  float64x2_t __rev2_109;  __rev2_109 = __builtin_shufflevector(__s2_109, __s2_109, 1, 0); \
46837  float64x1_t __ret_109; \
46838  __ret_109 = __noswap_vfma_laneq_f64(__s0_109, __s1_109, -__rev2_109, __p3_109); \
46839  __ret_109; \
46840})
46841#endif
46842
46843#ifdef __LITTLE_ENDIAN__
46844#define vfms_laneq_f32(__p0_110, __p1_110, __p2_110, __p3_110) __extension__ ({ \
46845  float32x2_t __s0_110 = __p0_110; \
46846  float32x2_t __s1_110 = __p1_110; \
46847  float32x4_t __s2_110 = __p2_110; \
46848  float32x2_t __ret_110; \
46849  __ret_110 = vfma_laneq_f32(__s0_110, __s1_110, -__s2_110, __p3_110); \
46850  __ret_110; \
46851})
46852#else
46853#define vfms_laneq_f32(__p0_111, __p1_111, __p2_111, __p3_111) __extension__ ({ \
46854  float32x2_t __s0_111 = __p0_111; \
46855  float32x2_t __s1_111 = __p1_111; \
46856  float32x4_t __s2_111 = __p2_111; \
46857  float32x2_t __rev0_111;  __rev0_111 = __builtin_shufflevector(__s0_111, __s0_111, 1, 0); \
46858  float32x2_t __rev1_111;  __rev1_111 = __builtin_shufflevector(__s1_111, __s1_111, 1, 0); \
46859  float32x4_t __rev2_111;  __rev2_111 = __builtin_shufflevector(__s2_111, __s2_111, 3, 2, 1, 0); \
46860  float32x2_t __ret_111; \
46861  __ret_111 = __noswap_vfma_laneq_f32(__rev0_111, __rev1_111, -__rev2_111, __p3_111); \
46862  __ret_111 = __builtin_shufflevector(__ret_111, __ret_111, 1, 0); \
46863  __ret_111; \
46864})
46865#endif
46866
46867#ifdef __LITTLE_ENDIAN__
46868__ai float64x2_t vfmsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
46869  float64x2_t __ret;
46870  __ret = vfmsq_f64(__p0, __p1, (float64x2_t) {__p2, __p2});
46871  return __ret;
46872}
46873#else
46874__ai float64x2_t vfmsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
46875  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46876  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46877  float64x2_t __ret;
46878  __ret = __noswap_vfmsq_f64(__rev0, __rev1, (float64x2_t) {__p2, __p2});
46879  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46880  return __ret;
46881}
46882#endif
46883
46884#ifdef __LITTLE_ENDIAN__
46885__ai float32x4_t vfmsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
46886  float32x4_t __ret;
46887  __ret = vfmsq_f32(__p0, __p1, (float32x4_t) {__p2, __p2, __p2, __p2});
46888  return __ret;
46889}
46890#else
46891__ai float32x4_t vfmsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
46892  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46893  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
46894  float32x4_t __ret;
46895  __ret = __noswap_vfmsq_f32(__rev0, __rev1, (float32x4_t) {__p2, __p2, __p2, __p2});
46896  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46897  return __ret;
46898}
46899#endif
46900
46901#ifdef __LITTLE_ENDIAN__
46902__ai float32x2_t vfms_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
46903  float32x2_t __ret;
46904  __ret = vfms_f32(__p0, __p1, (float32x2_t) {__p2, __p2});
46905  return __ret;
46906}
46907#else
46908__ai float32x2_t vfms_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
46909  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46910  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46911  float32x2_t __ret;
46912  __ret = __noswap_vfms_f32(__rev0, __rev1, (float32x2_t) {__p2, __p2});
46913  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46914  return __ret;
46915}
46916#endif
46917
46918#ifdef __LITTLE_ENDIAN__
46919__ai poly64x1_t vget_high_p64(poly64x2_t __p0) {
46920  poly64x1_t __ret;
46921  __ret = __builtin_shufflevector(__p0, __p0, 1);
46922  return __ret;
46923}
46924#else
46925__ai poly64x1_t vget_high_p64(poly64x2_t __p0) {
46926  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46927  poly64x1_t __ret;
46928  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
46929  return __ret;
46930}
46931__ai poly64x1_t __noswap_vget_high_p64(poly64x2_t __p0) {
46932  poly64x1_t __ret;
46933  __ret = __builtin_shufflevector(__p0, __p0, 1);
46934  return __ret;
46935}
46936#endif
46937
46938#ifdef __LITTLE_ENDIAN__
46939__ai float64x1_t vget_high_f64(float64x2_t __p0) {
46940  float64x1_t __ret;
46941  __ret = __builtin_shufflevector(__p0, __p0, 1);
46942  return __ret;
46943}
46944#else
46945__ai float64x1_t vget_high_f64(float64x2_t __p0) {
46946  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46947  float64x1_t __ret;
46948  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
46949  return __ret;
46950}
46951#endif
46952
46953#ifdef __LITTLE_ENDIAN__
46954#define vget_lane_p64(__p0, __p1) __extension__ ({ \
46955  poly64x1_t __s0 = __p0; \
46956  poly64_t __ret; \
46957  __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
46958  __ret; \
46959})
46960#else
46961#define vget_lane_p64(__p0, __p1) __extension__ ({ \
46962  poly64x1_t __s0 = __p0; \
46963  poly64_t __ret; \
46964  __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
46965  __ret; \
46966})
46967#define __noswap_vget_lane_p64(__p0, __p1) __extension__ ({ \
46968  poly64x1_t __s0 = __p0; \
46969  poly64_t __ret; \
46970  __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
46971  __ret; \
46972})
46973#endif
46974
46975#ifdef __LITTLE_ENDIAN__
46976#define vgetq_lane_p64(__p0, __p1) __extension__ ({ \
46977  poly64x2_t __s0 = __p0; \
46978  poly64_t __ret; \
46979  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
46980  __ret; \
46981})
46982#else
46983#define vgetq_lane_p64(__p0, __p1) __extension__ ({ \
46984  poly64x2_t __s0 = __p0; \
46985  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46986  poly64_t __ret; \
46987  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
46988  __ret; \
46989})
46990#define __noswap_vgetq_lane_p64(__p0, __p1) __extension__ ({ \
46991  poly64x2_t __s0 = __p0; \
46992  poly64_t __ret; \
46993  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
46994  __ret; \
46995})
46996#endif
46997
46998#ifdef __LITTLE_ENDIAN__
46999#define vgetq_lane_f64(__p0, __p1) __extension__ ({ \
47000  float64x2_t __s0 = __p0; \
47001  float64_t __ret; \
47002  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__s0, __p1); \
47003  __ret; \
47004})
47005#else
47006#define vgetq_lane_f64(__p0, __p1) __extension__ ({ \
47007  float64x2_t __s0 = __p0; \
47008  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47009  float64_t __ret; \
47010  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__rev0, __p1); \
47011  __ret; \
47012})
47013#define __noswap_vgetq_lane_f64(__p0, __p1) __extension__ ({ \
47014  float64x2_t __s0 = __p0; \
47015  float64_t __ret; \
47016  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__s0, __p1); \
47017  __ret; \
47018})
47019#endif
47020
47021#ifdef __LITTLE_ENDIAN__
47022#define vget_lane_f64(__p0, __p1) __extension__ ({ \
47023  float64x1_t __s0 = __p0; \
47024  float64_t __ret; \
47025  __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
47026  __ret; \
47027})
47028#else
47029#define vget_lane_f64(__p0, __p1) __extension__ ({ \
47030  float64x1_t __s0 = __p0; \
47031  float64_t __ret; \
47032  __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
47033  __ret; \
47034})
47035#define __noswap_vget_lane_f64(__p0, __p1) __extension__ ({ \
47036  float64x1_t __s0 = __p0; \
47037  float64_t __ret; \
47038  __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
47039  __ret; \
47040})
47041#endif
47042
47043#ifdef __LITTLE_ENDIAN__
47044#define vget_lane_f16(__p0_112, __p1_112) __extension__ ({ \
47045  float16x4_t __s0_112 = __p0_112; \
47046  float16_t __ret_112; \
47047float16x4_t __reint_112 = __s0_112; \
47048int16_t __reint1_112 = vget_lane_s16(*(int16x4_t *) &__reint_112, __p1_112); \
47049  __ret_112 = *(float16_t *) &__reint1_112; \
47050  __ret_112; \
47051})
47052#else
47053#define vget_lane_f16(__p0_113, __p1_113) __extension__ ({ \
47054  float16x4_t __s0_113 = __p0_113; \
47055  float16x4_t __rev0_113;  __rev0_113 = __builtin_shufflevector(__s0_113, __s0_113, 3, 2, 1, 0); \
47056  float16_t __ret_113; \
47057float16x4_t __reint_113 = __rev0_113; \
47058int16_t __reint1_113 = __noswap_vget_lane_s16(*(int16x4_t *) &__reint_113, __p1_113); \
47059  __ret_113 = *(float16_t *) &__reint1_113; \
47060  __ret_113; \
47061})
47062#endif
47063
47064#ifdef __LITTLE_ENDIAN__
47065#define vgetq_lane_f16(__p0_114, __p1_114) __extension__ ({ \
47066  float16x8_t __s0_114 = __p0_114; \
47067  float16_t __ret_114; \
47068float16x8_t __reint_114 = __s0_114; \
47069int16_t __reint1_114 = vgetq_lane_s16(*(int16x8_t *) &__reint_114, __p1_114); \
47070  __ret_114 = *(float16_t *) &__reint1_114; \
47071  __ret_114; \
47072})
47073#else
47074#define vgetq_lane_f16(__p0_115, __p1_115) __extension__ ({ \
47075  float16x8_t __s0_115 = __p0_115; \
47076  float16x8_t __rev0_115;  __rev0_115 = __builtin_shufflevector(__s0_115, __s0_115, 7, 6, 5, 4, 3, 2, 1, 0); \
47077  float16_t __ret_115; \
47078float16x8_t __reint_115 = __rev0_115; \
47079int16_t __reint1_115 = __noswap_vgetq_lane_s16(*(int16x8_t *) &__reint_115, __p1_115); \
47080  __ret_115 = *(float16_t *) &__reint1_115; \
47081  __ret_115; \
47082})
47083#endif
47084
47085#ifdef __LITTLE_ENDIAN__
47086__ai poly64x1_t vget_low_p64(poly64x2_t __p0) {
47087  poly64x1_t __ret;
47088  __ret = __builtin_shufflevector(__p0, __p0, 0);
47089  return __ret;
47090}
47091#else
47092__ai poly64x1_t vget_low_p64(poly64x2_t __p0) {
47093  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47094  poly64x1_t __ret;
47095  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
47096  return __ret;
47097}
47098#endif
47099
47100#ifdef __LITTLE_ENDIAN__
47101__ai float64x1_t vget_low_f64(float64x2_t __p0) {
47102  float64x1_t __ret;
47103  __ret = __builtin_shufflevector(__p0, __p0, 0);
47104  return __ret;
47105}
47106#else
47107__ai float64x1_t vget_low_f64(float64x2_t __p0) {
47108  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47109  float64x1_t __ret;
47110  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
47111  return __ret;
47112}
47113#endif
47114
47115#ifdef __LITTLE_ENDIAN__
47116#define vld1_p64(__p0) __extension__ ({ \
47117  poly64x1_t __ret; \
47118  __ret = (poly64x1_t) __builtin_neon_vld1_v(__p0, 6); \
47119  __ret; \
47120})
47121#else
47122#define vld1_p64(__p0) __extension__ ({ \
47123  poly64x1_t __ret; \
47124  __ret = (poly64x1_t) __builtin_neon_vld1_v(__p0, 6); \
47125  __ret; \
47126})
47127#endif
47128
47129#ifdef __LITTLE_ENDIAN__
47130#define vld1q_p64(__p0) __extension__ ({ \
47131  poly64x2_t __ret; \
47132  __ret = (poly64x2_t) __builtin_neon_vld1q_v(__p0, 38); \
47133  __ret; \
47134})
47135#else
47136#define vld1q_p64(__p0) __extension__ ({ \
47137  poly64x2_t __ret; \
47138  __ret = (poly64x2_t) __builtin_neon_vld1q_v(__p0, 38); \
47139  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47140  __ret; \
47141})
47142#endif
47143
47144#ifdef __LITTLE_ENDIAN__
47145#define vld1q_f64(__p0) __extension__ ({ \
47146  float64x2_t __ret; \
47147  __ret = (float64x2_t) __builtin_neon_vld1q_v(__p0, 42); \
47148  __ret; \
47149})
47150#else
47151#define vld1q_f64(__p0) __extension__ ({ \
47152  float64x2_t __ret; \
47153  __ret = (float64x2_t) __builtin_neon_vld1q_v(__p0, 42); \
47154  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47155  __ret; \
47156})
47157#endif
47158
47159#ifdef __LITTLE_ENDIAN__
47160#define vld1_f64(__p0) __extension__ ({ \
47161  float64x1_t __ret; \
47162  __ret = (float64x1_t) __builtin_neon_vld1_v(__p0, 10); \
47163  __ret; \
47164})
47165#else
47166#define vld1_f64(__p0) __extension__ ({ \
47167  float64x1_t __ret; \
47168  __ret = (float64x1_t) __builtin_neon_vld1_v(__p0, 10); \
47169  __ret; \
47170})
47171#endif
47172
47173#ifdef __LITTLE_ENDIAN__
47174#define vld1_dup_p64(__p0) __extension__ ({ \
47175  poly64x1_t __ret; \
47176  __ret = (poly64x1_t) __builtin_neon_vld1_dup_v(__p0, 6); \
47177  __ret; \
47178})
47179#else
47180#define vld1_dup_p64(__p0) __extension__ ({ \
47181  poly64x1_t __ret; \
47182  __ret = (poly64x1_t) __builtin_neon_vld1_dup_v(__p0, 6); \
47183  __ret; \
47184})
47185#endif
47186
47187#ifdef __LITTLE_ENDIAN__
47188#define vld1q_dup_p64(__p0) __extension__ ({ \
47189  poly64x2_t __ret; \
47190  __ret = (poly64x2_t) __builtin_neon_vld1q_dup_v(__p0, 38); \
47191  __ret; \
47192})
47193#else
47194#define vld1q_dup_p64(__p0) __extension__ ({ \
47195  poly64x2_t __ret; \
47196  __ret = (poly64x2_t) __builtin_neon_vld1q_dup_v(__p0, 38); \
47197  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47198  __ret; \
47199})
47200#endif
47201
47202#ifdef __LITTLE_ENDIAN__
47203#define vld1q_dup_f64(__p0) __extension__ ({ \
47204  float64x2_t __ret; \
47205  __ret = (float64x2_t) __builtin_neon_vld1q_dup_v(__p0, 42); \
47206  __ret; \
47207})
47208#else
47209#define vld1q_dup_f64(__p0) __extension__ ({ \
47210  float64x2_t __ret; \
47211  __ret = (float64x2_t) __builtin_neon_vld1q_dup_v(__p0, 42); \
47212  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47213  __ret; \
47214})
47215#endif
47216
47217#ifdef __LITTLE_ENDIAN__
47218#define vld1_dup_f64(__p0) __extension__ ({ \
47219  float64x1_t __ret; \
47220  __ret = (float64x1_t) __builtin_neon_vld1_dup_v(__p0, 10); \
47221  __ret; \
47222})
47223#else
47224#define vld1_dup_f64(__p0) __extension__ ({ \
47225  float64x1_t __ret; \
47226  __ret = (float64x1_t) __builtin_neon_vld1_dup_v(__p0, 10); \
47227  __ret; \
47228})
47229#endif
47230
47231#ifdef __LITTLE_ENDIAN__
47232#define vld1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
47233  poly64x1_t __s1 = __p1; \
47234  poly64x1_t __ret; \
47235  __ret = (poly64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
47236  __ret; \
47237})
47238#else
47239#define vld1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
47240  poly64x1_t __s1 = __p1; \
47241  poly64x1_t __ret; \
47242  __ret = (poly64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
47243  __ret; \
47244})
47245#endif
47246
47247#ifdef __LITTLE_ENDIAN__
47248#define vld1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
47249  poly64x2_t __s1 = __p1; \
47250  poly64x2_t __ret; \
47251  __ret = (poly64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 38); \
47252  __ret; \
47253})
47254#else
47255#define vld1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
47256  poly64x2_t __s1 = __p1; \
47257  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47258  poly64x2_t __ret; \
47259  __ret = (poly64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 38); \
47260  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47261  __ret; \
47262})
47263#endif
47264
47265#ifdef __LITTLE_ENDIAN__
47266#define vld1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
47267  float64x2_t __s1 = __p1; \
47268  float64x2_t __ret; \
47269  __ret = (float64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 42); \
47270  __ret; \
47271})
47272#else
47273#define vld1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
47274  float64x2_t __s1 = __p1; \
47275  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47276  float64x2_t __ret; \
47277  __ret = (float64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 42); \
47278  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47279  __ret; \
47280})
47281#endif
47282
47283#ifdef __LITTLE_ENDIAN__
47284#define vld1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
47285  float64x1_t __s1 = __p1; \
47286  float64x1_t __ret; \
47287  __ret = (float64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
47288  __ret; \
47289})
47290#else
47291#define vld1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
47292  float64x1_t __s1 = __p1; \
47293  float64x1_t __ret; \
47294  __ret = (float64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
47295  __ret; \
47296})
47297#endif
47298
47299#ifdef __LITTLE_ENDIAN__
47300#define vld1_p8_x2(__p0) __extension__ ({ \
47301  poly8x8x2_t __ret; \
47302  __builtin_neon_vld1_x2_v(&__ret, __p0, 4); \
47303  __ret; \
47304})
47305#else
47306#define vld1_p8_x2(__p0) __extension__ ({ \
47307  poly8x8x2_t __ret; \
47308  __builtin_neon_vld1_x2_v(&__ret, __p0, 4); \
47309 \
47310  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
47311  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
47312  __ret; \
47313})
47314#endif
47315
47316#ifdef __LITTLE_ENDIAN__
47317#define vld1_p64_x2(__p0) __extension__ ({ \
47318  poly64x1x2_t __ret; \
47319  __builtin_neon_vld1_x2_v(&__ret, __p0, 6); \
47320  __ret; \
47321})
47322#else
47323#define vld1_p64_x2(__p0) __extension__ ({ \
47324  poly64x1x2_t __ret; \
47325  __builtin_neon_vld1_x2_v(&__ret, __p0, 6); \
47326  __ret; \
47327})
47328#endif
47329
47330#ifdef __LITTLE_ENDIAN__
47331#define vld1_p16_x2(__p0) __extension__ ({ \
47332  poly16x4x2_t __ret; \
47333  __builtin_neon_vld1_x2_v(&__ret, __p0, 5); \
47334  __ret; \
47335})
47336#else
47337#define vld1_p16_x2(__p0) __extension__ ({ \
47338  poly16x4x2_t __ret; \
47339  __builtin_neon_vld1_x2_v(&__ret, __p0, 5); \
47340 \
47341  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
47342  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
47343  __ret; \
47344})
47345#endif
47346
47347#ifdef __LITTLE_ENDIAN__
47348#define vld1q_p8_x2(__p0) __extension__ ({ \
47349  poly8x16x2_t __ret; \
47350  __builtin_neon_vld1q_x2_v(&__ret, __p0, 36); \
47351  __ret; \
47352})
47353#else
47354#define vld1q_p8_x2(__p0) __extension__ ({ \
47355  poly8x16x2_t __ret; \
47356  __builtin_neon_vld1q_x2_v(&__ret, __p0, 36); \
47357 \
47358  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47359  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47360  __ret; \
47361})
47362#endif
47363
47364#ifdef __LITTLE_ENDIAN__
47365#define vld1q_p64_x2(__p0) __extension__ ({ \
47366  poly64x2x2_t __ret; \
47367  __builtin_neon_vld1q_x2_v(&__ret, __p0, 38); \
47368  __ret; \
47369})
47370#else
47371#define vld1q_p64_x2(__p0) __extension__ ({ \
47372  poly64x2x2_t __ret; \
47373  __builtin_neon_vld1q_x2_v(&__ret, __p0, 38); \
47374 \
47375  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47376  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47377  __ret; \
47378})
47379#endif
47380
47381#ifdef __LITTLE_ENDIAN__
47382#define vld1q_p16_x2(__p0) __extension__ ({ \
47383  poly16x8x2_t __ret; \
47384  __builtin_neon_vld1q_x2_v(&__ret, __p0, 37); \
47385  __ret; \
47386})
47387#else
47388#define vld1q_p16_x2(__p0) __extension__ ({ \
47389  poly16x8x2_t __ret; \
47390  __builtin_neon_vld1q_x2_v(&__ret, __p0, 37); \
47391 \
47392  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
47393  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
47394  __ret; \
47395})
47396#endif
47397
47398#ifdef __LITTLE_ENDIAN__
47399#define vld1q_u8_x2(__p0) __extension__ ({ \
47400  uint8x16x2_t __ret; \
47401  __builtin_neon_vld1q_x2_v(&__ret, __p0, 48); \
47402  __ret; \
47403})
47404#else
47405#define vld1q_u8_x2(__p0) __extension__ ({ \
47406  uint8x16x2_t __ret; \
47407  __builtin_neon_vld1q_x2_v(&__ret, __p0, 48); \
47408 \
47409  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47410  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47411  __ret; \
47412})
47413#endif
47414
47415#ifdef __LITTLE_ENDIAN__
47416#define vld1q_u32_x2(__p0) __extension__ ({ \
47417  uint32x4x2_t __ret; \
47418  __builtin_neon_vld1q_x2_v(&__ret, __p0, 50); \
47419  __ret; \
47420})
47421#else
47422#define vld1q_u32_x2(__p0) __extension__ ({ \
47423  uint32x4x2_t __ret; \
47424  __builtin_neon_vld1q_x2_v(&__ret, __p0, 50); \
47425 \
47426  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
47427  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
47428  __ret; \
47429})
47430#endif
47431
47432#ifdef __LITTLE_ENDIAN__
47433#define vld1q_u64_x2(__p0) __extension__ ({ \
47434  uint64x2x2_t __ret; \
47435  __builtin_neon_vld1q_x2_v(&__ret, __p0, 51); \
47436  __ret; \
47437})
47438#else
47439#define vld1q_u64_x2(__p0) __extension__ ({ \
47440  uint64x2x2_t __ret; \
47441  __builtin_neon_vld1q_x2_v(&__ret, __p0, 51); \
47442 \
47443  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47444  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47445  __ret; \
47446})
47447#endif
47448
47449#ifdef __LITTLE_ENDIAN__
47450#define vld1q_u16_x2(__p0) __extension__ ({ \
47451  uint16x8x2_t __ret; \
47452  __builtin_neon_vld1q_x2_v(&__ret, __p0, 49); \
47453  __ret; \
47454})
47455#else
47456#define vld1q_u16_x2(__p0) __extension__ ({ \
47457  uint16x8x2_t __ret; \
47458  __builtin_neon_vld1q_x2_v(&__ret, __p0, 49); \
47459 \
47460  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
47461  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
47462  __ret; \
47463})
47464#endif
47465
47466#ifdef __LITTLE_ENDIAN__
47467#define vld1q_s8_x2(__p0) __extension__ ({ \
47468  int8x16x2_t __ret; \
47469  __builtin_neon_vld1q_x2_v(&__ret, __p0, 32); \
47470  __ret; \
47471})
47472#else
47473#define vld1q_s8_x2(__p0) __extension__ ({ \
47474  int8x16x2_t __ret; \
47475  __builtin_neon_vld1q_x2_v(&__ret, __p0, 32); \
47476 \
47477  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47478  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47479  __ret; \
47480})
47481#endif
47482
47483#ifdef __LITTLE_ENDIAN__
47484#define vld1q_f64_x2(__p0) __extension__ ({ \
47485  float64x2x2_t __ret; \
47486  __builtin_neon_vld1q_x2_v(&__ret, __p0, 42); \
47487  __ret; \
47488})
47489#else
47490#define vld1q_f64_x2(__p0) __extension__ ({ \
47491  float64x2x2_t __ret; \
47492  __builtin_neon_vld1q_x2_v(&__ret, __p0, 42); \
47493 \
47494  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47495  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47496  __ret; \
47497})
47498#endif
47499
47500#ifdef __LITTLE_ENDIAN__
47501#define vld1q_f32_x2(__p0) __extension__ ({ \
47502  float32x4x2_t __ret; \
47503  __builtin_neon_vld1q_x2_v(&__ret, __p0, 41); \
47504  __ret; \
47505})
47506#else
47507#define vld1q_f32_x2(__p0) __extension__ ({ \
47508  float32x4x2_t __ret; \
47509  __builtin_neon_vld1q_x2_v(&__ret, __p0, 41); \
47510 \
47511  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
47512  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
47513  __ret; \
47514})
47515#endif
47516
47517#ifdef __LITTLE_ENDIAN__
47518#define vld1q_f16_x2(__p0) __extension__ ({ \
47519  float16x8x2_t __ret; \
47520  __builtin_neon_vld1q_x2_v(&__ret, __p0, 40); \
47521  __ret; \
47522})
47523#else
47524#define vld1q_f16_x2(__p0) __extension__ ({ \
47525  float16x8x2_t __ret; \
47526  __builtin_neon_vld1q_x2_v(&__ret, __p0, 40); \
47527 \
47528  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
47529  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
47530  __ret; \
47531})
47532#endif
47533
47534#ifdef __LITTLE_ENDIAN__
47535#define vld1q_s32_x2(__p0) __extension__ ({ \
47536  int32x4x2_t __ret; \
47537  __builtin_neon_vld1q_x2_v(&__ret, __p0, 34); \
47538  __ret; \
47539})
47540#else
47541#define vld1q_s32_x2(__p0) __extension__ ({ \
47542  int32x4x2_t __ret; \
47543  __builtin_neon_vld1q_x2_v(&__ret, __p0, 34); \
47544 \
47545  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
47546  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
47547  __ret; \
47548})
47549#endif
47550
47551#ifdef __LITTLE_ENDIAN__
47552#define vld1q_s64_x2(__p0) __extension__ ({ \
47553  int64x2x2_t __ret; \
47554  __builtin_neon_vld1q_x2_v(&__ret, __p0, 35); \
47555  __ret; \
47556})
47557#else
47558#define vld1q_s64_x2(__p0) __extension__ ({ \
47559  int64x2x2_t __ret; \
47560  __builtin_neon_vld1q_x2_v(&__ret, __p0, 35); \
47561 \
47562  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47563  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47564  __ret; \
47565})
47566#endif
47567
47568#ifdef __LITTLE_ENDIAN__
47569#define vld1q_s16_x2(__p0) __extension__ ({ \
47570  int16x8x2_t __ret; \
47571  __builtin_neon_vld1q_x2_v(&__ret, __p0, 33); \
47572  __ret; \
47573})
47574#else
47575#define vld1q_s16_x2(__p0) __extension__ ({ \
47576  int16x8x2_t __ret; \
47577  __builtin_neon_vld1q_x2_v(&__ret, __p0, 33); \
47578 \
47579  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
47580  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
47581  __ret; \
47582})
47583#endif
47584
47585#ifdef __LITTLE_ENDIAN__
47586#define vld1_u8_x2(__p0) __extension__ ({ \
47587  uint8x8x2_t __ret; \
47588  __builtin_neon_vld1_x2_v(&__ret, __p0, 16); \
47589  __ret; \
47590})
47591#else
47592#define vld1_u8_x2(__p0) __extension__ ({ \
47593  uint8x8x2_t __ret; \
47594  __builtin_neon_vld1_x2_v(&__ret, __p0, 16); \
47595 \
47596  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
47597  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
47598  __ret; \
47599})
47600#endif
47601
47602#ifdef __LITTLE_ENDIAN__
47603#define vld1_u32_x2(__p0) __extension__ ({ \
47604  uint32x2x2_t __ret; \
47605  __builtin_neon_vld1_x2_v(&__ret, __p0, 18); \
47606  __ret; \
47607})
47608#else
47609#define vld1_u32_x2(__p0) __extension__ ({ \
47610  uint32x2x2_t __ret; \
47611  __builtin_neon_vld1_x2_v(&__ret, __p0, 18); \
47612 \
47613  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47614  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47615  __ret; \
47616})
47617#endif
47618
47619#ifdef __LITTLE_ENDIAN__
47620#define vld1_u64_x2(__p0) __extension__ ({ \
47621  uint64x1x2_t __ret; \
47622  __builtin_neon_vld1_x2_v(&__ret, __p0, 19); \
47623  __ret; \
47624})
47625#else
47626#define vld1_u64_x2(__p0) __extension__ ({ \
47627  uint64x1x2_t __ret; \
47628  __builtin_neon_vld1_x2_v(&__ret, __p0, 19); \
47629  __ret; \
47630})
47631#endif
47632
47633#ifdef __LITTLE_ENDIAN__
47634#define vld1_u16_x2(__p0) __extension__ ({ \
47635  uint16x4x2_t __ret; \
47636  __builtin_neon_vld1_x2_v(&__ret, __p0, 17); \
47637  __ret; \
47638})
47639#else
47640#define vld1_u16_x2(__p0) __extension__ ({ \
47641  uint16x4x2_t __ret; \
47642  __builtin_neon_vld1_x2_v(&__ret, __p0, 17); \
47643 \
47644  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
47645  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
47646  __ret; \
47647})
47648#endif
47649
47650#ifdef __LITTLE_ENDIAN__
47651#define vld1_s8_x2(__p0) __extension__ ({ \
47652  int8x8x2_t __ret; \
47653  __builtin_neon_vld1_x2_v(&__ret, __p0, 0); \
47654  __ret; \
47655})
47656#else
47657#define vld1_s8_x2(__p0) __extension__ ({ \
47658  int8x8x2_t __ret; \
47659  __builtin_neon_vld1_x2_v(&__ret, __p0, 0); \
47660 \
47661  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
47662  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
47663  __ret; \
47664})
47665#endif
47666
47667#ifdef __LITTLE_ENDIAN__
47668#define vld1_f64_x2(__p0) __extension__ ({ \
47669  float64x1x2_t __ret; \
47670  __builtin_neon_vld1_x2_v(&__ret, __p0, 10); \
47671  __ret; \
47672})
47673#else
47674#define vld1_f64_x2(__p0) __extension__ ({ \
47675  float64x1x2_t __ret; \
47676  __builtin_neon_vld1_x2_v(&__ret, __p0, 10); \
47677  __ret; \
47678})
47679#endif
47680
47681#ifdef __LITTLE_ENDIAN__
47682#define vld1_f32_x2(__p0) __extension__ ({ \
47683  float32x2x2_t __ret; \
47684  __builtin_neon_vld1_x2_v(&__ret, __p0, 9); \
47685  __ret; \
47686})
47687#else
47688#define vld1_f32_x2(__p0) __extension__ ({ \
47689  float32x2x2_t __ret; \
47690  __builtin_neon_vld1_x2_v(&__ret, __p0, 9); \
47691 \
47692  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47693  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47694  __ret; \
47695})
47696#endif
47697
47698#ifdef __LITTLE_ENDIAN__
47699#define vld1_f16_x2(__p0) __extension__ ({ \
47700  float16x4x2_t __ret; \
47701  __builtin_neon_vld1_x2_v(&__ret, __p0, 8); \
47702  __ret; \
47703})
47704#else
47705#define vld1_f16_x2(__p0) __extension__ ({ \
47706  float16x4x2_t __ret; \
47707  __builtin_neon_vld1_x2_v(&__ret, __p0, 8); \
47708 \
47709  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
47710  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
47711  __ret; \
47712})
47713#endif
47714
47715#ifdef __LITTLE_ENDIAN__
47716#define vld1_s32_x2(__p0) __extension__ ({ \
47717  int32x2x2_t __ret; \
47718  __builtin_neon_vld1_x2_v(&__ret, __p0, 2); \
47719  __ret; \
47720})
47721#else
47722#define vld1_s32_x2(__p0) __extension__ ({ \
47723  int32x2x2_t __ret; \
47724  __builtin_neon_vld1_x2_v(&__ret, __p0, 2); \
47725 \
47726  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47727  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47728  __ret; \
47729})
47730#endif
47731
47732#ifdef __LITTLE_ENDIAN__
47733#define vld1_s64_x2(__p0) __extension__ ({ \
47734  int64x1x2_t __ret; \
47735  __builtin_neon_vld1_x2_v(&__ret, __p0, 3); \
47736  __ret; \
47737})
47738#else
47739#define vld1_s64_x2(__p0) __extension__ ({ \
47740  int64x1x2_t __ret; \
47741  __builtin_neon_vld1_x2_v(&__ret, __p0, 3); \
47742  __ret; \
47743})
47744#endif
47745
47746#ifdef __LITTLE_ENDIAN__
47747#define vld1_s16_x2(__p0) __extension__ ({ \
47748  int16x4x2_t __ret; \
47749  __builtin_neon_vld1_x2_v(&__ret, __p0, 1); \
47750  __ret; \
47751})
47752#else
47753#define vld1_s16_x2(__p0) __extension__ ({ \
47754  int16x4x2_t __ret; \
47755  __builtin_neon_vld1_x2_v(&__ret, __p0, 1); \
47756 \
47757  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
47758  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
47759  __ret; \
47760})
47761#endif
47762
47763#ifdef __LITTLE_ENDIAN__
47764#define vld1_p8_x3(__p0) __extension__ ({ \
47765  poly8x8x3_t __ret; \
47766  __builtin_neon_vld1_x3_v(&__ret, __p0, 4); \
47767  __ret; \
47768})
47769#else
47770#define vld1_p8_x3(__p0) __extension__ ({ \
47771  poly8x8x3_t __ret; \
47772  __builtin_neon_vld1_x3_v(&__ret, __p0, 4); \
47773 \
47774  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
47775  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
47776  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
47777  __ret; \
47778})
47779#endif
47780
47781#ifdef __LITTLE_ENDIAN__
47782#define vld1_p64_x3(__p0) __extension__ ({ \
47783  poly64x1x3_t __ret; \
47784  __builtin_neon_vld1_x3_v(&__ret, __p0, 6); \
47785  __ret; \
47786})
47787#else
47788#define vld1_p64_x3(__p0) __extension__ ({ \
47789  poly64x1x3_t __ret; \
47790  __builtin_neon_vld1_x3_v(&__ret, __p0, 6); \
47791  __ret; \
47792})
47793#endif
47794
47795#ifdef __LITTLE_ENDIAN__
47796#define vld1_p16_x3(__p0) __extension__ ({ \
47797  poly16x4x3_t __ret; \
47798  __builtin_neon_vld1_x3_v(&__ret, __p0, 5); \
47799  __ret; \
47800})
47801#else
47802#define vld1_p16_x3(__p0) __extension__ ({ \
47803  poly16x4x3_t __ret; \
47804  __builtin_neon_vld1_x3_v(&__ret, __p0, 5); \
47805 \
47806  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
47807  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
47808  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
47809  __ret; \
47810})
47811#endif
47812
47813#ifdef __LITTLE_ENDIAN__
47814#define vld1q_p8_x3(__p0) __extension__ ({ \
47815  poly8x16x3_t __ret; \
47816  __builtin_neon_vld1q_x3_v(&__ret, __p0, 36); \
47817  __ret; \
47818})
47819#else
47820#define vld1q_p8_x3(__p0) __extension__ ({ \
47821  poly8x16x3_t __ret; \
47822  __builtin_neon_vld1q_x3_v(&__ret, __p0, 36); \
47823 \
47824  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47825  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47826  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47827  __ret; \
47828})
47829#endif
47830
47831#ifdef __LITTLE_ENDIAN__
47832#define vld1q_p64_x3(__p0) __extension__ ({ \
47833  poly64x2x3_t __ret; \
47834  __builtin_neon_vld1q_x3_v(&__ret, __p0, 38); \
47835  __ret; \
47836})
47837#else
47838#define vld1q_p64_x3(__p0) __extension__ ({ \
47839  poly64x2x3_t __ret; \
47840  __builtin_neon_vld1q_x3_v(&__ret, __p0, 38); \
47841 \
47842  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47843  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47844  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
47845  __ret; \
47846})
47847#endif
47848
47849#ifdef __LITTLE_ENDIAN__
47850#define vld1q_p16_x3(__p0) __extension__ ({ \
47851  poly16x8x3_t __ret; \
47852  __builtin_neon_vld1q_x3_v(&__ret, __p0, 37); \
47853  __ret; \
47854})
47855#else
47856#define vld1q_p16_x3(__p0) __extension__ ({ \
47857  poly16x8x3_t __ret; \
47858  __builtin_neon_vld1q_x3_v(&__ret, __p0, 37); \
47859 \
47860  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
47861  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
47862  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
47863  __ret; \
47864})
47865#endif
47866
47867#ifdef __LITTLE_ENDIAN__
47868#define vld1q_u8_x3(__p0) __extension__ ({ \
47869  uint8x16x3_t __ret; \
47870  __builtin_neon_vld1q_x3_v(&__ret, __p0, 48); \
47871  __ret; \
47872})
47873#else
47874#define vld1q_u8_x3(__p0) __extension__ ({ \
47875  uint8x16x3_t __ret; \
47876  __builtin_neon_vld1q_x3_v(&__ret, __p0, 48); \
47877 \
47878  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47879  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47880  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47881  __ret; \
47882})
47883#endif
47884
47885#ifdef __LITTLE_ENDIAN__
47886#define vld1q_u32_x3(__p0) __extension__ ({ \
47887  uint32x4x3_t __ret; \
47888  __builtin_neon_vld1q_x3_v(&__ret, __p0, 50); \
47889  __ret; \
47890})
47891#else
47892#define vld1q_u32_x3(__p0) __extension__ ({ \
47893  uint32x4x3_t __ret; \
47894  __builtin_neon_vld1q_x3_v(&__ret, __p0, 50); \
47895 \
47896  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
47897  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
47898  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
47899  __ret; \
47900})
47901#endif
47902
47903#ifdef __LITTLE_ENDIAN__
47904#define vld1q_u64_x3(__p0) __extension__ ({ \
47905  uint64x2x3_t __ret; \
47906  __builtin_neon_vld1q_x3_v(&__ret, __p0, 51); \
47907  __ret; \
47908})
47909#else
47910#define vld1q_u64_x3(__p0) __extension__ ({ \
47911  uint64x2x3_t __ret; \
47912  __builtin_neon_vld1q_x3_v(&__ret, __p0, 51); \
47913 \
47914  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47915  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47916  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
47917  __ret; \
47918})
47919#endif
47920
47921#ifdef __LITTLE_ENDIAN__
47922#define vld1q_u16_x3(__p0) __extension__ ({ \
47923  uint16x8x3_t __ret; \
47924  __builtin_neon_vld1q_x3_v(&__ret, __p0, 49); \
47925  __ret; \
47926})
47927#else
47928#define vld1q_u16_x3(__p0) __extension__ ({ \
47929  uint16x8x3_t __ret; \
47930  __builtin_neon_vld1q_x3_v(&__ret, __p0, 49); \
47931 \
47932  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
47933  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
47934  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
47935  __ret; \
47936})
47937#endif
47938
47939#ifdef __LITTLE_ENDIAN__
47940#define vld1q_s8_x3(__p0) __extension__ ({ \
47941  int8x16x3_t __ret; \
47942  __builtin_neon_vld1q_x3_v(&__ret, __p0, 32); \
47943  __ret; \
47944})
47945#else
47946#define vld1q_s8_x3(__p0) __extension__ ({ \
47947  int8x16x3_t __ret; \
47948  __builtin_neon_vld1q_x3_v(&__ret, __p0, 32); \
47949 \
47950  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47951  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47952  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
47953  __ret; \
47954})
47955#endif
47956
47957#ifdef __LITTLE_ENDIAN__
47958#define vld1q_f64_x3(__p0) __extension__ ({ \
47959  float64x2x3_t __ret; \
47960  __builtin_neon_vld1q_x3_v(&__ret, __p0, 42); \
47961  __ret; \
47962})
47963#else
47964#define vld1q_f64_x3(__p0) __extension__ ({ \
47965  float64x2x3_t __ret; \
47966  __builtin_neon_vld1q_x3_v(&__ret, __p0, 42); \
47967 \
47968  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
47969  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
47970  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
47971  __ret; \
47972})
47973#endif
47974
47975#ifdef __LITTLE_ENDIAN__
47976#define vld1q_f32_x3(__p0) __extension__ ({ \
47977  float32x4x3_t __ret; \
47978  __builtin_neon_vld1q_x3_v(&__ret, __p0, 41); \
47979  __ret; \
47980})
47981#else
47982#define vld1q_f32_x3(__p0) __extension__ ({ \
47983  float32x4x3_t __ret; \
47984  __builtin_neon_vld1q_x3_v(&__ret, __p0, 41); \
47985 \
47986  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
47987  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
47988  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
47989  __ret; \
47990})
47991#endif
47992
47993#ifdef __LITTLE_ENDIAN__
47994#define vld1q_f16_x3(__p0) __extension__ ({ \
47995  float16x8x3_t __ret; \
47996  __builtin_neon_vld1q_x3_v(&__ret, __p0, 40); \
47997  __ret; \
47998})
47999#else
48000#define vld1q_f16_x3(__p0) __extension__ ({ \
48001  float16x8x3_t __ret; \
48002  __builtin_neon_vld1q_x3_v(&__ret, __p0, 40); \
48003 \
48004  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48005  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48006  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48007  __ret; \
48008})
48009#endif
48010
48011#ifdef __LITTLE_ENDIAN__
48012#define vld1q_s32_x3(__p0) __extension__ ({ \
48013  int32x4x3_t __ret; \
48014  __builtin_neon_vld1q_x3_v(&__ret, __p0, 34); \
48015  __ret; \
48016})
48017#else
48018#define vld1q_s32_x3(__p0) __extension__ ({ \
48019  int32x4x3_t __ret; \
48020  __builtin_neon_vld1q_x3_v(&__ret, __p0, 34); \
48021 \
48022  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48023  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48024  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48025  __ret; \
48026})
48027#endif
48028
48029#ifdef __LITTLE_ENDIAN__
48030#define vld1q_s64_x3(__p0) __extension__ ({ \
48031  int64x2x3_t __ret; \
48032  __builtin_neon_vld1q_x3_v(&__ret, __p0, 35); \
48033  __ret; \
48034})
48035#else
48036#define vld1q_s64_x3(__p0) __extension__ ({ \
48037  int64x2x3_t __ret; \
48038  __builtin_neon_vld1q_x3_v(&__ret, __p0, 35); \
48039 \
48040  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48041  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48042  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48043  __ret; \
48044})
48045#endif
48046
48047#ifdef __LITTLE_ENDIAN__
48048#define vld1q_s16_x3(__p0) __extension__ ({ \
48049  int16x8x3_t __ret; \
48050  __builtin_neon_vld1q_x3_v(&__ret, __p0, 33); \
48051  __ret; \
48052})
48053#else
48054#define vld1q_s16_x3(__p0) __extension__ ({ \
48055  int16x8x3_t __ret; \
48056  __builtin_neon_vld1q_x3_v(&__ret, __p0, 33); \
48057 \
48058  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48059  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48060  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48061  __ret; \
48062})
48063#endif
48064
48065#ifdef __LITTLE_ENDIAN__
48066#define vld1_u8_x3(__p0) __extension__ ({ \
48067  uint8x8x3_t __ret; \
48068  __builtin_neon_vld1_x3_v(&__ret, __p0, 16); \
48069  __ret; \
48070})
48071#else
48072#define vld1_u8_x3(__p0) __extension__ ({ \
48073  uint8x8x3_t __ret; \
48074  __builtin_neon_vld1_x3_v(&__ret, __p0, 16); \
48075 \
48076  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48077  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48078  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48079  __ret; \
48080})
48081#endif
48082
48083#ifdef __LITTLE_ENDIAN__
48084#define vld1_u32_x3(__p0) __extension__ ({ \
48085  uint32x2x3_t __ret; \
48086  __builtin_neon_vld1_x3_v(&__ret, __p0, 18); \
48087  __ret; \
48088})
48089#else
48090#define vld1_u32_x3(__p0) __extension__ ({ \
48091  uint32x2x3_t __ret; \
48092  __builtin_neon_vld1_x3_v(&__ret, __p0, 18); \
48093 \
48094  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48095  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48096  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48097  __ret; \
48098})
48099#endif
48100
48101#ifdef __LITTLE_ENDIAN__
48102#define vld1_u64_x3(__p0) __extension__ ({ \
48103  uint64x1x3_t __ret; \
48104  __builtin_neon_vld1_x3_v(&__ret, __p0, 19); \
48105  __ret; \
48106})
48107#else
48108#define vld1_u64_x3(__p0) __extension__ ({ \
48109  uint64x1x3_t __ret; \
48110  __builtin_neon_vld1_x3_v(&__ret, __p0, 19); \
48111  __ret; \
48112})
48113#endif
48114
48115#ifdef __LITTLE_ENDIAN__
48116#define vld1_u16_x3(__p0) __extension__ ({ \
48117  uint16x4x3_t __ret; \
48118  __builtin_neon_vld1_x3_v(&__ret, __p0, 17); \
48119  __ret; \
48120})
48121#else
48122#define vld1_u16_x3(__p0) __extension__ ({ \
48123  uint16x4x3_t __ret; \
48124  __builtin_neon_vld1_x3_v(&__ret, __p0, 17); \
48125 \
48126  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48127  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48128  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48129  __ret; \
48130})
48131#endif
48132
48133#ifdef __LITTLE_ENDIAN__
48134#define vld1_s8_x3(__p0) __extension__ ({ \
48135  int8x8x3_t __ret; \
48136  __builtin_neon_vld1_x3_v(&__ret, __p0, 0); \
48137  __ret; \
48138})
48139#else
48140#define vld1_s8_x3(__p0) __extension__ ({ \
48141  int8x8x3_t __ret; \
48142  __builtin_neon_vld1_x3_v(&__ret, __p0, 0); \
48143 \
48144  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48145  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48146  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48147  __ret; \
48148})
48149#endif
48150
48151#ifdef __LITTLE_ENDIAN__
48152#define vld1_f64_x3(__p0) __extension__ ({ \
48153  float64x1x3_t __ret; \
48154  __builtin_neon_vld1_x3_v(&__ret, __p0, 10); \
48155  __ret; \
48156})
48157#else
48158#define vld1_f64_x3(__p0) __extension__ ({ \
48159  float64x1x3_t __ret; \
48160  __builtin_neon_vld1_x3_v(&__ret, __p0, 10); \
48161  __ret; \
48162})
48163#endif
48164
48165#ifdef __LITTLE_ENDIAN__
48166#define vld1_f32_x3(__p0) __extension__ ({ \
48167  float32x2x3_t __ret; \
48168  __builtin_neon_vld1_x3_v(&__ret, __p0, 9); \
48169  __ret; \
48170})
48171#else
48172#define vld1_f32_x3(__p0) __extension__ ({ \
48173  float32x2x3_t __ret; \
48174  __builtin_neon_vld1_x3_v(&__ret, __p0, 9); \
48175 \
48176  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48177  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48178  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48179  __ret; \
48180})
48181#endif
48182
48183#ifdef __LITTLE_ENDIAN__
48184#define vld1_f16_x3(__p0) __extension__ ({ \
48185  float16x4x3_t __ret; \
48186  __builtin_neon_vld1_x3_v(&__ret, __p0, 8); \
48187  __ret; \
48188})
48189#else
48190#define vld1_f16_x3(__p0) __extension__ ({ \
48191  float16x4x3_t __ret; \
48192  __builtin_neon_vld1_x3_v(&__ret, __p0, 8); \
48193 \
48194  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48195  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48196  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48197  __ret; \
48198})
48199#endif
48200
48201#ifdef __LITTLE_ENDIAN__
48202#define vld1_s32_x3(__p0) __extension__ ({ \
48203  int32x2x3_t __ret; \
48204  __builtin_neon_vld1_x3_v(&__ret, __p0, 2); \
48205  __ret; \
48206})
48207#else
48208#define vld1_s32_x3(__p0) __extension__ ({ \
48209  int32x2x3_t __ret; \
48210  __builtin_neon_vld1_x3_v(&__ret, __p0, 2); \
48211 \
48212  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48213  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48214  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48215  __ret; \
48216})
48217#endif
48218
48219#ifdef __LITTLE_ENDIAN__
48220#define vld1_s64_x3(__p0) __extension__ ({ \
48221  int64x1x3_t __ret; \
48222  __builtin_neon_vld1_x3_v(&__ret, __p0, 3); \
48223  __ret; \
48224})
48225#else
48226#define vld1_s64_x3(__p0) __extension__ ({ \
48227  int64x1x3_t __ret; \
48228  __builtin_neon_vld1_x3_v(&__ret, __p0, 3); \
48229  __ret; \
48230})
48231#endif
48232
48233#ifdef __LITTLE_ENDIAN__
48234#define vld1_s16_x3(__p0) __extension__ ({ \
48235  int16x4x3_t __ret; \
48236  __builtin_neon_vld1_x3_v(&__ret, __p0, 1); \
48237  __ret; \
48238})
48239#else
48240#define vld1_s16_x3(__p0) __extension__ ({ \
48241  int16x4x3_t __ret; \
48242  __builtin_neon_vld1_x3_v(&__ret, __p0, 1); \
48243 \
48244  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48245  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48246  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48247  __ret; \
48248})
48249#endif
48250
48251#ifdef __LITTLE_ENDIAN__
48252#define vld1_p8_x4(__p0) __extension__ ({ \
48253  poly8x8x4_t __ret; \
48254  __builtin_neon_vld1_x4_v(&__ret, __p0, 4); \
48255  __ret; \
48256})
48257#else
48258#define vld1_p8_x4(__p0) __extension__ ({ \
48259  poly8x8x4_t __ret; \
48260  __builtin_neon_vld1_x4_v(&__ret, __p0, 4); \
48261 \
48262  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48263  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48264  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48265  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
48266  __ret; \
48267})
48268#endif
48269
48270#ifdef __LITTLE_ENDIAN__
48271#define vld1_p64_x4(__p0) __extension__ ({ \
48272  poly64x1x4_t __ret; \
48273  __builtin_neon_vld1_x4_v(&__ret, __p0, 6); \
48274  __ret; \
48275})
48276#else
48277#define vld1_p64_x4(__p0) __extension__ ({ \
48278  poly64x1x4_t __ret; \
48279  __builtin_neon_vld1_x4_v(&__ret, __p0, 6); \
48280  __ret; \
48281})
48282#endif
48283
48284#ifdef __LITTLE_ENDIAN__
48285#define vld1_p16_x4(__p0) __extension__ ({ \
48286  poly16x4x4_t __ret; \
48287  __builtin_neon_vld1_x4_v(&__ret, __p0, 5); \
48288  __ret; \
48289})
48290#else
48291#define vld1_p16_x4(__p0) __extension__ ({ \
48292  poly16x4x4_t __ret; \
48293  __builtin_neon_vld1_x4_v(&__ret, __p0, 5); \
48294 \
48295  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48296  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48297  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48298  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
48299  __ret; \
48300})
48301#endif
48302
48303#ifdef __LITTLE_ENDIAN__
48304#define vld1q_p8_x4(__p0) __extension__ ({ \
48305  poly8x16x4_t __ret; \
48306  __builtin_neon_vld1q_x4_v(&__ret, __p0, 36); \
48307  __ret; \
48308})
48309#else
48310#define vld1q_p8_x4(__p0) __extension__ ({ \
48311  poly8x16x4_t __ret; \
48312  __builtin_neon_vld1q_x4_v(&__ret, __p0, 36); \
48313 \
48314  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48315  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48316  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48317  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48318  __ret; \
48319})
48320#endif
48321
48322#ifdef __LITTLE_ENDIAN__
48323#define vld1q_p64_x4(__p0) __extension__ ({ \
48324  poly64x2x4_t __ret; \
48325  __builtin_neon_vld1q_x4_v(&__ret, __p0, 38); \
48326  __ret; \
48327})
48328#else
48329#define vld1q_p64_x4(__p0) __extension__ ({ \
48330  poly64x2x4_t __ret; \
48331  __builtin_neon_vld1q_x4_v(&__ret, __p0, 38); \
48332 \
48333  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48334  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48335  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48336  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48337  __ret; \
48338})
48339#endif
48340
48341#ifdef __LITTLE_ENDIAN__
48342#define vld1q_p16_x4(__p0) __extension__ ({ \
48343  poly16x8x4_t __ret; \
48344  __builtin_neon_vld1q_x4_v(&__ret, __p0, 37); \
48345  __ret; \
48346})
48347#else
48348#define vld1q_p16_x4(__p0) __extension__ ({ \
48349  poly16x8x4_t __ret; \
48350  __builtin_neon_vld1q_x4_v(&__ret, __p0, 37); \
48351 \
48352  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48353  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48354  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48355  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
48356  __ret; \
48357})
48358#endif
48359
48360#ifdef __LITTLE_ENDIAN__
48361#define vld1q_u8_x4(__p0) __extension__ ({ \
48362  uint8x16x4_t __ret; \
48363  __builtin_neon_vld1q_x4_v(&__ret, __p0, 48); \
48364  __ret; \
48365})
48366#else
48367#define vld1q_u8_x4(__p0) __extension__ ({ \
48368  uint8x16x4_t __ret; \
48369  __builtin_neon_vld1q_x4_v(&__ret, __p0, 48); \
48370 \
48371  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48372  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48373  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48374  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48375  __ret; \
48376})
48377#endif
48378
48379#ifdef __LITTLE_ENDIAN__
48380#define vld1q_u32_x4(__p0) __extension__ ({ \
48381  uint32x4x4_t __ret; \
48382  __builtin_neon_vld1q_x4_v(&__ret, __p0, 50); \
48383  __ret; \
48384})
48385#else
48386#define vld1q_u32_x4(__p0) __extension__ ({ \
48387  uint32x4x4_t __ret; \
48388  __builtin_neon_vld1q_x4_v(&__ret, __p0, 50); \
48389 \
48390  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48391  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48392  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48393  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
48394  __ret; \
48395})
48396#endif
48397
48398#ifdef __LITTLE_ENDIAN__
48399#define vld1q_u64_x4(__p0) __extension__ ({ \
48400  uint64x2x4_t __ret; \
48401  __builtin_neon_vld1q_x4_v(&__ret, __p0, 51); \
48402  __ret; \
48403})
48404#else
48405#define vld1q_u64_x4(__p0) __extension__ ({ \
48406  uint64x2x4_t __ret; \
48407  __builtin_neon_vld1q_x4_v(&__ret, __p0, 51); \
48408 \
48409  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48410  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48411  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48412  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48413  __ret; \
48414})
48415#endif
48416
48417#ifdef __LITTLE_ENDIAN__
48418#define vld1q_u16_x4(__p0) __extension__ ({ \
48419  uint16x8x4_t __ret; \
48420  __builtin_neon_vld1q_x4_v(&__ret, __p0, 49); \
48421  __ret; \
48422})
48423#else
48424#define vld1q_u16_x4(__p0) __extension__ ({ \
48425  uint16x8x4_t __ret; \
48426  __builtin_neon_vld1q_x4_v(&__ret, __p0, 49); \
48427 \
48428  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48429  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48430  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48431  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
48432  __ret; \
48433})
48434#endif
48435
48436#ifdef __LITTLE_ENDIAN__
48437#define vld1q_s8_x4(__p0) __extension__ ({ \
48438  int8x16x4_t __ret; \
48439  __builtin_neon_vld1q_x4_v(&__ret, __p0, 32); \
48440  __ret; \
48441})
48442#else
48443#define vld1q_s8_x4(__p0) __extension__ ({ \
48444  int8x16x4_t __ret; \
48445  __builtin_neon_vld1q_x4_v(&__ret, __p0, 32); \
48446 \
48447  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48448  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48449  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48450  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48451  __ret; \
48452})
48453#endif
48454
48455#ifdef __LITTLE_ENDIAN__
48456#define vld1q_f64_x4(__p0) __extension__ ({ \
48457  float64x2x4_t __ret; \
48458  __builtin_neon_vld1q_x4_v(&__ret, __p0, 42); \
48459  __ret; \
48460})
48461#else
48462#define vld1q_f64_x4(__p0) __extension__ ({ \
48463  float64x2x4_t __ret; \
48464  __builtin_neon_vld1q_x4_v(&__ret, __p0, 42); \
48465 \
48466  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48467  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48468  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48469  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48470  __ret; \
48471})
48472#endif
48473
48474#ifdef __LITTLE_ENDIAN__
48475#define vld1q_f32_x4(__p0) __extension__ ({ \
48476  float32x4x4_t __ret; \
48477  __builtin_neon_vld1q_x4_v(&__ret, __p0, 41); \
48478  __ret; \
48479})
48480#else
48481#define vld1q_f32_x4(__p0) __extension__ ({ \
48482  float32x4x4_t __ret; \
48483  __builtin_neon_vld1q_x4_v(&__ret, __p0, 41); \
48484 \
48485  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48486  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48487  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48488  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
48489  __ret; \
48490})
48491#endif
48492
48493#ifdef __LITTLE_ENDIAN__
48494#define vld1q_f16_x4(__p0) __extension__ ({ \
48495  float16x8x4_t __ret; \
48496  __builtin_neon_vld1q_x4_v(&__ret, __p0, 40); \
48497  __ret; \
48498})
48499#else
48500#define vld1q_f16_x4(__p0) __extension__ ({ \
48501  float16x8x4_t __ret; \
48502  __builtin_neon_vld1q_x4_v(&__ret, __p0, 40); \
48503 \
48504  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48505  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48506  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48507  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
48508  __ret; \
48509})
48510#endif
48511
48512#ifdef __LITTLE_ENDIAN__
48513#define vld1q_s32_x4(__p0) __extension__ ({ \
48514  int32x4x4_t __ret; \
48515  __builtin_neon_vld1q_x4_v(&__ret, __p0, 34); \
48516  __ret; \
48517})
48518#else
48519#define vld1q_s32_x4(__p0) __extension__ ({ \
48520  int32x4x4_t __ret; \
48521  __builtin_neon_vld1q_x4_v(&__ret, __p0, 34); \
48522 \
48523  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48524  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48525  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48526  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
48527  __ret; \
48528})
48529#endif
48530
48531#ifdef __LITTLE_ENDIAN__
48532#define vld1q_s64_x4(__p0) __extension__ ({ \
48533  int64x2x4_t __ret; \
48534  __builtin_neon_vld1q_x4_v(&__ret, __p0, 35); \
48535  __ret; \
48536})
48537#else
48538#define vld1q_s64_x4(__p0) __extension__ ({ \
48539  int64x2x4_t __ret; \
48540  __builtin_neon_vld1q_x4_v(&__ret, __p0, 35); \
48541 \
48542  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48543  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48544  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48545  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48546  __ret; \
48547})
48548#endif
48549
48550#ifdef __LITTLE_ENDIAN__
48551#define vld1q_s16_x4(__p0) __extension__ ({ \
48552  int16x8x4_t __ret; \
48553  __builtin_neon_vld1q_x4_v(&__ret, __p0, 33); \
48554  __ret; \
48555})
48556#else
48557#define vld1q_s16_x4(__p0) __extension__ ({ \
48558  int16x8x4_t __ret; \
48559  __builtin_neon_vld1q_x4_v(&__ret, __p0, 33); \
48560 \
48561  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48562  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48563  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48564  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
48565  __ret; \
48566})
48567#endif
48568
48569#ifdef __LITTLE_ENDIAN__
48570#define vld1_u8_x4(__p0) __extension__ ({ \
48571  uint8x8x4_t __ret; \
48572  __builtin_neon_vld1_x4_v(&__ret, __p0, 16); \
48573  __ret; \
48574})
48575#else
48576#define vld1_u8_x4(__p0) __extension__ ({ \
48577  uint8x8x4_t __ret; \
48578  __builtin_neon_vld1_x4_v(&__ret, __p0, 16); \
48579 \
48580  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48581  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48582  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48583  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
48584  __ret; \
48585})
48586#endif
48587
48588#ifdef __LITTLE_ENDIAN__
48589#define vld1_u32_x4(__p0) __extension__ ({ \
48590  uint32x2x4_t __ret; \
48591  __builtin_neon_vld1_x4_v(&__ret, __p0, 18); \
48592  __ret; \
48593})
48594#else
48595#define vld1_u32_x4(__p0) __extension__ ({ \
48596  uint32x2x4_t __ret; \
48597  __builtin_neon_vld1_x4_v(&__ret, __p0, 18); \
48598 \
48599  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48600  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48601  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48602  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48603  __ret; \
48604})
48605#endif
48606
48607#ifdef __LITTLE_ENDIAN__
48608#define vld1_u64_x4(__p0) __extension__ ({ \
48609  uint64x1x4_t __ret; \
48610  __builtin_neon_vld1_x4_v(&__ret, __p0, 19); \
48611  __ret; \
48612})
48613#else
48614#define vld1_u64_x4(__p0) __extension__ ({ \
48615  uint64x1x4_t __ret; \
48616  __builtin_neon_vld1_x4_v(&__ret, __p0, 19); \
48617  __ret; \
48618})
48619#endif
48620
48621#ifdef __LITTLE_ENDIAN__
48622#define vld1_u16_x4(__p0) __extension__ ({ \
48623  uint16x4x4_t __ret; \
48624  __builtin_neon_vld1_x4_v(&__ret, __p0, 17); \
48625  __ret; \
48626})
48627#else
48628#define vld1_u16_x4(__p0) __extension__ ({ \
48629  uint16x4x4_t __ret; \
48630  __builtin_neon_vld1_x4_v(&__ret, __p0, 17); \
48631 \
48632  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48633  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48634  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48635  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
48636  __ret; \
48637})
48638#endif
48639
48640#ifdef __LITTLE_ENDIAN__
48641#define vld1_s8_x4(__p0) __extension__ ({ \
48642  int8x8x4_t __ret; \
48643  __builtin_neon_vld1_x4_v(&__ret, __p0, 0); \
48644  __ret; \
48645})
48646#else
48647#define vld1_s8_x4(__p0) __extension__ ({ \
48648  int8x8x4_t __ret; \
48649  __builtin_neon_vld1_x4_v(&__ret, __p0, 0); \
48650 \
48651  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48652  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48653  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48654  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
48655  __ret; \
48656})
48657#endif
48658
48659#ifdef __LITTLE_ENDIAN__
48660#define vld1_f64_x4(__p0) __extension__ ({ \
48661  float64x1x4_t __ret; \
48662  __builtin_neon_vld1_x4_v(&__ret, __p0, 10); \
48663  __ret; \
48664})
48665#else
48666#define vld1_f64_x4(__p0) __extension__ ({ \
48667  float64x1x4_t __ret; \
48668  __builtin_neon_vld1_x4_v(&__ret, __p0, 10); \
48669  __ret; \
48670})
48671#endif
48672
48673#ifdef __LITTLE_ENDIAN__
48674#define vld1_f32_x4(__p0) __extension__ ({ \
48675  float32x2x4_t __ret; \
48676  __builtin_neon_vld1_x4_v(&__ret, __p0, 9); \
48677  __ret; \
48678})
48679#else
48680#define vld1_f32_x4(__p0) __extension__ ({ \
48681  float32x2x4_t __ret; \
48682  __builtin_neon_vld1_x4_v(&__ret, __p0, 9); \
48683 \
48684  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48685  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48686  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48687  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48688  __ret; \
48689})
48690#endif
48691
48692#ifdef __LITTLE_ENDIAN__
48693#define vld1_f16_x4(__p0) __extension__ ({ \
48694  float16x4x4_t __ret; \
48695  __builtin_neon_vld1_x4_v(&__ret, __p0, 8); \
48696  __ret; \
48697})
48698#else
48699#define vld1_f16_x4(__p0) __extension__ ({ \
48700  float16x4x4_t __ret; \
48701  __builtin_neon_vld1_x4_v(&__ret, __p0, 8); \
48702 \
48703  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48704  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48705  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48706  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
48707  __ret; \
48708})
48709#endif
48710
48711#ifdef __LITTLE_ENDIAN__
48712#define vld1_s32_x4(__p0) __extension__ ({ \
48713  int32x2x4_t __ret; \
48714  __builtin_neon_vld1_x4_v(&__ret, __p0, 2); \
48715  __ret; \
48716})
48717#else
48718#define vld1_s32_x4(__p0) __extension__ ({ \
48719  int32x2x4_t __ret; \
48720  __builtin_neon_vld1_x4_v(&__ret, __p0, 2); \
48721 \
48722  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48723  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48724  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48725  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
48726  __ret; \
48727})
48728#endif
48729
48730#ifdef __LITTLE_ENDIAN__
48731#define vld1_s64_x4(__p0) __extension__ ({ \
48732  int64x1x4_t __ret; \
48733  __builtin_neon_vld1_x4_v(&__ret, __p0, 3); \
48734  __ret; \
48735})
48736#else
48737#define vld1_s64_x4(__p0) __extension__ ({ \
48738  int64x1x4_t __ret; \
48739  __builtin_neon_vld1_x4_v(&__ret, __p0, 3); \
48740  __ret; \
48741})
48742#endif
48743
48744#ifdef __LITTLE_ENDIAN__
48745#define vld1_s16_x4(__p0) __extension__ ({ \
48746  int16x4x4_t __ret; \
48747  __builtin_neon_vld1_x4_v(&__ret, __p0, 1); \
48748  __ret; \
48749})
48750#else
48751#define vld1_s16_x4(__p0) __extension__ ({ \
48752  int16x4x4_t __ret; \
48753  __builtin_neon_vld1_x4_v(&__ret, __p0, 1); \
48754 \
48755  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48756  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48757  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48758  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
48759  __ret; \
48760})
48761#endif
48762
48763#ifdef __LITTLE_ENDIAN__
48764#define vld2_p64(__p0) __extension__ ({ \
48765  poly64x1x2_t __ret; \
48766  __builtin_neon_vld2_v(&__ret, __p0, 6); \
48767  __ret; \
48768})
48769#else
48770#define vld2_p64(__p0) __extension__ ({ \
48771  poly64x1x2_t __ret; \
48772  __builtin_neon_vld2_v(&__ret, __p0, 6); \
48773  __ret; \
48774})
48775#endif
48776
48777#ifdef __LITTLE_ENDIAN__
48778#define vld2q_p64(__p0) __extension__ ({ \
48779  poly64x2x2_t __ret; \
48780  __builtin_neon_vld2q_v(&__ret, __p0, 38); \
48781  __ret; \
48782})
48783#else
48784#define vld2q_p64(__p0) __extension__ ({ \
48785  poly64x2x2_t __ret; \
48786  __builtin_neon_vld2q_v(&__ret, __p0, 38); \
48787 \
48788  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48789  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48790  __ret; \
48791})
48792#endif
48793
48794#ifdef __LITTLE_ENDIAN__
48795#define vld2q_u64(__p0) __extension__ ({ \
48796  uint64x2x2_t __ret; \
48797  __builtin_neon_vld2q_v(&__ret, __p0, 51); \
48798  __ret; \
48799})
48800#else
48801#define vld2q_u64(__p0) __extension__ ({ \
48802  uint64x2x2_t __ret; \
48803  __builtin_neon_vld2q_v(&__ret, __p0, 51); \
48804 \
48805  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48806  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48807  __ret; \
48808})
48809#endif
48810
48811#ifdef __LITTLE_ENDIAN__
48812#define vld2q_f64(__p0) __extension__ ({ \
48813  float64x2x2_t __ret; \
48814  __builtin_neon_vld2q_v(&__ret, __p0, 42); \
48815  __ret; \
48816})
48817#else
48818#define vld2q_f64(__p0) __extension__ ({ \
48819  float64x2x2_t __ret; \
48820  __builtin_neon_vld2q_v(&__ret, __p0, 42); \
48821 \
48822  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48823  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48824  __ret; \
48825})
48826#endif
48827
48828#ifdef __LITTLE_ENDIAN__
48829#define vld2q_s64(__p0) __extension__ ({ \
48830  int64x2x2_t __ret; \
48831  __builtin_neon_vld2q_v(&__ret, __p0, 35); \
48832  __ret; \
48833})
48834#else
48835#define vld2q_s64(__p0) __extension__ ({ \
48836  int64x2x2_t __ret; \
48837  __builtin_neon_vld2q_v(&__ret, __p0, 35); \
48838 \
48839  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48840  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48841  __ret; \
48842})
48843#endif
48844
48845#ifdef __LITTLE_ENDIAN__
48846#define vld2_f64(__p0) __extension__ ({ \
48847  float64x1x2_t __ret; \
48848  __builtin_neon_vld2_v(&__ret, __p0, 10); \
48849  __ret; \
48850})
48851#else
48852#define vld2_f64(__p0) __extension__ ({ \
48853  float64x1x2_t __ret; \
48854  __builtin_neon_vld2_v(&__ret, __p0, 10); \
48855  __ret; \
48856})
48857#endif
48858
48859#ifdef __LITTLE_ENDIAN__
48860#define vld2_dup_p64(__p0) __extension__ ({ \
48861  poly64x1x2_t __ret; \
48862  __builtin_neon_vld2_dup_v(&__ret, __p0, 6); \
48863  __ret; \
48864})
48865#else
48866#define vld2_dup_p64(__p0) __extension__ ({ \
48867  poly64x1x2_t __ret; \
48868  __builtin_neon_vld2_dup_v(&__ret, __p0, 6); \
48869  __ret; \
48870})
48871#endif
48872
48873#ifdef __LITTLE_ENDIAN__
48874#define vld2q_dup_p8(__p0) __extension__ ({ \
48875  poly8x16x2_t __ret; \
48876  __builtin_neon_vld2q_dup_v(&__ret, __p0, 36); \
48877  __ret; \
48878})
48879#else
48880#define vld2q_dup_p8(__p0) __extension__ ({ \
48881  poly8x16x2_t __ret; \
48882  __builtin_neon_vld2q_dup_v(&__ret, __p0, 36); \
48883 \
48884  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48885  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48886  __ret; \
48887})
48888#endif
48889
48890#ifdef __LITTLE_ENDIAN__
48891#define vld2q_dup_p64(__p0) __extension__ ({ \
48892  poly64x2x2_t __ret; \
48893  __builtin_neon_vld2q_dup_v(&__ret, __p0, 38); \
48894  __ret; \
48895})
48896#else
48897#define vld2q_dup_p64(__p0) __extension__ ({ \
48898  poly64x2x2_t __ret; \
48899  __builtin_neon_vld2q_dup_v(&__ret, __p0, 38); \
48900 \
48901  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48902  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48903  __ret; \
48904})
48905#endif
48906
48907#ifdef __LITTLE_ENDIAN__
48908#define vld2q_dup_p16(__p0) __extension__ ({ \
48909  poly16x8x2_t __ret; \
48910  __builtin_neon_vld2q_dup_v(&__ret, __p0, 37); \
48911  __ret; \
48912})
48913#else
48914#define vld2q_dup_p16(__p0) __extension__ ({ \
48915  poly16x8x2_t __ret; \
48916  __builtin_neon_vld2q_dup_v(&__ret, __p0, 37); \
48917 \
48918  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48919  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48920  __ret; \
48921})
48922#endif
48923
48924#ifdef __LITTLE_ENDIAN__
48925#define vld2q_dup_u8(__p0) __extension__ ({ \
48926  uint8x16x2_t __ret; \
48927  __builtin_neon_vld2q_dup_v(&__ret, __p0, 48); \
48928  __ret; \
48929})
48930#else
48931#define vld2q_dup_u8(__p0) __extension__ ({ \
48932  uint8x16x2_t __ret; \
48933  __builtin_neon_vld2q_dup_v(&__ret, __p0, 48); \
48934 \
48935  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48936  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48937  __ret; \
48938})
48939#endif
48940
48941#ifdef __LITTLE_ENDIAN__
48942#define vld2q_dup_u32(__p0) __extension__ ({ \
48943  uint32x4x2_t __ret; \
48944  __builtin_neon_vld2q_dup_v(&__ret, __p0, 50); \
48945  __ret; \
48946})
48947#else
48948#define vld2q_dup_u32(__p0) __extension__ ({ \
48949  uint32x4x2_t __ret; \
48950  __builtin_neon_vld2q_dup_v(&__ret, __p0, 50); \
48951 \
48952  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48953  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48954  __ret; \
48955})
48956#endif
48957
48958#ifdef __LITTLE_ENDIAN__
48959#define vld2q_dup_u64(__p0) __extension__ ({ \
48960  uint64x2x2_t __ret; \
48961  __builtin_neon_vld2q_dup_v(&__ret, __p0, 51); \
48962  __ret; \
48963})
48964#else
48965#define vld2q_dup_u64(__p0) __extension__ ({ \
48966  uint64x2x2_t __ret; \
48967  __builtin_neon_vld2q_dup_v(&__ret, __p0, 51); \
48968 \
48969  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48970  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48971  __ret; \
48972})
48973#endif
48974
48975#ifdef __LITTLE_ENDIAN__
48976#define vld2q_dup_u16(__p0) __extension__ ({ \
48977  uint16x8x2_t __ret; \
48978  __builtin_neon_vld2q_dup_v(&__ret, __p0, 49); \
48979  __ret; \
48980})
48981#else
48982#define vld2q_dup_u16(__p0) __extension__ ({ \
48983  uint16x8x2_t __ret; \
48984  __builtin_neon_vld2q_dup_v(&__ret, __p0, 49); \
48985 \
48986  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48987  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48988  __ret; \
48989})
48990#endif
48991
48992#ifdef __LITTLE_ENDIAN__
48993#define vld2q_dup_s8(__p0) __extension__ ({ \
48994  int8x16x2_t __ret; \
48995  __builtin_neon_vld2q_dup_v(&__ret, __p0, 32); \
48996  __ret; \
48997})
48998#else
48999#define vld2q_dup_s8(__p0) __extension__ ({ \
49000  int8x16x2_t __ret; \
49001  __builtin_neon_vld2q_dup_v(&__ret, __p0, 32); \
49002 \
49003  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49004  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49005  __ret; \
49006})
49007#endif
49008
49009#ifdef __LITTLE_ENDIAN__
49010#define vld2q_dup_f64(__p0) __extension__ ({ \
49011  float64x2x2_t __ret; \
49012  __builtin_neon_vld2q_dup_v(&__ret, __p0, 42); \
49013  __ret; \
49014})
49015#else
49016#define vld2q_dup_f64(__p0) __extension__ ({ \
49017  float64x2x2_t __ret; \
49018  __builtin_neon_vld2q_dup_v(&__ret, __p0, 42); \
49019 \
49020  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49021  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49022  __ret; \
49023})
49024#endif
49025
49026#ifdef __LITTLE_ENDIAN__
49027#define vld2q_dup_f32(__p0) __extension__ ({ \
49028  float32x4x2_t __ret; \
49029  __builtin_neon_vld2q_dup_v(&__ret, __p0, 41); \
49030  __ret; \
49031})
49032#else
49033#define vld2q_dup_f32(__p0) __extension__ ({ \
49034  float32x4x2_t __ret; \
49035  __builtin_neon_vld2q_dup_v(&__ret, __p0, 41); \
49036 \
49037  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49038  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49039  __ret; \
49040})
49041#endif
49042
49043#ifdef __LITTLE_ENDIAN__
49044#define vld2q_dup_f16(__p0) __extension__ ({ \
49045  float16x8x2_t __ret; \
49046  __builtin_neon_vld2q_dup_v(&__ret, __p0, 40); \
49047  __ret; \
49048})
49049#else
49050#define vld2q_dup_f16(__p0) __extension__ ({ \
49051  float16x8x2_t __ret; \
49052  __builtin_neon_vld2q_dup_v(&__ret, __p0, 40); \
49053 \
49054  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49055  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49056  __ret; \
49057})
49058#endif
49059
49060#ifdef __LITTLE_ENDIAN__
49061#define vld2q_dup_s32(__p0) __extension__ ({ \
49062  int32x4x2_t __ret; \
49063  __builtin_neon_vld2q_dup_v(&__ret, __p0, 34); \
49064  __ret; \
49065})
49066#else
49067#define vld2q_dup_s32(__p0) __extension__ ({ \
49068  int32x4x2_t __ret; \
49069  __builtin_neon_vld2q_dup_v(&__ret, __p0, 34); \
49070 \
49071  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49072  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49073  __ret; \
49074})
49075#endif
49076
49077#ifdef __LITTLE_ENDIAN__
49078#define vld2q_dup_s64(__p0) __extension__ ({ \
49079  int64x2x2_t __ret; \
49080  __builtin_neon_vld2q_dup_v(&__ret, __p0, 35); \
49081  __ret; \
49082})
49083#else
49084#define vld2q_dup_s64(__p0) __extension__ ({ \
49085  int64x2x2_t __ret; \
49086  __builtin_neon_vld2q_dup_v(&__ret, __p0, 35); \
49087 \
49088  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49089  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49090  __ret; \
49091})
49092#endif
49093
49094#ifdef __LITTLE_ENDIAN__
49095#define vld2q_dup_s16(__p0) __extension__ ({ \
49096  int16x8x2_t __ret; \
49097  __builtin_neon_vld2q_dup_v(&__ret, __p0, 33); \
49098  __ret; \
49099})
49100#else
49101#define vld2q_dup_s16(__p0) __extension__ ({ \
49102  int16x8x2_t __ret; \
49103  __builtin_neon_vld2q_dup_v(&__ret, __p0, 33); \
49104 \
49105  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49106  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49107  __ret; \
49108})
49109#endif
49110
49111#ifdef __LITTLE_ENDIAN__
49112#define vld2_dup_f64(__p0) __extension__ ({ \
49113  float64x1x2_t __ret; \
49114  __builtin_neon_vld2_dup_v(&__ret, __p0, 10); \
49115  __ret; \
49116})
49117#else
49118#define vld2_dup_f64(__p0) __extension__ ({ \
49119  float64x1x2_t __ret; \
49120  __builtin_neon_vld2_dup_v(&__ret, __p0, 10); \
49121  __ret; \
49122})
49123#endif
49124
49125#ifdef __LITTLE_ENDIAN__
49126#define vld2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
49127  poly64x1x2_t __s1 = __p1; \
49128  poly64x1x2_t __ret; \
49129  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
49130  __ret; \
49131})
49132#else
49133#define vld2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
49134  poly64x1x2_t __s1 = __p1; \
49135  poly64x1x2_t __ret; \
49136  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
49137  __ret; \
49138})
49139#endif
49140
49141#ifdef __LITTLE_ENDIAN__
49142#define vld2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
49143  poly8x16x2_t __s1 = __p1; \
49144  poly8x16x2_t __ret; \
49145  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 36); \
49146  __ret; \
49147})
49148#else
49149#define vld2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
49150  poly8x16x2_t __s1 = __p1; \
49151  poly8x16x2_t __rev1; \
49152  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49153  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49154  poly8x16x2_t __ret; \
49155  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 36); \
49156 \
49157  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49158  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49159  __ret; \
49160})
49161#endif
49162
49163#ifdef __LITTLE_ENDIAN__
49164#define vld2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
49165  poly64x2x2_t __s1 = __p1; \
49166  poly64x2x2_t __ret; \
49167  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 38); \
49168  __ret; \
49169})
49170#else
49171#define vld2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
49172  poly64x2x2_t __s1 = __p1; \
49173  poly64x2x2_t __rev1; \
49174  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
49175  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
49176  poly64x2x2_t __ret; \
49177  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 38); \
49178 \
49179  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49180  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49181  __ret; \
49182})
49183#endif
49184
49185#ifdef __LITTLE_ENDIAN__
49186#define vld2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
49187  uint8x16x2_t __s1 = __p1; \
49188  uint8x16x2_t __ret; \
49189  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 48); \
49190  __ret; \
49191})
49192#else
49193#define vld2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
49194  uint8x16x2_t __s1 = __p1; \
49195  uint8x16x2_t __rev1; \
49196  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49197  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49198  uint8x16x2_t __ret; \
49199  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 48); \
49200 \
49201  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49202  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49203  __ret; \
49204})
49205#endif
49206
49207#ifdef __LITTLE_ENDIAN__
49208#define vld2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
49209  uint64x2x2_t __s1 = __p1; \
49210  uint64x2x2_t __ret; \
49211  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 51); \
49212  __ret; \
49213})
49214#else
49215#define vld2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
49216  uint64x2x2_t __s1 = __p1; \
49217  uint64x2x2_t __rev1; \
49218  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
49219  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
49220  uint64x2x2_t __ret; \
49221  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 51); \
49222 \
49223  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49224  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49225  __ret; \
49226})
49227#endif
49228
49229#ifdef __LITTLE_ENDIAN__
49230#define vld2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
49231  int8x16x2_t __s1 = __p1; \
49232  int8x16x2_t __ret; \
49233  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 32); \
49234  __ret; \
49235})
49236#else
49237#define vld2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
49238  int8x16x2_t __s1 = __p1; \
49239  int8x16x2_t __rev1; \
49240  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49241  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49242  int8x16x2_t __ret; \
49243  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 32); \
49244 \
49245  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49246  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49247  __ret; \
49248})
49249#endif
49250
49251#ifdef __LITTLE_ENDIAN__
49252#define vld2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
49253  float64x2x2_t __s1 = __p1; \
49254  float64x2x2_t __ret; \
49255  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 42); \
49256  __ret; \
49257})
49258#else
49259#define vld2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
49260  float64x2x2_t __s1 = __p1; \
49261  float64x2x2_t __rev1; \
49262  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
49263  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
49264  float64x2x2_t __ret; \
49265  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 42); \
49266 \
49267  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49268  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49269  __ret; \
49270})
49271#endif
49272
49273#ifdef __LITTLE_ENDIAN__
49274#define vld2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
49275  int64x2x2_t __s1 = __p1; \
49276  int64x2x2_t __ret; \
49277  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 35); \
49278  __ret; \
49279})
49280#else
49281#define vld2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
49282  int64x2x2_t __s1 = __p1; \
49283  int64x2x2_t __rev1; \
49284  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
49285  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
49286  int64x2x2_t __ret; \
49287  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 35); \
49288 \
49289  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49290  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49291  __ret; \
49292})
49293#endif
49294
49295#ifdef __LITTLE_ENDIAN__
49296#define vld2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
49297  uint64x1x2_t __s1 = __p1; \
49298  uint64x1x2_t __ret; \
49299  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
49300  __ret; \
49301})
49302#else
49303#define vld2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
49304  uint64x1x2_t __s1 = __p1; \
49305  uint64x1x2_t __ret; \
49306  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
49307  __ret; \
49308})
49309#endif
49310
49311#ifdef __LITTLE_ENDIAN__
49312#define vld2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
49313  float64x1x2_t __s1 = __p1; \
49314  float64x1x2_t __ret; \
49315  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 10); \
49316  __ret; \
49317})
49318#else
49319#define vld2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
49320  float64x1x2_t __s1 = __p1; \
49321  float64x1x2_t __ret; \
49322  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 10); \
49323  __ret; \
49324})
49325#endif
49326
49327#ifdef __LITTLE_ENDIAN__
49328#define vld2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
49329  int64x1x2_t __s1 = __p1; \
49330  int64x1x2_t __ret; \
49331  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 3); \
49332  __ret; \
49333})
49334#else
49335#define vld2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
49336  int64x1x2_t __s1 = __p1; \
49337  int64x1x2_t __ret; \
49338  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 3); \
49339  __ret; \
49340})
49341#endif
49342
49343#ifdef __LITTLE_ENDIAN__
49344#define vld3_p64(__p0) __extension__ ({ \
49345  poly64x1x3_t __ret; \
49346  __builtin_neon_vld3_v(&__ret, __p0, 6); \
49347  __ret; \
49348})
49349#else
49350#define vld3_p64(__p0) __extension__ ({ \
49351  poly64x1x3_t __ret; \
49352  __builtin_neon_vld3_v(&__ret, __p0, 6); \
49353  __ret; \
49354})
49355#endif
49356
49357#ifdef __LITTLE_ENDIAN__
49358#define vld3q_p64(__p0) __extension__ ({ \
49359  poly64x2x3_t __ret; \
49360  __builtin_neon_vld3q_v(&__ret, __p0, 38); \
49361  __ret; \
49362})
49363#else
49364#define vld3q_p64(__p0) __extension__ ({ \
49365  poly64x2x3_t __ret; \
49366  __builtin_neon_vld3q_v(&__ret, __p0, 38); \
49367 \
49368  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49369  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49370  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49371  __ret; \
49372})
49373#endif
49374
49375#ifdef __LITTLE_ENDIAN__
49376#define vld3q_u64(__p0) __extension__ ({ \
49377  uint64x2x3_t __ret; \
49378  __builtin_neon_vld3q_v(&__ret, __p0, 51); \
49379  __ret; \
49380})
49381#else
49382#define vld3q_u64(__p0) __extension__ ({ \
49383  uint64x2x3_t __ret; \
49384  __builtin_neon_vld3q_v(&__ret, __p0, 51); \
49385 \
49386  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49387  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49388  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49389  __ret; \
49390})
49391#endif
49392
49393#ifdef __LITTLE_ENDIAN__
49394#define vld3q_f64(__p0) __extension__ ({ \
49395  float64x2x3_t __ret; \
49396  __builtin_neon_vld3q_v(&__ret, __p0, 42); \
49397  __ret; \
49398})
49399#else
49400#define vld3q_f64(__p0) __extension__ ({ \
49401  float64x2x3_t __ret; \
49402  __builtin_neon_vld3q_v(&__ret, __p0, 42); \
49403 \
49404  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49405  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49406  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49407  __ret; \
49408})
49409#endif
49410
49411#ifdef __LITTLE_ENDIAN__
49412#define vld3q_s64(__p0) __extension__ ({ \
49413  int64x2x3_t __ret; \
49414  __builtin_neon_vld3q_v(&__ret, __p0, 35); \
49415  __ret; \
49416})
49417#else
49418#define vld3q_s64(__p0) __extension__ ({ \
49419  int64x2x3_t __ret; \
49420  __builtin_neon_vld3q_v(&__ret, __p0, 35); \
49421 \
49422  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49423  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49424  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49425  __ret; \
49426})
49427#endif
49428
49429#ifdef __LITTLE_ENDIAN__
49430#define vld3_f64(__p0) __extension__ ({ \
49431  float64x1x3_t __ret; \
49432  __builtin_neon_vld3_v(&__ret, __p0, 10); \
49433  __ret; \
49434})
49435#else
49436#define vld3_f64(__p0) __extension__ ({ \
49437  float64x1x3_t __ret; \
49438  __builtin_neon_vld3_v(&__ret, __p0, 10); \
49439  __ret; \
49440})
49441#endif
49442
49443#ifdef __LITTLE_ENDIAN__
49444#define vld3_dup_p64(__p0) __extension__ ({ \
49445  poly64x1x3_t __ret; \
49446  __builtin_neon_vld3_dup_v(&__ret, __p0, 6); \
49447  __ret; \
49448})
49449#else
49450#define vld3_dup_p64(__p0) __extension__ ({ \
49451  poly64x1x3_t __ret; \
49452  __builtin_neon_vld3_dup_v(&__ret, __p0, 6); \
49453  __ret; \
49454})
49455#endif
49456
49457#ifdef __LITTLE_ENDIAN__
49458#define vld3q_dup_p8(__p0) __extension__ ({ \
49459  poly8x16x3_t __ret; \
49460  __builtin_neon_vld3q_dup_v(&__ret, __p0, 36); \
49461  __ret; \
49462})
49463#else
49464#define vld3q_dup_p8(__p0) __extension__ ({ \
49465  poly8x16x3_t __ret; \
49466  __builtin_neon_vld3q_dup_v(&__ret, __p0, 36); \
49467 \
49468  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49469  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49470  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49471  __ret; \
49472})
49473#endif
49474
49475#ifdef __LITTLE_ENDIAN__
49476#define vld3q_dup_p64(__p0) __extension__ ({ \
49477  poly64x2x3_t __ret; \
49478  __builtin_neon_vld3q_dup_v(&__ret, __p0, 38); \
49479  __ret; \
49480})
49481#else
49482#define vld3q_dup_p64(__p0) __extension__ ({ \
49483  poly64x2x3_t __ret; \
49484  __builtin_neon_vld3q_dup_v(&__ret, __p0, 38); \
49485 \
49486  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49487  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49488  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49489  __ret; \
49490})
49491#endif
49492
49493#ifdef __LITTLE_ENDIAN__
49494#define vld3q_dup_p16(__p0) __extension__ ({ \
49495  poly16x8x3_t __ret; \
49496  __builtin_neon_vld3q_dup_v(&__ret, __p0, 37); \
49497  __ret; \
49498})
49499#else
49500#define vld3q_dup_p16(__p0) __extension__ ({ \
49501  poly16x8x3_t __ret; \
49502  __builtin_neon_vld3q_dup_v(&__ret, __p0, 37); \
49503 \
49504  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49505  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49506  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49507  __ret; \
49508})
49509#endif
49510
49511#ifdef __LITTLE_ENDIAN__
49512#define vld3q_dup_u8(__p0) __extension__ ({ \
49513  uint8x16x3_t __ret; \
49514  __builtin_neon_vld3q_dup_v(&__ret, __p0, 48); \
49515  __ret; \
49516})
49517#else
49518#define vld3q_dup_u8(__p0) __extension__ ({ \
49519  uint8x16x3_t __ret; \
49520  __builtin_neon_vld3q_dup_v(&__ret, __p0, 48); \
49521 \
49522  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49523  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49524  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49525  __ret; \
49526})
49527#endif
49528
49529#ifdef __LITTLE_ENDIAN__
49530#define vld3q_dup_u32(__p0) __extension__ ({ \
49531  uint32x4x3_t __ret; \
49532  __builtin_neon_vld3q_dup_v(&__ret, __p0, 50); \
49533  __ret; \
49534})
49535#else
49536#define vld3q_dup_u32(__p0) __extension__ ({ \
49537  uint32x4x3_t __ret; \
49538  __builtin_neon_vld3q_dup_v(&__ret, __p0, 50); \
49539 \
49540  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49541  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49542  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49543  __ret; \
49544})
49545#endif
49546
49547#ifdef __LITTLE_ENDIAN__
49548#define vld3q_dup_u64(__p0) __extension__ ({ \
49549  uint64x2x3_t __ret; \
49550  __builtin_neon_vld3q_dup_v(&__ret, __p0, 51); \
49551  __ret; \
49552})
49553#else
49554#define vld3q_dup_u64(__p0) __extension__ ({ \
49555  uint64x2x3_t __ret; \
49556  __builtin_neon_vld3q_dup_v(&__ret, __p0, 51); \
49557 \
49558  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49559  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49560  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49561  __ret; \
49562})
49563#endif
49564
49565#ifdef __LITTLE_ENDIAN__
49566#define vld3q_dup_u16(__p0) __extension__ ({ \
49567  uint16x8x3_t __ret; \
49568  __builtin_neon_vld3q_dup_v(&__ret, __p0, 49); \
49569  __ret; \
49570})
49571#else
49572#define vld3q_dup_u16(__p0) __extension__ ({ \
49573  uint16x8x3_t __ret; \
49574  __builtin_neon_vld3q_dup_v(&__ret, __p0, 49); \
49575 \
49576  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49577  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49578  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49579  __ret; \
49580})
49581#endif
49582
49583#ifdef __LITTLE_ENDIAN__
49584#define vld3q_dup_s8(__p0) __extension__ ({ \
49585  int8x16x3_t __ret; \
49586  __builtin_neon_vld3q_dup_v(&__ret, __p0, 32); \
49587  __ret; \
49588})
49589#else
49590#define vld3q_dup_s8(__p0) __extension__ ({ \
49591  int8x16x3_t __ret; \
49592  __builtin_neon_vld3q_dup_v(&__ret, __p0, 32); \
49593 \
49594  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49595  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49596  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49597  __ret; \
49598})
49599#endif
49600
49601#ifdef __LITTLE_ENDIAN__
49602#define vld3q_dup_f64(__p0) __extension__ ({ \
49603  float64x2x3_t __ret; \
49604  __builtin_neon_vld3q_dup_v(&__ret, __p0, 42); \
49605  __ret; \
49606})
49607#else
49608#define vld3q_dup_f64(__p0) __extension__ ({ \
49609  float64x2x3_t __ret; \
49610  __builtin_neon_vld3q_dup_v(&__ret, __p0, 42); \
49611 \
49612  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49613  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49614  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49615  __ret; \
49616})
49617#endif
49618
49619#ifdef __LITTLE_ENDIAN__
49620#define vld3q_dup_f32(__p0) __extension__ ({ \
49621  float32x4x3_t __ret; \
49622  __builtin_neon_vld3q_dup_v(&__ret, __p0, 41); \
49623  __ret; \
49624})
49625#else
49626#define vld3q_dup_f32(__p0) __extension__ ({ \
49627  float32x4x3_t __ret; \
49628  __builtin_neon_vld3q_dup_v(&__ret, __p0, 41); \
49629 \
49630  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49631  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49632  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49633  __ret; \
49634})
49635#endif
49636
49637#ifdef __LITTLE_ENDIAN__
49638#define vld3q_dup_f16(__p0) __extension__ ({ \
49639  float16x8x3_t __ret; \
49640  __builtin_neon_vld3q_dup_v(&__ret, __p0, 40); \
49641  __ret; \
49642})
49643#else
49644#define vld3q_dup_f16(__p0) __extension__ ({ \
49645  float16x8x3_t __ret; \
49646  __builtin_neon_vld3q_dup_v(&__ret, __p0, 40); \
49647 \
49648  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49649  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49650  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49651  __ret; \
49652})
49653#endif
49654
49655#ifdef __LITTLE_ENDIAN__
49656#define vld3q_dup_s32(__p0) __extension__ ({ \
49657  int32x4x3_t __ret; \
49658  __builtin_neon_vld3q_dup_v(&__ret, __p0, 34); \
49659  __ret; \
49660})
49661#else
49662#define vld3q_dup_s32(__p0) __extension__ ({ \
49663  int32x4x3_t __ret; \
49664  __builtin_neon_vld3q_dup_v(&__ret, __p0, 34); \
49665 \
49666  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49667  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49668  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49669  __ret; \
49670})
49671#endif
49672
49673#ifdef __LITTLE_ENDIAN__
49674#define vld3q_dup_s64(__p0) __extension__ ({ \
49675  int64x2x3_t __ret; \
49676  __builtin_neon_vld3q_dup_v(&__ret, __p0, 35); \
49677  __ret; \
49678})
49679#else
49680#define vld3q_dup_s64(__p0) __extension__ ({ \
49681  int64x2x3_t __ret; \
49682  __builtin_neon_vld3q_dup_v(&__ret, __p0, 35); \
49683 \
49684  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49685  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49686  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49687  __ret; \
49688})
49689#endif
49690
49691#ifdef __LITTLE_ENDIAN__
49692#define vld3q_dup_s16(__p0) __extension__ ({ \
49693  int16x8x3_t __ret; \
49694  __builtin_neon_vld3q_dup_v(&__ret, __p0, 33); \
49695  __ret; \
49696})
49697#else
49698#define vld3q_dup_s16(__p0) __extension__ ({ \
49699  int16x8x3_t __ret; \
49700  __builtin_neon_vld3q_dup_v(&__ret, __p0, 33); \
49701 \
49702  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49703  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49704  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49705  __ret; \
49706})
49707#endif
49708
49709#ifdef __LITTLE_ENDIAN__
49710#define vld3_dup_f64(__p0) __extension__ ({ \
49711  float64x1x3_t __ret; \
49712  __builtin_neon_vld3_dup_v(&__ret, __p0, 10); \
49713  __ret; \
49714})
49715#else
49716#define vld3_dup_f64(__p0) __extension__ ({ \
49717  float64x1x3_t __ret; \
49718  __builtin_neon_vld3_dup_v(&__ret, __p0, 10); \
49719  __ret; \
49720})
49721#endif
49722
49723#ifdef __LITTLE_ENDIAN__
49724#define vld3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
49725  poly64x1x3_t __s1 = __p1; \
49726  poly64x1x3_t __ret; \
49727  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
49728  __ret; \
49729})
49730#else
49731#define vld3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
49732  poly64x1x3_t __s1 = __p1; \
49733  poly64x1x3_t __ret; \
49734  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
49735  __ret; \
49736})
49737#endif
49738
49739#ifdef __LITTLE_ENDIAN__
49740#define vld3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
49741  poly8x16x3_t __s1 = __p1; \
49742  poly8x16x3_t __ret; \
49743  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 36); \
49744  __ret; \
49745})
49746#else
49747#define vld3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
49748  poly8x16x3_t __s1 = __p1; \
49749  poly8x16x3_t __rev1; \
49750  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49751  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49752  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49753  poly8x16x3_t __ret; \
49754  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 36); \
49755 \
49756  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49757  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49758  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49759  __ret; \
49760})
49761#endif
49762
49763#ifdef __LITTLE_ENDIAN__
49764#define vld3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
49765  poly64x2x3_t __s1 = __p1; \
49766  poly64x2x3_t __ret; \
49767  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 38); \
49768  __ret; \
49769})
49770#else
49771#define vld3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
49772  poly64x2x3_t __s1 = __p1; \
49773  poly64x2x3_t __rev1; \
49774  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
49775  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
49776  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
49777  poly64x2x3_t __ret; \
49778  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 38); \
49779 \
49780  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49781  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49782  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49783  __ret; \
49784})
49785#endif
49786
49787#ifdef __LITTLE_ENDIAN__
49788#define vld3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
49789  uint8x16x3_t __s1 = __p1; \
49790  uint8x16x3_t __ret; \
49791  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 48); \
49792  __ret; \
49793})
49794#else
49795#define vld3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
49796  uint8x16x3_t __s1 = __p1; \
49797  uint8x16x3_t __rev1; \
49798  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49799  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49800  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49801  uint8x16x3_t __ret; \
49802  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 48); \
49803 \
49804  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49805  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49806  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49807  __ret; \
49808})
49809#endif
49810
49811#ifdef __LITTLE_ENDIAN__
49812#define vld3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
49813  uint64x2x3_t __s1 = __p1; \
49814  uint64x2x3_t __ret; \
49815  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 51); \
49816  __ret; \
49817})
49818#else
49819#define vld3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
49820  uint64x2x3_t __s1 = __p1; \
49821  uint64x2x3_t __rev1; \
49822  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
49823  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
49824  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
49825  uint64x2x3_t __ret; \
49826  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 51); \
49827 \
49828  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49829  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49830  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49831  __ret; \
49832})
49833#endif
49834
49835#ifdef __LITTLE_ENDIAN__
49836#define vld3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
49837  int8x16x3_t __s1 = __p1; \
49838  int8x16x3_t __ret; \
49839  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 32); \
49840  __ret; \
49841})
49842#else
49843#define vld3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
49844  int8x16x3_t __s1 = __p1; \
49845  int8x16x3_t __rev1; \
49846  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49847  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49848  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49849  int8x16x3_t __ret; \
49850  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 32); \
49851 \
49852  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49853  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49854  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49855  __ret; \
49856})
49857#endif
49858
49859#ifdef __LITTLE_ENDIAN__
49860#define vld3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
49861  float64x2x3_t __s1 = __p1; \
49862  float64x2x3_t __ret; \
49863  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 42); \
49864  __ret; \
49865})
49866#else
49867#define vld3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
49868  float64x2x3_t __s1 = __p1; \
49869  float64x2x3_t __rev1; \
49870  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
49871  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
49872  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
49873  float64x2x3_t __ret; \
49874  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 42); \
49875 \
49876  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49877  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49878  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49879  __ret; \
49880})
49881#endif
49882
49883#ifdef __LITTLE_ENDIAN__
49884#define vld3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
49885  int64x2x3_t __s1 = __p1; \
49886  int64x2x3_t __ret; \
49887  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 35); \
49888  __ret; \
49889})
49890#else
49891#define vld3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
49892  int64x2x3_t __s1 = __p1; \
49893  int64x2x3_t __rev1; \
49894  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
49895  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
49896  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
49897  int64x2x3_t __ret; \
49898  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 35); \
49899 \
49900  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49901  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49902  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49903  __ret; \
49904})
49905#endif
49906
49907#ifdef __LITTLE_ENDIAN__
49908#define vld3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
49909  uint64x1x3_t __s1 = __p1; \
49910  uint64x1x3_t __ret; \
49911  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
49912  __ret; \
49913})
49914#else
49915#define vld3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
49916  uint64x1x3_t __s1 = __p1; \
49917  uint64x1x3_t __ret; \
49918  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
49919  __ret; \
49920})
49921#endif
49922
49923#ifdef __LITTLE_ENDIAN__
49924#define vld3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
49925  float64x1x3_t __s1 = __p1; \
49926  float64x1x3_t __ret; \
49927  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
49928  __ret; \
49929})
49930#else
49931#define vld3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
49932  float64x1x3_t __s1 = __p1; \
49933  float64x1x3_t __ret; \
49934  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
49935  __ret; \
49936})
49937#endif
49938
49939#ifdef __LITTLE_ENDIAN__
49940#define vld3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
49941  int64x1x3_t __s1 = __p1; \
49942  int64x1x3_t __ret; \
49943  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
49944  __ret; \
49945})
49946#else
49947#define vld3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
49948  int64x1x3_t __s1 = __p1; \
49949  int64x1x3_t __ret; \
49950  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
49951  __ret; \
49952})
49953#endif
49954
49955#ifdef __LITTLE_ENDIAN__
49956#define vld4_p64(__p0) __extension__ ({ \
49957  poly64x1x4_t __ret; \
49958  __builtin_neon_vld4_v(&__ret, __p0, 6); \
49959  __ret; \
49960})
49961#else
49962#define vld4_p64(__p0) __extension__ ({ \
49963  poly64x1x4_t __ret; \
49964  __builtin_neon_vld4_v(&__ret, __p0, 6); \
49965  __ret; \
49966})
49967#endif
49968
49969#ifdef __LITTLE_ENDIAN__
49970#define vld4q_p64(__p0) __extension__ ({ \
49971  poly64x2x4_t __ret; \
49972  __builtin_neon_vld4q_v(&__ret, __p0, 38); \
49973  __ret; \
49974})
49975#else
49976#define vld4q_p64(__p0) __extension__ ({ \
49977  poly64x2x4_t __ret; \
49978  __builtin_neon_vld4q_v(&__ret, __p0, 38); \
49979 \
49980  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49981  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49982  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49983  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49984  __ret; \
49985})
49986#endif
49987
49988#ifdef __LITTLE_ENDIAN__
49989#define vld4q_u64(__p0) __extension__ ({ \
49990  uint64x2x4_t __ret; \
49991  __builtin_neon_vld4q_v(&__ret, __p0, 51); \
49992  __ret; \
49993})
49994#else
49995#define vld4q_u64(__p0) __extension__ ({ \
49996  uint64x2x4_t __ret; \
49997  __builtin_neon_vld4q_v(&__ret, __p0, 51); \
49998 \
49999  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50000  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50001  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50002  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
50003  __ret; \
50004})
50005#endif
50006
50007#ifdef __LITTLE_ENDIAN__
50008#define vld4q_f64(__p0) __extension__ ({ \
50009  float64x2x4_t __ret; \
50010  __builtin_neon_vld4q_v(&__ret, __p0, 42); \
50011  __ret; \
50012})
50013#else
50014#define vld4q_f64(__p0) __extension__ ({ \
50015  float64x2x4_t __ret; \
50016  __builtin_neon_vld4q_v(&__ret, __p0, 42); \
50017 \
50018  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50019  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50020  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50021  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
50022  __ret; \
50023})
50024#endif
50025
50026#ifdef __LITTLE_ENDIAN__
50027#define vld4q_s64(__p0) __extension__ ({ \
50028  int64x2x4_t __ret; \
50029  __builtin_neon_vld4q_v(&__ret, __p0, 35); \
50030  __ret; \
50031})
50032#else
50033#define vld4q_s64(__p0) __extension__ ({ \
50034  int64x2x4_t __ret; \
50035  __builtin_neon_vld4q_v(&__ret, __p0, 35); \
50036 \
50037  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50038  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50039  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50040  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
50041  __ret; \
50042})
50043#endif
50044
50045#ifdef __LITTLE_ENDIAN__
50046#define vld4_f64(__p0) __extension__ ({ \
50047  float64x1x4_t __ret; \
50048  __builtin_neon_vld4_v(&__ret, __p0, 10); \
50049  __ret; \
50050})
50051#else
50052#define vld4_f64(__p0) __extension__ ({ \
50053  float64x1x4_t __ret; \
50054  __builtin_neon_vld4_v(&__ret, __p0, 10); \
50055  __ret; \
50056})
50057#endif
50058
50059#ifdef __LITTLE_ENDIAN__
50060#define vld4_dup_p64(__p0) __extension__ ({ \
50061  poly64x1x4_t __ret; \
50062  __builtin_neon_vld4_dup_v(&__ret, __p0, 6); \
50063  __ret; \
50064})
50065#else
50066#define vld4_dup_p64(__p0) __extension__ ({ \
50067  poly64x1x4_t __ret; \
50068  __builtin_neon_vld4_dup_v(&__ret, __p0, 6); \
50069  __ret; \
50070})
50071#endif
50072
50073#ifdef __LITTLE_ENDIAN__
50074#define vld4q_dup_p8(__p0) __extension__ ({ \
50075  poly8x16x4_t __ret; \
50076  __builtin_neon_vld4q_dup_v(&__ret, __p0, 36); \
50077  __ret; \
50078})
50079#else
50080#define vld4q_dup_p8(__p0) __extension__ ({ \
50081  poly8x16x4_t __ret; \
50082  __builtin_neon_vld4q_dup_v(&__ret, __p0, 36); \
50083 \
50084  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50085  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50086  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50087  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50088  __ret; \
50089})
50090#endif
50091
50092#ifdef __LITTLE_ENDIAN__
50093#define vld4q_dup_p64(__p0) __extension__ ({ \
50094  poly64x2x4_t __ret; \
50095  __builtin_neon_vld4q_dup_v(&__ret, __p0, 38); \
50096  __ret; \
50097})
50098#else
50099#define vld4q_dup_p64(__p0) __extension__ ({ \
50100  poly64x2x4_t __ret; \
50101  __builtin_neon_vld4q_dup_v(&__ret, __p0, 38); \
50102 \
50103  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50104  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50105  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50106  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
50107  __ret; \
50108})
50109#endif
50110
50111#ifdef __LITTLE_ENDIAN__
50112#define vld4q_dup_p16(__p0) __extension__ ({ \
50113  poly16x8x4_t __ret; \
50114  __builtin_neon_vld4q_dup_v(&__ret, __p0, 37); \
50115  __ret; \
50116})
50117#else
50118#define vld4q_dup_p16(__p0) __extension__ ({ \
50119  poly16x8x4_t __ret; \
50120  __builtin_neon_vld4q_dup_v(&__ret, __p0, 37); \
50121 \
50122  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50123  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50124  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50125  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
50126  __ret; \
50127})
50128#endif
50129
50130#ifdef __LITTLE_ENDIAN__
50131#define vld4q_dup_u8(__p0) __extension__ ({ \
50132  uint8x16x4_t __ret; \
50133  __builtin_neon_vld4q_dup_v(&__ret, __p0, 48); \
50134  __ret; \
50135})
50136#else
50137#define vld4q_dup_u8(__p0) __extension__ ({ \
50138  uint8x16x4_t __ret; \
50139  __builtin_neon_vld4q_dup_v(&__ret, __p0, 48); \
50140 \
50141  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50142  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50143  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50144  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50145  __ret; \
50146})
50147#endif
50148
50149#ifdef __LITTLE_ENDIAN__
50150#define vld4q_dup_u32(__p0) __extension__ ({ \
50151  uint32x4x4_t __ret; \
50152  __builtin_neon_vld4q_dup_v(&__ret, __p0, 50); \
50153  __ret; \
50154})
50155#else
50156#define vld4q_dup_u32(__p0) __extension__ ({ \
50157  uint32x4x4_t __ret; \
50158  __builtin_neon_vld4q_dup_v(&__ret, __p0, 50); \
50159 \
50160  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50161  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50162  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
50163  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
50164  __ret; \
50165})
50166#endif
50167
50168#ifdef __LITTLE_ENDIAN__
50169#define vld4q_dup_u64(__p0) __extension__ ({ \
50170  uint64x2x4_t __ret; \
50171  __builtin_neon_vld4q_dup_v(&__ret, __p0, 51); \
50172  __ret; \
50173})
50174#else
50175#define vld4q_dup_u64(__p0) __extension__ ({ \
50176  uint64x2x4_t __ret; \
50177  __builtin_neon_vld4q_dup_v(&__ret, __p0, 51); \
50178 \
50179  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50180  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50181  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50182  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
50183  __ret; \
50184})
50185#endif
50186
50187#ifdef __LITTLE_ENDIAN__
50188#define vld4q_dup_u16(__p0) __extension__ ({ \
50189  uint16x8x4_t __ret; \
50190  __builtin_neon_vld4q_dup_v(&__ret, __p0, 49); \
50191  __ret; \
50192})
50193#else
50194#define vld4q_dup_u16(__p0) __extension__ ({ \
50195  uint16x8x4_t __ret; \
50196  __builtin_neon_vld4q_dup_v(&__ret, __p0, 49); \
50197 \
50198  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50199  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50200  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50201  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
50202  __ret; \
50203})
50204#endif
50205
50206#ifdef __LITTLE_ENDIAN__
50207#define vld4q_dup_s8(__p0) __extension__ ({ \
50208  int8x16x4_t __ret; \
50209  __builtin_neon_vld4q_dup_v(&__ret, __p0, 32); \
50210  __ret; \
50211})
50212#else
50213#define vld4q_dup_s8(__p0) __extension__ ({ \
50214  int8x16x4_t __ret; \
50215  __builtin_neon_vld4q_dup_v(&__ret, __p0, 32); \
50216 \
50217  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50218  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50219  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50220  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50221  __ret; \
50222})
50223#endif
50224
50225#ifdef __LITTLE_ENDIAN__
50226#define vld4q_dup_f64(__p0) __extension__ ({ \
50227  float64x2x4_t __ret; \
50228  __builtin_neon_vld4q_dup_v(&__ret, __p0, 42); \
50229  __ret; \
50230})
50231#else
50232#define vld4q_dup_f64(__p0) __extension__ ({ \
50233  float64x2x4_t __ret; \
50234  __builtin_neon_vld4q_dup_v(&__ret, __p0, 42); \
50235 \
50236  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50237  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50238  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50239  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
50240  __ret; \
50241})
50242#endif
50243
50244#ifdef __LITTLE_ENDIAN__
50245#define vld4q_dup_f32(__p0) __extension__ ({ \
50246  float32x4x4_t __ret; \
50247  __builtin_neon_vld4q_dup_v(&__ret, __p0, 41); \
50248  __ret; \
50249})
50250#else
50251#define vld4q_dup_f32(__p0) __extension__ ({ \
50252  float32x4x4_t __ret; \
50253  __builtin_neon_vld4q_dup_v(&__ret, __p0, 41); \
50254 \
50255  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50256  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50257  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
50258  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
50259  __ret; \
50260})
50261#endif
50262
50263#ifdef __LITTLE_ENDIAN__
50264#define vld4q_dup_f16(__p0) __extension__ ({ \
50265  float16x8x4_t __ret; \
50266  __builtin_neon_vld4q_dup_v(&__ret, __p0, 40); \
50267  __ret; \
50268})
50269#else
50270#define vld4q_dup_f16(__p0) __extension__ ({ \
50271  float16x8x4_t __ret; \
50272  __builtin_neon_vld4q_dup_v(&__ret, __p0, 40); \
50273 \
50274  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50275  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50276  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50277  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
50278  __ret; \
50279})
50280#endif
50281
50282#ifdef __LITTLE_ENDIAN__
50283#define vld4q_dup_s32(__p0) __extension__ ({ \
50284  int32x4x4_t __ret; \
50285  __builtin_neon_vld4q_dup_v(&__ret, __p0, 34); \
50286  __ret; \
50287})
50288#else
50289#define vld4q_dup_s32(__p0) __extension__ ({ \
50290  int32x4x4_t __ret; \
50291  __builtin_neon_vld4q_dup_v(&__ret, __p0, 34); \
50292 \
50293  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50294  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50295  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
50296  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
50297  __ret; \
50298})
50299#endif
50300
50301#ifdef __LITTLE_ENDIAN__
50302#define vld4q_dup_s64(__p0) __extension__ ({ \
50303  int64x2x4_t __ret; \
50304  __builtin_neon_vld4q_dup_v(&__ret, __p0, 35); \
50305  __ret; \
50306})
50307#else
50308#define vld4q_dup_s64(__p0) __extension__ ({ \
50309  int64x2x4_t __ret; \
50310  __builtin_neon_vld4q_dup_v(&__ret, __p0, 35); \
50311 \
50312  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50313  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50314  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50315  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
50316  __ret; \
50317})
50318#endif
50319
50320#ifdef __LITTLE_ENDIAN__
50321#define vld4q_dup_s16(__p0) __extension__ ({ \
50322  int16x8x4_t __ret; \
50323  __builtin_neon_vld4q_dup_v(&__ret, __p0, 33); \
50324  __ret; \
50325})
50326#else
50327#define vld4q_dup_s16(__p0) __extension__ ({ \
50328  int16x8x4_t __ret; \
50329  __builtin_neon_vld4q_dup_v(&__ret, __p0, 33); \
50330 \
50331  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50332  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50333  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50334  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
50335  __ret; \
50336})
50337#endif
50338
50339#ifdef __LITTLE_ENDIAN__
50340#define vld4_dup_f64(__p0) __extension__ ({ \
50341  float64x1x4_t __ret; \
50342  __builtin_neon_vld4_dup_v(&__ret, __p0, 10); \
50343  __ret; \
50344})
50345#else
50346#define vld4_dup_f64(__p0) __extension__ ({ \
50347  float64x1x4_t __ret; \
50348  __builtin_neon_vld4_dup_v(&__ret, __p0, 10); \
50349  __ret; \
50350})
50351#endif
50352
50353#ifdef __LITTLE_ENDIAN__
50354#define vld4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50355  poly64x1x4_t __s1 = __p1; \
50356  poly64x1x4_t __ret; \
50357  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
50358  __ret; \
50359})
50360#else
50361#define vld4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50362  poly64x1x4_t __s1 = __p1; \
50363  poly64x1x4_t __ret; \
50364  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
50365  __ret; \
50366})
50367#endif
50368
50369#ifdef __LITTLE_ENDIAN__
50370#define vld4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
50371  poly8x16x4_t __s1 = __p1; \
50372  poly8x16x4_t __ret; \
50373  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 36); \
50374  __ret; \
50375})
50376#else
50377#define vld4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
50378  poly8x16x4_t __s1 = __p1; \
50379  poly8x16x4_t __rev1; \
50380  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50381  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50382  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50383  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50384  poly8x16x4_t __ret; \
50385  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 36); \
50386 \
50387  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50388  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50389  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50390  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50391  __ret; \
50392})
50393#endif
50394
50395#ifdef __LITTLE_ENDIAN__
50396#define vld4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50397  poly64x2x4_t __s1 = __p1; \
50398  poly64x2x4_t __ret; \
50399  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 38); \
50400  __ret; \
50401})
50402#else
50403#define vld4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50404  poly64x2x4_t __s1 = __p1; \
50405  poly64x2x4_t __rev1; \
50406  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50407  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50408  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50409  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
50410  poly64x2x4_t __ret; \
50411  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 38); \
50412 \
50413  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50414  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50415  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50416  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
50417  __ret; \
50418})
50419#endif
50420
50421#ifdef __LITTLE_ENDIAN__
50422#define vld4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
50423  uint8x16x4_t __s1 = __p1; \
50424  uint8x16x4_t __ret; \
50425  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 48); \
50426  __ret; \
50427})
50428#else
50429#define vld4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
50430  uint8x16x4_t __s1 = __p1; \
50431  uint8x16x4_t __rev1; \
50432  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50433  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50434  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50435  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50436  uint8x16x4_t __ret; \
50437  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 48); \
50438 \
50439  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50440  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50441  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50442  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50443  __ret; \
50444})
50445#endif
50446
50447#ifdef __LITTLE_ENDIAN__
50448#define vld4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50449  uint64x2x4_t __s1 = __p1; \
50450  uint64x2x4_t __ret; \
50451  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 51); \
50452  __ret; \
50453})
50454#else
50455#define vld4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50456  uint64x2x4_t __s1 = __p1; \
50457  uint64x2x4_t __rev1; \
50458  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50459  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50460  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50461  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
50462  uint64x2x4_t __ret; \
50463  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 51); \
50464 \
50465  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50466  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50467  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50468  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
50469  __ret; \
50470})
50471#endif
50472
50473#ifdef __LITTLE_ENDIAN__
50474#define vld4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
50475  int8x16x4_t __s1 = __p1; \
50476  int8x16x4_t __ret; \
50477  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 32); \
50478  __ret; \
50479})
50480#else
50481#define vld4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
50482  int8x16x4_t __s1 = __p1; \
50483  int8x16x4_t __rev1; \
50484  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50485  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50486  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50487  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50488  int8x16x4_t __ret; \
50489  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 32); \
50490 \
50491  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50492  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50493  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50494  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50495  __ret; \
50496})
50497#endif
50498
50499#ifdef __LITTLE_ENDIAN__
50500#define vld4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50501  float64x2x4_t __s1 = __p1; \
50502  float64x2x4_t __ret; \
50503  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 42); \
50504  __ret; \
50505})
50506#else
50507#define vld4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50508  float64x2x4_t __s1 = __p1; \
50509  float64x2x4_t __rev1; \
50510  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50511  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50512  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50513  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
50514  float64x2x4_t __ret; \
50515  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 42); \
50516 \
50517  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50518  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50519  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50520  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
50521  __ret; \
50522})
50523#endif
50524
50525#ifdef __LITTLE_ENDIAN__
50526#define vld4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50527  int64x2x4_t __s1 = __p1; \
50528  int64x2x4_t __ret; \
50529  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 35); \
50530  __ret; \
50531})
50532#else
50533#define vld4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50534  int64x2x4_t __s1 = __p1; \
50535  int64x2x4_t __rev1; \
50536  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50537  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50538  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50539  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
50540  int64x2x4_t __ret; \
50541  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 35); \
50542 \
50543  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50544  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50545  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50546  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
50547  __ret; \
50548})
50549#endif
50550
50551#ifdef __LITTLE_ENDIAN__
50552#define vld4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50553  uint64x1x4_t __s1 = __p1; \
50554  uint64x1x4_t __ret; \
50555  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
50556  __ret; \
50557})
50558#else
50559#define vld4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50560  uint64x1x4_t __s1 = __p1; \
50561  uint64x1x4_t __ret; \
50562  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
50563  __ret; \
50564})
50565#endif
50566
50567#ifdef __LITTLE_ENDIAN__
50568#define vld4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50569  float64x1x4_t __s1 = __p1; \
50570  float64x1x4_t __ret; \
50571  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
50572  __ret; \
50573})
50574#else
50575#define vld4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50576  float64x1x4_t __s1 = __p1; \
50577  float64x1x4_t __ret; \
50578  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
50579  __ret; \
50580})
50581#endif
50582
50583#ifdef __LITTLE_ENDIAN__
50584#define vld4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50585  int64x1x4_t __s1 = __p1; \
50586  int64x1x4_t __ret; \
50587  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
50588  __ret; \
50589})
50590#else
50591#define vld4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50592  int64x1x4_t __s1 = __p1; \
50593  int64x1x4_t __ret; \
50594  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
50595  __ret; \
50596})
50597#endif
50598
50599#ifdef __LITTLE_ENDIAN__
50600#define vldrq_p128(__p0) __extension__ ({ \
50601  poly128_t __ret; \
50602  __ret = (poly128_t) __builtin_neon_vldrq_p128(__p0); \
50603  __ret; \
50604})
50605#else
50606#define vldrq_p128(__p0) __extension__ ({ \
50607  poly128_t __ret; \
50608  __ret = (poly128_t) __builtin_neon_vldrq_p128(__p0); \
50609  __ret; \
50610})
50611#endif
50612
50613#ifdef __LITTLE_ENDIAN__
50614__ai float64x2_t vmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
50615  float64x2_t __ret;
50616  __ret = (float64x2_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
50617  return __ret;
50618}
50619#else
50620__ai float64x2_t vmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
50621  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50622  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
50623  float64x2_t __ret;
50624  __ret = (float64x2_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
50625  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
50626  return __ret;
50627}
50628#endif
50629
50630#ifdef __LITTLE_ENDIAN__
50631__ai float64x1_t vmax_f64(float64x1_t __p0, float64x1_t __p1) {
50632  float64x1_t __ret;
50633  __ret = (float64x1_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
50634  return __ret;
50635}
50636#else
50637__ai float64x1_t vmax_f64(float64x1_t __p0, float64x1_t __p1) {
50638  float64x1_t __ret;
50639  __ret = (float64x1_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
50640  return __ret;
50641}
50642#endif
50643
50644#ifdef __LITTLE_ENDIAN__
50645__ai float64x2_t vmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
50646  float64x2_t __ret;
50647  __ret = (float64x2_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
50648  return __ret;
50649}
50650#else
50651__ai float64x2_t vmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
50652  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50653  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
50654  float64x2_t __ret;
50655  __ret = (float64x2_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
50656  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
50657  return __ret;
50658}
50659#endif
50660
50661#ifdef __LITTLE_ENDIAN__
50662__ai float32x4_t vmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
50663  float32x4_t __ret;
50664  __ret = (float32x4_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
50665  return __ret;
50666}
50667#else
50668__ai float32x4_t vmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
50669  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
50670  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
50671  float32x4_t __ret;
50672  __ret = (float32x4_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
50673  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
50674  return __ret;
50675}
50676#endif
50677
50678#ifdef __LITTLE_ENDIAN__
50679__ai float64x1_t vmaxnm_f64(float64x1_t __p0, float64x1_t __p1) {
50680  float64x1_t __ret;
50681  __ret = (float64x1_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
50682  return __ret;
50683}
50684#else
50685__ai float64x1_t vmaxnm_f64(float64x1_t __p0, float64x1_t __p1) {
50686  float64x1_t __ret;
50687  __ret = (float64x1_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
50688  return __ret;
50689}
50690#endif
50691
50692#ifdef __LITTLE_ENDIAN__
50693__ai float32x2_t vmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
50694  float32x2_t __ret;
50695  __ret = (float32x2_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
50696  return __ret;
50697}
50698#else
50699__ai float32x2_t vmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
50700  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50701  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
50702  float32x2_t __ret;
50703  __ret = (float32x2_t) __builtin_neon_vmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
50704  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
50705  return __ret;
50706}
50707#endif
50708
50709#ifdef __LITTLE_ENDIAN__
50710__ai float64_t vmaxnmvq_f64(float64x2_t __p0) {
50711  float64_t __ret;
50712  __ret = (float64_t) __builtin_neon_vmaxnmvq_f64((int8x16_t)__p0);
50713  return __ret;
50714}
50715#else
50716__ai float64_t vmaxnmvq_f64(float64x2_t __p0) {
50717  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50718  float64_t __ret;
50719  __ret = (float64_t) __builtin_neon_vmaxnmvq_f64((int8x16_t)__rev0);
50720  return __ret;
50721}
50722#endif
50723
50724#ifdef __LITTLE_ENDIAN__
50725__ai float32_t vmaxnmvq_f32(float32x4_t __p0) {
50726  float32_t __ret;
50727  __ret = (float32_t) __builtin_neon_vmaxnmvq_f32((int8x16_t)__p0);
50728  return __ret;
50729}
50730#else
50731__ai float32_t vmaxnmvq_f32(float32x4_t __p0) {
50732  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
50733  float32_t __ret;
50734  __ret = (float32_t) __builtin_neon_vmaxnmvq_f32((int8x16_t)__rev0);
50735  return __ret;
50736}
50737#endif
50738
50739#ifdef __LITTLE_ENDIAN__
50740__ai float32_t vmaxnmv_f32(float32x2_t __p0) {
50741  float32_t __ret;
50742  __ret = (float32_t) __builtin_neon_vmaxnmv_f32((int8x8_t)__p0);
50743  return __ret;
50744}
50745#else
50746__ai float32_t vmaxnmv_f32(float32x2_t __p0) {
50747  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50748  float32_t __ret;
50749  __ret = (float32_t) __builtin_neon_vmaxnmv_f32((int8x8_t)__rev0);
50750  return __ret;
50751}
50752#endif
50753
50754#ifdef __LITTLE_ENDIAN__
50755__ai uint8_t vmaxvq_u8(uint8x16_t __p0) {
50756  uint8_t __ret;
50757  __ret = (uint8_t) __builtin_neon_vmaxvq_u8((int8x16_t)__p0);
50758  return __ret;
50759}
50760#else
50761__ai uint8_t vmaxvq_u8(uint8x16_t __p0) {
50762  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
50763  uint8_t __ret;
50764  __ret = (uint8_t) __builtin_neon_vmaxvq_u8((int8x16_t)__rev0);
50765  return __ret;
50766}
50767#endif
50768
50769#ifdef __LITTLE_ENDIAN__
50770__ai uint32_t vmaxvq_u32(uint32x4_t __p0) {
50771  uint32_t __ret;
50772  __ret = (uint32_t) __builtin_neon_vmaxvq_u32((int8x16_t)__p0);
50773  return __ret;
50774}
50775#else
50776__ai uint32_t vmaxvq_u32(uint32x4_t __p0) {
50777  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
50778  uint32_t __ret;
50779  __ret = (uint32_t) __builtin_neon_vmaxvq_u32((int8x16_t)__rev0);
50780  return __ret;
50781}
50782#endif
50783
50784#ifdef __LITTLE_ENDIAN__
50785__ai uint16_t vmaxvq_u16(uint16x8_t __p0) {
50786  uint16_t __ret;
50787  __ret = (uint16_t) __builtin_neon_vmaxvq_u16((int8x16_t)__p0);
50788  return __ret;
50789}
50790#else
50791__ai uint16_t vmaxvq_u16(uint16x8_t __p0) {
50792  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
50793  uint16_t __ret;
50794  __ret = (uint16_t) __builtin_neon_vmaxvq_u16((int8x16_t)__rev0);
50795  return __ret;
50796}
50797#endif
50798
50799#ifdef __LITTLE_ENDIAN__
50800__ai int8_t vmaxvq_s8(int8x16_t __p0) {
50801  int8_t __ret;
50802  __ret = (int8_t) __builtin_neon_vmaxvq_s8((int8x16_t)__p0);
50803  return __ret;
50804}
50805#else
50806__ai int8_t vmaxvq_s8(int8x16_t __p0) {
50807  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
50808  int8_t __ret;
50809  __ret = (int8_t) __builtin_neon_vmaxvq_s8((int8x16_t)__rev0);
50810  return __ret;
50811}
50812#endif
50813
50814#ifdef __LITTLE_ENDIAN__
50815__ai float64_t vmaxvq_f64(float64x2_t __p0) {
50816  float64_t __ret;
50817  __ret = (float64_t) __builtin_neon_vmaxvq_f64((int8x16_t)__p0);
50818  return __ret;
50819}
50820#else
50821__ai float64_t vmaxvq_f64(float64x2_t __p0) {
50822  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50823  float64_t __ret;
50824  __ret = (float64_t) __builtin_neon_vmaxvq_f64((int8x16_t)__rev0);
50825  return __ret;
50826}
50827#endif
50828
50829#ifdef __LITTLE_ENDIAN__
50830__ai float32_t vmaxvq_f32(float32x4_t __p0) {
50831  float32_t __ret;
50832  __ret = (float32_t) __builtin_neon_vmaxvq_f32((int8x16_t)__p0);
50833  return __ret;
50834}
50835#else
50836__ai float32_t vmaxvq_f32(float32x4_t __p0) {
50837  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
50838  float32_t __ret;
50839  __ret = (float32_t) __builtin_neon_vmaxvq_f32((int8x16_t)__rev0);
50840  return __ret;
50841}
50842#endif
50843
50844#ifdef __LITTLE_ENDIAN__
50845__ai int32_t vmaxvq_s32(int32x4_t __p0) {
50846  int32_t __ret;
50847  __ret = (int32_t) __builtin_neon_vmaxvq_s32((int8x16_t)__p0);
50848  return __ret;
50849}
50850#else
50851__ai int32_t vmaxvq_s32(int32x4_t __p0) {
50852  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
50853  int32_t __ret;
50854  __ret = (int32_t) __builtin_neon_vmaxvq_s32((int8x16_t)__rev0);
50855  return __ret;
50856}
50857#endif
50858
50859#ifdef __LITTLE_ENDIAN__
50860__ai int16_t vmaxvq_s16(int16x8_t __p0) {
50861  int16_t __ret;
50862  __ret = (int16_t) __builtin_neon_vmaxvq_s16((int8x16_t)__p0);
50863  return __ret;
50864}
50865#else
50866__ai int16_t vmaxvq_s16(int16x8_t __p0) {
50867  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
50868  int16_t __ret;
50869  __ret = (int16_t) __builtin_neon_vmaxvq_s16((int8x16_t)__rev0);
50870  return __ret;
50871}
50872#endif
50873
50874#ifdef __LITTLE_ENDIAN__
50875__ai uint8_t vmaxv_u8(uint8x8_t __p0) {
50876  uint8_t __ret;
50877  __ret = (uint8_t) __builtin_neon_vmaxv_u8((int8x8_t)__p0);
50878  return __ret;
50879}
50880#else
50881__ai uint8_t vmaxv_u8(uint8x8_t __p0) {
50882  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
50883  uint8_t __ret;
50884  __ret = (uint8_t) __builtin_neon_vmaxv_u8((int8x8_t)__rev0);
50885  return __ret;
50886}
50887#endif
50888
50889#ifdef __LITTLE_ENDIAN__
50890__ai uint32_t vmaxv_u32(uint32x2_t __p0) {
50891  uint32_t __ret;
50892  __ret = (uint32_t) __builtin_neon_vmaxv_u32((int8x8_t)__p0);
50893  return __ret;
50894}
50895#else
50896__ai uint32_t vmaxv_u32(uint32x2_t __p0) {
50897  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50898  uint32_t __ret;
50899  __ret = (uint32_t) __builtin_neon_vmaxv_u32((int8x8_t)__rev0);
50900  return __ret;
50901}
50902#endif
50903
50904#ifdef __LITTLE_ENDIAN__
50905__ai uint16_t vmaxv_u16(uint16x4_t __p0) {
50906  uint16_t __ret;
50907  __ret = (uint16_t) __builtin_neon_vmaxv_u16((int8x8_t)__p0);
50908  return __ret;
50909}
50910#else
50911__ai uint16_t vmaxv_u16(uint16x4_t __p0) {
50912  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
50913  uint16_t __ret;
50914  __ret = (uint16_t) __builtin_neon_vmaxv_u16((int8x8_t)__rev0);
50915  return __ret;
50916}
50917#endif
50918
50919#ifdef __LITTLE_ENDIAN__
50920__ai int8_t vmaxv_s8(int8x8_t __p0) {
50921  int8_t __ret;
50922  __ret = (int8_t) __builtin_neon_vmaxv_s8((int8x8_t)__p0);
50923  return __ret;
50924}
50925#else
50926__ai int8_t vmaxv_s8(int8x8_t __p0) {
50927  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
50928  int8_t __ret;
50929  __ret = (int8_t) __builtin_neon_vmaxv_s8((int8x8_t)__rev0);
50930  return __ret;
50931}
50932#endif
50933
50934#ifdef __LITTLE_ENDIAN__
50935__ai float32_t vmaxv_f32(float32x2_t __p0) {
50936  float32_t __ret;
50937  __ret = (float32_t) __builtin_neon_vmaxv_f32((int8x8_t)__p0);
50938  return __ret;
50939}
50940#else
50941__ai float32_t vmaxv_f32(float32x2_t __p0) {
50942  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50943  float32_t __ret;
50944  __ret = (float32_t) __builtin_neon_vmaxv_f32((int8x8_t)__rev0);
50945  return __ret;
50946}
50947#endif
50948
50949#ifdef __LITTLE_ENDIAN__
50950__ai int32_t vmaxv_s32(int32x2_t __p0) {
50951  int32_t __ret;
50952  __ret = (int32_t) __builtin_neon_vmaxv_s32((int8x8_t)__p0);
50953  return __ret;
50954}
50955#else
50956__ai int32_t vmaxv_s32(int32x2_t __p0) {
50957  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50958  int32_t __ret;
50959  __ret = (int32_t) __builtin_neon_vmaxv_s32((int8x8_t)__rev0);
50960  return __ret;
50961}
50962#endif
50963
50964#ifdef __LITTLE_ENDIAN__
50965__ai int16_t vmaxv_s16(int16x4_t __p0) {
50966  int16_t __ret;
50967  __ret = (int16_t) __builtin_neon_vmaxv_s16((int8x8_t)__p0);
50968  return __ret;
50969}
50970#else
50971__ai int16_t vmaxv_s16(int16x4_t __p0) {
50972  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
50973  int16_t __ret;
50974  __ret = (int16_t) __builtin_neon_vmaxv_s16((int8x8_t)__rev0);
50975  return __ret;
50976}
50977#endif
50978
50979#ifdef __LITTLE_ENDIAN__
50980__ai float64x2_t vminq_f64(float64x2_t __p0, float64x2_t __p1) {
50981  float64x2_t __ret;
50982  __ret = (float64x2_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
50983  return __ret;
50984}
50985#else
50986__ai float64x2_t vminq_f64(float64x2_t __p0, float64x2_t __p1) {
50987  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
50988  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
50989  float64x2_t __ret;
50990  __ret = (float64x2_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
50991  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
50992  return __ret;
50993}
50994#endif
50995
50996#ifdef __LITTLE_ENDIAN__
50997__ai float64x1_t vmin_f64(float64x1_t __p0, float64x1_t __p1) {
50998  float64x1_t __ret;
50999  __ret = (float64x1_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
51000  return __ret;
51001}
51002#else
51003__ai float64x1_t vmin_f64(float64x1_t __p0, float64x1_t __p1) {
51004  float64x1_t __ret;
51005  __ret = (float64x1_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
51006  return __ret;
51007}
51008#endif
51009
51010#ifdef __LITTLE_ENDIAN__
51011__ai float64x2_t vminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
51012  float64x2_t __ret;
51013  __ret = (float64x2_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
51014  return __ret;
51015}
51016#else
51017__ai float64x2_t vminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
51018  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51019  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51020  float64x2_t __ret;
51021  __ret = (float64x2_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
51022  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51023  return __ret;
51024}
51025#endif
51026
51027#ifdef __LITTLE_ENDIAN__
51028__ai float32x4_t vminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
51029  float32x4_t __ret;
51030  __ret = (float32x4_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
51031  return __ret;
51032}
51033#else
51034__ai float32x4_t vminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
51035  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51036  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
51037  float32x4_t __ret;
51038  __ret = (float32x4_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
51039  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
51040  return __ret;
51041}
51042#endif
51043
51044#ifdef __LITTLE_ENDIAN__
51045__ai float64x1_t vminnm_f64(float64x1_t __p0, float64x1_t __p1) {
51046  float64x1_t __ret;
51047  __ret = (float64x1_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
51048  return __ret;
51049}
51050#else
51051__ai float64x1_t vminnm_f64(float64x1_t __p0, float64x1_t __p1) {
51052  float64x1_t __ret;
51053  __ret = (float64x1_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
51054  return __ret;
51055}
51056#endif
51057
51058#ifdef __LITTLE_ENDIAN__
51059__ai float32x2_t vminnm_f32(float32x2_t __p0, float32x2_t __p1) {
51060  float32x2_t __ret;
51061  __ret = (float32x2_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
51062  return __ret;
51063}
51064#else
51065__ai float32x2_t vminnm_f32(float32x2_t __p0, float32x2_t __p1) {
51066  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51067  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51068  float32x2_t __ret;
51069  __ret = (float32x2_t) __builtin_neon_vminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
51070  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51071  return __ret;
51072}
51073#endif
51074
51075#ifdef __LITTLE_ENDIAN__
51076__ai float64_t vminnmvq_f64(float64x2_t __p0) {
51077  float64_t __ret;
51078  __ret = (float64_t) __builtin_neon_vminnmvq_f64((int8x16_t)__p0);
51079  return __ret;
51080}
51081#else
51082__ai float64_t vminnmvq_f64(float64x2_t __p0) {
51083  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51084  float64_t __ret;
51085  __ret = (float64_t) __builtin_neon_vminnmvq_f64((int8x16_t)__rev0);
51086  return __ret;
51087}
51088#endif
51089
51090#ifdef __LITTLE_ENDIAN__
51091__ai float32_t vminnmvq_f32(float32x4_t __p0) {
51092  float32_t __ret;
51093  __ret = (float32_t) __builtin_neon_vminnmvq_f32((int8x16_t)__p0);
51094  return __ret;
51095}
51096#else
51097__ai float32_t vminnmvq_f32(float32x4_t __p0) {
51098  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51099  float32_t __ret;
51100  __ret = (float32_t) __builtin_neon_vminnmvq_f32((int8x16_t)__rev0);
51101  return __ret;
51102}
51103#endif
51104
51105#ifdef __LITTLE_ENDIAN__
51106__ai float32_t vminnmv_f32(float32x2_t __p0) {
51107  float32_t __ret;
51108  __ret = (float32_t) __builtin_neon_vminnmv_f32((int8x8_t)__p0);
51109  return __ret;
51110}
51111#else
51112__ai float32_t vminnmv_f32(float32x2_t __p0) {
51113  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51114  float32_t __ret;
51115  __ret = (float32_t) __builtin_neon_vminnmv_f32((int8x8_t)__rev0);
51116  return __ret;
51117}
51118#endif
51119
51120#ifdef __LITTLE_ENDIAN__
51121__ai uint8_t vminvq_u8(uint8x16_t __p0) {
51122  uint8_t __ret;
51123  __ret = (uint8_t) __builtin_neon_vminvq_u8((int8x16_t)__p0);
51124  return __ret;
51125}
51126#else
51127__ai uint8_t vminvq_u8(uint8x16_t __p0) {
51128  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51129  uint8_t __ret;
51130  __ret = (uint8_t) __builtin_neon_vminvq_u8((int8x16_t)__rev0);
51131  return __ret;
51132}
51133#endif
51134
51135#ifdef __LITTLE_ENDIAN__
51136__ai uint32_t vminvq_u32(uint32x4_t __p0) {
51137  uint32_t __ret;
51138  __ret = (uint32_t) __builtin_neon_vminvq_u32((int8x16_t)__p0);
51139  return __ret;
51140}
51141#else
51142__ai uint32_t vminvq_u32(uint32x4_t __p0) {
51143  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51144  uint32_t __ret;
51145  __ret = (uint32_t) __builtin_neon_vminvq_u32((int8x16_t)__rev0);
51146  return __ret;
51147}
51148#endif
51149
51150#ifdef __LITTLE_ENDIAN__
51151__ai uint16_t vminvq_u16(uint16x8_t __p0) {
51152  uint16_t __ret;
51153  __ret = (uint16_t) __builtin_neon_vminvq_u16((int8x16_t)__p0);
51154  return __ret;
51155}
51156#else
51157__ai uint16_t vminvq_u16(uint16x8_t __p0) {
51158  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51159  uint16_t __ret;
51160  __ret = (uint16_t) __builtin_neon_vminvq_u16((int8x16_t)__rev0);
51161  return __ret;
51162}
51163#endif
51164
51165#ifdef __LITTLE_ENDIAN__
51166__ai int8_t vminvq_s8(int8x16_t __p0) {
51167  int8_t __ret;
51168  __ret = (int8_t) __builtin_neon_vminvq_s8((int8x16_t)__p0);
51169  return __ret;
51170}
51171#else
51172__ai int8_t vminvq_s8(int8x16_t __p0) {
51173  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51174  int8_t __ret;
51175  __ret = (int8_t) __builtin_neon_vminvq_s8((int8x16_t)__rev0);
51176  return __ret;
51177}
51178#endif
51179
51180#ifdef __LITTLE_ENDIAN__
51181__ai float64_t vminvq_f64(float64x2_t __p0) {
51182  float64_t __ret;
51183  __ret = (float64_t) __builtin_neon_vminvq_f64((int8x16_t)__p0);
51184  return __ret;
51185}
51186#else
51187__ai float64_t vminvq_f64(float64x2_t __p0) {
51188  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51189  float64_t __ret;
51190  __ret = (float64_t) __builtin_neon_vminvq_f64((int8x16_t)__rev0);
51191  return __ret;
51192}
51193#endif
51194
51195#ifdef __LITTLE_ENDIAN__
51196__ai float32_t vminvq_f32(float32x4_t __p0) {
51197  float32_t __ret;
51198  __ret = (float32_t) __builtin_neon_vminvq_f32((int8x16_t)__p0);
51199  return __ret;
51200}
51201#else
51202__ai float32_t vminvq_f32(float32x4_t __p0) {
51203  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51204  float32_t __ret;
51205  __ret = (float32_t) __builtin_neon_vminvq_f32((int8x16_t)__rev0);
51206  return __ret;
51207}
51208#endif
51209
51210#ifdef __LITTLE_ENDIAN__
51211__ai int32_t vminvq_s32(int32x4_t __p0) {
51212  int32_t __ret;
51213  __ret = (int32_t) __builtin_neon_vminvq_s32((int8x16_t)__p0);
51214  return __ret;
51215}
51216#else
51217__ai int32_t vminvq_s32(int32x4_t __p0) {
51218  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51219  int32_t __ret;
51220  __ret = (int32_t) __builtin_neon_vminvq_s32((int8x16_t)__rev0);
51221  return __ret;
51222}
51223#endif
51224
51225#ifdef __LITTLE_ENDIAN__
51226__ai int16_t vminvq_s16(int16x8_t __p0) {
51227  int16_t __ret;
51228  __ret = (int16_t) __builtin_neon_vminvq_s16((int8x16_t)__p0);
51229  return __ret;
51230}
51231#else
51232__ai int16_t vminvq_s16(int16x8_t __p0) {
51233  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51234  int16_t __ret;
51235  __ret = (int16_t) __builtin_neon_vminvq_s16((int8x16_t)__rev0);
51236  return __ret;
51237}
51238#endif
51239
51240#ifdef __LITTLE_ENDIAN__
51241__ai uint8_t vminv_u8(uint8x8_t __p0) {
51242  uint8_t __ret;
51243  __ret = (uint8_t) __builtin_neon_vminv_u8((int8x8_t)__p0);
51244  return __ret;
51245}
51246#else
51247__ai uint8_t vminv_u8(uint8x8_t __p0) {
51248  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51249  uint8_t __ret;
51250  __ret = (uint8_t) __builtin_neon_vminv_u8((int8x8_t)__rev0);
51251  return __ret;
51252}
51253#endif
51254
51255#ifdef __LITTLE_ENDIAN__
51256__ai uint32_t vminv_u32(uint32x2_t __p0) {
51257  uint32_t __ret;
51258  __ret = (uint32_t) __builtin_neon_vminv_u32((int8x8_t)__p0);
51259  return __ret;
51260}
51261#else
51262__ai uint32_t vminv_u32(uint32x2_t __p0) {
51263  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51264  uint32_t __ret;
51265  __ret = (uint32_t) __builtin_neon_vminv_u32((int8x8_t)__rev0);
51266  return __ret;
51267}
51268#endif
51269
51270#ifdef __LITTLE_ENDIAN__
51271__ai uint16_t vminv_u16(uint16x4_t __p0) {
51272  uint16_t __ret;
51273  __ret = (uint16_t) __builtin_neon_vminv_u16((int8x8_t)__p0);
51274  return __ret;
51275}
51276#else
51277__ai uint16_t vminv_u16(uint16x4_t __p0) {
51278  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51279  uint16_t __ret;
51280  __ret = (uint16_t) __builtin_neon_vminv_u16((int8x8_t)__rev0);
51281  return __ret;
51282}
51283#endif
51284
51285#ifdef __LITTLE_ENDIAN__
51286__ai int8_t vminv_s8(int8x8_t __p0) {
51287  int8_t __ret;
51288  __ret = (int8_t) __builtin_neon_vminv_s8((int8x8_t)__p0);
51289  return __ret;
51290}
51291#else
51292__ai int8_t vminv_s8(int8x8_t __p0) {
51293  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51294  int8_t __ret;
51295  __ret = (int8_t) __builtin_neon_vminv_s8((int8x8_t)__rev0);
51296  return __ret;
51297}
51298#endif
51299
51300#ifdef __LITTLE_ENDIAN__
51301__ai float32_t vminv_f32(float32x2_t __p0) {
51302  float32_t __ret;
51303  __ret = (float32_t) __builtin_neon_vminv_f32((int8x8_t)__p0);
51304  return __ret;
51305}
51306#else
51307__ai float32_t vminv_f32(float32x2_t __p0) {
51308  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51309  float32_t __ret;
51310  __ret = (float32_t) __builtin_neon_vminv_f32((int8x8_t)__rev0);
51311  return __ret;
51312}
51313#endif
51314
51315#ifdef __LITTLE_ENDIAN__
51316__ai int32_t vminv_s32(int32x2_t __p0) {
51317  int32_t __ret;
51318  __ret = (int32_t) __builtin_neon_vminv_s32((int8x8_t)__p0);
51319  return __ret;
51320}
51321#else
51322__ai int32_t vminv_s32(int32x2_t __p0) {
51323  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51324  int32_t __ret;
51325  __ret = (int32_t) __builtin_neon_vminv_s32((int8x8_t)__rev0);
51326  return __ret;
51327}
51328#endif
51329
51330#ifdef __LITTLE_ENDIAN__
51331__ai int16_t vminv_s16(int16x4_t __p0) {
51332  int16_t __ret;
51333  __ret = (int16_t) __builtin_neon_vminv_s16((int8x8_t)__p0);
51334  return __ret;
51335}
51336#else
51337__ai int16_t vminv_s16(int16x4_t __p0) {
51338  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51339  int16_t __ret;
51340  __ret = (int16_t) __builtin_neon_vminv_s16((int8x8_t)__rev0);
51341  return __ret;
51342}
51343#endif
51344
51345#ifdef __LITTLE_ENDIAN__
51346__ai float64x2_t vmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
51347  float64x2_t __ret;
51348  __ret = __p0 + __p1 * __p2;
51349  return __ret;
51350}
51351#else
51352__ai float64x2_t vmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
51353  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51354  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51355  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
51356  float64x2_t __ret;
51357  __ret = __rev0 + __rev1 * __rev2;
51358  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51359  return __ret;
51360}
51361#endif
51362
51363#ifdef __LITTLE_ENDIAN__
51364__ai float64x1_t vmla_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
51365  float64x1_t __ret;
51366  __ret = __p0 + __p1 * __p2;
51367  return __ret;
51368}
51369#else
51370__ai float64x1_t vmla_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
51371  float64x1_t __ret;
51372  __ret = __p0 + __p1 * __p2;
51373  return __ret;
51374}
51375#endif
51376
51377#ifdef __LITTLE_ENDIAN__
51378#define vmlaq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
51379  uint32x4_t __s0 = __p0; \
51380  uint32x4_t __s1 = __p1; \
51381  uint32x4_t __s2 = __p2; \
51382  uint32x4_t __ret; \
51383  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
51384  __ret; \
51385})
51386#else
51387#define vmlaq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
51388  uint32x4_t __s0 = __p0; \
51389  uint32x4_t __s1 = __p1; \
51390  uint32x4_t __s2 = __p2; \
51391  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
51392  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
51393  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
51394  uint32x4_t __ret; \
51395  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
51396  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
51397  __ret; \
51398})
51399#endif
51400
51401#ifdef __LITTLE_ENDIAN__
51402#define vmlaq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
51403  uint16x8_t __s0 = __p0; \
51404  uint16x8_t __s1 = __p1; \
51405  uint16x8_t __s2 = __p2; \
51406  uint16x8_t __ret; \
51407  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
51408  __ret; \
51409})
51410#else
51411#define vmlaq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
51412  uint16x8_t __s0 = __p0; \
51413  uint16x8_t __s1 = __p1; \
51414  uint16x8_t __s2 = __p2; \
51415  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
51416  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
51417  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
51418  uint16x8_t __ret; \
51419  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
51420  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
51421  __ret; \
51422})
51423#endif
51424
51425#ifdef __LITTLE_ENDIAN__
51426#define vmlaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
51427  float32x4_t __s0 = __p0; \
51428  float32x4_t __s1 = __p1; \
51429  float32x4_t __s2 = __p2; \
51430  float32x4_t __ret; \
51431  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
51432  __ret; \
51433})
51434#else
51435#define vmlaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
51436  float32x4_t __s0 = __p0; \
51437  float32x4_t __s1 = __p1; \
51438  float32x4_t __s2 = __p2; \
51439  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
51440  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
51441  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
51442  float32x4_t __ret; \
51443  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
51444  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
51445  __ret; \
51446})
51447#endif
51448
51449#ifdef __LITTLE_ENDIAN__
51450#define vmlaq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
51451  int32x4_t __s0 = __p0; \
51452  int32x4_t __s1 = __p1; \
51453  int32x4_t __s2 = __p2; \
51454  int32x4_t __ret; \
51455  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
51456  __ret; \
51457})
51458#else
51459#define vmlaq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
51460  int32x4_t __s0 = __p0; \
51461  int32x4_t __s1 = __p1; \
51462  int32x4_t __s2 = __p2; \
51463  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
51464  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
51465  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
51466  int32x4_t __ret; \
51467  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
51468  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
51469  __ret; \
51470})
51471#endif
51472
51473#ifdef __LITTLE_ENDIAN__
51474#define vmlaq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
51475  int16x8_t __s0 = __p0; \
51476  int16x8_t __s1 = __p1; \
51477  int16x8_t __s2 = __p2; \
51478  int16x8_t __ret; \
51479  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
51480  __ret; \
51481})
51482#else
51483#define vmlaq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
51484  int16x8_t __s0 = __p0; \
51485  int16x8_t __s1 = __p1; \
51486  int16x8_t __s2 = __p2; \
51487  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
51488  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
51489  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
51490  int16x8_t __ret; \
51491  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
51492  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
51493  __ret; \
51494})
51495#endif
51496
51497#ifdef __LITTLE_ENDIAN__
51498#define vmla_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
51499  uint32x2_t __s0 = __p0; \
51500  uint32x2_t __s1 = __p1; \
51501  uint32x4_t __s2 = __p2; \
51502  uint32x2_t __ret; \
51503  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
51504  __ret; \
51505})
51506#else
51507#define vmla_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
51508  uint32x2_t __s0 = __p0; \
51509  uint32x2_t __s1 = __p1; \
51510  uint32x4_t __s2 = __p2; \
51511  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
51512  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
51513  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
51514  uint32x2_t __ret; \
51515  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
51516  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
51517  __ret; \
51518})
51519#endif
51520
51521#ifdef __LITTLE_ENDIAN__
51522#define vmla_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
51523  uint16x4_t __s0 = __p0; \
51524  uint16x4_t __s1 = __p1; \
51525  uint16x8_t __s2 = __p2; \
51526  uint16x4_t __ret; \
51527  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
51528  __ret; \
51529})
51530#else
51531#define vmla_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
51532  uint16x4_t __s0 = __p0; \
51533  uint16x4_t __s1 = __p1; \
51534  uint16x8_t __s2 = __p2; \
51535  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
51536  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
51537  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
51538  uint16x4_t __ret; \
51539  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
51540  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
51541  __ret; \
51542})
51543#endif
51544
51545#ifdef __LITTLE_ENDIAN__
51546#define vmla_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
51547  float32x2_t __s0 = __p0; \
51548  float32x2_t __s1 = __p1; \
51549  float32x4_t __s2 = __p2; \
51550  float32x2_t __ret; \
51551  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
51552  __ret; \
51553})
51554#else
51555#define vmla_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
51556  float32x2_t __s0 = __p0; \
51557  float32x2_t __s1 = __p1; \
51558  float32x4_t __s2 = __p2; \
51559  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
51560  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
51561  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
51562  float32x2_t __ret; \
51563  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
51564  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
51565  __ret; \
51566})
51567#endif
51568
51569#ifdef __LITTLE_ENDIAN__
51570#define vmla_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
51571  int32x2_t __s0 = __p0; \
51572  int32x2_t __s1 = __p1; \
51573  int32x4_t __s2 = __p2; \
51574  int32x2_t __ret; \
51575  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
51576  __ret; \
51577})
51578#else
51579#define vmla_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
51580  int32x2_t __s0 = __p0; \
51581  int32x2_t __s1 = __p1; \
51582  int32x4_t __s2 = __p2; \
51583  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
51584  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
51585  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
51586  int32x2_t __ret; \
51587  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
51588  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
51589  __ret; \
51590})
51591#endif
51592
51593#ifdef __LITTLE_ENDIAN__
51594#define vmla_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
51595  int16x4_t __s0 = __p0; \
51596  int16x4_t __s1 = __p1; \
51597  int16x8_t __s2 = __p2; \
51598  int16x4_t __ret; \
51599  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
51600  __ret; \
51601})
51602#else
51603#define vmla_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
51604  int16x4_t __s0 = __p0; \
51605  int16x4_t __s1 = __p1; \
51606  int16x8_t __s2 = __p2; \
51607  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
51608  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
51609  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
51610  int16x4_t __ret; \
51611  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
51612  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
51613  __ret; \
51614})
51615#endif
51616
51617#ifdef __LITTLE_ENDIAN__
51618__ai float64x2_t vmlaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
51619  float64x2_t __ret;
51620  __ret = __p0 + __p1 * (float64x2_t) {__p2, __p2};
51621  return __ret;
51622}
51623#else
51624__ai float64x2_t vmlaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
51625  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51626  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51627  float64x2_t __ret;
51628  __ret = __rev0 + __rev1 * (float64x2_t) {__p2, __p2};
51629  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51630  return __ret;
51631}
51632#endif
51633
51634#ifdef __LITTLE_ENDIAN__
51635#define vmlal_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
51636  uint64x2_t __s0 = __p0; \
51637  uint32x4_t __s1 = __p1; \
51638  uint32x2_t __s2 = __p2; \
51639  uint64x2_t __ret; \
51640  __ret = __s0 + vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
51641  __ret; \
51642})
51643#else
51644#define vmlal_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
51645  uint64x2_t __s0 = __p0; \
51646  uint32x4_t __s1 = __p1; \
51647  uint32x2_t __s2 = __p2; \
51648  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
51649  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
51650  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
51651  uint64x2_t __ret; \
51652  __ret = __rev0 + __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
51653  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
51654  __ret; \
51655})
51656#endif
51657
51658#ifdef __LITTLE_ENDIAN__
51659#define vmlal_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
51660  uint32x4_t __s0 = __p0; \
51661  uint16x8_t __s1 = __p1; \
51662  uint16x4_t __s2 = __p2; \
51663  uint32x4_t __ret; \
51664  __ret = __s0 + vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
51665  __ret; \
51666})
51667#else
51668#define vmlal_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
51669  uint32x4_t __s0 = __p0; \
51670  uint16x8_t __s1 = __p1; \
51671  uint16x4_t __s2 = __p2; \
51672  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
51673  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
51674  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
51675  uint32x4_t __ret; \
51676  __ret = __rev0 + __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
51677  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
51678  __ret; \
51679})
51680#endif
51681
51682#ifdef __LITTLE_ENDIAN__
51683#define vmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
51684  int64x2_t __s0 = __p0; \
51685  int32x4_t __s1 = __p1; \
51686  int32x2_t __s2 = __p2; \
51687  int64x2_t __ret; \
51688  __ret = __s0 + vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
51689  __ret; \
51690})
51691#else
51692#define vmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
51693  int64x2_t __s0 = __p0; \
51694  int32x4_t __s1 = __p1; \
51695  int32x2_t __s2 = __p2; \
51696  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
51697  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
51698  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
51699  int64x2_t __ret; \
51700  __ret = __rev0 + __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
51701  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
51702  __ret; \
51703})
51704#endif
51705
51706#ifdef __LITTLE_ENDIAN__
51707#define vmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
51708  int32x4_t __s0 = __p0; \
51709  int16x8_t __s1 = __p1; \
51710  int16x4_t __s2 = __p2; \
51711  int32x4_t __ret; \
51712  __ret = __s0 + vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
51713  __ret; \
51714})
51715#else
51716#define vmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
51717  int32x4_t __s0 = __p0; \
51718  int16x8_t __s1 = __p1; \
51719  int16x4_t __s2 = __p2; \
51720  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
51721  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
51722  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
51723  int32x4_t __ret; \
51724  __ret = __rev0 + __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
51725  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
51726  __ret; \
51727})
51728#endif
51729
51730#ifdef __LITTLE_ENDIAN__
51731#define vmlal_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
51732  uint64x2_t __s0 = __p0; \
51733  uint32x4_t __s1 = __p1; \
51734  uint32x4_t __s2 = __p2; \
51735  uint64x2_t __ret; \
51736  __ret = __s0 + vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
51737  __ret; \
51738})
51739#else
51740#define vmlal_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
51741  uint64x2_t __s0 = __p0; \
51742  uint32x4_t __s1 = __p1; \
51743  uint32x4_t __s2 = __p2; \
51744  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
51745  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
51746  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
51747  uint64x2_t __ret; \
51748  __ret = __rev0 + __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
51749  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
51750  __ret; \
51751})
51752#endif
51753
51754#ifdef __LITTLE_ENDIAN__
51755#define vmlal_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
51756  uint32x4_t __s0 = __p0; \
51757  uint16x8_t __s1 = __p1; \
51758  uint16x8_t __s2 = __p2; \
51759  uint32x4_t __ret; \
51760  __ret = __s0 + vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
51761  __ret; \
51762})
51763#else
51764#define vmlal_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
51765  uint32x4_t __s0 = __p0; \
51766  uint16x8_t __s1 = __p1; \
51767  uint16x8_t __s2 = __p2; \
51768  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
51769  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
51770  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
51771  uint32x4_t __ret; \
51772  __ret = __rev0 + __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
51773  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
51774  __ret; \
51775})
51776#endif
51777
51778#ifdef __LITTLE_ENDIAN__
51779#define vmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
51780  int64x2_t __s0 = __p0; \
51781  int32x4_t __s1 = __p1; \
51782  int32x4_t __s2 = __p2; \
51783  int64x2_t __ret; \
51784  __ret = __s0 + vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
51785  __ret; \
51786})
51787#else
51788#define vmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
51789  int64x2_t __s0 = __p0; \
51790  int32x4_t __s1 = __p1; \
51791  int32x4_t __s2 = __p2; \
51792  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
51793  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
51794  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
51795  int64x2_t __ret; \
51796  __ret = __rev0 + __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
51797  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
51798  __ret; \
51799})
51800#endif
51801
51802#ifdef __LITTLE_ENDIAN__
51803#define vmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
51804  int32x4_t __s0 = __p0; \
51805  int16x8_t __s1 = __p1; \
51806  int16x8_t __s2 = __p2; \
51807  int32x4_t __ret; \
51808  __ret = __s0 + vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
51809  __ret; \
51810})
51811#else
51812#define vmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
51813  int32x4_t __s0 = __p0; \
51814  int16x8_t __s1 = __p1; \
51815  int16x8_t __s2 = __p2; \
51816  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
51817  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
51818  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
51819  int32x4_t __ret; \
51820  __ret = __rev0 + __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
51821  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
51822  __ret; \
51823})
51824#endif
51825
51826#ifdef __LITTLE_ENDIAN__
51827#define vmlal_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
51828  uint64x2_t __s0 = __p0; \
51829  uint32x2_t __s1 = __p1; \
51830  uint32x4_t __s2 = __p2; \
51831  uint64x2_t __ret; \
51832  __ret = __s0 + vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
51833  __ret; \
51834})
51835#else
51836#define vmlal_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
51837  uint64x2_t __s0 = __p0; \
51838  uint32x2_t __s1 = __p1; \
51839  uint32x4_t __s2 = __p2; \
51840  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
51841  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
51842  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
51843  uint64x2_t __ret; \
51844  __ret = __rev0 + __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
51845  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
51846  __ret; \
51847})
51848#endif
51849
51850#ifdef __LITTLE_ENDIAN__
51851#define vmlal_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
51852  uint32x4_t __s0 = __p0; \
51853  uint16x4_t __s1 = __p1; \
51854  uint16x8_t __s2 = __p2; \
51855  uint32x4_t __ret; \
51856  __ret = __s0 + vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
51857  __ret; \
51858})
51859#else
51860#define vmlal_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
51861  uint32x4_t __s0 = __p0; \
51862  uint16x4_t __s1 = __p1; \
51863  uint16x8_t __s2 = __p2; \
51864  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
51865  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
51866  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
51867  uint32x4_t __ret; \
51868  __ret = __rev0 + __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
51869  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
51870  __ret; \
51871})
51872#endif
51873
51874#ifdef __LITTLE_ENDIAN__
51875#define vmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
51876  int64x2_t __s0 = __p0; \
51877  int32x2_t __s1 = __p1; \
51878  int32x4_t __s2 = __p2; \
51879  int64x2_t __ret; \
51880  __ret = __s0 + vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
51881  __ret; \
51882})
51883#else
51884#define vmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
51885  int64x2_t __s0 = __p0; \
51886  int32x2_t __s1 = __p1; \
51887  int32x4_t __s2 = __p2; \
51888  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
51889  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
51890  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
51891  int64x2_t __ret; \
51892  __ret = __rev0 + __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
51893  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
51894  __ret; \
51895})
51896#endif
51897
51898#ifdef __LITTLE_ENDIAN__
51899#define vmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
51900  int32x4_t __s0 = __p0; \
51901  int16x4_t __s1 = __p1; \
51902  int16x8_t __s2 = __p2; \
51903  int32x4_t __ret; \
51904  __ret = __s0 + vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
51905  __ret; \
51906})
51907#else
51908#define vmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
51909  int32x4_t __s0 = __p0; \
51910  int16x4_t __s1 = __p1; \
51911  int16x8_t __s2 = __p2; \
51912  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
51913  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
51914  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
51915  int32x4_t __ret; \
51916  __ret = __rev0 + __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
51917  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
51918  __ret; \
51919})
51920#endif
51921
51922#ifdef __LITTLE_ENDIAN__
51923__ai float64x2_t vmlsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
51924  float64x2_t __ret;
51925  __ret = __p0 - __p1 * __p2;
51926  return __ret;
51927}
51928#else
51929__ai float64x2_t vmlsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
51930  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51931  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51932  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
51933  float64x2_t __ret;
51934  __ret = __rev0 - __rev1 * __rev2;
51935  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51936  return __ret;
51937}
51938#endif
51939
51940#ifdef __LITTLE_ENDIAN__
51941__ai float64x1_t vmls_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
51942  float64x1_t __ret;
51943  __ret = __p0 - __p1 * __p2;
51944  return __ret;
51945}
51946#else
51947__ai float64x1_t vmls_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
51948  float64x1_t __ret;
51949  __ret = __p0 - __p1 * __p2;
51950  return __ret;
51951}
51952#endif
51953
51954#ifdef __LITTLE_ENDIAN__
51955#define vmlsq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
51956  uint32x4_t __s0 = __p0; \
51957  uint32x4_t __s1 = __p1; \
51958  uint32x4_t __s2 = __p2; \
51959  uint32x4_t __ret; \
51960  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
51961  __ret; \
51962})
51963#else
51964#define vmlsq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
51965  uint32x4_t __s0 = __p0; \
51966  uint32x4_t __s1 = __p1; \
51967  uint32x4_t __s2 = __p2; \
51968  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
51969  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
51970  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
51971  uint32x4_t __ret; \
51972  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
51973  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
51974  __ret; \
51975})
51976#endif
51977
51978#ifdef __LITTLE_ENDIAN__
51979#define vmlsq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
51980  uint16x8_t __s0 = __p0; \
51981  uint16x8_t __s1 = __p1; \
51982  uint16x8_t __s2 = __p2; \
51983  uint16x8_t __ret; \
51984  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
51985  __ret; \
51986})
51987#else
51988#define vmlsq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
51989  uint16x8_t __s0 = __p0; \
51990  uint16x8_t __s1 = __p1; \
51991  uint16x8_t __s2 = __p2; \
51992  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
51993  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
51994  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
51995  uint16x8_t __ret; \
51996  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
51997  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
51998  __ret; \
51999})
52000#endif
52001
52002#ifdef __LITTLE_ENDIAN__
52003#define vmlsq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52004  float32x4_t __s0 = __p0; \
52005  float32x4_t __s1 = __p1; \
52006  float32x4_t __s2 = __p2; \
52007  float32x4_t __ret; \
52008  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52009  __ret; \
52010})
52011#else
52012#define vmlsq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52013  float32x4_t __s0 = __p0; \
52014  float32x4_t __s1 = __p1; \
52015  float32x4_t __s2 = __p2; \
52016  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52017  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52018  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52019  float32x4_t __ret; \
52020  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52021  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52022  __ret; \
52023})
52024#endif
52025
52026#ifdef __LITTLE_ENDIAN__
52027#define vmlsq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52028  int32x4_t __s0 = __p0; \
52029  int32x4_t __s1 = __p1; \
52030  int32x4_t __s2 = __p2; \
52031  int32x4_t __ret; \
52032  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52033  __ret; \
52034})
52035#else
52036#define vmlsq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52037  int32x4_t __s0 = __p0; \
52038  int32x4_t __s1 = __p1; \
52039  int32x4_t __s2 = __p2; \
52040  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52041  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52042  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52043  int32x4_t __ret; \
52044  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52045  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52046  __ret; \
52047})
52048#endif
52049
52050#ifdef __LITTLE_ENDIAN__
52051#define vmlsq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52052  int16x8_t __s0 = __p0; \
52053  int16x8_t __s1 = __p1; \
52054  int16x8_t __s2 = __p2; \
52055  int16x8_t __ret; \
52056  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52057  __ret; \
52058})
52059#else
52060#define vmlsq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52061  int16x8_t __s0 = __p0; \
52062  int16x8_t __s1 = __p1; \
52063  int16x8_t __s2 = __p2; \
52064  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
52065  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52066  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52067  int16x8_t __ret; \
52068  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52069  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
52070  __ret; \
52071})
52072#endif
52073
52074#ifdef __LITTLE_ENDIAN__
52075#define vmls_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52076  uint32x2_t __s0 = __p0; \
52077  uint32x2_t __s1 = __p1; \
52078  uint32x4_t __s2 = __p2; \
52079  uint32x2_t __ret; \
52080  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
52081  __ret; \
52082})
52083#else
52084#define vmls_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52085  uint32x2_t __s0 = __p0; \
52086  uint32x2_t __s1 = __p1; \
52087  uint32x4_t __s2 = __p2; \
52088  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52089  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52090  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52091  uint32x2_t __ret; \
52092  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
52093  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52094  __ret; \
52095})
52096#endif
52097
52098#ifdef __LITTLE_ENDIAN__
52099#define vmls_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52100  uint16x4_t __s0 = __p0; \
52101  uint16x4_t __s1 = __p1; \
52102  uint16x8_t __s2 = __p2; \
52103  uint16x4_t __ret; \
52104  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52105  __ret; \
52106})
52107#else
52108#define vmls_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52109  uint16x4_t __s0 = __p0; \
52110  uint16x4_t __s1 = __p1; \
52111  uint16x8_t __s2 = __p2; \
52112  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52113  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52114  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52115  uint16x4_t __ret; \
52116  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52117  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52118  __ret; \
52119})
52120#endif
52121
52122#ifdef __LITTLE_ENDIAN__
52123#define vmls_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52124  float32x2_t __s0 = __p0; \
52125  float32x2_t __s1 = __p1; \
52126  float32x4_t __s2 = __p2; \
52127  float32x2_t __ret; \
52128  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
52129  __ret; \
52130})
52131#else
52132#define vmls_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52133  float32x2_t __s0 = __p0; \
52134  float32x2_t __s1 = __p1; \
52135  float32x4_t __s2 = __p2; \
52136  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52137  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52138  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52139  float32x2_t __ret; \
52140  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
52141  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52142  __ret; \
52143})
52144#endif
52145
52146#ifdef __LITTLE_ENDIAN__
52147#define vmls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52148  int32x2_t __s0 = __p0; \
52149  int32x2_t __s1 = __p1; \
52150  int32x4_t __s2 = __p2; \
52151  int32x2_t __ret; \
52152  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
52153  __ret; \
52154})
52155#else
52156#define vmls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52157  int32x2_t __s0 = __p0; \
52158  int32x2_t __s1 = __p1; \
52159  int32x4_t __s2 = __p2; \
52160  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52161  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52162  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52163  int32x2_t __ret; \
52164  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
52165  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52166  __ret; \
52167})
52168#endif
52169
52170#ifdef __LITTLE_ENDIAN__
52171#define vmls_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52172  int16x4_t __s0 = __p0; \
52173  int16x4_t __s1 = __p1; \
52174  int16x8_t __s2 = __p2; \
52175  int16x4_t __ret; \
52176  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52177  __ret; \
52178})
52179#else
52180#define vmls_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52181  int16x4_t __s0 = __p0; \
52182  int16x4_t __s1 = __p1; \
52183  int16x8_t __s2 = __p2; \
52184  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52185  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52186  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52187  int16x4_t __ret; \
52188  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52189  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52190  __ret; \
52191})
52192#endif
52193
52194#ifdef __LITTLE_ENDIAN__
52195__ai float64x2_t vmlsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
52196  float64x2_t __ret;
52197  __ret = __p0 - __p1 * (float64x2_t) {__p2, __p2};
52198  return __ret;
52199}
52200#else
52201__ai float64x2_t vmlsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
52202  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52203  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52204  float64x2_t __ret;
52205  __ret = __rev0 - __rev1 * (float64x2_t) {__p2, __p2};
52206  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52207  return __ret;
52208}
52209#endif
52210
52211#ifdef __LITTLE_ENDIAN__
52212#define vmlsl_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52213  uint64x2_t __s0 = __p0; \
52214  uint32x4_t __s1 = __p1; \
52215  uint32x2_t __s2 = __p2; \
52216  uint64x2_t __ret; \
52217  __ret = __s0 - vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52218  __ret; \
52219})
52220#else
52221#define vmlsl_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52222  uint64x2_t __s0 = __p0; \
52223  uint32x4_t __s1 = __p1; \
52224  uint32x2_t __s2 = __p2; \
52225  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52226  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52227  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
52228  uint64x2_t __ret; \
52229  __ret = __rev0 - __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52230  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52231  __ret; \
52232})
52233#endif
52234
52235#ifdef __LITTLE_ENDIAN__
52236#define vmlsl_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52237  uint32x4_t __s0 = __p0; \
52238  uint16x8_t __s1 = __p1; \
52239  uint16x4_t __s2 = __p2; \
52240  uint32x4_t __ret; \
52241  __ret = __s0 - vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52242  __ret; \
52243})
52244#else
52245#define vmlsl_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52246  uint32x4_t __s0 = __p0; \
52247  uint16x8_t __s1 = __p1; \
52248  uint16x4_t __s2 = __p2; \
52249  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52250  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52251  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52252  uint32x4_t __ret; \
52253  __ret = __rev0 - __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52254  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52255  __ret; \
52256})
52257#endif
52258
52259#ifdef __LITTLE_ENDIAN__
52260#define vmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52261  int64x2_t __s0 = __p0; \
52262  int32x4_t __s1 = __p1; \
52263  int32x2_t __s2 = __p2; \
52264  int64x2_t __ret; \
52265  __ret = __s0 - vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52266  __ret; \
52267})
52268#else
52269#define vmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52270  int64x2_t __s0 = __p0; \
52271  int32x4_t __s1 = __p1; \
52272  int32x2_t __s2 = __p2; \
52273  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52274  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52275  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
52276  int64x2_t __ret; \
52277  __ret = __rev0 - __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52278  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52279  __ret; \
52280})
52281#endif
52282
52283#ifdef __LITTLE_ENDIAN__
52284#define vmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52285  int32x4_t __s0 = __p0; \
52286  int16x8_t __s1 = __p1; \
52287  int16x4_t __s2 = __p2; \
52288  int32x4_t __ret; \
52289  __ret = __s0 - vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52290  __ret; \
52291})
52292#else
52293#define vmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52294  int32x4_t __s0 = __p0; \
52295  int16x8_t __s1 = __p1; \
52296  int16x4_t __s2 = __p2; \
52297  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52298  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52299  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52300  int32x4_t __ret; \
52301  __ret = __rev0 - __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52302  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52303  __ret; \
52304})
52305#endif
52306
52307#ifdef __LITTLE_ENDIAN__
52308#define vmlsl_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52309  uint64x2_t __s0 = __p0; \
52310  uint32x4_t __s1 = __p1; \
52311  uint32x4_t __s2 = __p2; \
52312  uint64x2_t __ret; \
52313  __ret = __s0 - vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52314  __ret; \
52315})
52316#else
52317#define vmlsl_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52318  uint64x2_t __s0 = __p0; \
52319  uint32x4_t __s1 = __p1; \
52320  uint32x4_t __s2 = __p2; \
52321  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52322  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52323  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52324  uint64x2_t __ret; \
52325  __ret = __rev0 - __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52326  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52327  __ret; \
52328})
52329#endif
52330
52331#ifdef __LITTLE_ENDIAN__
52332#define vmlsl_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52333  uint32x4_t __s0 = __p0; \
52334  uint16x8_t __s1 = __p1; \
52335  uint16x8_t __s2 = __p2; \
52336  uint32x4_t __ret; \
52337  __ret = __s0 - vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52338  __ret; \
52339})
52340#else
52341#define vmlsl_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52342  uint32x4_t __s0 = __p0; \
52343  uint16x8_t __s1 = __p1; \
52344  uint16x8_t __s2 = __p2; \
52345  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52346  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52347  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52348  uint32x4_t __ret; \
52349  __ret = __rev0 - __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52350  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52351  __ret; \
52352})
52353#endif
52354
52355#ifdef __LITTLE_ENDIAN__
52356#define vmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52357  int64x2_t __s0 = __p0; \
52358  int32x4_t __s1 = __p1; \
52359  int32x4_t __s2 = __p2; \
52360  int64x2_t __ret; \
52361  __ret = __s0 - vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52362  __ret; \
52363})
52364#else
52365#define vmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52366  int64x2_t __s0 = __p0; \
52367  int32x4_t __s1 = __p1; \
52368  int32x4_t __s2 = __p2; \
52369  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52370  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52371  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52372  int64x2_t __ret; \
52373  __ret = __rev0 - __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52374  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52375  __ret; \
52376})
52377#endif
52378
52379#ifdef __LITTLE_ENDIAN__
52380#define vmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52381  int32x4_t __s0 = __p0; \
52382  int16x8_t __s1 = __p1; \
52383  int16x8_t __s2 = __p2; \
52384  int32x4_t __ret; \
52385  __ret = __s0 - vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52386  __ret; \
52387})
52388#else
52389#define vmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52390  int32x4_t __s0 = __p0; \
52391  int16x8_t __s1 = __p1; \
52392  int16x8_t __s2 = __p2; \
52393  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52394  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52395  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52396  int32x4_t __ret; \
52397  __ret = __rev0 - __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52398  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52399  __ret; \
52400})
52401#endif
52402
52403#ifdef __LITTLE_ENDIAN__
52404#define vmlsl_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52405  uint64x2_t __s0 = __p0; \
52406  uint32x2_t __s1 = __p1; \
52407  uint32x4_t __s2 = __p2; \
52408  uint64x2_t __ret; \
52409  __ret = __s0 - vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52410  __ret; \
52411})
52412#else
52413#define vmlsl_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52414  uint64x2_t __s0 = __p0; \
52415  uint32x2_t __s1 = __p1; \
52416  uint32x4_t __s2 = __p2; \
52417  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52418  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52419  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52420  uint64x2_t __ret; \
52421  __ret = __rev0 - __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52422  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52423  __ret; \
52424})
52425#endif
52426
52427#ifdef __LITTLE_ENDIAN__
52428#define vmlsl_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52429  uint32x4_t __s0 = __p0; \
52430  uint16x4_t __s1 = __p1; \
52431  uint16x8_t __s2 = __p2; \
52432  uint32x4_t __ret; \
52433  __ret = __s0 - vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52434  __ret; \
52435})
52436#else
52437#define vmlsl_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52438  uint32x4_t __s0 = __p0; \
52439  uint16x4_t __s1 = __p1; \
52440  uint16x8_t __s2 = __p2; \
52441  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52442  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52443  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52444  uint32x4_t __ret; \
52445  __ret = __rev0 - __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52446  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52447  __ret; \
52448})
52449#endif
52450
52451#ifdef __LITTLE_ENDIAN__
52452#define vmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52453  int64x2_t __s0 = __p0; \
52454  int32x2_t __s1 = __p1; \
52455  int32x4_t __s2 = __p2; \
52456  int64x2_t __ret; \
52457  __ret = __s0 - vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52458  __ret; \
52459})
52460#else
52461#define vmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52462  int64x2_t __s0 = __p0; \
52463  int32x2_t __s1 = __p1; \
52464  int32x4_t __s2 = __p2; \
52465  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52466  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52467  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52468  int64x2_t __ret; \
52469  __ret = __rev0 - __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52470  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52471  __ret; \
52472})
52473#endif
52474
52475#ifdef __LITTLE_ENDIAN__
52476#define vmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52477  int32x4_t __s0 = __p0; \
52478  int16x4_t __s1 = __p1; \
52479  int16x8_t __s2 = __p2; \
52480  int32x4_t __ret; \
52481  __ret = __s0 - vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52482  __ret; \
52483})
52484#else
52485#define vmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52486  int32x4_t __s0 = __p0; \
52487  int16x4_t __s1 = __p1; \
52488  int16x8_t __s2 = __p2; \
52489  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52490  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52491  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52492  int32x4_t __ret; \
52493  __ret = __rev0 - __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52494  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52495  __ret; \
52496})
52497#endif
52498
52499#ifdef __LITTLE_ENDIAN__
52500__ai float64x2_t vmovq_n_f64(float64_t __p0) {
52501  float64x2_t __ret;
52502  __ret = (float64x2_t) {__p0, __p0};
52503  return __ret;
52504}
52505#else
52506__ai float64x2_t vmovq_n_f64(float64_t __p0) {
52507  float64x2_t __ret;
52508  __ret = (float64x2_t) {__p0, __p0};
52509  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52510  return __ret;
52511}
52512#endif
52513
52514#ifdef __LITTLE_ENDIAN__
52515__ai float64x1_t vmov_n_f64(float64_t __p0) {
52516  float64x1_t __ret;
52517  __ret = (float64x1_t) {__p0};
52518  return __ret;
52519}
52520#else
52521__ai float64x1_t vmov_n_f64(float64_t __p0) {
52522  float64x1_t __ret;
52523  __ret = (float64x1_t) {__p0};
52524  return __ret;
52525}
52526#endif
52527
52528#ifdef __LITTLE_ENDIAN__
52529__ai uint16x8_t vmovl_high_u8(uint8x16_t __p0_116) {
52530  uint16x8_t __ret_116;
52531  uint8x8_t __a1_116 = vget_high_u8(__p0_116);
52532  __ret_116 = (uint16x8_t)(vshll_n_u8(__a1_116, 0));
52533  return __ret_116;
52534}
52535#else
52536__ai uint16x8_t vmovl_high_u8(uint8x16_t __p0_117) {
52537  uint8x16_t __rev0_117;  __rev0_117 = __builtin_shufflevector(__p0_117, __p0_117, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52538  uint16x8_t __ret_117;
52539  uint8x8_t __a1_117 = __noswap_vget_high_u8(__rev0_117);
52540  __ret_117 = (uint16x8_t)(__noswap_vshll_n_u8(__a1_117, 0));
52541  __ret_117 = __builtin_shufflevector(__ret_117, __ret_117, 7, 6, 5, 4, 3, 2, 1, 0);
52542  return __ret_117;
52543}
52544__ai uint16x8_t __noswap_vmovl_high_u8(uint8x16_t __p0_118) {
52545  uint16x8_t __ret_118;
52546  uint8x8_t __a1_118 = __noswap_vget_high_u8(__p0_118);
52547  __ret_118 = (uint16x8_t)(__noswap_vshll_n_u8(__a1_118, 0));
52548  return __ret_118;
52549}
52550#endif
52551
52552#ifdef __LITTLE_ENDIAN__
52553__ai uint64x2_t vmovl_high_u32(uint32x4_t __p0_119) {
52554  uint64x2_t __ret_119;
52555  uint32x2_t __a1_119 = vget_high_u32(__p0_119);
52556  __ret_119 = (uint64x2_t)(vshll_n_u32(__a1_119, 0));
52557  return __ret_119;
52558}
52559#else
52560__ai uint64x2_t vmovl_high_u32(uint32x4_t __p0_120) {
52561  uint32x4_t __rev0_120;  __rev0_120 = __builtin_shufflevector(__p0_120, __p0_120, 3, 2, 1, 0);
52562  uint64x2_t __ret_120;
52563  uint32x2_t __a1_120 = __noswap_vget_high_u32(__rev0_120);
52564  __ret_120 = (uint64x2_t)(__noswap_vshll_n_u32(__a1_120, 0));
52565  __ret_120 = __builtin_shufflevector(__ret_120, __ret_120, 1, 0);
52566  return __ret_120;
52567}
52568__ai uint64x2_t __noswap_vmovl_high_u32(uint32x4_t __p0_121) {
52569  uint64x2_t __ret_121;
52570  uint32x2_t __a1_121 = __noswap_vget_high_u32(__p0_121);
52571  __ret_121 = (uint64x2_t)(__noswap_vshll_n_u32(__a1_121, 0));
52572  return __ret_121;
52573}
52574#endif
52575
52576#ifdef __LITTLE_ENDIAN__
52577__ai uint32x4_t vmovl_high_u16(uint16x8_t __p0_122) {
52578  uint32x4_t __ret_122;
52579  uint16x4_t __a1_122 = vget_high_u16(__p0_122);
52580  __ret_122 = (uint32x4_t)(vshll_n_u16(__a1_122, 0));
52581  return __ret_122;
52582}
52583#else
52584__ai uint32x4_t vmovl_high_u16(uint16x8_t __p0_123) {
52585  uint16x8_t __rev0_123;  __rev0_123 = __builtin_shufflevector(__p0_123, __p0_123, 7, 6, 5, 4, 3, 2, 1, 0);
52586  uint32x4_t __ret_123;
52587  uint16x4_t __a1_123 = __noswap_vget_high_u16(__rev0_123);
52588  __ret_123 = (uint32x4_t)(__noswap_vshll_n_u16(__a1_123, 0));
52589  __ret_123 = __builtin_shufflevector(__ret_123, __ret_123, 3, 2, 1, 0);
52590  return __ret_123;
52591}
52592__ai uint32x4_t __noswap_vmovl_high_u16(uint16x8_t __p0_124) {
52593  uint32x4_t __ret_124;
52594  uint16x4_t __a1_124 = __noswap_vget_high_u16(__p0_124);
52595  __ret_124 = (uint32x4_t)(__noswap_vshll_n_u16(__a1_124, 0));
52596  return __ret_124;
52597}
52598#endif
52599
52600#ifdef __LITTLE_ENDIAN__
52601__ai int16x8_t vmovl_high_s8(int8x16_t __p0_125) {
52602  int16x8_t __ret_125;
52603  int8x8_t __a1_125 = vget_high_s8(__p0_125);
52604  __ret_125 = (int16x8_t)(vshll_n_s8(__a1_125, 0));
52605  return __ret_125;
52606}
52607#else
52608__ai int16x8_t vmovl_high_s8(int8x16_t __p0_126) {
52609  int8x16_t __rev0_126;  __rev0_126 = __builtin_shufflevector(__p0_126, __p0_126, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52610  int16x8_t __ret_126;
52611  int8x8_t __a1_126 = __noswap_vget_high_s8(__rev0_126);
52612  __ret_126 = (int16x8_t)(__noswap_vshll_n_s8(__a1_126, 0));
52613  __ret_126 = __builtin_shufflevector(__ret_126, __ret_126, 7, 6, 5, 4, 3, 2, 1, 0);
52614  return __ret_126;
52615}
52616__ai int16x8_t __noswap_vmovl_high_s8(int8x16_t __p0_127) {
52617  int16x8_t __ret_127;
52618  int8x8_t __a1_127 = __noswap_vget_high_s8(__p0_127);
52619  __ret_127 = (int16x8_t)(__noswap_vshll_n_s8(__a1_127, 0));
52620  return __ret_127;
52621}
52622#endif
52623
52624#ifdef __LITTLE_ENDIAN__
52625__ai int64x2_t vmovl_high_s32(int32x4_t __p0_128) {
52626  int64x2_t __ret_128;
52627  int32x2_t __a1_128 = vget_high_s32(__p0_128);
52628  __ret_128 = (int64x2_t)(vshll_n_s32(__a1_128, 0));
52629  return __ret_128;
52630}
52631#else
52632__ai int64x2_t vmovl_high_s32(int32x4_t __p0_129) {
52633  int32x4_t __rev0_129;  __rev0_129 = __builtin_shufflevector(__p0_129, __p0_129, 3, 2, 1, 0);
52634  int64x2_t __ret_129;
52635  int32x2_t __a1_129 = __noswap_vget_high_s32(__rev0_129);
52636  __ret_129 = (int64x2_t)(__noswap_vshll_n_s32(__a1_129, 0));
52637  __ret_129 = __builtin_shufflevector(__ret_129, __ret_129, 1, 0);
52638  return __ret_129;
52639}
52640__ai int64x2_t __noswap_vmovl_high_s32(int32x4_t __p0_130) {
52641  int64x2_t __ret_130;
52642  int32x2_t __a1_130 = __noswap_vget_high_s32(__p0_130);
52643  __ret_130 = (int64x2_t)(__noswap_vshll_n_s32(__a1_130, 0));
52644  return __ret_130;
52645}
52646#endif
52647
52648#ifdef __LITTLE_ENDIAN__
52649__ai int32x4_t vmovl_high_s16(int16x8_t __p0_131) {
52650  int32x4_t __ret_131;
52651  int16x4_t __a1_131 = vget_high_s16(__p0_131);
52652  __ret_131 = (int32x4_t)(vshll_n_s16(__a1_131, 0));
52653  return __ret_131;
52654}
52655#else
52656__ai int32x4_t vmovl_high_s16(int16x8_t __p0_132) {
52657  int16x8_t __rev0_132;  __rev0_132 = __builtin_shufflevector(__p0_132, __p0_132, 7, 6, 5, 4, 3, 2, 1, 0);
52658  int32x4_t __ret_132;
52659  int16x4_t __a1_132 = __noswap_vget_high_s16(__rev0_132);
52660  __ret_132 = (int32x4_t)(__noswap_vshll_n_s16(__a1_132, 0));
52661  __ret_132 = __builtin_shufflevector(__ret_132, __ret_132, 3, 2, 1, 0);
52662  return __ret_132;
52663}
52664__ai int32x4_t __noswap_vmovl_high_s16(int16x8_t __p0_133) {
52665  int32x4_t __ret_133;
52666  int16x4_t __a1_133 = __noswap_vget_high_s16(__p0_133);
52667  __ret_133 = (int32x4_t)(__noswap_vshll_n_s16(__a1_133, 0));
52668  return __ret_133;
52669}
52670#endif
52671
52672#ifdef __LITTLE_ENDIAN__
52673__ai uint16x8_t vmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
52674  uint16x8_t __ret;
52675  __ret = vcombine_u16(__p0, vmovn_u32(__p1));
52676  return __ret;
52677}
52678#else
52679__ai uint16x8_t vmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
52680  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52681  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
52682  uint16x8_t __ret;
52683  __ret = __noswap_vcombine_u16(__rev0, __noswap_vmovn_u32(__rev1));
52684  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
52685  return __ret;
52686}
52687#endif
52688
52689#ifdef __LITTLE_ENDIAN__
52690__ai uint32x4_t vmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
52691  uint32x4_t __ret;
52692  __ret = vcombine_u32(__p0, vmovn_u64(__p1));
52693  return __ret;
52694}
52695#else
52696__ai uint32x4_t vmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
52697  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52698  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52699  uint32x4_t __ret;
52700  __ret = __noswap_vcombine_u32(__rev0, __noswap_vmovn_u64(__rev1));
52701  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
52702  return __ret;
52703}
52704#endif
52705
52706#ifdef __LITTLE_ENDIAN__
52707__ai uint8x16_t vmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
52708  uint8x16_t __ret;
52709  __ret = vcombine_u8(__p0, vmovn_u16(__p1));
52710  return __ret;
52711}
52712#else
52713__ai uint8x16_t vmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
52714  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52715  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
52716  uint8x16_t __ret;
52717  __ret = __noswap_vcombine_u8(__rev0, __noswap_vmovn_u16(__rev1));
52718  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52719  return __ret;
52720}
52721#endif
52722
52723#ifdef __LITTLE_ENDIAN__
52724__ai int16x8_t vmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
52725  int16x8_t __ret;
52726  __ret = vcombine_s16(__p0, vmovn_s32(__p1));
52727  return __ret;
52728}
52729#else
52730__ai int16x8_t vmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
52731  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52732  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
52733  int16x8_t __ret;
52734  __ret = __noswap_vcombine_s16(__rev0, __noswap_vmovn_s32(__rev1));
52735  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
52736  return __ret;
52737}
52738#endif
52739
52740#ifdef __LITTLE_ENDIAN__
52741__ai int32x4_t vmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
52742  int32x4_t __ret;
52743  __ret = vcombine_s32(__p0, vmovn_s64(__p1));
52744  return __ret;
52745}
52746#else
52747__ai int32x4_t vmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
52748  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52749  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52750  int32x4_t __ret;
52751  __ret = __noswap_vcombine_s32(__rev0, __noswap_vmovn_s64(__rev1));
52752  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
52753  return __ret;
52754}
52755#endif
52756
52757#ifdef __LITTLE_ENDIAN__
52758__ai int8x16_t vmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
52759  int8x16_t __ret;
52760  __ret = vcombine_s8(__p0, vmovn_s16(__p1));
52761  return __ret;
52762}
52763#else
52764__ai int8x16_t vmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
52765  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52766  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
52767  int8x16_t __ret;
52768  __ret = __noswap_vcombine_s8(__rev0, __noswap_vmovn_s16(__rev1));
52769  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52770  return __ret;
52771}
52772#endif
52773
52774#ifdef __LITTLE_ENDIAN__
52775__ai float64x2_t vmulq_f64(float64x2_t __p0, float64x2_t __p1) {
52776  float64x2_t __ret;
52777  __ret = __p0 * __p1;
52778  return __ret;
52779}
52780#else
52781__ai float64x2_t vmulq_f64(float64x2_t __p0, float64x2_t __p1) {
52782  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52783  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52784  float64x2_t __ret;
52785  __ret = __rev0 * __rev1;
52786  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52787  return __ret;
52788}
52789#endif
52790
52791#ifdef __LITTLE_ENDIAN__
52792__ai float64x1_t vmul_f64(float64x1_t __p0, float64x1_t __p1) {
52793  float64x1_t __ret;
52794  __ret = __p0 * __p1;
52795  return __ret;
52796}
52797#else
52798__ai float64x1_t vmul_f64(float64x1_t __p0, float64x1_t __p1) {
52799  float64x1_t __ret;
52800  __ret = __p0 * __p1;
52801  return __ret;
52802}
52803#endif
52804
52805#ifdef __LITTLE_ENDIAN__
52806#define vmuld_lane_f64(__p0_134, __p1_134, __p2_134) __extension__ ({ \
52807  float64_t __s0_134 = __p0_134; \
52808  float64x1_t __s1_134 = __p1_134; \
52809  float64_t __ret_134; \
52810  __ret_134 = __s0_134 * vget_lane_f64(__s1_134, __p2_134); \
52811  __ret_134; \
52812})
52813#else
52814#define vmuld_lane_f64(__p0_135, __p1_135, __p2_135) __extension__ ({ \
52815  float64_t __s0_135 = __p0_135; \
52816  float64x1_t __s1_135 = __p1_135; \
52817  float64_t __ret_135; \
52818  __ret_135 = __s0_135 * __noswap_vget_lane_f64(__s1_135, __p2_135); \
52819  __ret_135; \
52820})
52821#endif
52822
52823#ifdef __LITTLE_ENDIAN__
52824#define vmuls_lane_f32(__p0_136, __p1_136, __p2_136) __extension__ ({ \
52825  float32_t __s0_136 = __p0_136; \
52826  float32x2_t __s1_136 = __p1_136; \
52827  float32_t __ret_136; \
52828  __ret_136 = __s0_136 * vget_lane_f32(__s1_136, __p2_136); \
52829  __ret_136; \
52830})
52831#else
52832#define vmuls_lane_f32(__p0_137, __p1_137, __p2_137) __extension__ ({ \
52833  float32_t __s0_137 = __p0_137; \
52834  float32x2_t __s1_137 = __p1_137; \
52835  float32x2_t __rev1_137;  __rev1_137 = __builtin_shufflevector(__s1_137, __s1_137, 1, 0); \
52836  float32_t __ret_137; \
52837  __ret_137 = __s0_137 * __noswap_vget_lane_f32(__rev1_137, __p2_137); \
52838  __ret_137; \
52839})
52840#endif
52841
52842#ifdef __LITTLE_ENDIAN__
52843#define vmul_lane_f64(__p0, __p1, __p2) __extension__ ({ \
52844  float64x1_t __s0 = __p0; \
52845  float64x1_t __s1 = __p1; \
52846  float64x1_t __ret; \
52847  __ret = (float64x1_t) __builtin_neon_vmul_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
52848  __ret; \
52849})
52850#else
52851#define vmul_lane_f64(__p0, __p1, __p2) __extension__ ({ \
52852  float64x1_t __s0 = __p0; \
52853  float64x1_t __s1 = __p1; \
52854  float64x1_t __ret; \
52855  __ret = (float64x1_t) __builtin_neon_vmul_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
52856  __ret; \
52857})
52858#endif
52859
52860#ifdef __LITTLE_ENDIAN__
52861#define vmulq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
52862  float64x2_t __s0 = __p0; \
52863  float64x1_t __s1 = __p1; \
52864  float64x2_t __ret; \
52865  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
52866  __ret; \
52867})
52868#else
52869#define vmulq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
52870  float64x2_t __s0 = __p0; \
52871  float64x1_t __s1 = __p1; \
52872  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52873  float64x2_t __ret; \
52874  __ret = __rev0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
52875  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52876  __ret; \
52877})
52878#endif
52879
52880#ifdef __LITTLE_ENDIAN__
52881#define vmuld_laneq_f64(__p0_138, __p1_138, __p2_138) __extension__ ({ \
52882  float64_t __s0_138 = __p0_138; \
52883  float64x2_t __s1_138 = __p1_138; \
52884  float64_t __ret_138; \
52885  __ret_138 = __s0_138 * vgetq_lane_f64(__s1_138, __p2_138); \
52886  __ret_138; \
52887})
52888#else
52889#define vmuld_laneq_f64(__p0_139, __p1_139, __p2_139) __extension__ ({ \
52890  float64_t __s0_139 = __p0_139; \
52891  float64x2_t __s1_139 = __p1_139; \
52892  float64x2_t __rev1_139;  __rev1_139 = __builtin_shufflevector(__s1_139, __s1_139, 1, 0); \
52893  float64_t __ret_139; \
52894  __ret_139 = __s0_139 * __noswap_vgetq_lane_f64(__rev1_139, __p2_139); \
52895  __ret_139; \
52896})
52897#endif
52898
52899#ifdef __LITTLE_ENDIAN__
52900#define vmuls_laneq_f32(__p0_140, __p1_140, __p2_140) __extension__ ({ \
52901  float32_t __s0_140 = __p0_140; \
52902  float32x4_t __s1_140 = __p1_140; \
52903  float32_t __ret_140; \
52904  __ret_140 = __s0_140 * vgetq_lane_f32(__s1_140, __p2_140); \
52905  __ret_140; \
52906})
52907#else
52908#define vmuls_laneq_f32(__p0_141, __p1_141, __p2_141) __extension__ ({ \
52909  float32_t __s0_141 = __p0_141; \
52910  float32x4_t __s1_141 = __p1_141; \
52911  float32x4_t __rev1_141;  __rev1_141 = __builtin_shufflevector(__s1_141, __s1_141, 3, 2, 1, 0); \
52912  float32_t __ret_141; \
52913  __ret_141 = __s0_141 * __noswap_vgetq_lane_f32(__rev1_141, __p2_141); \
52914  __ret_141; \
52915})
52916#endif
52917
52918#ifdef __LITTLE_ENDIAN__
52919#define vmul_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
52920  float64x1_t __s0 = __p0; \
52921  float64x2_t __s1 = __p1; \
52922  float64x1_t __ret; \
52923  __ret = (float64x1_t) __builtin_neon_vmul_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 10); \
52924  __ret; \
52925})
52926#else
52927#define vmul_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
52928  float64x1_t __s0 = __p0; \
52929  float64x2_t __s1 = __p1; \
52930  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52931  float64x1_t __ret; \
52932  __ret = (float64x1_t) __builtin_neon_vmul_laneq_v((int8x8_t)__s0, (int8x16_t)__rev1, __p2, 10); \
52933  __ret; \
52934})
52935#endif
52936
52937#ifdef __LITTLE_ENDIAN__
52938#define vmulq_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
52939  uint32x4_t __s0 = __p0; \
52940  uint32x4_t __s1 = __p1; \
52941  uint32x4_t __ret; \
52942  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
52943  __ret; \
52944})
52945#else
52946#define vmulq_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
52947  uint32x4_t __s0 = __p0; \
52948  uint32x4_t __s1 = __p1; \
52949  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52950  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52951  uint32x4_t __ret; \
52952  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
52953  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52954  __ret; \
52955})
52956#endif
52957
52958#ifdef __LITTLE_ENDIAN__
52959#define vmulq_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
52960  uint16x8_t __s0 = __p0; \
52961  uint16x8_t __s1 = __p1; \
52962  uint16x8_t __ret; \
52963  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
52964  __ret; \
52965})
52966#else
52967#define vmulq_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
52968  uint16x8_t __s0 = __p0; \
52969  uint16x8_t __s1 = __p1; \
52970  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
52971  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52972  uint16x8_t __ret; \
52973  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
52974  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
52975  __ret; \
52976})
52977#endif
52978
52979#ifdef __LITTLE_ENDIAN__
52980#define vmulq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
52981  float64x2_t __s0 = __p0; \
52982  float64x2_t __s1 = __p1; \
52983  float64x2_t __ret; \
52984  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
52985  __ret; \
52986})
52987#else
52988#define vmulq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
52989  float64x2_t __s0 = __p0; \
52990  float64x2_t __s1 = __p1; \
52991  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52992  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52993  float64x2_t __ret; \
52994  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
52995  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52996  __ret; \
52997})
52998#endif
52999
53000#ifdef __LITTLE_ENDIAN__
53001#define vmulq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
53002  float32x4_t __s0 = __p0; \
53003  float32x4_t __s1 = __p1; \
53004  float32x4_t __ret; \
53005  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
53006  __ret; \
53007})
53008#else
53009#define vmulq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
53010  float32x4_t __s0 = __p0; \
53011  float32x4_t __s1 = __p1; \
53012  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53013  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53014  float32x4_t __ret; \
53015  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
53016  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53017  __ret; \
53018})
53019#endif
53020
53021#ifdef __LITTLE_ENDIAN__
53022#define vmulq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
53023  int32x4_t __s0 = __p0; \
53024  int32x4_t __s1 = __p1; \
53025  int32x4_t __ret; \
53026  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
53027  __ret; \
53028})
53029#else
53030#define vmulq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
53031  int32x4_t __s0 = __p0; \
53032  int32x4_t __s1 = __p1; \
53033  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53034  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53035  int32x4_t __ret; \
53036  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
53037  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53038  __ret; \
53039})
53040#endif
53041
53042#ifdef __LITTLE_ENDIAN__
53043#define vmulq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
53044  int16x8_t __s0 = __p0; \
53045  int16x8_t __s1 = __p1; \
53046  int16x8_t __ret; \
53047  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
53048  __ret; \
53049})
53050#else
53051#define vmulq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
53052  int16x8_t __s0 = __p0; \
53053  int16x8_t __s1 = __p1; \
53054  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
53055  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53056  int16x8_t __ret; \
53057  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
53058  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
53059  __ret; \
53060})
53061#endif
53062
53063#ifdef __LITTLE_ENDIAN__
53064#define vmul_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
53065  uint32x2_t __s0 = __p0; \
53066  uint32x4_t __s1 = __p1; \
53067  uint32x2_t __ret; \
53068  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
53069  __ret; \
53070})
53071#else
53072#define vmul_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
53073  uint32x2_t __s0 = __p0; \
53074  uint32x4_t __s1 = __p1; \
53075  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53076  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53077  uint32x2_t __ret; \
53078  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
53079  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53080  __ret; \
53081})
53082#endif
53083
53084#ifdef __LITTLE_ENDIAN__
53085#define vmul_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
53086  uint16x4_t __s0 = __p0; \
53087  uint16x8_t __s1 = __p1; \
53088  uint16x4_t __ret; \
53089  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
53090  __ret; \
53091})
53092#else
53093#define vmul_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
53094  uint16x4_t __s0 = __p0; \
53095  uint16x8_t __s1 = __p1; \
53096  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53097  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53098  uint16x4_t __ret; \
53099  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
53100  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53101  __ret; \
53102})
53103#endif
53104
53105#ifdef __LITTLE_ENDIAN__
53106#define vmul_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
53107  float32x2_t __s0 = __p0; \
53108  float32x4_t __s1 = __p1; \
53109  float32x2_t __ret; \
53110  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
53111  __ret; \
53112})
53113#else
53114#define vmul_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
53115  float32x2_t __s0 = __p0; \
53116  float32x4_t __s1 = __p1; \
53117  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53118  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53119  float32x2_t __ret; \
53120  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
53121  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53122  __ret; \
53123})
53124#endif
53125
53126#ifdef __LITTLE_ENDIAN__
53127#define vmul_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
53128  int32x2_t __s0 = __p0; \
53129  int32x4_t __s1 = __p1; \
53130  int32x2_t __ret; \
53131  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
53132  __ret; \
53133})
53134#else
53135#define vmul_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
53136  int32x2_t __s0 = __p0; \
53137  int32x4_t __s1 = __p1; \
53138  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53139  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53140  int32x2_t __ret; \
53141  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
53142  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53143  __ret; \
53144})
53145#endif
53146
53147#ifdef __LITTLE_ENDIAN__
53148#define vmul_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
53149  int16x4_t __s0 = __p0; \
53150  int16x8_t __s1 = __p1; \
53151  int16x4_t __ret; \
53152  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
53153  __ret; \
53154})
53155#else
53156#define vmul_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
53157  int16x4_t __s0 = __p0; \
53158  int16x8_t __s1 = __p1; \
53159  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53160  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53161  int16x4_t __ret; \
53162  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
53163  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53164  __ret; \
53165})
53166#endif
53167
53168#ifdef __LITTLE_ENDIAN__
53169__ai float64x1_t vmul_n_f64(float64x1_t __p0, float64_t __p1) {
53170  float64x1_t __ret;
53171  __ret = (float64x1_t) __builtin_neon_vmul_n_f64((int8x8_t)__p0, __p1);
53172  return __ret;
53173}
53174#else
53175__ai float64x1_t vmul_n_f64(float64x1_t __p0, float64_t __p1) {
53176  float64x1_t __ret;
53177  __ret = (float64x1_t) __builtin_neon_vmul_n_f64((int8x8_t)__p0, __p1);
53178  return __ret;
53179}
53180#endif
53181
53182#ifdef __LITTLE_ENDIAN__
53183__ai float64x2_t vmulq_n_f64(float64x2_t __p0, float64_t __p1) {
53184  float64x2_t __ret;
53185  __ret = __p0 * (float64x2_t) {__p1, __p1};
53186  return __ret;
53187}
53188#else
53189__ai float64x2_t vmulq_n_f64(float64x2_t __p0, float64_t __p1) {
53190  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53191  float64x2_t __ret;
53192  __ret = __rev0 * (float64x2_t) {__p1, __p1};
53193  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53194  return __ret;
53195}
53196#endif
53197
53198#ifdef __LITTLE_ENDIAN__
53199__ai poly128_t vmull_p64(poly64_t __p0, poly64_t __p1) {
53200  poly128_t __ret;
53201  __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
53202  return __ret;
53203}
53204#else
53205__ai poly128_t vmull_p64(poly64_t __p0, poly64_t __p1) {
53206  poly128_t __ret;
53207  __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
53208  return __ret;
53209}
53210__ai poly128_t __noswap_vmull_p64(poly64_t __p0, poly64_t __p1) {
53211  poly128_t __ret;
53212  __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
53213  return __ret;
53214}
53215#endif
53216
53217#ifdef __LITTLE_ENDIAN__
53218__ai poly16x8_t vmull_high_p8(poly8x16_t __p0, poly8x16_t __p1) {
53219  poly16x8_t __ret;
53220  __ret = vmull_p8(vget_high_p8(__p0), vget_high_p8(__p1));
53221  return __ret;
53222}
53223#else
53224__ai poly16x8_t vmull_high_p8(poly8x16_t __p0, poly8x16_t __p1) {
53225  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53226  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53227  poly16x8_t __ret;
53228  __ret = __noswap_vmull_p8(__noswap_vget_high_p8(__rev0), __noswap_vget_high_p8(__rev1));
53229  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
53230  return __ret;
53231}
53232#endif
53233
53234#ifdef __LITTLE_ENDIAN__
53235__ai uint16x8_t vmull_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
53236  uint16x8_t __ret;
53237  __ret = vmull_u8(vget_high_u8(__p0), vget_high_u8(__p1));
53238  return __ret;
53239}
53240#else
53241__ai uint16x8_t vmull_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
53242  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53243  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53244  uint16x8_t __ret;
53245  __ret = __noswap_vmull_u8(__noswap_vget_high_u8(__rev0), __noswap_vget_high_u8(__rev1));
53246  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
53247  return __ret;
53248}
53249#endif
53250
53251#ifdef __LITTLE_ENDIAN__
53252__ai uint64x2_t vmull_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
53253  uint64x2_t __ret;
53254  __ret = vmull_u32(vget_high_u32(__p0), vget_high_u32(__p1));
53255  return __ret;
53256}
53257#else
53258__ai uint64x2_t vmull_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
53259  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53260  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53261  uint64x2_t __ret;
53262  __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __noswap_vget_high_u32(__rev1));
53263  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53264  return __ret;
53265}
53266#endif
53267
53268#ifdef __LITTLE_ENDIAN__
53269__ai uint32x4_t vmull_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
53270  uint32x4_t __ret;
53271  __ret = vmull_u16(vget_high_u16(__p0), vget_high_u16(__p1));
53272  return __ret;
53273}
53274#else
53275__ai uint32x4_t vmull_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
53276  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53277  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
53278  uint32x4_t __ret;
53279  __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __noswap_vget_high_u16(__rev1));
53280  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53281  return __ret;
53282}
53283#endif
53284
53285#ifdef __LITTLE_ENDIAN__
53286__ai int16x8_t vmull_high_s8(int8x16_t __p0, int8x16_t __p1) {
53287  int16x8_t __ret;
53288  __ret = vmull_s8(vget_high_s8(__p0), vget_high_s8(__p1));
53289  return __ret;
53290}
53291#else
53292__ai int16x8_t vmull_high_s8(int8x16_t __p0, int8x16_t __p1) {
53293  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53294  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53295  int16x8_t __ret;
53296  __ret = __noswap_vmull_s8(__noswap_vget_high_s8(__rev0), __noswap_vget_high_s8(__rev1));
53297  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
53298  return __ret;
53299}
53300#endif
53301
53302#ifdef __LITTLE_ENDIAN__
53303__ai int64x2_t vmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
53304  int64x2_t __ret;
53305  __ret = vmull_s32(vget_high_s32(__p0), vget_high_s32(__p1));
53306  return __ret;
53307}
53308#else
53309__ai int64x2_t vmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
53310  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53311  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53312  int64x2_t __ret;
53313  __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
53314  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53315  return __ret;
53316}
53317#endif
53318
53319#ifdef __LITTLE_ENDIAN__
53320__ai int32x4_t vmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
53321  int32x4_t __ret;
53322  __ret = vmull_s16(vget_high_s16(__p0), vget_high_s16(__p1));
53323  return __ret;
53324}
53325#else
53326__ai int32x4_t vmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
53327  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53328  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
53329  int32x4_t __ret;
53330  __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
53331  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53332  return __ret;
53333}
53334#endif
53335
53336#ifdef __LITTLE_ENDIAN__
53337__ai poly128_t vmull_high_p64(poly64x2_t __p0, poly64x2_t __p1) {
53338  poly128_t __ret;
53339  __ret = vmull_p64((poly64_t)(vget_high_p64(__p0)), (poly64_t)(vget_high_p64(__p1)));
53340  return __ret;
53341}
53342#else
53343__ai poly128_t vmull_high_p64(poly64x2_t __p0, poly64x2_t __p1) {
53344  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53345  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53346  poly128_t __ret;
53347  __ret = __noswap_vmull_p64((poly64_t)(__noswap_vget_high_p64(__rev0)), (poly64_t)(__noswap_vget_high_p64(__rev1)));
53348  return __ret;
53349}
53350#endif
53351
53352#ifdef __LITTLE_ENDIAN__
53353#define vmull_high_lane_u32(__p0, __p1, __p2) __extension__ ({ \
53354  uint32x4_t __s0 = __p0; \
53355  uint32x2_t __s1 = __p1; \
53356  uint64x2_t __ret; \
53357  __ret = vmull_u32(vget_high_u32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
53358  __ret; \
53359})
53360#else
53361#define vmull_high_lane_u32(__p0, __p1, __p2) __extension__ ({ \
53362  uint32x4_t __s0 = __p0; \
53363  uint32x2_t __s1 = __p1; \
53364  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53365  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53366  uint64x2_t __ret; \
53367  __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
53368  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53369  __ret; \
53370})
53371#endif
53372
53373#ifdef __LITTLE_ENDIAN__
53374#define vmull_high_lane_u16(__p0, __p1, __p2) __extension__ ({ \
53375  uint16x8_t __s0 = __p0; \
53376  uint16x4_t __s1 = __p1; \
53377  uint32x4_t __ret; \
53378  __ret = vmull_u16(vget_high_u16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
53379  __ret; \
53380})
53381#else
53382#define vmull_high_lane_u16(__p0, __p1, __p2) __extension__ ({ \
53383  uint16x8_t __s0 = __p0; \
53384  uint16x4_t __s1 = __p1; \
53385  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
53386  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53387  uint32x4_t __ret; \
53388  __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
53389  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53390  __ret; \
53391})
53392#endif
53393
53394#ifdef __LITTLE_ENDIAN__
53395#define vmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
53396  int32x4_t __s0 = __p0; \
53397  int32x2_t __s1 = __p1; \
53398  int64x2_t __ret; \
53399  __ret = vmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
53400  __ret; \
53401})
53402#else
53403#define vmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
53404  int32x4_t __s0 = __p0; \
53405  int32x2_t __s1 = __p1; \
53406  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53407  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53408  int64x2_t __ret; \
53409  __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
53410  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53411  __ret; \
53412})
53413#endif
53414
53415#ifdef __LITTLE_ENDIAN__
53416#define vmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
53417  int16x8_t __s0 = __p0; \
53418  int16x4_t __s1 = __p1; \
53419  int32x4_t __ret; \
53420  __ret = vmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
53421  __ret; \
53422})
53423#else
53424#define vmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
53425  int16x8_t __s0 = __p0; \
53426  int16x4_t __s1 = __p1; \
53427  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
53428  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53429  int32x4_t __ret; \
53430  __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
53431  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53432  __ret; \
53433})
53434#endif
53435
53436#ifdef __LITTLE_ENDIAN__
53437#define vmull_high_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
53438  uint32x4_t __s0 = __p0; \
53439  uint32x4_t __s1 = __p1; \
53440  uint64x2_t __ret; \
53441  __ret = vmull_u32(vget_high_u32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
53442  __ret; \
53443})
53444#else
53445#define vmull_high_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
53446  uint32x4_t __s0 = __p0; \
53447  uint32x4_t __s1 = __p1; \
53448  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53449  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53450  uint64x2_t __ret; \
53451  __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
53452  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53453  __ret; \
53454})
53455#endif
53456
53457#ifdef __LITTLE_ENDIAN__
53458#define vmull_high_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
53459  uint16x8_t __s0 = __p0; \
53460  uint16x8_t __s1 = __p1; \
53461  uint32x4_t __ret; \
53462  __ret = vmull_u16(vget_high_u16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
53463  __ret; \
53464})
53465#else
53466#define vmull_high_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
53467  uint16x8_t __s0 = __p0; \
53468  uint16x8_t __s1 = __p1; \
53469  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
53470  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53471  uint32x4_t __ret; \
53472  __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
53473  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53474  __ret; \
53475})
53476#endif
53477
53478#ifdef __LITTLE_ENDIAN__
53479#define vmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
53480  int32x4_t __s0 = __p0; \
53481  int32x4_t __s1 = __p1; \
53482  int64x2_t __ret; \
53483  __ret = vmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
53484  __ret; \
53485})
53486#else
53487#define vmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
53488  int32x4_t __s0 = __p0; \
53489  int32x4_t __s1 = __p1; \
53490  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53491  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53492  int64x2_t __ret; \
53493  __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
53494  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53495  __ret; \
53496})
53497#endif
53498
53499#ifdef __LITTLE_ENDIAN__
53500#define vmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
53501  int16x8_t __s0 = __p0; \
53502  int16x8_t __s1 = __p1; \
53503  int32x4_t __ret; \
53504  __ret = vmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
53505  __ret; \
53506})
53507#else
53508#define vmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
53509  int16x8_t __s0 = __p0; \
53510  int16x8_t __s1 = __p1; \
53511  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
53512  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53513  int32x4_t __ret; \
53514  __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
53515  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53516  __ret; \
53517})
53518#endif
53519
53520#ifdef __LITTLE_ENDIAN__
53521__ai uint64x2_t vmull_high_n_u32(uint32x4_t __p0, uint32_t __p1) {
53522  uint64x2_t __ret;
53523  __ret = vmull_n_u32(vget_high_u32(__p0), __p1);
53524  return __ret;
53525}
53526#else
53527__ai uint64x2_t vmull_high_n_u32(uint32x4_t __p0, uint32_t __p1) {
53528  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53529  uint64x2_t __ret;
53530  __ret = __noswap_vmull_n_u32(__noswap_vget_high_u32(__rev0), __p1);
53531  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53532  return __ret;
53533}
53534#endif
53535
53536#ifdef __LITTLE_ENDIAN__
53537__ai uint32x4_t vmull_high_n_u16(uint16x8_t __p0, uint16_t __p1) {
53538  uint32x4_t __ret;
53539  __ret = vmull_n_u16(vget_high_u16(__p0), __p1);
53540  return __ret;
53541}
53542#else
53543__ai uint32x4_t vmull_high_n_u16(uint16x8_t __p0, uint16_t __p1) {
53544  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53545  uint32x4_t __ret;
53546  __ret = __noswap_vmull_n_u16(__noswap_vget_high_u16(__rev0), __p1);
53547  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53548  return __ret;
53549}
53550#endif
53551
53552#ifdef __LITTLE_ENDIAN__
53553__ai int64x2_t vmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
53554  int64x2_t __ret;
53555  __ret = vmull_n_s32(vget_high_s32(__p0), __p1);
53556  return __ret;
53557}
53558#else
53559__ai int64x2_t vmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
53560  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53561  int64x2_t __ret;
53562  __ret = __noswap_vmull_n_s32(__noswap_vget_high_s32(__rev0), __p1);
53563  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53564  return __ret;
53565}
53566#endif
53567
53568#ifdef __LITTLE_ENDIAN__
53569__ai int32x4_t vmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
53570  int32x4_t __ret;
53571  __ret = vmull_n_s16(vget_high_s16(__p0), __p1);
53572  return __ret;
53573}
53574#else
53575__ai int32x4_t vmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
53576  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53577  int32x4_t __ret;
53578  __ret = __noswap_vmull_n_s16(__noswap_vget_high_s16(__rev0), __p1);
53579  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53580  return __ret;
53581}
53582#endif
53583
53584#ifdef __LITTLE_ENDIAN__
53585#define vmull_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
53586  uint32x2_t __s0 = __p0; \
53587  uint32x4_t __s1 = __p1; \
53588  uint64x2_t __ret; \
53589  __ret = vmull_u32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
53590  __ret; \
53591})
53592#else
53593#define vmull_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
53594  uint32x2_t __s0 = __p0; \
53595  uint32x4_t __s1 = __p1; \
53596  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53597  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53598  uint64x2_t __ret; \
53599  __ret = __noswap_vmull_u32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
53600  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53601  __ret; \
53602})
53603#endif
53604
53605#ifdef __LITTLE_ENDIAN__
53606#define vmull_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
53607  uint16x4_t __s0 = __p0; \
53608  uint16x8_t __s1 = __p1; \
53609  uint32x4_t __ret; \
53610  __ret = vmull_u16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
53611  __ret; \
53612})
53613#else
53614#define vmull_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
53615  uint16x4_t __s0 = __p0; \
53616  uint16x8_t __s1 = __p1; \
53617  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53618  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53619  uint32x4_t __ret; \
53620  __ret = __noswap_vmull_u16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
53621  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53622  __ret; \
53623})
53624#endif
53625
53626#ifdef __LITTLE_ENDIAN__
53627#define vmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
53628  int32x2_t __s0 = __p0; \
53629  int32x4_t __s1 = __p1; \
53630  int64x2_t __ret; \
53631  __ret = vmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
53632  __ret; \
53633})
53634#else
53635#define vmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
53636  int32x2_t __s0 = __p0; \
53637  int32x4_t __s1 = __p1; \
53638  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53639  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53640  int64x2_t __ret; \
53641  __ret = __noswap_vmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
53642  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53643  __ret; \
53644})
53645#endif
53646
53647#ifdef __LITTLE_ENDIAN__
53648#define vmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
53649  int16x4_t __s0 = __p0; \
53650  int16x8_t __s1 = __p1; \
53651  int32x4_t __ret; \
53652  __ret = vmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
53653  __ret; \
53654})
53655#else
53656#define vmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
53657  int16x4_t __s0 = __p0; \
53658  int16x8_t __s1 = __p1; \
53659  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53660  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53661  int32x4_t __ret; \
53662  __ret = __noswap_vmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
53663  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53664  __ret; \
53665})
53666#endif
53667
53668#ifdef __LITTLE_ENDIAN__
53669__ai float64x2_t vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
53670  float64x2_t __ret;
53671  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
53672  return __ret;
53673}
53674#else
53675__ai float64x2_t vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
53676  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53677  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53678  float64x2_t __ret;
53679  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
53680  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53681  return __ret;
53682}
53683__ai float64x2_t __noswap_vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
53684  float64x2_t __ret;
53685  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
53686  return __ret;
53687}
53688#endif
53689
53690#ifdef __LITTLE_ENDIAN__
53691__ai float32x4_t vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
53692  float32x4_t __ret;
53693  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
53694  return __ret;
53695}
53696#else
53697__ai float32x4_t vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
53698  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53699  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53700  float32x4_t __ret;
53701  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
53702  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53703  return __ret;
53704}
53705__ai float32x4_t __noswap_vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
53706  float32x4_t __ret;
53707  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
53708  return __ret;
53709}
53710#endif
53711
53712#ifdef __LITTLE_ENDIAN__
53713__ai float64x1_t vmulx_f64(float64x1_t __p0, float64x1_t __p1) {
53714  float64x1_t __ret;
53715  __ret = (float64x1_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
53716  return __ret;
53717}
53718#else
53719__ai float64x1_t vmulx_f64(float64x1_t __p0, float64x1_t __p1) {
53720  float64x1_t __ret;
53721  __ret = (float64x1_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
53722  return __ret;
53723}
53724#endif
53725
53726#ifdef __LITTLE_ENDIAN__
53727__ai float32x2_t vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
53728  float32x2_t __ret;
53729  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
53730  return __ret;
53731}
53732#else
53733__ai float32x2_t vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
53734  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53735  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53736  float32x2_t __ret;
53737  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
53738  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53739  return __ret;
53740}
53741__ai float32x2_t __noswap_vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
53742  float32x2_t __ret;
53743  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
53744  return __ret;
53745}
53746#endif
53747
53748#ifdef __LITTLE_ENDIAN__
53749__ai float64_t vmulxd_f64(float64_t __p0, float64_t __p1) {
53750  float64_t __ret;
53751  __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
53752  return __ret;
53753}
53754#else
53755__ai float64_t vmulxd_f64(float64_t __p0, float64_t __p1) {
53756  float64_t __ret;
53757  __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
53758  return __ret;
53759}
53760__ai float64_t __noswap_vmulxd_f64(float64_t __p0, float64_t __p1) {
53761  float64_t __ret;
53762  __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
53763  return __ret;
53764}
53765#endif
53766
53767#ifdef __LITTLE_ENDIAN__
53768__ai float32_t vmulxs_f32(float32_t __p0, float32_t __p1) {
53769  float32_t __ret;
53770  __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
53771  return __ret;
53772}
53773#else
53774__ai float32_t vmulxs_f32(float32_t __p0, float32_t __p1) {
53775  float32_t __ret;
53776  __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
53777  return __ret;
53778}
53779__ai float32_t __noswap_vmulxs_f32(float32_t __p0, float32_t __p1) {
53780  float32_t __ret;
53781  __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
53782  return __ret;
53783}
53784#endif
53785
53786#ifdef __LITTLE_ENDIAN__
53787#define vmulxd_lane_f64(__p0_142, __p1_142, __p2_142) __extension__ ({ \
53788  float64_t __s0_142 = __p0_142; \
53789  float64x1_t __s1_142 = __p1_142; \
53790  float64_t __ret_142; \
53791  __ret_142 = vmulxd_f64(__s0_142, vget_lane_f64(__s1_142, __p2_142)); \
53792  __ret_142; \
53793})
53794#else
53795#define vmulxd_lane_f64(__p0_143, __p1_143, __p2_143) __extension__ ({ \
53796  float64_t __s0_143 = __p0_143; \
53797  float64x1_t __s1_143 = __p1_143; \
53798  float64_t __ret_143; \
53799  __ret_143 = __noswap_vmulxd_f64(__s0_143, __noswap_vget_lane_f64(__s1_143, __p2_143)); \
53800  __ret_143; \
53801})
53802#endif
53803
53804#ifdef __LITTLE_ENDIAN__
53805#define vmulxs_lane_f32(__p0_144, __p1_144, __p2_144) __extension__ ({ \
53806  float32_t __s0_144 = __p0_144; \
53807  float32x2_t __s1_144 = __p1_144; \
53808  float32_t __ret_144; \
53809  __ret_144 = vmulxs_f32(__s0_144, vget_lane_f32(__s1_144, __p2_144)); \
53810  __ret_144; \
53811})
53812#else
53813#define vmulxs_lane_f32(__p0_145, __p1_145, __p2_145) __extension__ ({ \
53814  float32_t __s0_145 = __p0_145; \
53815  float32x2_t __s1_145 = __p1_145; \
53816  float32x2_t __rev1_145;  __rev1_145 = __builtin_shufflevector(__s1_145, __s1_145, 1, 0); \
53817  float32_t __ret_145; \
53818  __ret_145 = __noswap_vmulxs_f32(__s0_145, __noswap_vget_lane_f32(__rev1_145, __p2_145)); \
53819  __ret_145; \
53820})
53821#endif
53822
53823#ifdef __LITTLE_ENDIAN__
53824#define vmulxq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53825  float64x2_t __s0 = __p0; \
53826  float64x1_t __s1 = __p1; \
53827  float64x2_t __ret; \
53828  __ret = vmulxq_f64(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
53829  __ret; \
53830})
53831#else
53832#define vmulxq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53833  float64x2_t __s0 = __p0; \
53834  float64x1_t __s1 = __p1; \
53835  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53836  float64x2_t __ret; \
53837  __ret = __noswap_vmulxq_f64(__rev0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
53838  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53839  __ret; \
53840})
53841#endif
53842
53843#ifdef __LITTLE_ENDIAN__
53844#define vmulxq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
53845  float32x4_t __s0 = __p0; \
53846  float32x2_t __s1 = __p1; \
53847  float32x4_t __ret; \
53848  __ret = vmulxq_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
53849  __ret; \
53850})
53851#else
53852#define vmulxq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
53853  float32x4_t __s0 = __p0; \
53854  float32x2_t __s1 = __p1; \
53855  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53856  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53857  float32x4_t __ret; \
53858  __ret = __noswap_vmulxq_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
53859  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53860  __ret; \
53861})
53862#endif
53863
53864#ifdef __LITTLE_ENDIAN__
53865#define vmulx_lane_f32(__p0, __p1, __p2) __extension__ ({ \
53866  float32x2_t __s0 = __p0; \
53867  float32x2_t __s1 = __p1; \
53868  float32x2_t __ret; \
53869  __ret = vmulx_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
53870  __ret; \
53871})
53872#else
53873#define vmulx_lane_f32(__p0, __p1, __p2) __extension__ ({ \
53874  float32x2_t __s0 = __p0; \
53875  float32x2_t __s1 = __p1; \
53876  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53877  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53878  float32x2_t __ret; \
53879  __ret = __noswap_vmulx_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
53880  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53881  __ret; \
53882})
53883#endif
53884
53885#ifdef __LITTLE_ENDIAN__
53886#define vmulxd_laneq_f64(__p0_146, __p1_146, __p2_146) __extension__ ({ \
53887  float64_t __s0_146 = __p0_146; \
53888  float64x2_t __s1_146 = __p1_146; \
53889  float64_t __ret_146; \
53890  __ret_146 = vmulxd_f64(__s0_146, vgetq_lane_f64(__s1_146, __p2_146)); \
53891  __ret_146; \
53892})
53893#else
53894#define vmulxd_laneq_f64(__p0_147, __p1_147, __p2_147) __extension__ ({ \
53895  float64_t __s0_147 = __p0_147; \
53896  float64x2_t __s1_147 = __p1_147; \
53897  float64x2_t __rev1_147;  __rev1_147 = __builtin_shufflevector(__s1_147, __s1_147, 1, 0); \
53898  float64_t __ret_147; \
53899  __ret_147 = __noswap_vmulxd_f64(__s0_147, __noswap_vgetq_lane_f64(__rev1_147, __p2_147)); \
53900  __ret_147; \
53901})
53902#endif
53903
53904#ifdef __LITTLE_ENDIAN__
53905#define vmulxs_laneq_f32(__p0_148, __p1_148, __p2_148) __extension__ ({ \
53906  float32_t __s0_148 = __p0_148; \
53907  float32x4_t __s1_148 = __p1_148; \
53908  float32_t __ret_148; \
53909  __ret_148 = vmulxs_f32(__s0_148, vgetq_lane_f32(__s1_148, __p2_148)); \
53910  __ret_148; \
53911})
53912#else
53913#define vmulxs_laneq_f32(__p0_149, __p1_149, __p2_149) __extension__ ({ \
53914  float32_t __s0_149 = __p0_149; \
53915  float32x4_t __s1_149 = __p1_149; \
53916  float32x4_t __rev1_149;  __rev1_149 = __builtin_shufflevector(__s1_149, __s1_149, 3, 2, 1, 0); \
53917  float32_t __ret_149; \
53918  __ret_149 = __noswap_vmulxs_f32(__s0_149, __noswap_vgetq_lane_f32(__rev1_149, __p2_149)); \
53919  __ret_149; \
53920})
53921#endif
53922
53923#ifdef __LITTLE_ENDIAN__
53924#define vmulxq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
53925  float64x2_t __s0 = __p0; \
53926  float64x2_t __s1 = __p1; \
53927  float64x2_t __ret; \
53928  __ret = vmulxq_f64(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
53929  __ret; \
53930})
53931#else
53932#define vmulxq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
53933  float64x2_t __s0 = __p0; \
53934  float64x2_t __s1 = __p1; \
53935  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53936  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53937  float64x2_t __ret; \
53938  __ret = __noswap_vmulxq_f64(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
53939  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53940  __ret; \
53941})
53942#endif
53943
53944#ifdef __LITTLE_ENDIAN__
53945#define vmulxq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
53946  float32x4_t __s0 = __p0; \
53947  float32x4_t __s1 = __p1; \
53948  float32x4_t __ret; \
53949  __ret = vmulxq_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
53950  __ret; \
53951})
53952#else
53953#define vmulxq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
53954  float32x4_t __s0 = __p0; \
53955  float32x4_t __s1 = __p1; \
53956  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53957  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53958  float32x4_t __ret; \
53959  __ret = __noswap_vmulxq_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
53960  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53961  __ret; \
53962})
53963#endif
53964
53965#ifdef __LITTLE_ENDIAN__
53966#define vmulx_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
53967  float32x2_t __s0 = __p0; \
53968  float32x4_t __s1 = __p1; \
53969  float32x2_t __ret; \
53970  __ret = vmulx_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
53971  __ret; \
53972})
53973#else
53974#define vmulx_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
53975  float32x2_t __s0 = __p0; \
53976  float32x4_t __s1 = __p1; \
53977  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53978  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53979  float32x2_t __ret; \
53980  __ret = __noswap_vmulx_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
53981  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53982  __ret; \
53983})
53984#endif
53985
53986#ifdef __LITTLE_ENDIAN__
53987__ai float64x2_t vnegq_f64(float64x2_t __p0) {
53988  float64x2_t __ret;
53989  __ret = -__p0;
53990  return __ret;
53991}
53992#else
53993__ai float64x2_t vnegq_f64(float64x2_t __p0) {
53994  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53995  float64x2_t __ret;
53996  __ret = -__rev0;
53997  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53998  return __ret;
53999}
54000#endif
54001
54002#ifdef __LITTLE_ENDIAN__
54003__ai int64x2_t vnegq_s64(int64x2_t __p0) {
54004  int64x2_t __ret;
54005  __ret = -__p0;
54006  return __ret;
54007}
54008#else
54009__ai int64x2_t vnegq_s64(int64x2_t __p0) {
54010  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54011  int64x2_t __ret;
54012  __ret = -__rev0;
54013  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54014  return __ret;
54015}
54016#endif
54017
54018#ifdef __LITTLE_ENDIAN__
54019__ai float64x1_t vneg_f64(float64x1_t __p0) {
54020  float64x1_t __ret;
54021  __ret = -__p0;
54022  return __ret;
54023}
54024#else
54025__ai float64x1_t vneg_f64(float64x1_t __p0) {
54026  float64x1_t __ret;
54027  __ret = -__p0;
54028  return __ret;
54029}
54030#endif
54031
54032#ifdef __LITTLE_ENDIAN__
54033__ai int64x1_t vneg_s64(int64x1_t __p0) {
54034  int64x1_t __ret;
54035  __ret = -__p0;
54036  return __ret;
54037}
54038#else
54039__ai int64x1_t vneg_s64(int64x1_t __p0) {
54040  int64x1_t __ret;
54041  __ret = -__p0;
54042  return __ret;
54043}
54044#endif
54045
54046#ifdef __LITTLE_ENDIAN__
54047__ai int64_t vnegd_s64(int64_t __p0) {
54048  int64_t __ret;
54049  __ret = (int64_t) __builtin_neon_vnegd_s64(__p0);
54050  return __ret;
54051}
54052#else
54053__ai int64_t vnegd_s64(int64_t __p0) {
54054  int64_t __ret;
54055  __ret = (int64_t) __builtin_neon_vnegd_s64(__p0);
54056  return __ret;
54057}
54058#endif
54059
54060#ifdef __LITTLE_ENDIAN__
54061__ai uint8x16_t vpaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
54062  uint8x16_t __ret;
54063  __ret = (uint8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
54064  return __ret;
54065}
54066#else
54067__ai uint8x16_t vpaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
54068  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54069  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54070  uint8x16_t __ret;
54071  __ret = (uint8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
54072  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54073  return __ret;
54074}
54075#endif
54076
54077#ifdef __LITTLE_ENDIAN__
54078__ai uint32x4_t vpaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
54079  uint32x4_t __ret;
54080  __ret = (uint32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
54081  return __ret;
54082}
54083#else
54084__ai uint32x4_t vpaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
54085  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54086  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54087  uint32x4_t __ret;
54088  __ret = (uint32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
54089  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54090  return __ret;
54091}
54092#endif
54093
54094#ifdef __LITTLE_ENDIAN__
54095__ai uint64x2_t vpaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
54096  uint64x2_t __ret;
54097  __ret = (uint64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
54098  return __ret;
54099}
54100#else
54101__ai uint64x2_t vpaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
54102  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54103  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54104  uint64x2_t __ret;
54105  __ret = (uint64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
54106  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54107  return __ret;
54108}
54109#endif
54110
54111#ifdef __LITTLE_ENDIAN__
54112__ai uint16x8_t vpaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
54113  uint16x8_t __ret;
54114  __ret = (uint16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
54115  return __ret;
54116}
54117#else
54118__ai uint16x8_t vpaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
54119  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54120  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54121  uint16x8_t __ret;
54122  __ret = (uint16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
54123  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54124  return __ret;
54125}
54126#endif
54127
54128#ifdef __LITTLE_ENDIAN__
54129__ai int8x16_t vpaddq_s8(int8x16_t __p0, int8x16_t __p1) {
54130  int8x16_t __ret;
54131  __ret = (int8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
54132  return __ret;
54133}
54134#else
54135__ai int8x16_t vpaddq_s8(int8x16_t __p0, int8x16_t __p1) {
54136  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54137  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54138  int8x16_t __ret;
54139  __ret = (int8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
54140  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54141  return __ret;
54142}
54143#endif
54144
54145#ifdef __LITTLE_ENDIAN__
54146__ai float64x2_t vpaddq_f64(float64x2_t __p0, float64x2_t __p1) {
54147  float64x2_t __ret;
54148  __ret = (float64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
54149  return __ret;
54150}
54151#else
54152__ai float64x2_t vpaddq_f64(float64x2_t __p0, float64x2_t __p1) {
54153  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54154  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54155  float64x2_t __ret;
54156  __ret = (float64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
54157  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54158  return __ret;
54159}
54160#endif
54161
54162#ifdef __LITTLE_ENDIAN__
54163__ai float32x4_t vpaddq_f32(float32x4_t __p0, float32x4_t __p1) {
54164  float32x4_t __ret;
54165  __ret = (float32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
54166  return __ret;
54167}
54168#else
54169__ai float32x4_t vpaddq_f32(float32x4_t __p0, float32x4_t __p1) {
54170  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54171  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54172  float32x4_t __ret;
54173  __ret = (float32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
54174  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54175  return __ret;
54176}
54177#endif
54178
54179#ifdef __LITTLE_ENDIAN__
54180__ai int32x4_t vpaddq_s32(int32x4_t __p0, int32x4_t __p1) {
54181  int32x4_t __ret;
54182  __ret = (int32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
54183  return __ret;
54184}
54185#else
54186__ai int32x4_t vpaddq_s32(int32x4_t __p0, int32x4_t __p1) {
54187  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54188  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54189  int32x4_t __ret;
54190  __ret = (int32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
54191  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54192  return __ret;
54193}
54194#endif
54195
54196#ifdef __LITTLE_ENDIAN__
54197__ai int64x2_t vpaddq_s64(int64x2_t __p0, int64x2_t __p1) {
54198  int64x2_t __ret;
54199  __ret = (int64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
54200  return __ret;
54201}
54202#else
54203__ai int64x2_t vpaddq_s64(int64x2_t __p0, int64x2_t __p1) {
54204  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54205  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54206  int64x2_t __ret;
54207  __ret = (int64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
54208  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54209  return __ret;
54210}
54211#endif
54212
54213#ifdef __LITTLE_ENDIAN__
54214__ai int16x8_t vpaddq_s16(int16x8_t __p0, int16x8_t __p1) {
54215  int16x8_t __ret;
54216  __ret = (int16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
54217  return __ret;
54218}
54219#else
54220__ai int16x8_t vpaddq_s16(int16x8_t __p0, int16x8_t __p1) {
54221  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54222  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54223  int16x8_t __ret;
54224  __ret = (int16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
54225  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54226  return __ret;
54227}
54228#endif
54229
54230#ifdef __LITTLE_ENDIAN__
54231__ai uint64_t vpaddd_u64(uint64x2_t __p0) {
54232  uint64_t __ret;
54233  __ret = (uint64_t) __builtin_neon_vpaddd_u64((int8x16_t)__p0);
54234  return __ret;
54235}
54236#else
54237__ai uint64_t vpaddd_u64(uint64x2_t __p0) {
54238  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54239  uint64_t __ret;
54240  __ret = (uint64_t) __builtin_neon_vpaddd_u64((int8x16_t)__rev0);
54241  return __ret;
54242}
54243#endif
54244
54245#ifdef __LITTLE_ENDIAN__
54246__ai float64_t vpaddd_f64(float64x2_t __p0) {
54247  float64_t __ret;
54248  __ret = (float64_t) __builtin_neon_vpaddd_f64((int8x16_t)__p0);
54249  return __ret;
54250}
54251#else
54252__ai float64_t vpaddd_f64(float64x2_t __p0) {
54253  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54254  float64_t __ret;
54255  __ret = (float64_t) __builtin_neon_vpaddd_f64((int8x16_t)__rev0);
54256  return __ret;
54257}
54258#endif
54259
54260#ifdef __LITTLE_ENDIAN__
54261__ai int64_t vpaddd_s64(int64x2_t __p0) {
54262  int64_t __ret;
54263  __ret = (int64_t) __builtin_neon_vpaddd_s64((int8x16_t)__p0);
54264  return __ret;
54265}
54266#else
54267__ai int64_t vpaddd_s64(int64x2_t __p0) {
54268  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54269  int64_t __ret;
54270  __ret = (int64_t) __builtin_neon_vpaddd_s64((int8x16_t)__rev0);
54271  return __ret;
54272}
54273#endif
54274
54275#ifdef __LITTLE_ENDIAN__
54276__ai float32_t vpadds_f32(float32x2_t __p0) {
54277  float32_t __ret;
54278  __ret = (float32_t) __builtin_neon_vpadds_f32((int8x8_t)__p0);
54279  return __ret;
54280}
54281#else
54282__ai float32_t vpadds_f32(float32x2_t __p0) {
54283  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54284  float32_t __ret;
54285  __ret = (float32_t) __builtin_neon_vpadds_f32((int8x8_t)__rev0);
54286  return __ret;
54287}
54288#endif
54289
54290#ifdef __LITTLE_ENDIAN__
54291__ai uint8x16_t vpmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
54292  uint8x16_t __ret;
54293  __ret = (uint8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
54294  return __ret;
54295}
54296#else
54297__ai uint8x16_t vpmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
54298  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54299  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54300  uint8x16_t __ret;
54301  __ret = (uint8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
54302  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54303  return __ret;
54304}
54305#endif
54306
54307#ifdef __LITTLE_ENDIAN__
54308__ai uint32x4_t vpmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
54309  uint32x4_t __ret;
54310  __ret = (uint32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
54311  return __ret;
54312}
54313#else
54314__ai uint32x4_t vpmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
54315  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54316  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54317  uint32x4_t __ret;
54318  __ret = (uint32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
54319  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54320  return __ret;
54321}
54322#endif
54323
54324#ifdef __LITTLE_ENDIAN__
54325__ai uint16x8_t vpmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
54326  uint16x8_t __ret;
54327  __ret = (uint16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
54328  return __ret;
54329}
54330#else
54331__ai uint16x8_t vpmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
54332  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54333  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54334  uint16x8_t __ret;
54335  __ret = (uint16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
54336  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54337  return __ret;
54338}
54339#endif
54340
54341#ifdef __LITTLE_ENDIAN__
54342__ai int8x16_t vpmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
54343  int8x16_t __ret;
54344  __ret = (int8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
54345  return __ret;
54346}
54347#else
54348__ai int8x16_t vpmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
54349  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54350  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54351  int8x16_t __ret;
54352  __ret = (int8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
54353  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54354  return __ret;
54355}
54356#endif
54357
54358#ifdef __LITTLE_ENDIAN__
54359__ai float64x2_t vpmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
54360  float64x2_t __ret;
54361  __ret = (float64x2_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
54362  return __ret;
54363}
54364#else
54365__ai float64x2_t vpmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
54366  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54367  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54368  float64x2_t __ret;
54369  __ret = (float64x2_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
54370  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54371  return __ret;
54372}
54373#endif
54374
54375#ifdef __LITTLE_ENDIAN__
54376__ai float32x4_t vpmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
54377  float32x4_t __ret;
54378  __ret = (float32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
54379  return __ret;
54380}
54381#else
54382__ai float32x4_t vpmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
54383  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54384  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54385  float32x4_t __ret;
54386  __ret = (float32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
54387  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54388  return __ret;
54389}
54390#endif
54391
54392#ifdef __LITTLE_ENDIAN__
54393__ai int32x4_t vpmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
54394  int32x4_t __ret;
54395  __ret = (int32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
54396  return __ret;
54397}
54398#else
54399__ai int32x4_t vpmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
54400  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54401  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54402  int32x4_t __ret;
54403  __ret = (int32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
54404  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54405  return __ret;
54406}
54407#endif
54408
54409#ifdef __LITTLE_ENDIAN__
54410__ai int16x8_t vpmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
54411  int16x8_t __ret;
54412  __ret = (int16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
54413  return __ret;
54414}
54415#else
54416__ai int16x8_t vpmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
54417  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54418  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54419  int16x8_t __ret;
54420  __ret = (int16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
54421  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54422  return __ret;
54423}
54424#endif
54425
54426#ifdef __LITTLE_ENDIAN__
54427__ai float64_t vpmaxqd_f64(float64x2_t __p0) {
54428  float64_t __ret;
54429  __ret = (float64_t) __builtin_neon_vpmaxqd_f64((int8x16_t)__p0);
54430  return __ret;
54431}
54432#else
54433__ai float64_t vpmaxqd_f64(float64x2_t __p0) {
54434  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54435  float64_t __ret;
54436  __ret = (float64_t) __builtin_neon_vpmaxqd_f64((int8x16_t)__rev0);
54437  return __ret;
54438}
54439#endif
54440
54441#ifdef __LITTLE_ENDIAN__
54442__ai float32_t vpmaxs_f32(float32x2_t __p0) {
54443  float32_t __ret;
54444  __ret = (float32_t) __builtin_neon_vpmaxs_f32((int8x8_t)__p0);
54445  return __ret;
54446}
54447#else
54448__ai float32_t vpmaxs_f32(float32x2_t __p0) {
54449  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54450  float32_t __ret;
54451  __ret = (float32_t) __builtin_neon_vpmaxs_f32((int8x8_t)__rev0);
54452  return __ret;
54453}
54454#endif
54455
54456#ifdef __LITTLE_ENDIAN__
54457__ai float64x2_t vpmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
54458  float64x2_t __ret;
54459  __ret = (float64x2_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
54460  return __ret;
54461}
54462#else
54463__ai float64x2_t vpmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
54464  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54465  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54466  float64x2_t __ret;
54467  __ret = (float64x2_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
54468  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54469  return __ret;
54470}
54471#endif
54472
54473#ifdef __LITTLE_ENDIAN__
54474__ai float32x4_t vpmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
54475  float32x4_t __ret;
54476  __ret = (float32x4_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
54477  return __ret;
54478}
54479#else
54480__ai float32x4_t vpmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
54481  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54482  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54483  float32x4_t __ret;
54484  __ret = (float32x4_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
54485  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54486  return __ret;
54487}
54488#endif
54489
54490#ifdef __LITTLE_ENDIAN__
54491__ai float32x2_t vpmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
54492  float32x2_t __ret;
54493  __ret = (float32x2_t) __builtin_neon_vpmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
54494  return __ret;
54495}
54496#else
54497__ai float32x2_t vpmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
54498  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54499  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54500  float32x2_t __ret;
54501  __ret = (float32x2_t) __builtin_neon_vpmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
54502  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54503  return __ret;
54504}
54505#endif
54506
54507#ifdef __LITTLE_ENDIAN__
54508__ai float64_t vpmaxnmqd_f64(float64x2_t __p0) {
54509  float64_t __ret;
54510  __ret = (float64_t) __builtin_neon_vpmaxnmqd_f64((int8x16_t)__p0);
54511  return __ret;
54512}
54513#else
54514__ai float64_t vpmaxnmqd_f64(float64x2_t __p0) {
54515  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54516  float64_t __ret;
54517  __ret = (float64_t) __builtin_neon_vpmaxnmqd_f64((int8x16_t)__rev0);
54518  return __ret;
54519}
54520#endif
54521
54522#ifdef __LITTLE_ENDIAN__
54523__ai float32_t vpmaxnms_f32(float32x2_t __p0) {
54524  float32_t __ret;
54525  __ret = (float32_t) __builtin_neon_vpmaxnms_f32((int8x8_t)__p0);
54526  return __ret;
54527}
54528#else
54529__ai float32_t vpmaxnms_f32(float32x2_t __p0) {
54530  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54531  float32_t __ret;
54532  __ret = (float32_t) __builtin_neon_vpmaxnms_f32((int8x8_t)__rev0);
54533  return __ret;
54534}
54535#endif
54536
54537#ifdef __LITTLE_ENDIAN__
54538__ai uint8x16_t vpminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
54539  uint8x16_t __ret;
54540  __ret = (uint8x16_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
54541  return __ret;
54542}
54543#else
54544__ai uint8x16_t vpminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
54545  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54546  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54547  uint8x16_t __ret;
54548  __ret = (uint8x16_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
54549  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54550  return __ret;
54551}
54552#endif
54553
54554#ifdef __LITTLE_ENDIAN__
54555__ai uint32x4_t vpminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
54556  uint32x4_t __ret;
54557  __ret = (uint32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
54558  return __ret;
54559}
54560#else
54561__ai uint32x4_t vpminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
54562  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54563  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54564  uint32x4_t __ret;
54565  __ret = (uint32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
54566  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54567  return __ret;
54568}
54569#endif
54570
54571#ifdef __LITTLE_ENDIAN__
54572__ai uint16x8_t vpminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
54573  uint16x8_t __ret;
54574  __ret = (uint16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
54575  return __ret;
54576}
54577#else
54578__ai uint16x8_t vpminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
54579  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54580  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54581  uint16x8_t __ret;
54582  __ret = (uint16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
54583  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54584  return __ret;
54585}
54586#endif
54587
54588#ifdef __LITTLE_ENDIAN__
54589__ai int8x16_t vpminq_s8(int8x16_t __p0, int8x16_t __p1) {
54590  int8x16_t __ret;
54591  __ret = (int8x16_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
54592  return __ret;
54593}
54594#else
54595__ai int8x16_t vpminq_s8(int8x16_t __p0, int8x16_t __p1) {
54596  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54597  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54598  int8x16_t __ret;
54599  __ret = (int8x16_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
54600  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54601  return __ret;
54602}
54603#endif
54604
54605#ifdef __LITTLE_ENDIAN__
54606__ai float64x2_t vpminq_f64(float64x2_t __p0, float64x2_t __p1) {
54607  float64x2_t __ret;
54608  __ret = (float64x2_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
54609  return __ret;
54610}
54611#else
54612__ai float64x2_t vpminq_f64(float64x2_t __p0, float64x2_t __p1) {
54613  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54614  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54615  float64x2_t __ret;
54616  __ret = (float64x2_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
54617  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54618  return __ret;
54619}
54620#endif
54621
54622#ifdef __LITTLE_ENDIAN__
54623__ai float32x4_t vpminq_f32(float32x4_t __p0, float32x4_t __p1) {
54624  float32x4_t __ret;
54625  __ret = (float32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
54626  return __ret;
54627}
54628#else
54629__ai float32x4_t vpminq_f32(float32x4_t __p0, float32x4_t __p1) {
54630  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54631  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54632  float32x4_t __ret;
54633  __ret = (float32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
54634  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54635  return __ret;
54636}
54637#endif
54638
54639#ifdef __LITTLE_ENDIAN__
54640__ai int32x4_t vpminq_s32(int32x4_t __p0, int32x4_t __p1) {
54641  int32x4_t __ret;
54642  __ret = (int32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
54643  return __ret;
54644}
54645#else
54646__ai int32x4_t vpminq_s32(int32x4_t __p0, int32x4_t __p1) {
54647  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54648  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54649  int32x4_t __ret;
54650  __ret = (int32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
54651  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54652  return __ret;
54653}
54654#endif
54655
54656#ifdef __LITTLE_ENDIAN__
54657__ai int16x8_t vpminq_s16(int16x8_t __p0, int16x8_t __p1) {
54658  int16x8_t __ret;
54659  __ret = (int16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
54660  return __ret;
54661}
54662#else
54663__ai int16x8_t vpminq_s16(int16x8_t __p0, int16x8_t __p1) {
54664  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54665  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54666  int16x8_t __ret;
54667  __ret = (int16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
54668  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54669  return __ret;
54670}
54671#endif
54672
54673#ifdef __LITTLE_ENDIAN__
54674__ai float64_t vpminqd_f64(float64x2_t __p0) {
54675  float64_t __ret;
54676  __ret = (float64_t) __builtin_neon_vpminqd_f64((int8x16_t)__p0);
54677  return __ret;
54678}
54679#else
54680__ai float64_t vpminqd_f64(float64x2_t __p0) {
54681  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54682  float64_t __ret;
54683  __ret = (float64_t) __builtin_neon_vpminqd_f64((int8x16_t)__rev0);
54684  return __ret;
54685}
54686#endif
54687
54688#ifdef __LITTLE_ENDIAN__
54689__ai float32_t vpmins_f32(float32x2_t __p0) {
54690  float32_t __ret;
54691  __ret = (float32_t) __builtin_neon_vpmins_f32((int8x8_t)__p0);
54692  return __ret;
54693}
54694#else
54695__ai float32_t vpmins_f32(float32x2_t __p0) {
54696  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54697  float32_t __ret;
54698  __ret = (float32_t) __builtin_neon_vpmins_f32((int8x8_t)__rev0);
54699  return __ret;
54700}
54701#endif
54702
54703#ifdef __LITTLE_ENDIAN__
54704__ai float64x2_t vpminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
54705  float64x2_t __ret;
54706  __ret = (float64x2_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
54707  return __ret;
54708}
54709#else
54710__ai float64x2_t vpminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
54711  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54712  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54713  float64x2_t __ret;
54714  __ret = (float64x2_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
54715  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54716  return __ret;
54717}
54718#endif
54719
54720#ifdef __LITTLE_ENDIAN__
54721__ai float32x4_t vpminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
54722  float32x4_t __ret;
54723  __ret = (float32x4_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
54724  return __ret;
54725}
54726#else
54727__ai float32x4_t vpminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
54728  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54729  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54730  float32x4_t __ret;
54731  __ret = (float32x4_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
54732  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54733  return __ret;
54734}
54735#endif
54736
54737#ifdef __LITTLE_ENDIAN__
54738__ai float32x2_t vpminnm_f32(float32x2_t __p0, float32x2_t __p1) {
54739  float32x2_t __ret;
54740  __ret = (float32x2_t) __builtin_neon_vpminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
54741  return __ret;
54742}
54743#else
54744__ai float32x2_t vpminnm_f32(float32x2_t __p0, float32x2_t __p1) {
54745  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54746  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54747  float32x2_t __ret;
54748  __ret = (float32x2_t) __builtin_neon_vpminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
54749  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54750  return __ret;
54751}
54752#endif
54753
54754#ifdef __LITTLE_ENDIAN__
54755__ai float64_t vpminnmqd_f64(float64x2_t __p0) {
54756  float64_t __ret;
54757  __ret = (float64_t) __builtin_neon_vpminnmqd_f64((int8x16_t)__p0);
54758  return __ret;
54759}
54760#else
54761__ai float64_t vpminnmqd_f64(float64x2_t __p0) {
54762  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54763  float64_t __ret;
54764  __ret = (float64_t) __builtin_neon_vpminnmqd_f64((int8x16_t)__rev0);
54765  return __ret;
54766}
54767#endif
54768
54769#ifdef __LITTLE_ENDIAN__
54770__ai float32_t vpminnms_f32(float32x2_t __p0) {
54771  float32_t __ret;
54772  __ret = (float32_t) __builtin_neon_vpminnms_f32((int8x8_t)__p0);
54773  return __ret;
54774}
54775#else
54776__ai float32_t vpminnms_f32(float32x2_t __p0) {
54777  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54778  float32_t __ret;
54779  __ret = (float32_t) __builtin_neon_vpminnms_f32((int8x8_t)__rev0);
54780  return __ret;
54781}
54782#endif
54783
54784#ifdef __LITTLE_ENDIAN__
54785__ai int64x2_t vqabsq_s64(int64x2_t __p0) {
54786  int64x2_t __ret;
54787  __ret = (int64x2_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 35);
54788  return __ret;
54789}
54790#else
54791__ai int64x2_t vqabsq_s64(int64x2_t __p0) {
54792  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54793  int64x2_t __ret;
54794  __ret = (int64x2_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 35);
54795  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54796  return __ret;
54797}
54798#endif
54799
54800#ifdef __LITTLE_ENDIAN__
54801__ai int64x1_t vqabs_s64(int64x1_t __p0) {
54802  int64x1_t __ret;
54803  __ret = (int64x1_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 3);
54804  return __ret;
54805}
54806#else
54807__ai int64x1_t vqabs_s64(int64x1_t __p0) {
54808  int64x1_t __ret;
54809  __ret = (int64x1_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 3);
54810  return __ret;
54811}
54812#endif
54813
54814#ifdef __LITTLE_ENDIAN__
54815__ai int8_t vqabsb_s8(int8_t __p0) {
54816  int8_t __ret;
54817  __ret = (int8_t) __builtin_neon_vqabsb_s8(__p0);
54818  return __ret;
54819}
54820#else
54821__ai int8_t vqabsb_s8(int8_t __p0) {
54822  int8_t __ret;
54823  __ret = (int8_t) __builtin_neon_vqabsb_s8(__p0);
54824  return __ret;
54825}
54826#endif
54827
54828#ifdef __LITTLE_ENDIAN__
54829__ai int32_t vqabss_s32(int32_t __p0) {
54830  int32_t __ret;
54831  __ret = (int32_t) __builtin_neon_vqabss_s32(__p0);
54832  return __ret;
54833}
54834#else
54835__ai int32_t vqabss_s32(int32_t __p0) {
54836  int32_t __ret;
54837  __ret = (int32_t) __builtin_neon_vqabss_s32(__p0);
54838  return __ret;
54839}
54840#endif
54841
54842#ifdef __LITTLE_ENDIAN__
54843__ai int64_t vqabsd_s64(int64_t __p0) {
54844  int64_t __ret;
54845  __ret = (int64_t) __builtin_neon_vqabsd_s64(__p0);
54846  return __ret;
54847}
54848#else
54849__ai int64_t vqabsd_s64(int64_t __p0) {
54850  int64_t __ret;
54851  __ret = (int64_t) __builtin_neon_vqabsd_s64(__p0);
54852  return __ret;
54853}
54854#endif
54855
54856#ifdef __LITTLE_ENDIAN__
54857__ai int16_t vqabsh_s16(int16_t __p0) {
54858  int16_t __ret;
54859  __ret = (int16_t) __builtin_neon_vqabsh_s16(__p0);
54860  return __ret;
54861}
54862#else
54863__ai int16_t vqabsh_s16(int16_t __p0) {
54864  int16_t __ret;
54865  __ret = (int16_t) __builtin_neon_vqabsh_s16(__p0);
54866  return __ret;
54867}
54868#endif
54869
54870#ifdef __LITTLE_ENDIAN__
54871__ai uint8_t vqaddb_u8(uint8_t __p0, uint8_t __p1) {
54872  uint8_t __ret;
54873  __ret = (uint8_t) __builtin_neon_vqaddb_u8(__p0, __p1);
54874  return __ret;
54875}
54876#else
54877__ai uint8_t vqaddb_u8(uint8_t __p0, uint8_t __p1) {
54878  uint8_t __ret;
54879  __ret = (uint8_t) __builtin_neon_vqaddb_u8(__p0, __p1);
54880  return __ret;
54881}
54882#endif
54883
54884#ifdef __LITTLE_ENDIAN__
54885__ai uint32_t vqadds_u32(uint32_t __p0, uint32_t __p1) {
54886  uint32_t __ret;
54887  __ret = (uint32_t) __builtin_neon_vqadds_u32(__p0, __p1);
54888  return __ret;
54889}
54890#else
54891__ai uint32_t vqadds_u32(uint32_t __p0, uint32_t __p1) {
54892  uint32_t __ret;
54893  __ret = (uint32_t) __builtin_neon_vqadds_u32(__p0, __p1);
54894  return __ret;
54895}
54896#endif
54897
54898#ifdef __LITTLE_ENDIAN__
54899__ai uint64_t vqaddd_u64(uint64_t __p0, uint64_t __p1) {
54900  uint64_t __ret;
54901  __ret = (uint64_t) __builtin_neon_vqaddd_u64(__p0, __p1);
54902  return __ret;
54903}
54904#else
54905__ai uint64_t vqaddd_u64(uint64_t __p0, uint64_t __p1) {
54906  uint64_t __ret;
54907  __ret = (uint64_t) __builtin_neon_vqaddd_u64(__p0, __p1);
54908  return __ret;
54909}
54910#endif
54911
54912#ifdef __LITTLE_ENDIAN__
54913__ai uint16_t vqaddh_u16(uint16_t __p0, uint16_t __p1) {
54914  uint16_t __ret;
54915  __ret = (uint16_t) __builtin_neon_vqaddh_u16(__p0, __p1);
54916  return __ret;
54917}
54918#else
54919__ai uint16_t vqaddh_u16(uint16_t __p0, uint16_t __p1) {
54920  uint16_t __ret;
54921  __ret = (uint16_t) __builtin_neon_vqaddh_u16(__p0, __p1);
54922  return __ret;
54923}
54924#endif
54925
54926#ifdef __LITTLE_ENDIAN__
54927__ai int8_t vqaddb_s8(int8_t __p0, int8_t __p1) {
54928  int8_t __ret;
54929  __ret = (int8_t) __builtin_neon_vqaddb_s8(__p0, __p1);
54930  return __ret;
54931}
54932#else
54933__ai int8_t vqaddb_s8(int8_t __p0, int8_t __p1) {
54934  int8_t __ret;
54935  __ret = (int8_t) __builtin_neon_vqaddb_s8(__p0, __p1);
54936  return __ret;
54937}
54938#endif
54939
54940#ifdef __LITTLE_ENDIAN__
54941__ai int32_t vqadds_s32(int32_t __p0, int32_t __p1) {
54942  int32_t __ret;
54943  __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
54944  return __ret;
54945}
54946#else
54947__ai int32_t vqadds_s32(int32_t __p0, int32_t __p1) {
54948  int32_t __ret;
54949  __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
54950  return __ret;
54951}
54952#endif
54953
54954#ifdef __LITTLE_ENDIAN__
54955__ai int64_t vqaddd_s64(int64_t __p0, int64_t __p1) {
54956  int64_t __ret;
54957  __ret = (int64_t) __builtin_neon_vqaddd_s64(__p0, __p1);
54958  return __ret;
54959}
54960#else
54961__ai int64_t vqaddd_s64(int64_t __p0, int64_t __p1) {
54962  int64_t __ret;
54963  __ret = (int64_t) __builtin_neon_vqaddd_s64(__p0, __p1);
54964  return __ret;
54965}
54966#endif
54967
54968#ifdef __LITTLE_ENDIAN__
54969__ai int16_t vqaddh_s16(int16_t __p0, int16_t __p1) {
54970  int16_t __ret;
54971  __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
54972  return __ret;
54973}
54974#else
54975__ai int16_t vqaddh_s16(int16_t __p0, int16_t __p1) {
54976  int16_t __ret;
54977  __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
54978  return __ret;
54979}
54980#endif
54981
54982#ifdef __LITTLE_ENDIAN__
54983__ai int64_t vqdmlals_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
54984  int64_t __ret;
54985  __ret = (int64_t) __builtin_neon_vqdmlals_s32(__p0, __p1, __p2);
54986  return __ret;
54987}
54988#else
54989__ai int64_t vqdmlals_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
54990  int64_t __ret;
54991  __ret = (int64_t) __builtin_neon_vqdmlals_s32(__p0, __p1, __p2);
54992  return __ret;
54993}
54994#endif
54995
54996#ifdef __LITTLE_ENDIAN__
54997__ai int32_t vqdmlalh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
54998  int32_t __ret;
54999  __ret = (int32_t) __builtin_neon_vqdmlalh_s16(__p0, __p1, __p2);
55000  return __ret;
55001}
55002#else
55003__ai int32_t vqdmlalh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
55004  int32_t __ret;
55005  __ret = (int32_t) __builtin_neon_vqdmlalh_s16(__p0, __p1, __p2);
55006  return __ret;
55007}
55008#endif
55009
55010#ifdef __LITTLE_ENDIAN__
55011__ai int64x2_t vqdmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
55012  int64x2_t __ret;
55013  __ret = vqdmlal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
55014  return __ret;
55015}
55016#else
55017__ai int64x2_t vqdmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
55018  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55019  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55020  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
55021  int64x2_t __ret;
55022  __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
55023  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55024  return __ret;
55025}
55026#endif
55027
55028#ifdef __LITTLE_ENDIAN__
55029__ai int32x4_t vqdmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
55030  int32x4_t __ret;
55031  __ret = vqdmlal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
55032  return __ret;
55033}
55034#else
55035__ai int32x4_t vqdmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
55036  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55037  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55038  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55039  int32x4_t __ret;
55040  __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
55041  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55042  return __ret;
55043}
55044#endif
55045
55046#ifdef __LITTLE_ENDIAN__
55047#define vqdmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55048  int64x2_t __s0 = __p0; \
55049  int32x4_t __s1 = __p1; \
55050  int32x2_t __s2 = __p2; \
55051  int64x2_t __ret; \
55052  __ret = vqdmlal_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
55053  __ret; \
55054})
55055#else
55056#define vqdmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55057  int64x2_t __s0 = __p0; \
55058  int32x4_t __s1 = __p1; \
55059  int32x2_t __s2 = __p2; \
55060  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
55061  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
55062  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
55063  int64x2_t __ret; \
55064  __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
55065  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
55066  __ret; \
55067})
55068#endif
55069
55070#ifdef __LITTLE_ENDIAN__
55071#define vqdmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55072  int32x4_t __s0 = __p0; \
55073  int16x8_t __s1 = __p1; \
55074  int16x4_t __s2 = __p2; \
55075  int32x4_t __ret; \
55076  __ret = vqdmlal_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
55077  __ret; \
55078})
55079#else
55080#define vqdmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55081  int32x4_t __s0 = __p0; \
55082  int16x8_t __s1 = __p1; \
55083  int16x4_t __s2 = __p2; \
55084  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
55085  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
55086  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
55087  int32x4_t __ret; \
55088  __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
55089  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
55090  __ret; \
55091})
55092#endif
55093
55094#ifdef __LITTLE_ENDIAN__
55095#define vqdmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55096  int64x2_t __s0 = __p0; \
55097  int32x4_t __s1 = __p1; \
55098  int32x4_t __s2 = __p2; \
55099  int64x2_t __ret; \
55100  __ret = vqdmlal_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
55101  __ret; \
55102})
55103#else
55104#define vqdmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55105  int64x2_t __s0 = __p0; \
55106  int32x4_t __s1 = __p1; \
55107  int32x4_t __s2 = __p2; \
55108  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
55109  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
55110  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
55111  int64x2_t __ret; \
55112  __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
55113  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
55114  __ret; \
55115})
55116#endif
55117
55118#ifdef __LITTLE_ENDIAN__
55119#define vqdmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55120  int32x4_t __s0 = __p0; \
55121  int16x8_t __s1 = __p1; \
55122  int16x8_t __s2 = __p2; \
55123  int32x4_t __ret; \
55124  __ret = vqdmlal_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
55125  __ret; \
55126})
55127#else
55128#define vqdmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55129  int32x4_t __s0 = __p0; \
55130  int16x8_t __s1 = __p1; \
55131  int16x8_t __s2 = __p2; \
55132  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
55133  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
55134  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
55135  int32x4_t __ret; \
55136  __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
55137  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
55138  __ret; \
55139})
55140#endif
55141
55142#ifdef __LITTLE_ENDIAN__
55143__ai int64x2_t vqdmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
55144  int64x2_t __ret;
55145  __ret = vqdmlal_n_s32(__p0, vget_high_s32(__p1), __p2);
55146  return __ret;
55147}
55148#else
55149__ai int64x2_t vqdmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
55150  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55151  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55152  int64x2_t __ret;
55153  __ret = __noswap_vqdmlal_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
55154  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55155  return __ret;
55156}
55157#endif
55158
55159#ifdef __LITTLE_ENDIAN__
55160__ai int32x4_t vqdmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
55161  int32x4_t __ret;
55162  __ret = vqdmlal_n_s16(__p0, vget_high_s16(__p1), __p2);
55163  return __ret;
55164}
55165#else
55166__ai int32x4_t vqdmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
55167  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55168  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55169  int32x4_t __ret;
55170  __ret = __noswap_vqdmlal_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
55171  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55172  return __ret;
55173}
55174#endif
55175
55176#ifdef __LITTLE_ENDIAN__
55177#define vqdmlals_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55178  int64_t __s0 = __p0; \
55179  int32_t __s1 = __p1; \
55180  int32x2_t __s2 = __p2; \
55181  int64_t __ret; \
55182  __ret = (int64_t) __builtin_neon_vqdmlals_lane_s32(__s0, __s1, (int8x8_t)__s2, __p3); \
55183  __ret; \
55184})
55185#else
55186#define vqdmlals_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55187  int64_t __s0 = __p0; \
55188  int32_t __s1 = __p1; \
55189  int32x2_t __s2 = __p2; \
55190  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
55191  int64_t __ret; \
55192  __ret = (int64_t) __builtin_neon_vqdmlals_lane_s32(__s0, __s1, (int8x8_t)__rev2, __p3); \
55193  __ret; \
55194})
55195#endif
55196
55197#ifdef __LITTLE_ENDIAN__
55198#define vqdmlalh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55199  int32_t __s0 = __p0; \
55200  int16_t __s1 = __p1; \
55201  int16x4_t __s2 = __p2; \
55202  int32_t __ret; \
55203  __ret = (int32_t) __builtin_neon_vqdmlalh_lane_s16(__s0, __s1, (int8x8_t)__s2, __p3); \
55204  __ret; \
55205})
55206#else
55207#define vqdmlalh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55208  int32_t __s0 = __p0; \
55209  int16_t __s1 = __p1; \
55210  int16x4_t __s2 = __p2; \
55211  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
55212  int32_t __ret; \
55213  __ret = (int32_t) __builtin_neon_vqdmlalh_lane_s16(__s0, __s1, (int8x8_t)__rev2, __p3); \
55214  __ret; \
55215})
55216#endif
55217
55218#ifdef __LITTLE_ENDIAN__
55219#define vqdmlals_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55220  int64_t __s0 = __p0; \
55221  int32_t __s1 = __p1; \
55222  int32x4_t __s2 = __p2; \
55223  int64_t __ret; \
55224  __ret = (int64_t) __builtin_neon_vqdmlals_laneq_s32(__s0, __s1, (int8x16_t)__s2, __p3); \
55225  __ret; \
55226})
55227#else
55228#define vqdmlals_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55229  int64_t __s0 = __p0; \
55230  int32_t __s1 = __p1; \
55231  int32x4_t __s2 = __p2; \
55232  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
55233  int64_t __ret; \
55234  __ret = (int64_t) __builtin_neon_vqdmlals_laneq_s32(__s0, __s1, (int8x16_t)__rev2, __p3); \
55235  __ret; \
55236})
55237#endif
55238
55239#ifdef __LITTLE_ENDIAN__
55240#define vqdmlalh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55241  int32_t __s0 = __p0; \
55242  int16_t __s1 = __p1; \
55243  int16x8_t __s2 = __p2; \
55244  int32_t __ret; \
55245  __ret = (int32_t) __builtin_neon_vqdmlalh_laneq_s16(__s0, __s1, (int8x16_t)__s2, __p3); \
55246  __ret; \
55247})
55248#else
55249#define vqdmlalh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55250  int32_t __s0 = __p0; \
55251  int16_t __s1 = __p1; \
55252  int16x8_t __s2 = __p2; \
55253  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
55254  int32_t __ret; \
55255  __ret = (int32_t) __builtin_neon_vqdmlalh_laneq_s16(__s0, __s1, (int8x16_t)__rev2, __p3); \
55256  __ret; \
55257})
55258#endif
55259
55260#ifdef __LITTLE_ENDIAN__
55261#define vqdmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55262  int64x2_t __s0 = __p0; \
55263  int32x2_t __s1 = __p1; \
55264  int32x4_t __s2 = __p2; \
55265  int64x2_t __ret; \
55266  __ret = vqdmlal_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
55267  __ret; \
55268})
55269#else
55270#define vqdmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55271  int64x2_t __s0 = __p0; \
55272  int32x2_t __s1 = __p1; \
55273  int32x4_t __s2 = __p2; \
55274  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
55275  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
55276  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
55277  int64x2_t __ret; \
55278  __ret = __noswap_vqdmlal_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
55279  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
55280  __ret; \
55281})
55282#endif
55283
55284#ifdef __LITTLE_ENDIAN__
55285#define vqdmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55286  int32x4_t __s0 = __p0; \
55287  int16x4_t __s1 = __p1; \
55288  int16x8_t __s2 = __p2; \
55289  int32x4_t __ret; \
55290  __ret = vqdmlal_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
55291  __ret; \
55292})
55293#else
55294#define vqdmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55295  int32x4_t __s0 = __p0; \
55296  int16x4_t __s1 = __p1; \
55297  int16x8_t __s2 = __p2; \
55298  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
55299  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
55300  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
55301  int32x4_t __ret; \
55302  __ret = __noswap_vqdmlal_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
55303  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
55304  __ret; \
55305})
55306#endif
55307
55308#ifdef __LITTLE_ENDIAN__
55309__ai int64_t vqdmlsls_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
55310  int64_t __ret;
55311  __ret = (int64_t) __builtin_neon_vqdmlsls_s32(__p0, __p1, __p2);
55312  return __ret;
55313}
55314#else
55315__ai int64_t vqdmlsls_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
55316  int64_t __ret;
55317  __ret = (int64_t) __builtin_neon_vqdmlsls_s32(__p0, __p1, __p2);
55318  return __ret;
55319}
55320#endif
55321
55322#ifdef __LITTLE_ENDIAN__
55323__ai int32_t vqdmlslh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
55324  int32_t __ret;
55325  __ret = (int32_t) __builtin_neon_vqdmlslh_s16(__p0, __p1, __p2);
55326  return __ret;
55327}
55328#else
55329__ai int32_t vqdmlslh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
55330  int32_t __ret;
55331  __ret = (int32_t) __builtin_neon_vqdmlslh_s16(__p0, __p1, __p2);
55332  return __ret;
55333}
55334#endif
55335
55336#ifdef __LITTLE_ENDIAN__
55337__ai int64x2_t vqdmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
55338  int64x2_t __ret;
55339  __ret = vqdmlsl_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
55340  return __ret;
55341}
55342#else
55343__ai int64x2_t vqdmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
55344  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55345  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55346  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
55347  int64x2_t __ret;
55348  __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
55349  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55350  return __ret;
55351}
55352#endif
55353
55354#ifdef __LITTLE_ENDIAN__
55355__ai int32x4_t vqdmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
55356  int32x4_t __ret;
55357  __ret = vqdmlsl_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
55358  return __ret;
55359}
55360#else
55361__ai int32x4_t vqdmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
55362  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55363  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55364  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
55365  int32x4_t __ret;
55366  __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
55367  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55368  return __ret;
55369}
55370#endif
55371
55372#ifdef __LITTLE_ENDIAN__
55373#define vqdmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55374  int64x2_t __s0 = __p0; \
55375  int32x4_t __s1 = __p1; \
55376  int32x2_t __s2 = __p2; \
55377  int64x2_t __ret; \
55378  __ret = vqdmlsl_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
55379  __ret; \
55380})
55381#else
55382#define vqdmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55383  int64x2_t __s0 = __p0; \
55384  int32x4_t __s1 = __p1; \
55385  int32x2_t __s2 = __p2; \
55386  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
55387  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
55388  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
55389  int64x2_t __ret; \
55390  __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
55391  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
55392  __ret; \
55393})
55394#endif
55395
55396#ifdef __LITTLE_ENDIAN__
55397#define vqdmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55398  int32x4_t __s0 = __p0; \
55399  int16x8_t __s1 = __p1; \
55400  int16x4_t __s2 = __p2; \
55401  int32x4_t __ret; \
55402  __ret = vqdmlsl_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
55403  __ret; \
55404})
55405#else
55406#define vqdmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55407  int32x4_t __s0 = __p0; \
55408  int16x8_t __s1 = __p1; \
55409  int16x4_t __s2 = __p2; \
55410  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
55411  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
55412  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
55413  int32x4_t __ret; \
55414  __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
55415  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
55416  __ret; \
55417})
55418#endif
55419
55420#ifdef __LITTLE_ENDIAN__
55421#define vqdmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55422  int64x2_t __s0 = __p0; \
55423  int32x4_t __s1 = __p1; \
55424  int32x4_t __s2 = __p2; \
55425  int64x2_t __ret; \
55426  __ret = vqdmlsl_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
55427  __ret; \
55428})
55429#else
55430#define vqdmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55431  int64x2_t __s0 = __p0; \
55432  int32x4_t __s1 = __p1; \
55433  int32x4_t __s2 = __p2; \
55434  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
55435  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
55436  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
55437  int64x2_t __ret; \
55438  __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
55439  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
55440  __ret; \
55441})
55442#endif
55443
55444#ifdef __LITTLE_ENDIAN__
55445#define vqdmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55446  int32x4_t __s0 = __p0; \
55447  int16x8_t __s1 = __p1; \
55448  int16x8_t __s2 = __p2; \
55449  int32x4_t __ret; \
55450  __ret = vqdmlsl_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
55451  __ret; \
55452})
55453#else
55454#define vqdmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55455  int32x4_t __s0 = __p0; \
55456  int16x8_t __s1 = __p1; \
55457  int16x8_t __s2 = __p2; \
55458  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
55459  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
55460  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
55461  int32x4_t __ret; \
55462  __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
55463  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
55464  __ret; \
55465})
55466#endif
55467
55468#ifdef __LITTLE_ENDIAN__
55469__ai int64x2_t vqdmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
55470  int64x2_t __ret;
55471  __ret = vqdmlsl_n_s32(__p0, vget_high_s32(__p1), __p2);
55472  return __ret;
55473}
55474#else
55475__ai int64x2_t vqdmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
55476  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55477  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55478  int64x2_t __ret;
55479  __ret = __noswap_vqdmlsl_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
55480  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55481  return __ret;
55482}
55483#endif
55484
55485#ifdef __LITTLE_ENDIAN__
55486__ai int32x4_t vqdmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
55487  int32x4_t __ret;
55488  __ret = vqdmlsl_n_s16(__p0, vget_high_s16(__p1), __p2);
55489  return __ret;
55490}
55491#else
55492__ai int32x4_t vqdmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
55493  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55494  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55495  int32x4_t __ret;
55496  __ret = __noswap_vqdmlsl_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
55497  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55498  return __ret;
55499}
55500#endif
55501
55502#ifdef __LITTLE_ENDIAN__
55503#define vqdmlsls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55504  int64_t __s0 = __p0; \
55505  int32_t __s1 = __p1; \
55506  int32x2_t __s2 = __p2; \
55507  int64_t __ret; \
55508  __ret = (int64_t) __builtin_neon_vqdmlsls_lane_s32(__s0, __s1, (int8x8_t)__s2, __p3); \
55509  __ret; \
55510})
55511#else
55512#define vqdmlsls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55513  int64_t __s0 = __p0; \
55514  int32_t __s1 = __p1; \
55515  int32x2_t __s2 = __p2; \
55516  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
55517  int64_t __ret; \
55518  __ret = (int64_t) __builtin_neon_vqdmlsls_lane_s32(__s0, __s1, (int8x8_t)__rev2, __p3); \
55519  __ret; \
55520})
55521#endif
55522
55523#ifdef __LITTLE_ENDIAN__
55524#define vqdmlslh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55525  int32_t __s0 = __p0; \
55526  int16_t __s1 = __p1; \
55527  int16x4_t __s2 = __p2; \
55528  int32_t __ret; \
55529  __ret = (int32_t) __builtin_neon_vqdmlslh_lane_s16(__s0, __s1, (int8x8_t)__s2, __p3); \
55530  __ret; \
55531})
55532#else
55533#define vqdmlslh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55534  int32_t __s0 = __p0; \
55535  int16_t __s1 = __p1; \
55536  int16x4_t __s2 = __p2; \
55537  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
55538  int32_t __ret; \
55539  __ret = (int32_t) __builtin_neon_vqdmlslh_lane_s16(__s0, __s1, (int8x8_t)__rev2, __p3); \
55540  __ret; \
55541})
55542#endif
55543
55544#ifdef __LITTLE_ENDIAN__
55545#define vqdmlsls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55546  int64_t __s0 = __p0; \
55547  int32_t __s1 = __p1; \
55548  int32x4_t __s2 = __p2; \
55549  int64_t __ret; \
55550  __ret = (int64_t) __builtin_neon_vqdmlsls_laneq_s32(__s0, __s1, (int8x16_t)__s2, __p3); \
55551  __ret; \
55552})
55553#else
55554#define vqdmlsls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55555  int64_t __s0 = __p0; \
55556  int32_t __s1 = __p1; \
55557  int32x4_t __s2 = __p2; \
55558  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
55559  int64_t __ret; \
55560  __ret = (int64_t) __builtin_neon_vqdmlsls_laneq_s32(__s0, __s1, (int8x16_t)__rev2, __p3); \
55561  __ret; \
55562})
55563#endif
55564
55565#ifdef __LITTLE_ENDIAN__
55566#define vqdmlslh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55567  int32_t __s0 = __p0; \
55568  int16_t __s1 = __p1; \
55569  int16x8_t __s2 = __p2; \
55570  int32_t __ret; \
55571  __ret = (int32_t) __builtin_neon_vqdmlslh_laneq_s16(__s0, __s1, (int8x16_t)__s2, __p3); \
55572  __ret; \
55573})
55574#else
55575#define vqdmlslh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55576  int32_t __s0 = __p0; \
55577  int16_t __s1 = __p1; \
55578  int16x8_t __s2 = __p2; \
55579  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
55580  int32_t __ret; \
55581  __ret = (int32_t) __builtin_neon_vqdmlslh_laneq_s16(__s0, __s1, (int8x16_t)__rev2, __p3); \
55582  __ret; \
55583})
55584#endif
55585
55586#ifdef __LITTLE_ENDIAN__
55587#define vqdmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55588  int64x2_t __s0 = __p0; \
55589  int32x2_t __s1 = __p1; \
55590  int32x4_t __s2 = __p2; \
55591  int64x2_t __ret; \
55592  __ret = vqdmlsl_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
55593  __ret; \
55594})
55595#else
55596#define vqdmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
55597  int64x2_t __s0 = __p0; \
55598  int32x2_t __s1 = __p1; \
55599  int32x4_t __s2 = __p2; \
55600  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
55601  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
55602  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
55603  int64x2_t __ret; \
55604  __ret = __noswap_vqdmlsl_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
55605  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
55606  __ret; \
55607})
55608#endif
55609
55610#ifdef __LITTLE_ENDIAN__
55611#define vqdmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55612  int32x4_t __s0 = __p0; \
55613  int16x4_t __s1 = __p1; \
55614  int16x8_t __s2 = __p2; \
55615  int32x4_t __ret; \
55616  __ret = vqdmlsl_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
55617  __ret; \
55618})
55619#else
55620#define vqdmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
55621  int32x4_t __s0 = __p0; \
55622  int16x4_t __s1 = __p1; \
55623  int16x8_t __s2 = __p2; \
55624  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
55625  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
55626  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
55627  int32x4_t __ret; \
55628  __ret = __noswap_vqdmlsl_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
55629  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
55630  __ret; \
55631})
55632#endif
55633
55634#ifdef __LITTLE_ENDIAN__
55635__ai int32_t vqdmulhs_s32(int32_t __p0, int32_t __p1) {
55636  int32_t __ret;
55637  __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
55638  return __ret;
55639}
55640#else
55641__ai int32_t vqdmulhs_s32(int32_t __p0, int32_t __p1) {
55642  int32_t __ret;
55643  __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
55644  return __ret;
55645}
55646__ai int32_t __noswap_vqdmulhs_s32(int32_t __p0, int32_t __p1) {
55647  int32_t __ret;
55648  __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
55649  return __ret;
55650}
55651#endif
55652
55653#ifdef __LITTLE_ENDIAN__
55654__ai int16_t vqdmulhh_s16(int16_t __p0, int16_t __p1) {
55655  int16_t __ret;
55656  __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
55657  return __ret;
55658}
55659#else
55660__ai int16_t vqdmulhh_s16(int16_t __p0, int16_t __p1) {
55661  int16_t __ret;
55662  __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
55663  return __ret;
55664}
55665__ai int16_t __noswap_vqdmulhh_s16(int16_t __p0, int16_t __p1) {
55666  int16_t __ret;
55667  __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
55668  return __ret;
55669}
55670#endif
55671
55672#ifdef __LITTLE_ENDIAN__
55673#define vqdmulhs_lane_s32(__p0_150, __p1_150, __p2_150) __extension__ ({ \
55674  int32_t __s0_150 = __p0_150; \
55675  int32x2_t __s1_150 = __p1_150; \
55676  int32_t __ret_150; \
55677  __ret_150 = vqdmulhs_s32(__s0_150, vget_lane_s32(__s1_150, __p2_150)); \
55678  __ret_150; \
55679})
55680#else
55681#define vqdmulhs_lane_s32(__p0_151, __p1_151, __p2_151) __extension__ ({ \
55682  int32_t __s0_151 = __p0_151; \
55683  int32x2_t __s1_151 = __p1_151; \
55684  int32x2_t __rev1_151;  __rev1_151 = __builtin_shufflevector(__s1_151, __s1_151, 1, 0); \
55685  int32_t __ret_151; \
55686  __ret_151 = __noswap_vqdmulhs_s32(__s0_151, __noswap_vget_lane_s32(__rev1_151, __p2_151)); \
55687  __ret_151; \
55688})
55689#endif
55690
55691#ifdef __LITTLE_ENDIAN__
55692#define vqdmulhh_lane_s16(__p0_152, __p1_152, __p2_152) __extension__ ({ \
55693  int16_t __s0_152 = __p0_152; \
55694  int16x4_t __s1_152 = __p1_152; \
55695  int16_t __ret_152; \
55696  __ret_152 = vqdmulhh_s16(__s0_152, vget_lane_s16(__s1_152, __p2_152)); \
55697  __ret_152; \
55698})
55699#else
55700#define vqdmulhh_lane_s16(__p0_153, __p1_153, __p2_153) __extension__ ({ \
55701  int16_t __s0_153 = __p0_153; \
55702  int16x4_t __s1_153 = __p1_153; \
55703  int16x4_t __rev1_153;  __rev1_153 = __builtin_shufflevector(__s1_153, __s1_153, 3, 2, 1, 0); \
55704  int16_t __ret_153; \
55705  __ret_153 = __noswap_vqdmulhh_s16(__s0_153, __noswap_vget_lane_s16(__rev1_153, __p2_153)); \
55706  __ret_153; \
55707})
55708#endif
55709
55710#ifdef __LITTLE_ENDIAN__
55711#define vqdmulhs_laneq_s32(__p0_154, __p1_154, __p2_154) __extension__ ({ \
55712  int32_t __s0_154 = __p0_154; \
55713  int32x4_t __s1_154 = __p1_154; \
55714  int32_t __ret_154; \
55715  __ret_154 = vqdmulhs_s32(__s0_154, vgetq_lane_s32(__s1_154, __p2_154)); \
55716  __ret_154; \
55717})
55718#else
55719#define vqdmulhs_laneq_s32(__p0_155, __p1_155, __p2_155) __extension__ ({ \
55720  int32_t __s0_155 = __p0_155; \
55721  int32x4_t __s1_155 = __p1_155; \
55722  int32x4_t __rev1_155;  __rev1_155 = __builtin_shufflevector(__s1_155, __s1_155, 3, 2, 1, 0); \
55723  int32_t __ret_155; \
55724  __ret_155 = __noswap_vqdmulhs_s32(__s0_155, __noswap_vgetq_lane_s32(__rev1_155, __p2_155)); \
55725  __ret_155; \
55726})
55727#endif
55728
55729#ifdef __LITTLE_ENDIAN__
55730#define vqdmulhh_laneq_s16(__p0_156, __p1_156, __p2_156) __extension__ ({ \
55731  int16_t __s0_156 = __p0_156; \
55732  int16x8_t __s1_156 = __p1_156; \
55733  int16_t __ret_156; \
55734  __ret_156 = vqdmulhh_s16(__s0_156, vgetq_lane_s16(__s1_156, __p2_156)); \
55735  __ret_156; \
55736})
55737#else
55738#define vqdmulhh_laneq_s16(__p0_157, __p1_157, __p2_157) __extension__ ({ \
55739  int16_t __s0_157 = __p0_157; \
55740  int16x8_t __s1_157 = __p1_157; \
55741  int16x8_t __rev1_157;  __rev1_157 = __builtin_shufflevector(__s1_157, __s1_157, 7, 6, 5, 4, 3, 2, 1, 0); \
55742  int16_t __ret_157; \
55743  __ret_157 = __noswap_vqdmulhh_s16(__s0_157, __noswap_vgetq_lane_s16(__rev1_157, __p2_157)); \
55744  __ret_157; \
55745})
55746#endif
55747
55748#ifdef __LITTLE_ENDIAN__
55749#define vqdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
55750  int32x4_t __s0 = __p0; \
55751  int32x4_t __s1 = __p1; \
55752  int32x4_t __ret; \
55753  __ret = vqdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
55754  __ret; \
55755})
55756#else
55757#define vqdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
55758  int32x4_t __s0 = __p0; \
55759  int32x4_t __s1 = __p1; \
55760  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
55761  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
55762  int32x4_t __ret; \
55763  __ret = __noswap_vqdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
55764  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
55765  __ret; \
55766})
55767#endif
55768
55769#ifdef __LITTLE_ENDIAN__
55770#define vqdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
55771  int16x8_t __s0 = __p0; \
55772  int16x8_t __s1 = __p1; \
55773  int16x8_t __ret; \
55774  __ret = vqdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
55775  __ret; \
55776})
55777#else
55778#define vqdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
55779  int16x8_t __s0 = __p0; \
55780  int16x8_t __s1 = __p1; \
55781  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
55782  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
55783  int16x8_t __ret; \
55784  __ret = __noswap_vqdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
55785  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
55786  __ret; \
55787})
55788#endif
55789
55790#ifdef __LITTLE_ENDIAN__
55791#define vqdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
55792  int32x2_t __s0 = __p0; \
55793  int32x4_t __s1 = __p1; \
55794  int32x2_t __ret; \
55795  __ret = vqdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
55796  __ret; \
55797})
55798#else
55799#define vqdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
55800  int32x2_t __s0 = __p0; \
55801  int32x4_t __s1 = __p1; \
55802  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
55803  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
55804  int32x2_t __ret; \
55805  __ret = __noswap_vqdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
55806  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
55807  __ret; \
55808})
55809#endif
55810
55811#ifdef __LITTLE_ENDIAN__
55812#define vqdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
55813  int16x4_t __s0 = __p0; \
55814  int16x8_t __s1 = __p1; \
55815  int16x4_t __ret; \
55816  __ret = vqdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
55817  __ret; \
55818})
55819#else
55820#define vqdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
55821  int16x4_t __s0 = __p0; \
55822  int16x8_t __s1 = __p1; \
55823  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
55824  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
55825  int16x4_t __ret; \
55826  __ret = __noswap_vqdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
55827  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
55828  __ret; \
55829})
55830#endif
55831
55832#ifdef __LITTLE_ENDIAN__
55833__ai int64_t vqdmulls_s32(int32_t __p0, int32_t __p1) {
55834  int64_t __ret;
55835  __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
55836  return __ret;
55837}
55838#else
55839__ai int64_t vqdmulls_s32(int32_t __p0, int32_t __p1) {
55840  int64_t __ret;
55841  __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
55842  return __ret;
55843}
55844__ai int64_t __noswap_vqdmulls_s32(int32_t __p0, int32_t __p1) {
55845  int64_t __ret;
55846  __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
55847  return __ret;
55848}
55849#endif
55850
55851#ifdef __LITTLE_ENDIAN__
55852__ai int32_t vqdmullh_s16(int16_t __p0, int16_t __p1) {
55853  int32_t __ret;
55854  __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
55855  return __ret;
55856}
55857#else
55858__ai int32_t vqdmullh_s16(int16_t __p0, int16_t __p1) {
55859  int32_t __ret;
55860  __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
55861  return __ret;
55862}
55863__ai int32_t __noswap_vqdmullh_s16(int16_t __p0, int16_t __p1) {
55864  int32_t __ret;
55865  __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
55866  return __ret;
55867}
55868#endif
55869
55870#ifdef __LITTLE_ENDIAN__
55871__ai int64x2_t vqdmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
55872  int64x2_t __ret;
55873  __ret = vqdmull_s32(vget_high_s32(__p0), vget_high_s32(__p1));
55874  return __ret;
55875}
55876#else
55877__ai int64x2_t vqdmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
55878  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55879  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55880  int64x2_t __ret;
55881  __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
55882  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55883  return __ret;
55884}
55885#endif
55886
55887#ifdef __LITTLE_ENDIAN__
55888__ai int32x4_t vqdmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
55889  int32x4_t __ret;
55890  __ret = vqdmull_s16(vget_high_s16(__p0), vget_high_s16(__p1));
55891  return __ret;
55892}
55893#else
55894__ai int32x4_t vqdmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
55895  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55896  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55897  int32x4_t __ret;
55898  __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
55899  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55900  return __ret;
55901}
55902#endif
55903
55904#ifdef __LITTLE_ENDIAN__
55905#define vqdmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
55906  int32x4_t __s0 = __p0; \
55907  int32x2_t __s1 = __p1; \
55908  int64x2_t __ret; \
55909  __ret = vqdmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
55910  __ret; \
55911})
55912#else
55913#define vqdmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
55914  int32x4_t __s0 = __p0; \
55915  int32x2_t __s1 = __p1; \
55916  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
55917  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
55918  int64x2_t __ret; \
55919  __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
55920  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
55921  __ret; \
55922})
55923#endif
55924
55925#ifdef __LITTLE_ENDIAN__
55926#define vqdmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
55927  int16x8_t __s0 = __p0; \
55928  int16x4_t __s1 = __p1; \
55929  int32x4_t __ret; \
55930  __ret = vqdmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
55931  __ret; \
55932})
55933#else
55934#define vqdmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
55935  int16x8_t __s0 = __p0; \
55936  int16x4_t __s1 = __p1; \
55937  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
55938  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
55939  int32x4_t __ret; \
55940  __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
55941  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
55942  __ret; \
55943})
55944#endif
55945
55946#ifdef __LITTLE_ENDIAN__
55947#define vqdmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
55948  int32x4_t __s0 = __p0; \
55949  int32x4_t __s1 = __p1; \
55950  int64x2_t __ret; \
55951  __ret = vqdmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
55952  __ret; \
55953})
55954#else
55955#define vqdmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
55956  int32x4_t __s0 = __p0; \
55957  int32x4_t __s1 = __p1; \
55958  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
55959  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
55960  int64x2_t __ret; \
55961  __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
55962  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
55963  __ret; \
55964})
55965#endif
55966
55967#ifdef __LITTLE_ENDIAN__
55968#define vqdmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
55969  int16x8_t __s0 = __p0; \
55970  int16x8_t __s1 = __p1; \
55971  int32x4_t __ret; \
55972  __ret = vqdmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
55973  __ret; \
55974})
55975#else
55976#define vqdmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
55977  int16x8_t __s0 = __p0; \
55978  int16x8_t __s1 = __p1; \
55979  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
55980  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
55981  int32x4_t __ret; \
55982  __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
55983  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
55984  __ret; \
55985})
55986#endif
55987
55988#ifdef __LITTLE_ENDIAN__
55989__ai int64x2_t vqdmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
55990  int64x2_t __ret;
55991  __ret = vqdmull_n_s32(vget_high_s32(__p0), __p1);
55992  return __ret;
55993}
55994#else
55995__ai int64x2_t vqdmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
55996  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55997  int64x2_t __ret;
55998  __ret = __noswap_vqdmull_n_s32(__noswap_vget_high_s32(__rev0), __p1);
55999  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56000  return __ret;
56001}
56002#endif
56003
56004#ifdef __LITTLE_ENDIAN__
56005__ai int32x4_t vqdmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
56006  int32x4_t __ret;
56007  __ret = vqdmull_n_s16(vget_high_s16(__p0), __p1);
56008  return __ret;
56009}
56010#else
56011__ai int32x4_t vqdmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
56012  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
56013  int32x4_t __ret;
56014  __ret = __noswap_vqdmull_n_s16(__noswap_vget_high_s16(__rev0), __p1);
56015  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56016  return __ret;
56017}
56018#endif
56019
56020#ifdef __LITTLE_ENDIAN__
56021#define vqdmulls_lane_s32(__p0_158, __p1_158, __p2_158) __extension__ ({ \
56022  int32_t __s0_158 = __p0_158; \
56023  int32x2_t __s1_158 = __p1_158; \
56024  int64_t __ret_158; \
56025  __ret_158 = vqdmulls_s32(__s0_158, vget_lane_s32(__s1_158, __p2_158)); \
56026  __ret_158; \
56027})
56028#else
56029#define vqdmulls_lane_s32(__p0_159, __p1_159, __p2_159) __extension__ ({ \
56030  int32_t __s0_159 = __p0_159; \
56031  int32x2_t __s1_159 = __p1_159; \
56032  int32x2_t __rev1_159;  __rev1_159 = __builtin_shufflevector(__s1_159, __s1_159, 1, 0); \
56033  int64_t __ret_159; \
56034  __ret_159 = __noswap_vqdmulls_s32(__s0_159, __noswap_vget_lane_s32(__rev1_159, __p2_159)); \
56035  __ret_159; \
56036})
56037#endif
56038
56039#ifdef __LITTLE_ENDIAN__
56040#define vqdmullh_lane_s16(__p0_160, __p1_160, __p2_160) __extension__ ({ \
56041  int16_t __s0_160 = __p0_160; \
56042  int16x4_t __s1_160 = __p1_160; \
56043  int32_t __ret_160; \
56044  __ret_160 = vqdmullh_s16(__s0_160, vget_lane_s16(__s1_160, __p2_160)); \
56045  __ret_160; \
56046})
56047#else
56048#define vqdmullh_lane_s16(__p0_161, __p1_161, __p2_161) __extension__ ({ \
56049  int16_t __s0_161 = __p0_161; \
56050  int16x4_t __s1_161 = __p1_161; \
56051  int16x4_t __rev1_161;  __rev1_161 = __builtin_shufflevector(__s1_161, __s1_161, 3, 2, 1, 0); \
56052  int32_t __ret_161; \
56053  __ret_161 = __noswap_vqdmullh_s16(__s0_161, __noswap_vget_lane_s16(__rev1_161, __p2_161)); \
56054  __ret_161; \
56055})
56056#endif
56057
56058#ifdef __LITTLE_ENDIAN__
56059#define vqdmulls_laneq_s32(__p0_162, __p1_162, __p2_162) __extension__ ({ \
56060  int32_t __s0_162 = __p0_162; \
56061  int32x4_t __s1_162 = __p1_162; \
56062  int64_t __ret_162; \
56063  __ret_162 = vqdmulls_s32(__s0_162, vgetq_lane_s32(__s1_162, __p2_162)); \
56064  __ret_162; \
56065})
56066#else
56067#define vqdmulls_laneq_s32(__p0_163, __p1_163, __p2_163) __extension__ ({ \
56068  int32_t __s0_163 = __p0_163; \
56069  int32x4_t __s1_163 = __p1_163; \
56070  int32x4_t __rev1_163;  __rev1_163 = __builtin_shufflevector(__s1_163, __s1_163, 3, 2, 1, 0); \
56071  int64_t __ret_163; \
56072  __ret_163 = __noswap_vqdmulls_s32(__s0_163, __noswap_vgetq_lane_s32(__rev1_163, __p2_163)); \
56073  __ret_163; \
56074})
56075#endif
56076
56077#ifdef __LITTLE_ENDIAN__
56078#define vqdmullh_laneq_s16(__p0_164, __p1_164, __p2_164) __extension__ ({ \
56079  int16_t __s0_164 = __p0_164; \
56080  int16x8_t __s1_164 = __p1_164; \
56081  int32_t __ret_164; \
56082  __ret_164 = vqdmullh_s16(__s0_164, vgetq_lane_s16(__s1_164, __p2_164)); \
56083  __ret_164; \
56084})
56085#else
56086#define vqdmullh_laneq_s16(__p0_165, __p1_165, __p2_165) __extension__ ({ \
56087  int16_t __s0_165 = __p0_165; \
56088  int16x8_t __s1_165 = __p1_165; \
56089  int16x8_t __rev1_165;  __rev1_165 = __builtin_shufflevector(__s1_165, __s1_165, 7, 6, 5, 4, 3, 2, 1, 0); \
56090  int32_t __ret_165; \
56091  __ret_165 = __noswap_vqdmullh_s16(__s0_165, __noswap_vgetq_lane_s16(__rev1_165, __p2_165)); \
56092  __ret_165; \
56093})
56094#endif
56095
56096#ifdef __LITTLE_ENDIAN__
56097#define vqdmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56098  int32x2_t __s0 = __p0; \
56099  int32x4_t __s1 = __p1; \
56100  int64x2_t __ret; \
56101  __ret = vqdmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
56102  __ret; \
56103})
56104#else
56105#define vqdmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56106  int32x2_t __s0 = __p0; \
56107  int32x4_t __s1 = __p1; \
56108  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56109  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56110  int64x2_t __ret; \
56111  __ret = __noswap_vqdmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
56112  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56113  __ret; \
56114})
56115#endif
56116
56117#ifdef __LITTLE_ENDIAN__
56118#define vqdmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56119  int16x4_t __s0 = __p0; \
56120  int16x8_t __s1 = __p1; \
56121  int32x4_t __ret; \
56122  __ret = vqdmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56123  __ret; \
56124})
56125#else
56126#define vqdmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56127  int16x4_t __s0 = __p0; \
56128  int16x8_t __s1 = __p1; \
56129  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56130  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56131  int32x4_t __ret; \
56132  __ret = __noswap_vqdmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56133  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56134  __ret; \
56135})
56136#endif
56137
56138#ifdef __LITTLE_ENDIAN__
56139__ai int16_t vqmovns_s32(int32_t __p0) {
56140  int16_t __ret;
56141  __ret = (int16_t) __builtin_neon_vqmovns_s32(__p0);
56142  return __ret;
56143}
56144#else
56145__ai int16_t vqmovns_s32(int32_t __p0) {
56146  int16_t __ret;
56147  __ret = (int16_t) __builtin_neon_vqmovns_s32(__p0);
56148  return __ret;
56149}
56150#endif
56151
56152#ifdef __LITTLE_ENDIAN__
56153__ai int32_t vqmovnd_s64(int64_t __p0) {
56154  int32_t __ret;
56155  __ret = (int32_t) __builtin_neon_vqmovnd_s64(__p0);
56156  return __ret;
56157}
56158#else
56159__ai int32_t vqmovnd_s64(int64_t __p0) {
56160  int32_t __ret;
56161  __ret = (int32_t) __builtin_neon_vqmovnd_s64(__p0);
56162  return __ret;
56163}
56164#endif
56165
56166#ifdef __LITTLE_ENDIAN__
56167__ai int8_t vqmovnh_s16(int16_t __p0) {
56168  int8_t __ret;
56169  __ret = (int8_t) __builtin_neon_vqmovnh_s16(__p0);
56170  return __ret;
56171}
56172#else
56173__ai int8_t vqmovnh_s16(int16_t __p0) {
56174  int8_t __ret;
56175  __ret = (int8_t) __builtin_neon_vqmovnh_s16(__p0);
56176  return __ret;
56177}
56178#endif
56179
56180#ifdef __LITTLE_ENDIAN__
56181__ai uint16_t vqmovns_u32(uint32_t __p0) {
56182  uint16_t __ret;
56183  __ret = (uint16_t) __builtin_neon_vqmovns_u32(__p0);
56184  return __ret;
56185}
56186#else
56187__ai uint16_t vqmovns_u32(uint32_t __p0) {
56188  uint16_t __ret;
56189  __ret = (uint16_t) __builtin_neon_vqmovns_u32(__p0);
56190  return __ret;
56191}
56192#endif
56193
56194#ifdef __LITTLE_ENDIAN__
56195__ai uint32_t vqmovnd_u64(uint64_t __p0) {
56196  uint32_t __ret;
56197  __ret = (uint32_t) __builtin_neon_vqmovnd_u64(__p0);
56198  return __ret;
56199}
56200#else
56201__ai uint32_t vqmovnd_u64(uint64_t __p0) {
56202  uint32_t __ret;
56203  __ret = (uint32_t) __builtin_neon_vqmovnd_u64(__p0);
56204  return __ret;
56205}
56206#endif
56207
56208#ifdef __LITTLE_ENDIAN__
56209__ai uint8_t vqmovnh_u16(uint16_t __p0) {
56210  uint8_t __ret;
56211  __ret = (uint8_t) __builtin_neon_vqmovnh_u16(__p0);
56212  return __ret;
56213}
56214#else
56215__ai uint8_t vqmovnh_u16(uint16_t __p0) {
56216  uint8_t __ret;
56217  __ret = (uint8_t) __builtin_neon_vqmovnh_u16(__p0);
56218  return __ret;
56219}
56220#endif
56221
56222#ifdef __LITTLE_ENDIAN__
56223__ai uint16x8_t vqmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
56224  uint16x8_t __ret;
56225  __ret = vcombine_u16(__p0, vqmovn_u32(__p1));
56226  return __ret;
56227}
56228#else
56229__ai uint16x8_t vqmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
56230  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56231  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56232  uint16x8_t __ret;
56233  __ret = __noswap_vcombine_u16(__rev0, __noswap_vqmovn_u32(__rev1));
56234  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
56235  return __ret;
56236}
56237#endif
56238
56239#ifdef __LITTLE_ENDIAN__
56240__ai uint32x4_t vqmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
56241  uint32x4_t __ret;
56242  __ret = vcombine_u32(__p0, vqmovn_u64(__p1));
56243  return __ret;
56244}
56245#else
56246__ai uint32x4_t vqmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
56247  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56248  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
56249  uint32x4_t __ret;
56250  __ret = __noswap_vcombine_u32(__rev0, __noswap_vqmovn_u64(__rev1));
56251  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56252  return __ret;
56253}
56254#endif
56255
56256#ifdef __LITTLE_ENDIAN__
56257__ai uint8x16_t vqmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
56258  uint8x16_t __ret;
56259  __ret = vcombine_u8(__p0, vqmovn_u16(__p1));
56260  return __ret;
56261}
56262#else
56263__ai uint8x16_t vqmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
56264  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
56265  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56266  uint8x16_t __ret;
56267  __ret = __noswap_vcombine_u8(__rev0, __noswap_vqmovn_u16(__rev1));
56268  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
56269  return __ret;
56270}
56271#endif
56272
56273#ifdef __LITTLE_ENDIAN__
56274__ai int16x8_t vqmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
56275  int16x8_t __ret;
56276  __ret = vcombine_s16(__p0, vqmovn_s32(__p1));
56277  return __ret;
56278}
56279#else
56280__ai int16x8_t vqmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
56281  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56282  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56283  int16x8_t __ret;
56284  __ret = __noswap_vcombine_s16(__rev0, __noswap_vqmovn_s32(__rev1));
56285  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
56286  return __ret;
56287}
56288#endif
56289
56290#ifdef __LITTLE_ENDIAN__
56291__ai int32x4_t vqmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
56292  int32x4_t __ret;
56293  __ret = vcombine_s32(__p0, vqmovn_s64(__p1));
56294  return __ret;
56295}
56296#else
56297__ai int32x4_t vqmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
56298  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56299  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
56300  int32x4_t __ret;
56301  __ret = __noswap_vcombine_s32(__rev0, __noswap_vqmovn_s64(__rev1));
56302  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56303  return __ret;
56304}
56305#endif
56306
56307#ifdef __LITTLE_ENDIAN__
56308__ai int8x16_t vqmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
56309  int8x16_t __ret;
56310  __ret = vcombine_s8(__p0, vqmovn_s16(__p1));
56311  return __ret;
56312}
56313#else
56314__ai int8x16_t vqmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
56315  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
56316  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56317  int8x16_t __ret;
56318  __ret = __noswap_vcombine_s8(__rev0, __noswap_vqmovn_s16(__rev1));
56319  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
56320  return __ret;
56321}
56322#endif
56323
56324#ifdef __LITTLE_ENDIAN__
56325__ai int16_t vqmovuns_s32(int32_t __p0) {
56326  int16_t __ret;
56327  __ret = (int16_t) __builtin_neon_vqmovuns_s32(__p0);
56328  return __ret;
56329}
56330#else
56331__ai int16_t vqmovuns_s32(int32_t __p0) {
56332  int16_t __ret;
56333  __ret = (int16_t) __builtin_neon_vqmovuns_s32(__p0);
56334  return __ret;
56335}
56336#endif
56337
56338#ifdef __LITTLE_ENDIAN__
56339__ai int32_t vqmovund_s64(int64_t __p0) {
56340  int32_t __ret;
56341  __ret = (int32_t) __builtin_neon_vqmovund_s64(__p0);
56342  return __ret;
56343}
56344#else
56345__ai int32_t vqmovund_s64(int64_t __p0) {
56346  int32_t __ret;
56347  __ret = (int32_t) __builtin_neon_vqmovund_s64(__p0);
56348  return __ret;
56349}
56350#endif
56351
56352#ifdef __LITTLE_ENDIAN__
56353__ai int8_t vqmovunh_s16(int16_t __p0) {
56354  int8_t __ret;
56355  __ret = (int8_t) __builtin_neon_vqmovunh_s16(__p0);
56356  return __ret;
56357}
56358#else
56359__ai int8_t vqmovunh_s16(int16_t __p0) {
56360  int8_t __ret;
56361  __ret = (int8_t) __builtin_neon_vqmovunh_s16(__p0);
56362  return __ret;
56363}
56364#endif
56365
56366#ifdef __LITTLE_ENDIAN__
56367__ai int16x8_t vqmovun_high_s32(int16x4_t __p0, int32x4_t __p1) {
56368  int16x8_t __ret;
56369  __ret = vcombine_u16((uint16x4_t)(__p0), vqmovun_s32(__p1));
56370  return __ret;
56371}
56372#else
56373__ai int16x8_t vqmovun_high_s32(int16x4_t __p0, int32x4_t __p1) {
56374  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56375  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56376  int16x8_t __ret;
56377  __ret = __noswap_vcombine_u16((uint16x4_t)(__rev0), __noswap_vqmovun_s32(__rev1));
56378  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
56379  return __ret;
56380}
56381#endif
56382
56383#ifdef __LITTLE_ENDIAN__
56384__ai int32x4_t vqmovun_high_s64(int32x2_t __p0, int64x2_t __p1) {
56385  int32x4_t __ret;
56386  __ret = vcombine_u32((uint32x2_t)(__p0), vqmovun_s64(__p1));
56387  return __ret;
56388}
56389#else
56390__ai int32x4_t vqmovun_high_s64(int32x2_t __p0, int64x2_t __p1) {
56391  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56392  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
56393  int32x4_t __ret;
56394  __ret = __noswap_vcombine_u32((uint32x2_t)(__rev0), __noswap_vqmovun_s64(__rev1));
56395  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56396  return __ret;
56397}
56398#endif
56399
56400#ifdef __LITTLE_ENDIAN__
56401__ai int8x16_t vqmovun_high_s16(int8x8_t __p0, int16x8_t __p1) {
56402  int8x16_t __ret;
56403  __ret = vcombine_u8((uint8x8_t)(__p0), vqmovun_s16(__p1));
56404  return __ret;
56405}
56406#else
56407__ai int8x16_t vqmovun_high_s16(int8x8_t __p0, int16x8_t __p1) {
56408  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
56409  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56410  int8x16_t __ret;
56411  __ret = __noswap_vcombine_u8((uint8x8_t)(__rev0), __noswap_vqmovun_s16(__rev1));
56412  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
56413  return __ret;
56414}
56415#endif
56416
56417#ifdef __LITTLE_ENDIAN__
56418__ai int64x2_t vqnegq_s64(int64x2_t __p0) {
56419  int64x2_t __ret;
56420  __ret = (int64x2_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 35);
56421  return __ret;
56422}
56423#else
56424__ai int64x2_t vqnegq_s64(int64x2_t __p0) {
56425  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56426  int64x2_t __ret;
56427  __ret = (int64x2_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 35);
56428  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56429  return __ret;
56430}
56431#endif
56432
56433#ifdef __LITTLE_ENDIAN__
56434__ai int64x1_t vqneg_s64(int64x1_t __p0) {
56435  int64x1_t __ret;
56436  __ret = (int64x1_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 3);
56437  return __ret;
56438}
56439#else
56440__ai int64x1_t vqneg_s64(int64x1_t __p0) {
56441  int64x1_t __ret;
56442  __ret = (int64x1_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 3);
56443  return __ret;
56444}
56445#endif
56446
56447#ifdef __LITTLE_ENDIAN__
56448__ai int8_t vqnegb_s8(int8_t __p0) {
56449  int8_t __ret;
56450  __ret = (int8_t) __builtin_neon_vqnegb_s8(__p0);
56451  return __ret;
56452}
56453#else
56454__ai int8_t vqnegb_s8(int8_t __p0) {
56455  int8_t __ret;
56456  __ret = (int8_t) __builtin_neon_vqnegb_s8(__p0);
56457  return __ret;
56458}
56459#endif
56460
56461#ifdef __LITTLE_ENDIAN__
56462__ai int32_t vqnegs_s32(int32_t __p0) {
56463  int32_t __ret;
56464  __ret = (int32_t) __builtin_neon_vqnegs_s32(__p0);
56465  return __ret;
56466}
56467#else
56468__ai int32_t vqnegs_s32(int32_t __p0) {
56469  int32_t __ret;
56470  __ret = (int32_t) __builtin_neon_vqnegs_s32(__p0);
56471  return __ret;
56472}
56473#endif
56474
56475#ifdef __LITTLE_ENDIAN__
56476__ai int64_t vqnegd_s64(int64_t __p0) {
56477  int64_t __ret;
56478  __ret = (int64_t) __builtin_neon_vqnegd_s64(__p0);
56479  return __ret;
56480}
56481#else
56482__ai int64_t vqnegd_s64(int64_t __p0) {
56483  int64_t __ret;
56484  __ret = (int64_t) __builtin_neon_vqnegd_s64(__p0);
56485  return __ret;
56486}
56487#endif
56488
56489#ifdef __LITTLE_ENDIAN__
56490__ai int16_t vqnegh_s16(int16_t __p0) {
56491  int16_t __ret;
56492  __ret = (int16_t) __builtin_neon_vqnegh_s16(__p0);
56493  return __ret;
56494}
56495#else
56496__ai int16_t vqnegh_s16(int16_t __p0) {
56497  int16_t __ret;
56498  __ret = (int16_t) __builtin_neon_vqnegh_s16(__p0);
56499  return __ret;
56500}
56501#endif
56502
56503#ifdef __LITTLE_ENDIAN__
56504__ai int32_t vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
56505  int32_t __ret;
56506  __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
56507  return __ret;
56508}
56509#else
56510__ai int32_t vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
56511  int32_t __ret;
56512  __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
56513  return __ret;
56514}
56515__ai int32_t __noswap_vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
56516  int32_t __ret;
56517  __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
56518  return __ret;
56519}
56520#endif
56521
56522#ifdef __LITTLE_ENDIAN__
56523__ai int16_t vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
56524  int16_t __ret;
56525  __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
56526  return __ret;
56527}
56528#else
56529__ai int16_t vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
56530  int16_t __ret;
56531  __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
56532  return __ret;
56533}
56534__ai int16_t __noswap_vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
56535  int16_t __ret;
56536  __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
56537  return __ret;
56538}
56539#endif
56540
56541#ifdef __LITTLE_ENDIAN__
56542#define vqrdmulhs_lane_s32(__p0_166, __p1_166, __p2_166) __extension__ ({ \
56543  int32_t __s0_166 = __p0_166; \
56544  int32x2_t __s1_166 = __p1_166; \
56545  int32_t __ret_166; \
56546  __ret_166 = vqrdmulhs_s32(__s0_166, vget_lane_s32(__s1_166, __p2_166)); \
56547  __ret_166; \
56548})
56549#else
56550#define vqrdmulhs_lane_s32(__p0_167, __p1_167, __p2_167) __extension__ ({ \
56551  int32_t __s0_167 = __p0_167; \
56552  int32x2_t __s1_167 = __p1_167; \
56553  int32x2_t __rev1_167;  __rev1_167 = __builtin_shufflevector(__s1_167, __s1_167, 1, 0); \
56554  int32_t __ret_167; \
56555  __ret_167 = __noswap_vqrdmulhs_s32(__s0_167, __noswap_vget_lane_s32(__rev1_167, __p2_167)); \
56556  __ret_167; \
56557})
56558#endif
56559
56560#ifdef __LITTLE_ENDIAN__
56561#define vqrdmulhh_lane_s16(__p0_168, __p1_168, __p2_168) __extension__ ({ \
56562  int16_t __s0_168 = __p0_168; \
56563  int16x4_t __s1_168 = __p1_168; \
56564  int16_t __ret_168; \
56565  __ret_168 = vqrdmulhh_s16(__s0_168, vget_lane_s16(__s1_168, __p2_168)); \
56566  __ret_168; \
56567})
56568#else
56569#define vqrdmulhh_lane_s16(__p0_169, __p1_169, __p2_169) __extension__ ({ \
56570  int16_t __s0_169 = __p0_169; \
56571  int16x4_t __s1_169 = __p1_169; \
56572  int16x4_t __rev1_169;  __rev1_169 = __builtin_shufflevector(__s1_169, __s1_169, 3, 2, 1, 0); \
56573  int16_t __ret_169; \
56574  __ret_169 = __noswap_vqrdmulhh_s16(__s0_169, __noswap_vget_lane_s16(__rev1_169, __p2_169)); \
56575  __ret_169; \
56576})
56577#endif
56578
56579#ifdef __LITTLE_ENDIAN__
56580#define vqrdmulhs_laneq_s32(__p0_170, __p1_170, __p2_170) __extension__ ({ \
56581  int32_t __s0_170 = __p0_170; \
56582  int32x4_t __s1_170 = __p1_170; \
56583  int32_t __ret_170; \
56584  __ret_170 = vqrdmulhs_s32(__s0_170, vgetq_lane_s32(__s1_170, __p2_170)); \
56585  __ret_170; \
56586})
56587#else
56588#define vqrdmulhs_laneq_s32(__p0_171, __p1_171, __p2_171) __extension__ ({ \
56589  int32_t __s0_171 = __p0_171; \
56590  int32x4_t __s1_171 = __p1_171; \
56591  int32x4_t __rev1_171;  __rev1_171 = __builtin_shufflevector(__s1_171, __s1_171, 3, 2, 1, 0); \
56592  int32_t __ret_171; \
56593  __ret_171 = __noswap_vqrdmulhs_s32(__s0_171, __noswap_vgetq_lane_s32(__rev1_171, __p2_171)); \
56594  __ret_171; \
56595})
56596#endif
56597
56598#ifdef __LITTLE_ENDIAN__
56599#define vqrdmulhh_laneq_s16(__p0_172, __p1_172, __p2_172) __extension__ ({ \
56600  int16_t __s0_172 = __p0_172; \
56601  int16x8_t __s1_172 = __p1_172; \
56602  int16_t __ret_172; \
56603  __ret_172 = vqrdmulhh_s16(__s0_172, vgetq_lane_s16(__s1_172, __p2_172)); \
56604  __ret_172; \
56605})
56606#else
56607#define vqrdmulhh_laneq_s16(__p0_173, __p1_173, __p2_173) __extension__ ({ \
56608  int16_t __s0_173 = __p0_173; \
56609  int16x8_t __s1_173 = __p1_173; \
56610  int16x8_t __rev1_173;  __rev1_173 = __builtin_shufflevector(__s1_173, __s1_173, 7, 6, 5, 4, 3, 2, 1, 0); \
56611  int16_t __ret_173; \
56612  __ret_173 = __noswap_vqrdmulhh_s16(__s0_173, __noswap_vgetq_lane_s16(__rev1_173, __p2_173)); \
56613  __ret_173; \
56614})
56615#endif
56616
56617#ifdef __LITTLE_ENDIAN__
56618#define vqrdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56619  int32x4_t __s0 = __p0; \
56620  int32x4_t __s1 = __p1; \
56621  int32x4_t __ret; \
56622  __ret = vqrdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56623  __ret; \
56624})
56625#else
56626#define vqrdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56627  int32x4_t __s0 = __p0; \
56628  int32x4_t __s1 = __p1; \
56629  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56630  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56631  int32x4_t __ret; \
56632  __ret = __noswap_vqrdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56633  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56634  __ret; \
56635})
56636#endif
56637
56638#ifdef __LITTLE_ENDIAN__
56639#define vqrdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56640  int16x8_t __s0 = __p0; \
56641  int16x8_t __s1 = __p1; \
56642  int16x8_t __ret; \
56643  __ret = vqrdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
56644  __ret; \
56645})
56646#else
56647#define vqrdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56648  int16x8_t __s0 = __p0; \
56649  int16x8_t __s1 = __p1; \
56650  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
56651  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56652  int16x8_t __ret; \
56653  __ret = __noswap_vqrdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
56654  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
56655  __ret; \
56656})
56657#endif
56658
56659#ifdef __LITTLE_ENDIAN__
56660#define vqrdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56661  int32x2_t __s0 = __p0; \
56662  int32x4_t __s1 = __p1; \
56663  int32x2_t __ret; \
56664  __ret = vqrdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
56665  __ret; \
56666})
56667#else
56668#define vqrdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56669  int32x2_t __s0 = __p0; \
56670  int32x4_t __s1 = __p1; \
56671  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56672  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56673  int32x2_t __ret; \
56674  __ret = __noswap_vqrdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
56675  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56676  __ret; \
56677})
56678#endif
56679
56680#ifdef __LITTLE_ENDIAN__
56681#define vqrdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56682  int16x4_t __s0 = __p0; \
56683  int16x8_t __s1 = __p1; \
56684  int16x4_t __ret; \
56685  __ret = vqrdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56686  __ret; \
56687})
56688#else
56689#define vqrdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56690  int16x4_t __s0 = __p0; \
56691  int16x8_t __s1 = __p1; \
56692  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56693  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56694  int16x4_t __ret; \
56695  __ret = __noswap_vqrdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56696  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56697  __ret; \
56698})
56699#endif
56700
56701#ifdef __LITTLE_ENDIAN__
56702__ai uint8_t vqrshlb_u8(uint8_t __p0, uint8_t __p1) {
56703  uint8_t __ret;
56704  __ret = (uint8_t) __builtin_neon_vqrshlb_u8(__p0, __p1);
56705  return __ret;
56706}
56707#else
56708__ai uint8_t vqrshlb_u8(uint8_t __p0, uint8_t __p1) {
56709  uint8_t __ret;
56710  __ret = (uint8_t) __builtin_neon_vqrshlb_u8(__p0, __p1);
56711  return __ret;
56712}
56713#endif
56714
56715#ifdef __LITTLE_ENDIAN__
56716__ai uint32_t vqrshls_u32(uint32_t __p0, uint32_t __p1) {
56717  uint32_t __ret;
56718  __ret = (uint32_t) __builtin_neon_vqrshls_u32(__p0, __p1);
56719  return __ret;
56720}
56721#else
56722__ai uint32_t vqrshls_u32(uint32_t __p0, uint32_t __p1) {
56723  uint32_t __ret;
56724  __ret = (uint32_t) __builtin_neon_vqrshls_u32(__p0, __p1);
56725  return __ret;
56726}
56727#endif
56728
56729#ifdef __LITTLE_ENDIAN__
56730__ai uint64_t vqrshld_u64(uint64_t __p0, uint64_t __p1) {
56731  uint64_t __ret;
56732  __ret = (uint64_t) __builtin_neon_vqrshld_u64(__p0, __p1);
56733  return __ret;
56734}
56735#else
56736__ai uint64_t vqrshld_u64(uint64_t __p0, uint64_t __p1) {
56737  uint64_t __ret;
56738  __ret = (uint64_t) __builtin_neon_vqrshld_u64(__p0, __p1);
56739  return __ret;
56740}
56741#endif
56742
56743#ifdef __LITTLE_ENDIAN__
56744__ai uint16_t vqrshlh_u16(uint16_t __p0, uint16_t __p1) {
56745  uint16_t __ret;
56746  __ret = (uint16_t) __builtin_neon_vqrshlh_u16(__p0, __p1);
56747  return __ret;
56748}
56749#else
56750__ai uint16_t vqrshlh_u16(uint16_t __p0, uint16_t __p1) {
56751  uint16_t __ret;
56752  __ret = (uint16_t) __builtin_neon_vqrshlh_u16(__p0, __p1);
56753  return __ret;
56754}
56755#endif
56756
56757#ifdef __LITTLE_ENDIAN__
56758__ai int8_t vqrshlb_s8(int8_t __p0, int8_t __p1) {
56759  int8_t __ret;
56760  __ret = (int8_t) __builtin_neon_vqrshlb_s8(__p0, __p1);
56761  return __ret;
56762}
56763#else
56764__ai int8_t vqrshlb_s8(int8_t __p0, int8_t __p1) {
56765  int8_t __ret;
56766  __ret = (int8_t) __builtin_neon_vqrshlb_s8(__p0, __p1);
56767  return __ret;
56768}
56769#endif
56770
56771#ifdef __LITTLE_ENDIAN__
56772__ai int32_t vqrshls_s32(int32_t __p0, int32_t __p1) {
56773  int32_t __ret;
56774  __ret = (int32_t) __builtin_neon_vqrshls_s32(__p0, __p1);
56775  return __ret;
56776}
56777#else
56778__ai int32_t vqrshls_s32(int32_t __p0, int32_t __p1) {
56779  int32_t __ret;
56780  __ret = (int32_t) __builtin_neon_vqrshls_s32(__p0, __p1);
56781  return __ret;
56782}
56783#endif
56784
56785#ifdef __LITTLE_ENDIAN__
56786__ai int64_t vqrshld_s64(int64_t __p0, int64_t __p1) {
56787  int64_t __ret;
56788  __ret = (int64_t) __builtin_neon_vqrshld_s64(__p0, __p1);
56789  return __ret;
56790}
56791#else
56792__ai int64_t vqrshld_s64(int64_t __p0, int64_t __p1) {
56793  int64_t __ret;
56794  __ret = (int64_t) __builtin_neon_vqrshld_s64(__p0, __p1);
56795  return __ret;
56796}
56797#endif
56798
56799#ifdef __LITTLE_ENDIAN__
56800__ai int16_t vqrshlh_s16(int16_t __p0, int16_t __p1) {
56801  int16_t __ret;
56802  __ret = (int16_t) __builtin_neon_vqrshlh_s16(__p0, __p1);
56803  return __ret;
56804}
56805#else
56806__ai int16_t vqrshlh_s16(int16_t __p0, int16_t __p1) {
56807  int16_t __ret;
56808  __ret = (int16_t) __builtin_neon_vqrshlh_s16(__p0, __p1);
56809  return __ret;
56810}
56811#endif
56812
56813#ifdef __LITTLE_ENDIAN__
56814#define vqrshrn_high_n_u32(__p0_174, __p1_174, __p2_174) __extension__ ({ \
56815  uint16x4_t __s0_174 = __p0_174; \
56816  uint32x4_t __s1_174 = __p1_174; \
56817  uint16x8_t __ret_174; \
56818  __ret_174 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_174), (uint16x4_t)(vqrshrn_n_u32(__s1_174, __p2_174)))); \
56819  __ret_174; \
56820})
56821#else
56822#define vqrshrn_high_n_u32(__p0_175, __p1_175, __p2_175) __extension__ ({ \
56823  uint16x4_t __s0_175 = __p0_175; \
56824  uint32x4_t __s1_175 = __p1_175; \
56825  uint16x4_t __rev0_175;  __rev0_175 = __builtin_shufflevector(__s0_175, __s0_175, 3, 2, 1, 0); \
56826  uint32x4_t __rev1_175;  __rev1_175 = __builtin_shufflevector(__s1_175, __s1_175, 3, 2, 1, 0); \
56827  uint16x8_t __ret_175; \
56828  __ret_175 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_175), (uint16x4_t)(__noswap_vqrshrn_n_u32(__rev1_175, __p2_175)))); \
56829  __ret_175 = __builtin_shufflevector(__ret_175, __ret_175, 7, 6, 5, 4, 3, 2, 1, 0); \
56830  __ret_175; \
56831})
56832#endif
56833
56834#ifdef __LITTLE_ENDIAN__
56835#define vqrshrn_high_n_u64(__p0_176, __p1_176, __p2_176) __extension__ ({ \
56836  uint32x2_t __s0_176 = __p0_176; \
56837  uint64x2_t __s1_176 = __p1_176; \
56838  uint32x4_t __ret_176; \
56839  __ret_176 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_176), (uint32x2_t)(vqrshrn_n_u64(__s1_176, __p2_176)))); \
56840  __ret_176; \
56841})
56842#else
56843#define vqrshrn_high_n_u64(__p0_177, __p1_177, __p2_177) __extension__ ({ \
56844  uint32x2_t __s0_177 = __p0_177; \
56845  uint64x2_t __s1_177 = __p1_177; \
56846  uint32x2_t __rev0_177;  __rev0_177 = __builtin_shufflevector(__s0_177, __s0_177, 1, 0); \
56847  uint64x2_t __rev1_177;  __rev1_177 = __builtin_shufflevector(__s1_177, __s1_177, 1, 0); \
56848  uint32x4_t __ret_177; \
56849  __ret_177 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_177), (uint32x2_t)(__noswap_vqrshrn_n_u64(__rev1_177, __p2_177)))); \
56850  __ret_177 = __builtin_shufflevector(__ret_177, __ret_177, 3, 2, 1, 0); \
56851  __ret_177; \
56852})
56853#endif
56854
56855#ifdef __LITTLE_ENDIAN__
56856#define vqrshrn_high_n_u16(__p0_178, __p1_178, __p2_178) __extension__ ({ \
56857  uint8x8_t __s0_178 = __p0_178; \
56858  uint16x8_t __s1_178 = __p1_178; \
56859  uint8x16_t __ret_178; \
56860  __ret_178 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_178), (uint8x8_t)(vqrshrn_n_u16(__s1_178, __p2_178)))); \
56861  __ret_178; \
56862})
56863#else
56864#define vqrshrn_high_n_u16(__p0_179, __p1_179, __p2_179) __extension__ ({ \
56865  uint8x8_t __s0_179 = __p0_179; \
56866  uint16x8_t __s1_179 = __p1_179; \
56867  uint8x8_t __rev0_179;  __rev0_179 = __builtin_shufflevector(__s0_179, __s0_179, 7, 6, 5, 4, 3, 2, 1, 0); \
56868  uint16x8_t __rev1_179;  __rev1_179 = __builtin_shufflevector(__s1_179, __s1_179, 7, 6, 5, 4, 3, 2, 1, 0); \
56869  uint8x16_t __ret_179; \
56870  __ret_179 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_179), (uint8x8_t)(__noswap_vqrshrn_n_u16(__rev1_179, __p2_179)))); \
56871  __ret_179 = __builtin_shufflevector(__ret_179, __ret_179, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
56872  __ret_179; \
56873})
56874#endif
56875
56876#ifdef __LITTLE_ENDIAN__
56877#define vqrshrn_high_n_s32(__p0_180, __p1_180, __p2_180) __extension__ ({ \
56878  int16x4_t __s0_180 = __p0_180; \
56879  int32x4_t __s1_180 = __p1_180; \
56880  int16x8_t __ret_180; \
56881  __ret_180 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_180), (int16x4_t)(vqrshrn_n_s32(__s1_180, __p2_180)))); \
56882  __ret_180; \
56883})
56884#else
56885#define vqrshrn_high_n_s32(__p0_181, __p1_181, __p2_181) __extension__ ({ \
56886  int16x4_t __s0_181 = __p0_181; \
56887  int32x4_t __s1_181 = __p1_181; \
56888  int16x4_t __rev0_181;  __rev0_181 = __builtin_shufflevector(__s0_181, __s0_181, 3, 2, 1, 0); \
56889  int32x4_t __rev1_181;  __rev1_181 = __builtin_shufflevector(__s1_181, __s1_181, 3, 2, 1, 0); \
56890  int16x8_t __ret_181; \
56891  __ret_181 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_181), (int16x4_t)(__noswap_vqrshrn_n_s32(__rev1_181, __p2_181)))); \
56892  __ret_181 = __builtin_shufflevector(__ret_181, __ret_181, 7, 6, 5, 4, 3, 2, 1, 0); \
56893  __ret_181; \
56894})
56895#endif
56896
56897#ifdef __LITTLE_ENDIAN__
56898#define vqrshrn_high_n_s64(__p0_182, __p1_182, __p2_182) __extension__ ({ \
56899  int32x2_t __s0_182 = __p0_182; \
56900  int64x2_t __s1_182 = __p1_182; \
56901  int32x4_t __ret_182; \
56902  __ret_182 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_182), (int32x2_t)(vqrshrn_n_s64(__s1_182, __p2_182)))); \
56903  __ret_182; \
56904})
56905#else
56906#define vqrshrn_high_n_s64(__p0_183, __p1_183, __p2_183) __extension__ ({ \
56907  int32x2_t __s0_183 = __p0_183; \
56908  int64x2_t __s1_183 = __p1_183; \
56909  int32x2_t __rev0_183;  __rev0_183 = __builtin_shufflevector(__s0_183, __s0_183, 1, 0); \
56910  int64x2_t __rev1_183;  __rev1_183 = __builtin_shufflevector(__s1_183, __s1_183, 1, 0); \
56911  int32x4_t __ret_183; \
56912  __ret_183 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_183), (int32x2_t)(__noswap_vqrshrn_n_s64(__rev1_183, __p2_183)))); \
56913  __ret_183 = __builtin_shufflevector(__ret_183, __ret_183, 3, 2, 1, 0); \
56914  __ret_183; \
56915})
56916#endif
56917
56918#ifdef __LITTLE_ENDIAN__
56919#define vqrshrn_high_n_s16(__p0_184, __p1_184, __p2_184) __extension__ ({ \
56920  int8x8_t __s0_184 = __p0_184; \
56921  int16x8_t __s1_184 = __p1_184; \
56922  int8x16_t __ret_184; \
56923  __ret_184 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_184), (int8x8_t)(vqrshrn_n_s16(__s1_184, __p2_184)))); \
56924  __ret_184; \
56925})
56926#else
56927#define vqrshrn_high_n_s16(__p0_185, __p1_185, __p2_185) __extension__ ({ \
56928  int8x8_t __s0_185 = __p0_185; \
56929  int16x8_t __s1_185 = __p1_185; \
56930  int8x8_t __rev0_185;  __rev0_185 = __builtin_shufflevector(__s0_185, __s0_185, 7, 6, 5, 4, 3, 2, 1, 0); \
56931  int16x8_t __rev1_185;  __rev1_185 = __builtin_shufflevector(__s1_185, __s1_185, 7, 6, 5, 4, 3, 2, 1, 0); \
56932  int8x16_t __ret_185; \
56933  __ret_185 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_185), (int8x8_t)(__noswap_vqrshrn_n_s16(__rev1_185, __p2_185)))); \
56934  __ret_185 = __builtin_shufflevector(__ret_185, __ret_185, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
56935  __ret_185; \
56936})
56937#endif
56938
56939#ifdef __LITTLE_ENDIAN__
56940#define vqrshrns_n_u32(__p0, __p1) __extension__ ({ \
56941  uint32_t __s0 = __p0; \
56942  uint16_t __ret; \
56943  __ret = (uint16_t) __builtin_neon_vqrshrns_n_u32(__s0, __p1); \
56944  __ret; \
56945})
56946#else
56947#define vqrshrns_n_u32(__p0, __p1) __extension__ ({ \
56948  uint32_t __s0 = __p0; \
56949  uint16_t __ret; \
56950  __ret = (uint16_t) __builtin_neon_vqrshrns_n_u32(__s0, __p1); \
56951  __ret; \
56952})
56953#endif
56954
56955#ifdef __LITTLE_ENDIAN__
56956#define vqrshrnd_n_u64(__p0, __p1) __extension__ ({ \
56957  uint64_t __s0 = __p0; \
56958  uint32_t __ret; \
56959  __ret = (uint32_t) __builtin_neon_vqrshrnd_n_u64(__s0, __p1); \
56960  __ret; \
56961})
56962#else
56963#define vqrshrnd_n_u64(__p0, __p1) __extension__ ({ \
56964  uint64_t __s0 = __p0; \
56965  uint32_t __ret; \
56966  __ret = (uint32_t) __builtin_neon_vqrshrnd_n_u64(__s0, __p1); \
56967  __ret; \
56968})
56969#endif
56970
56971#ifdef __LITTLE_ENDIAN__
56972#define vqrshrnh_n_u16(__p0, __p1) __extension__ ({ \
56973  uint16_t __s0 = __p0; \
56974  uint8_t __ret; \
56975  __ret = (uint8_t) __builtin_neon_vqrshrnh_n_u16(__s0, __p1); \
56976  __ret; \
56977})
56978#else
56979#define vqrshrnh_n_u16(__p0, __p1) __extension__ ({ \
56980  uint16_t __s0 = __p0; \
56981  uint8_t __ret; \
56982  __ret = (uint8_t) __builtin_neon_vqrshrnh_n_u16(__s0, __p1); \
56983  __ret; \
56984})
56985#endif
56986
56987#ifdef __LITTLE_ENDIAN__
56988#define vqrshrns_n_s32(__p0, __p1) __extension__ ({ \
56989  int32_t __s0 = __p0; \
56990  int16_t __ret; \
56991  __ret = (int16_t) __builtin_neon_vqrshrns_n_s32(__s0, __p1); \
56992  __ret; \
56993})
56994#else
56995#define vqrshrns_n_s32(__p0, __p1) __extension__ ({ \
56996  int32_t __s0 = __p0; \
56997  int16_t __ret; \
56998  __ret = (int16_t) __builtin_neon_vqrshrns_n_s32(__s0, __p1); \
56999  __ret; \
57000})
57001#endif
57002
57003#ifdef __LITTLE_ENDIAN__
57004#define vqrshrnd_n_s64(__p0, __p1) __extension__ ({ \
57005  int64_t __s0 = __p0; \
57006  int32_t __ret; \
57007  __ret = (int32_t) __builtin_neon_vqrshrnd_n_s64(__s0, __p1); \
57008  __ret; \
57009})
57010#else
57011#define vqrshrnd_n_s64(__p0, __p1) __extension__ ({ \
57012  int64_t __s0 = __p0; \
57013  int32_t __ret; \
57014  __ret = (int32_t) __builtin_neon_vqrshrnd_n_s64(__s0, __p1); \
57015  __ret; \
57016})
57017#endif
57018
57019#ifdef __LITTLE_ENDIAN__
57020#define vqrshrnh_n_s16(__p0, __p1) __extension__ ({ \
57021  int16_t __s0 = __p0; \
57022  int8_t __ret; \
57023  __ret = (int8_t) __builtin_neon_vqrshrnh_n_s16(__s0, __p1); \
57024  __ret; \
57025})
57026#else
57027#define vqrshrnh_n_s16(__p0, __p1) __extension__ ({ \
57028  int16_t __s0 = __p0; \
57029  int8_t __ret; \
57030  __ret = (int8_t) __builtin_neon_vqrshrnh_n_s16(__s0, __p1); \
57031  __ret; \
57032})
57033#endif
57034
57035#ifdef __LITTLE_ENDIAN__
57036#define vqrshrun_high_n_s32(__p0_186, __p1_186, __p2_186) __extension__ ({ \
57037  int16x4_t __s0_186 = __p0_186; \
57038  int32x4_t __s1_186 = __p1_186; \
57039  int16x8_t __ret_186; \
57040  __ret_186 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_186), (int16x4_t)(vqrshrun_n_s32(__s1_186, __p2_186)))); \
57041  __ret_186; \
57042})
57043#else
57044#define vqrshrun_high_n_s32(__p0_187, __p1_187, __p2_187) __extension__ ({ \
57045  int16x4_t __s0_187 = __p0_187; \
57046  int32x4_t __s1_187 = __p1_187; \
57047  int16x4_t __rev0_187;  __rev0_187 = __builtin_shufflevector(__s0_187, __s0_187, 3, 2, 1, 0); \
57048  int32x4_t __rev1_187;  __rev1_187 = __builtin_shufflevector(__s1_187, __s1_187, 3, 2, 1, 0); \
57049  int16x8_t __ret_187; \
57050  __ret_187 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_187), (int16x4_t)(__noswap_vqrshrun_n_s32(__rev1_187, __p2_187)))); \
57051  __ret_187 = __builtin_shufflevector(__ret_187, __ret_187, 7, 6, 5, 4, 3, 2, 1, 0); \
57052  __ret_187; \
57053})
57054#endif
57055
57056#ifdef __LITTLE_ENDIAN__
57057#define vqrshrun_high_n_s64(__p0_188, __p1_188, __p2_188) __extension__ ({ \
57058  int32x2_t __s0_188 = __p0_188; \
57059  int64x2_t __s1_188 = __p1_188; \
57060  int32x4_t __ret_188; \
57061  __ret_188 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_188), (int32x2_t)(vqrshrun_n_s64(__s1_188, __p2_188)))); \
57062  __ret_188; \
57063})
57064#else
57065#define vqrshrun_high_n_s64(__p0_189, __p1_189, __p2_189) __extension__ ({ \
57066  int32x2_t __s0_189 = __p0_189; \
57067  int64x2_t __s1_189 = __p1_189; \
57068  int32x2_t __rev0_189;  __rev0_189 = __builtin_shufflevector(__s0_189, __s0_189, 1, 0); \
57069  int64x2_t __rev1_189;  __rev1_189 = __builtin_shufflevector(__s1_189, __s1_189, 1, 0); \
57070  int32x4_t __ret_189; \
57071  __ret_189 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_189), (int32x2_t)(__noswap_vqrshrun_n_s64(__rev1_189, __p2_189)))); \
57072  __ret_189 = __builtin_shufflevector(__ret_189, __ret_189, 3, 2, 1, 0); \
57073  __ret_189; \
57074})
57075#endif
57076
57077#ifdef __LITTLE_ENDIAN__
57078#define vqrshrun_high_n_s16(__p0_190, __p1_190, __p2_190) __extension__ ({ \
57079  int8x8_t __s0_190 = __p0_190; \
57080  int16x8_t __s1_190 = __p1_190; \
57081  int8x16_t __ret_190; \
57082  __ret_190 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_190), (int8x8_t)(vqrshrun_n_s16(__s1_190, __p2_190)))); \
57083  __ret_190; \
57084})
57085#else
57086#define vqrshrun_high_n_s16(__p0_191, __p1_191, __p2_191) __extension__ ({ \
57087  int8x8_t __s0_191 = __p0_191; \
57088  int16x8_t __s1_191 = __p1_191; \
57089  int8x8_t __rev0_191;  __rev0_191 = __builtin_shufflevector(__s0_191, __s0_191, 7, 6, 5, 4, 3, 2, 1, 0); \
57090  int16x8_t __rev1_191;  __rev1_191 = __builtin_shufflevector(__s1_191, __s1_191, 7, 6, 5, 4, 3, 2, 1, 0); \
57091  int8x16_t __ret_191; \
57092  __ret_191 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_191), (int8x8_t)(__noswap_vqrshrun_n_s16(__rev1_191, __p2_191)))); \
57093  __ret_191 = __builtin_shufflevector(__ret_191, __ret_191, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
57094  __ret_191; \
57095})
57096#endif
57097
57098#ifdef __LITTLE_ENDIAN__
57099#define vqrshruns_n_s32(__p0, __p1) __extension__ ({ \
57100  int32_t __s0 = __p0; \
57101  int16_t __ret; \
57102  __ret = (int16_t) __builtin_neon_vqrshruns_n_s32(__s0, __p1); \
57103  __ret; \
57104})
57105#else
57106#define vqrshruns_n_s32(__p0, __p1) __extension__ ({ \
57107  int32_t __s0 = __p0; \
57108  int16_t __ret; \
57109  __ret = (int16_t) __builtin_neon_vqrshruns_n_s32(__s0, __p1); \
57110  __ret; \
57111})
57112#endif
57113
57114#ifdef __LITTLE_ENDIAN__
57115#define vqrshrund_n_s64(__p0, __p1) __extension__ ({ \
57116  int64_t __s0 = __p0; \
57117  int32_t __ret; \
57118  __ret = (int32_t) __builtin_neon_vqrshrund_n_s64(__s0, __p1); \
57119  __ret; \
57120})
57121#else
57122#define vqrshrund_n_s64(__p0, __p1) __extension__ ({ \
57123  int64_t __s0 = __p0; \
57124  int32_t __ret; \
57125  __ret = (int32_t) __builtin_neon_vqrshrund_n_s64(__s0, __p1); \
57126  __ret; \
57127})
57128#endif
57129
57130#ifdef __LITTLE_ENDIAN__
57131#define vqrshrunh_n_s16(__p0, __p1) __extension__ ({ \
57132  int16_t __s0 = __p0; \
57133  int8_t __ret; \
57134  __ret = (int8_t) __builtin_neon_vqrshrunh_n_s16(__s0, __p1); \
57135  __ret; \
57136})
57137#else
57138#define vqrshrunh_n_s16(__p0, __p1) __extension__ ({ \
57139  int16_t __s0 = __p0; \
57140  int8_t __ret; \
57141  __ret = (int8_t) __builtin_neon_vqrshrunh_n_s16(__s0, __p1); \
57142  __ret; \
57143})
57144#endif
57145
57146#ifdef __LITTLE_ENDIAN__
57147__ai uint8_t vqshlb_u8(uint8_t __p0, uint8_t __p1) {
57148  uint8_t __ret;
57149  __ret = (uint8_t) __builtin_neon_vqshlb_u8(__p0, __p1);
57150  return __ret;
57151}
57152#else
57153__ai uint8_t vqshlb_u8(uint8_t __p0, uint8_t __p1) {
57154  uint8_t __ret;
57155  __ret = (uint8_t) __builtin_neon_vqshlb_u8(__p0, __p1);
57156  return __ret;
57157}
57158#endif
57159
57160#ifdef __LITTLE_ENDIAN__
57161__ai uint32_t vqshls_u32(uint32_t __p0, uint32_t __p1) {
57162  uint32_t __ret;
57163  __ret = (uint32_t) __builtin_neon_vqshls_u32(__p0, __p1);
57164  return __ret;
57165}
57166#else
57167__ai uint32_t vqshls_u32(uint32_t __p0, uint32_t __p1) {
57168  uint32_t __ret;
57169  __ret = (uint32_t) __builtin_neon_vqshls_u32(__p0, __p1);
57170  return __ret;
57171}
57172#endif
57173
57174#ifdef __LITTLE_ENDIAN__
57175__ai uint64_t vqshld_u64(uint64_t __p0, uint64_t __p1) {
57176  uint64_t __ret;
57177  __ret = (uint64_t) __builtin_neon_vqshld_u64(__p0, __p1);
57178  return __ret;
57179}
57180#else
57181__ai uint64_t vqshld_u64(uint64_t __p0, uint64_t __p1) {
57182  uint64_t __ret;
57183  __ret = (uint64_t) __builtin_neon_vqshld_u64(__p0, __p1);
57184  return __ret;
57185}
57186#endif
57187
57188#ifdef __LITTLE_ENDIAN__
57189__ai uint16_t vqshlh_u16(uint16_t __p0, uint16_t __p1) {
57190  uint16_t __ret;
57191  __ret = (uint16_t) __builtin_neon_vqshlh_u16(__p0, __p1);
57192  return __ret;
57193}
57194#else
57195__ai uint16_t vqshlh_u16(uint16_t __p0, uint16_t __p1) {
57196  uint16_t __ret;
57197  __ret = (uint16_t) __builtin_neon_vqshlh_u16(__p0, __p1);
57198  return __ret;
57199}
57200#endif
57201
57202#ifdef __LITTLE_ENDIAN__
57203__ai int8_t vqshlb_s8(int8_t __p0, int8_t __p1) {
57204  int8_t __ret;
57205  __ret = (int8_t) __builtin_neon_vqshlb_s8(__p0, __p1);
57206  return __ret;
57207}
57208#else
57209__ai int8_t vqshlb_s8(int8_t __p0, int8_t __p1) {
57210  int8_t __ret;
57211  __ret = (int8_t) __builtin_neon_vqshlb_s8(__p0, __p1);
57212  return __ret;
57213}
57214#endif
57215
57216#ifdef __LITTLE_ENDIAN__
57217__ai int32_t vqshls_s32(int32_t __p0, int32_t __p1) {
57218  int32_t __ret;
57219  __ret = (int32_t) __builtin_neon_vqshls_s32(__p0, __p1);
57220  return __ret;
57221}
57222#else
57223__ai int32_t vqshls_s32(int32_t __p0, int32_t __p1) {
57224  int32_t __ret;
57225  __ret = (int32_t) __builtin_neon_vqshls_s32(__p0, __p1);
57226  return __ret;
57227}
57228#endif
57229
57230#ifdef __LITTLE_ENDIAN__
57231__ai int64_t vqshld_s64(int64_t __p0, int64_t __p1) {
57232  int64_t __ret;
57233  __ret = (int64_t) __builtin_neon_vqshld_s64(__p0, __p1);
57234  return __ret;
57235}
57236#else
57237__ai int64_t vqshld_s64(int64_t __p0, int64_t __p1) {
57238  int64_t __ret;
57239  __ret = (int64_t) __builtin_neon_vqshld_s64(__p0, __p1);
57240  return __ret;
57241}
57242#endif
57243
57244#ifdef __LITTLE_ENDIAN__
57245__ai int16_t vqshlh_s16(int16_t __p0, int16_t __p1) {
57246  int16_t __ret;
57247  __ret = (int16_t) __builtin_neon_vqshlh_s16(__p0, __p1);
57248  return __ret;
57249}
57250#else
57251__ai int16_t vqshlh_s16(int16_t __p0, int16_t __p1) {
57252  int16_t __ret;
57253  __ret = (int16_t) __builtin_neon_vqshlh_s16(__p0, __p1);
57254  return __ret;
57255}
57256#endif
57257
57258#ifdef __LITTLE_ENDIAN__
57259#define vqshlb_n_u8(__p0, __p1) __extension__ ({ \
57260  uint8_t __s0 = __p0; \
57261  uint8_t __ret; \
57262  __ret = (uint8_t) __builtin_neon_vqshlb_n_u8(__s0, __p1); \
57263  __ret; \
57264})
57265#else
57266#define vqshlb_n_u8(__p0, __p1) __extension__ ({ \
57267  uint8_t __s0 = __p0; \
57268  uint8_t __ret; \
57269  __ret = (uint8_t) __builtin_neon_vqshlb_n_u8(__s0, __p1); \
57270  __ret; \
57271})
57272#endif
57273
57274#ifdef __LITTLE_ENDIAN__
57275#define vqshls_n_u32(__p0, __p1) __extension__ ({ \
57276  uint32_t __s0 = __p0; \
57277  uint32_t __ret; \
57278  __ret = (uint32_t) __builtin_neon_vqshls_n_u32(__s0, __p1); \
57279  __ret; \
57280})
57281#else
57282#define vqshls_n_u32(__p0, __p1) __extension__ ({ \
57283  uint32_t __s0 = __p0; \
57284  uint32_t __ret; \
57285  __ret = (uint32_t) __builtin_neon_vqshls_n_u32(__s0, __p1); \
57286  __ret; \
57287})
57288#endif
57289
57290#ifdef __LITTLE_ENDIAN__
57291#define vqshld_n_u64(__p0, __p1) __extension__ ({ \
57292  uint64_t __s0 = __p0; \
57293  uint64_t __ret; \
57294  __ret = (uint64_t) __builtin_neon_vqshld_n_u64(__s0, __p1); \
57295  __ret; \
57296})
57297#else
57298#define vqshld_n_u64(__p0, __p1) __extension__ ({ \
57299  uint64_t __s0 = __p0; \
57300  uint64_t __ret; \
57301  __ret = (uint64_t) __builtin_neon_vqshld_n_u64(__s0, __p1); \
57302  __ret; \
57303})
57304#endif
57305
57306#ifdef __LITTLE_ENDIAN__
57307#define vqshlh_n_u16(__p0, __p1) __extension__ ({ \
57308  uint16_t __s0 = __p0; \
57309  uint16_t __ret; \
57310  __ret = (uint16_t) __builtin_neon_vqshlh_n_u16(__s0, __p1); \
57311  __ret; \
57312})
57313#else
57314#define vqshlh_n_u16(__p0, __p1) __extension__ ({ \
57315  uint16_t __s0 = __p0; \
57316  uint16_t __ret; \
57317  __ret = (uint16_t) __builtin_neon_vqshlh_n_u16(__s0, __p1); \
57318  __ret; \
57319})
57320#endif
57321
57322#ifdef __LITTLE_ENDIAN__
57323#define vqshlb_n_s8(__p0, __p1) __extension__ ({ \
57324  int8_t __s0 = __p0; \
57325  int8_t __ret; \
57326  __ret = (int8_t) __builtin_neon_vqshlb_n_s8(__s0, __p1); \
57327  __ret; \
57328})
57329#else
57330#define vqshlb_n_s8(__p0, __p1) __extension__ ({ \
57331  int8_t __s0 = __p0; \
57332  int8_t __ret; \
57333  __ret = (int8_t) __builtin_neon_vqshlb_n_s8(__s0, __p1); \
57334  __ret; \
57335})
57336#endif
57337
57338#ifdef __LITTLE_ENDIAN__
57339#define vqshls_n_s32(__p0, __p1) __extension__ ({ \
57340  int32_t __s0 = __p0; \
57341  int32_t __ret; \
57342  __ret = (int32_t) __builtin_neon_vqshls_n_s32(__s0, __p1); \
57343  __ret; \
57344})
57345#else
57346#define vqshls_n_s32(__p0, __p1) __extension__ ({ \
57347  int32_t __s0 = __p0; \
57348  int32_t __ret; \
57349  __ret = (int32_t) __builtin_neon_vqshls_n_s32(__s0, __p1); \
57350  __ret; \
57351})
57352#endif
57353
57354#ifdef __LITTLE_ENDIAN__
57355#define vqshld_n_s64(__p0, __p1) __extension__ ({ \
57356  int64_t __s0 = __p0; \
57357  int64_t __ret; \
57358  __ret = (int64_t) __builtin_neon_vqshld_n_s64(__s0, __p1); \
57359  __ret; \
57360})
57361#else
57362#define vqshld_n_s64(__p0, __p1) __extension__ ({ \
57363  int64_t __s0 = __p0; \
57364  int64_t __ret; \
57365  __ret = (int64_t) __builtin_neon_vqshld_n_s64(__s0, __p1); \
57366  __ret; \
57367})
57368#endif
57369
57370#ifdef __LITTLE_ENDIAN__
57371#define vqshlh_n_s16(__p0, __p1) __extension__ ({ \
57372  int16_t __s0 = __p0; \
57373  int16_t __ret; \
57374  __ret = (int16_t) __builtin_neon_vqshlh_n_s16(__s0, __p1); \
57375  __ret; \
57376})
57377#else
57378#define vqshlh_n_s16(__p0, __p1) __extension__ ({ \
57379  int16_t __s0 = __p0; \
57380  int16_t __ret; \
57381  __ret = (int16_t) __builtin_neon_vqshlh_n_s16(__s0, __p1); \
57382  __ret; \
57383})
57384#endif
57385
57386#ifdef __LITTLE_ENDIAN__
57387#define vqshlub_n_s8(__p0, __p1) __extension__ ({ \
57388  int8_t __s0 = __p0; \
57389  int8_t __ret; \
57390  __ret = (int8_t) __builtin_neon_vqshlub_n_s8(__s0, __p1); \
57391  __ret; \
57392})
57393#else
57394#define vqshlub_n_s8(__p0, __p1) __extension__ ({ \
57395  int8_t __s0 = __p0; \
57396  int8_t __ret; \
57397  __ret = (int8_t) __builtin_neon_vqshlub_n_s8(__s0, __p1); \
57398  __ret; \
57399})
57400#endif
57401
57402#ifdef __LITTLE_ENDIAN__
57403#define vqshlus_n_s32(__p0, __p1) __extension__ ({ \
57404  int32_t __s0 = __p0; \
57405  int32_t __ret; \
57406  __ret = (int32_t) __builtin_neon_vqshlus_n_s32(__s0, __p1); \
57407  __ret; \
57408})
57409#else
57410#define vqshlus_n_s32(__p0, __p1) __extension__ ({ \
57411  int32_t __s0 = __p0; \
57412  int32_t __ret; \
57413  __ret = (int32_t) __builtin_neon_vqshlus_n_s32(__s0, __p1); \
57414  __ret; \
57415})
57416#endif
57417
57418#ifdef __LITTLE_ENDIAN__
57419#define vqshlud_n_s64(__p0, __p1) __extension__ ({ \
57420  int64_t __s0 = __p0; \
57421  int64_t __ret; \
57422  __ret = (int64_t) __builtin_neon_vqshlud_n_s64(__s0, __p1); \
57423  __ret; \
57424})
57425#else
57426#define vqshlud_n_s64(__p0, __p1) __extension__ ({ \
57427  int64_t __s0 = __p0; \
57428  int64_t __ret; \
57429  __ret = (int64_t) __builtin_neon_vqshlud_n_s64(__s0, __p1); \
57430  __ret; \
57431})
57432#endif
57433
57434#ifdef __LITTLE_ENDIAN__
57435#define vqshluh_n_s16(__p0, __p1) __extension__ ({ \
57436  int16_t __s0 = __p0; \
57437  int16_t __ret; \
57438  __ret = (int16_t) __builtin_neon_vqshluh_n_s16(__s0, __p1); \
57439  __ret; \
57440})
57441#else
57442#define vqshluh_n_s16(__p0, __p1) __extension__ ({ \
57443  int16_t __s0 = __p0; \
57444  int16_t __ret; \
57445  __ret = (int16_t) __builtin_neon_vqshluh_n_s16(__s0, __p1); \
57446  __ret; \
57447})
57448#endif
57449
57450#ifdef __LITTLE_ENDIAN__
57451#define vqshrn_high_n_u32(__p0_192, __p1_192, __p2_192) __extension__ ({ \
57452  uint16x4_t __s0_192 = __p0_192; \
57453  uint32x4_t __s1_192 = __p1_192; \
57454  uint16x8_t __ret_192; \
57455  __ret_192 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_192), (uint16x4_t)(vqshrn_n_u32(__s1_192, __p2_192)))); \
57456  __ret_192; \
57457})
57458#else
57459#define vqshrn_high_n_u32(__p0_193, __p1_193, __p2_193) __extension__ ({ \
57460  uint16x4_t __s0_193 = __p0_193; \
57461  uint32x4_t __s1_193 = __p1_193; \
57462  uint16x4_t __rev0_193;  __rev0_193 = __builtin_shufflevector(__s0_193, __s0_193, 3, 2, 1, 0); \
57463  uint32x4_t __rev1_193;  __rev1_193 = __builtin_shufflevector(__s1_193, __s1_193, 3, 2, 1, 0); \
57464  uint16x8_t __ret_193; \
57465  __ret_193 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_193), (uint16x4_t)(__noswap_vqshrn_n_u32(__rev1_193, __p2_193)))); \
57466  __ret_193 = __builtin_shufflevector(__ret_193, __ret_193, 7, 6, 5, 4, 3, 2, 1, 0); \
57467  __ret_193; \
57468})
57469#endif
57470
57471#ifdef __LITTLE_ENDIAN__
57472#define vqshrn_high_n_u64(__p0_194, __p1_194, __p2_194) __extension__ ({ \
57473  uint32x2_t __s0_194 = __p0_194; \
57474  uint64x2_t __s1_194 = __p1_194; \
57475  uint32x4_t __ret_194; \
57476  __ret_194 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_194), (uint32x2_t)(vqshrn_n_u64(__s1_194, __p2_194)))); \
57477  __ret_194; \
57478})
57479#else
57480#define vqshrn_high_n_u64(__p0_195, __p1_195, __p2_195) __extension__ ({ \
57481  uint32x2_t __s0_195 = __p0_195; \
57482  uint64x2_t __s1_195 = __p1_195; \
57483  uint32x2_t __rev0_195;  __rev0_195 = __builtin_shufflevector(__s0_195, __s0_195, 1, 0); \
57484  uint64x2_t __rev1_195;  __rev1_195 = __builtin_shufflevector(__s1_195, __s1_195, 1, 0); \
57485  uint32x4_t __ret_195; \
57486  __ret_195 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_195), (uint32x2_t)(__noswap_vqshrn_n_u64(__rev1_195, __p2_195)))); \
57487  __ret_195 = __builtin_shufflevector(__ret_195, __ret_195, 3, 2, 1, 0); \
57488  __ret_195; \
57489})
57490#endif
57491
57492#ifdef __LITTLE_ENDIAN__
57493#define vqshrn_high_n_u16(__p0_196, __p1_196, __p2_196) __extension__ ({ \
57494  uint8x8_t __s0_196 = __p0_196; \
57495  uint16x8_t __s1_196 = __p1_196; \
57496  uint8x16_t __ret_196; \
57497  __ret_196 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_196), (uint8x8_t)(vqshrn_n_u16(__s1_196, __p2_196)))); \
57498  __ret_196; \
57499})
57500#else
57501#define vqshrn_high_n_u16(__p0_197, __p1_197, __p2_197) __extension__ ({ \
57502  uint8x8_t __s0_197 = __p0_197; \
57503  uint16x8_t __s1_197 = __p1_197; \
57504  uint8x8_t __rev0_197;  __rev0_197 = __builtin_shufflevector(__s0_197, __s0_197, 7, 6, 5, 4, 3, 2, 1, 0); \
57505  uint16x8_t __rev1_197;  __rev1_197 = __builtin_shufflevector(__s1_197, __s1_197, 7, 6, 5, 4, 3, 2, 1, 0); \
57506  uint8x16_t __ret_197; \
57507  __ret_197 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_197), (uint8x8_t)(__noswap_vqshrn_n_u16(__rev1_197, __p2_197)))); \
57508  __ret_197 = __builtin_shufflevector(__ret_197, __ret_197, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
57509  __ret_197; \
57510})
57511#endif
57512
57513#ifdef __LITTLE_ENDIAN__
57514#define vqshrn_high_n_s32(__p0_198, __p1_198, __p2_198) __extension__ ({ \
57515  int16x4_t __s0_198 = __p0_198; \
57516  int32x4_t __s1_198 = __p1_198; \
57517  int16x8_t __ret_198; \
57518  __ret_198 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_198), (int16x4_t)(vqshrn_n_s32(__s1_198, __p2_198)))); \
57519  __ret_198; \
57520})
57521#else
57522#define vqshrn_high_n_s32(__p0_199, __p1_199, __p2_199) __extension__ ({ \
57523  int16x4_t __s0_199 = __p0_199; \
57524  int32x4_t __s1_199 = __p1_199; \
57525  int16x4_t __rev0_199;  __rev0_199 = __builtin_shufflevector(__s0_199, __s0_199, 3, 2, 1, 0); \
57526  int32x4_t __rev1_199;  __rev1_199 = __builtin_shufflevector(__s1_199, __s1_199, 3, 2, 1, 0); \
57527  int16x8_t __ret_199; \
57528  __ret_199 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_199), (int16x4_t)(__noswap_vqshrn_n_s32(__rev1_199, __p2_199)))); \
57529  __ret_199 = __builtin_shufflevector(__ret_199, __ret_199, 7, 6, 5, 4, 3, 2, 1, 0); \
57530  __ret_199; \
57531})
57532#endif
57533
57534#ifdef __LITTLE_ENDIAN__
57535#define vqshrn_high_n_s64(__p0_200, __p1_200, __p2_200) __extension__ ({ \
57536  int32x2_t __s0_200 = __p0_200; \
57537  int64x2_t __s1_200 = __p1_200; \
57538  int32x4_t __ret_200; \
57539  __ret_200 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_200), (int32x2_t)(vqshrn_n_s64(__s1_200, __p2_200)))); \
57540  __ret_200; \
57541})
57542#else
57543#define vqshrn_high_n_s64(__p0_201, __p1_201, __p2_201) __extension__ ({ \
57544  int32x2_t __s0_201 = __p0_201; \
57545  int64x2_t __s1_201 = __p1_201; \
57546  int32x2_t __rev0_201;  __rev0_201 = __builtin_shufflevector(__s0_201, __s0_201, 1, 0); \
57547  int64x2_t __rev1_201;  __rev1_201 = __builtin_shufflevector(__s1_201, __s1_201, 1, 0); \
57548  int32x4_t __ret_201; \
57549  __ret_201 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_201), (int32x2_t)(__noswap_vqshrn_n_s64(__rev1_201, __p2_201)))); \
57550  __ret_201 = __builtin_shufflevector(__ret_201, __ret_201, 3, 2, 1, 0); \
57551  __ret_201; \
57552})
57553#endif
57554
57555#ifdef __LITTLE_ENDIAN__
57556#define vqshrn_high_n_s16(__p0_202, __p1_202, __p2_202) __extension__ ({ \
57557  int8x8_t __s0_202 = __p0_202; \
57558  int16x8_t __s1_202 = __p1_202; \
57559  int8x16_t __ret_202; \
57560  __ret_202 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_202), (int8x8_t)(vqshrn_n_s16(__s1_202, __p2_202)))); \
57561  __ret_202; \
57562})
57563#else
57564#define vqshrn_high_n_s16(__p0_203, __p1_203, __p2_203) __extension__ ({ \
57565  int8x8_t __s0_203 = __p0_203; \
57566  int16x8_t __s1_203 = __p1_203; \
57567  int8x8_t __rev0_203;  __rev0_203 = __builtin_shufflevector(__s0_203, __s0_203, 7, 6, 5, 4, 3, 2, 1, 0); \
57568  int16x8_t __rev1_203;  __rev1_203 = __builtin_shufflevector(__s1_203, __s1_203, 7, 6, 5, 4, 3, 2, 1, 0); \
57569  int8x16_t __ret_203; \
57570  __ret_203 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_203), (int8x8_t)(__noswap_vqshrn_n_s16(__rev1_203, __p2_203)))); \
57571  __ret_203 = __builtin_shufflevector(__ret_203, __ret_203, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
57572  __ret_203; \
57573})
57574#endif
57575
57576#ifdef __LITTLE_ENDIAN__
57577#define vqshrns_n_u32(__p0, __p1) __extension__ ({ \
57578  uint32_t __s0 = __p0; \
57579  uint16_t __ret; \
57580  __ret = (uint16_t) __builtin_neon_vqshrns_n_u32(__s0, __p1); \
57581  __ret; \
57582})
57583#else
57584#define vqshrns_n_u32(__p0, __p1) __extension__ ({ \
57585  uint32_t __s0 = __p0; \
57586  uint16_t __ret; \
57587  __ret = (uint16_t) __builtin_neon_vqshrns_n_u32(__s0, __p1); \
57588  __ret; \
57589})
57590#endif
57591
57592#ifdef __LITTLE_ENDIAN__
57593#define vqshrnd_n_u64(__p0, __p1) __extension__ ({ \
57594  uint64_t __s0 = __p0; \
57595  uint32_t __ret; \
57596  __ret = (uint32_t) __builtin_neon_vqshrnd_n_u64(__s0, __p1); \
57597  __ret; \
57598})
57599#else
57600#define vqshrnd_n_u64(__p0, __p1) __extension__ ({ \
57601  uint64_t __s0 = __p0; \
57602  uint32_t __ret; \
57603  __ret = (uint32_t) __builtin_neon_vqshrnd_n_u64(__s0, __p1); \
57604  __ret; \
57605})
57606#endif
57607
57608#ifdef __LITTLE_ENDIAN__
57609#define vqshrnh_n_u16(__p0, __p1) __extension__ ({ \
57610  uint16_t __s0 = __p0; \
57611  uint8_t __ret; \
57612  __ret = (uint8_t) __builtin_neon_vqshrnh_n_u16(__s0, __p1); \
57613  __ret; \
57614})
57615#else
57616#define vqshrnh_n_u16(__p0, __p1) __extension__ ({ \
57617  uint16_t __s0 = __p0; \
57618  uint8_t __ret; \
57619  __ret = (uint8_t) __builtin_neon_vqshrnh_n_u16(__s0, __p1); \
57620  __ret; \
57621})
57622#endif
57623
57624#ifdef __LITTLE_ENDIAN__
57625#define vqshrns_n_s32(__p0, __p1) __extension__ ({ \
57626  int32_t __s0 = __p0; \
57627  int16_t __ret; \
57628  __ret = (int16_t) __builtin_neon_vqshrns_n_s32(__s0, __p1); \
57629  __ret; \
57630})
57631#else
57632#define vqshrns_n_s32(__p0, __p1) __extension__ ({ \
57633  int32_t __s0 = __p0; \
57634  int16_t __ret; \
57635  __ret = (int16_t) __builtin_neon_vqshrns_n_s32(__s0, __p1); \
57636  __ret; \
57637})
57638#endif
57639
57640#ifdef __LITTLE_ENDIAN__
57641#define vqshrnd_n_s64(__p0, __p1) __extension__ ({ \
57642  int64_t __s0 = __p0; \
57643  int32_t __ret; \
57644  __ret = (int32_t) __builtin_neon_vqshrnd_n_s64(__s0, __p1); \
57645  __ret; \
57646})
57647#else
57648#define vqshrnd_n_s64(__p0, __p1) __extension__ ({ \
57649  int64_t __s0 = __p0; \
57650  int32_t __ret; \
57651  __ret = (int32_t) __builtin_neon_vqshrnd_n_s64(__s0, __p1); \
57652  __ret; \
57653})
57654#endif
57655
57656#ifdef __LITTLE_ENDIAN__
57657#define vqshrnh_n_s16(__p0, __p1) __extension__ ({ \
57658  int16_t __s0 = __p0; \
57659  int8_t __ret; \
57660  __ret = (int8_t) __builtin_neon_vqshrnh_n_s16(__s0, __p1); \
57661  __ret; \
57662})
57663#else
57664#define vqshrnh_n_s16(__p0, __p1) __extension__ ({ \
57665  int16_t __s0 = __p0; \
57666  int8_t __ret; \
57667  __ret = (int8_t) __builtin_neon_vqshrnh_n_s16(__s0, __p1); \
57668  __ret; \
57669})
57670#endif
57671
57672#ifdef __LITTLE_ENDIAN__
57673#define vqshrun_high_n_s32(__p0_204, __p1_204, __p2_204) __extension__ ({ \
57674  int16x4_t __s0_204 = __p0_204; \
57675  int32x4_t __s1_204 = __p1_204; \
57676  int16x8_t __ret_204; \
57677  __ret_204 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_204), (int16x4_t)(vqshrun_n_s32(__s1_204, __p2_204)))); \
57678  __ret_204; \
57679})
57680#else
57681#define vqshrun_high_n_s32(__p0_205, __p1_205, __p2_205) __extension__ ({ \
57682  int16x4_t __s0_205 = __p0_205; \
57683  int32x4_t __s1_205 = __p1_205; \
57684  int16x4_t __rev0_205;  __rev0_205 = __builtin_shufflevector(__s0_205, __s0_205, 3, 2, 1, 0); \
57685  int32x4_t __rev1_205;  __rev1_205 = __builtin_shufflevector(__s1_205, __s1_205, 3, 2, 1, 0); \
57686  int16x8_t __ret_205; \
57687  __ret_205 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_205), (int16x4_t)(__noswap_vqshrun_n_s32(__rev1_205, __p2_205)))); \
57688  __ret_205 = __builtin_shufflevector(__ret_205, __ret_205, 7, 6, 5, 4, 3, 2, 1, 0); \
57689  __ret_205; \
57690})
57691#endif
57692
57693#ifdef __LITTLE_ENDIAN__
57694#define vqshrun_high_n_s64(__p0_206, __p1_206, __p2_206) __extension__ ({ \
57695  int32x2_t __s0_206 = __p0_206; \
57696  int64x2_t __s1_206 = __p1_206; \
57697  int32x4_t __ret_206; \
57698  __ret_206 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_206), (int32x2_t)(vqshrun_n_s64(__s1_206, __p2_206)))); \
57699  __ret_206; \
57700})
57701#else
57702#define vqshrun_high_n_s64(__p0_207, __p1_207, __p2_207) __extension__ ({ \
57703  int32x2_t __s0_207 = __p0_207; \
57704  int64x2_t __s1_207 = __p1_207; \
57705  int32x2_t __rev0_207;  __rev0_207 = __builtin_shufflevector(__s0_207, __s0_207, 1, 0); \
57706  int64x2_t __rev1_207;  __rev1_207 = __builtin_shufflevector(__s1_207, __s1_207, 1, 0); \
57707  int32x4_t __ret_207; \
57708  __ret_207 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_207), (int32x2_t)(__noswap_vqshrun_n_s64(__rev1_207, __p2_207)))); \
57709  __ret_207 = __builtin_shufflevector(__ret_207, __ret_207, 3, 2, 1, 0); \
57710  __ret_207; \
57711})
57712#endif
57713
57714#ifdef __LITTLE_ENDIAN__
57715#define vqshrun_high_n_s16(__p0_208, __p1_208, __p2_208) __extension__ ({ \
57716  int8x8_t __s0_208 = __p0_208; \
57717  int16x8_t __s1_208 = __p1_208; \
57718  int8x16_t __ret_208; \
57719  __ret_208 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_208), (int8x8_t)(vqshrun_n_s16(__s1_208, __p2_208)))); \
57720  __ret_208; \
57721})
57722#else
57723#define vqshrun_high_n_s16(__p0_209, __p1_209, __p2_209) __extension__ ({ \
57724  int8x8_t __s0_209 = __p0_209; \
57725  int16x8_t __s1_209 = __p1_209; \
57726  int8x8_t __rev0_209;  __rev0_209 = __builtin_shufflevector(__s0_209, __s0_209, 7, 6, 5, 4, 3, 2, 1, 0); \
57727  int16x8_t __rev1_209;  __rev1_209 = __builtin_shufflevector(__s1_209, __s1_209, 7, 6, 5, 4, 3, 2, 1, 0); \
57728  int8x16_t __ret_209; \
57729  __ret_209 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_209), (int8x8_t)(__noswap_vqshrun_n_s16(__rev1_209, __p2_209)))); \
57730  __ret_209 = __builtin_shufflevector(__ret_209, __ret_209, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
57731  __ret_209; \
57732})
57733#endif
57734
57735#ifdef __LITTLE_ENDIAN__
57736#define vqshruns_n_s32(__p0, __p1) __extension__ ({ \
57737  int32_t __s0 = __p0; \
57738  int16_t __ret; \
57739  __ret = (int16_t) __builtin_neon_vqshruns_n_s32(__s0, __p1); \
57740  __ret; \
57741})
57742#else
57743#define vqshruns_n_s32(__p0, __p1) __extension__ ({ \
57744  int32_t __s0 = __p0; \
57745  int16_t __ret; \
57746  __ret = (int16_t) __builtin_neon_vqshruns_n_s32(__s0, __p1); \
57747  __ret; \
57748})
57749#endif
57750
57751#ifdef __LITTLE_ENDIAN__
57752#define vqshrund_n_s64(__p0, __p1) __extension__ ({ \
57753  int64_t __s0 = __p0; \
57754  int32_t __ret; \
57755  __ret = (int32_t) __builtin_neon_vqshrund_n_s64(__s0, __p1); \
57756  __ret; \
57757})
57758#else
57759#define vqshrund_n_s64(__p0, __p1) __extension__ ({ \
57760  int64_t __s0 = __p0; \
57761  int32_t __ret; \
57762  __ret = (int32_t) __builtin_neon_vqshrund_n_s64(__s0, __p1); \
57763  __ret; \
57764})
57765#endif
57766
57767#ifdef __LITTLE_ENDIAN__
57768#define vqshrunh_n_s16(__p0, __p1) __extension__ ({ \
57769  int16_t __s0 = __p0; \
57770  int8_t __ret; \
57771  __ret = (int8_t) __builtin_neon_vqshrunh_n_s16(__s0, __p1); \
57772  __ret; \
57773})
57774#else
57775#define vqshrunh_n_s16(__p0, __p1) __extension__ ({ \
57776  int16_t __s0 = __p0; \
57777  int8_t __ret; \
57778  __ret = (int8_t) __builtin_neon_vqshrunh_n_s16(__s0, __p1); \
57779  __ret; \
57780})
57781#endif
57782
57783#ifdef __LITTLE_ENDIAN__
57784__ai uint8_t vqsubb_u8(uint8_t __p0, uint8_t __p1) {
57785  uint8_t __ret;
57786  __ret = (uint8_t) __builtin_neon_vqsubb_u8(__p0, __p1);
57787  return __ret;
57788}
57789#else
57790__ai uint8_t vqsubb_u8(uint8_t __p0, uint8_t __p1) {
57791  uint8_t __ret;
57792  __ret = (uint8_t) __builtin_neon_vqsubb_u8(__p0, __p1);
57793  return __ret;
57794}
57795#endif
57796
57797#ifdef __LITTLE_ENDIAN__
57798__ai uint32_t vqsubs_u32(uint32_t __p0, uint32_t __p1) {
57799  uint32_t __ret;
57800  __ret = (uint32_t) __builtin_neon_vqsubs_u32(__p0, __p1);
57801  return __ret;
57802}
57803#else
57804__ai uint32_t vqsubs_u32(uint32_t __p0, uint32_t __p1) {
57805  uint32_t __ret;
57806  __ret = (uint32_t) __builtin_neon_vqsubs_u32(__p0, __p1);
57807  return __ret;
57808}
57809#endif
57810
57811#ifdef __LITTLE_ENDIAN__
57812__ai uint64_t vqsubd_u64(uint64_t __p0, uint64_t __p1) {
57813  uint64_t __ret;
57814  __ret = (uint64_t) __builtin_neon_vqsubd_u64(__p0, __p1);
57815  return __ret;
57816}
57817#else
57818__ai uint64_t vqsubd_u64(uint64_t __p0, uint64_t __p1) {
57819  uint64_t __ret;
57820  __ret = (uint64_t) __builtin_neon_vqsubd_u64(__p0, __p1);
57821  return __ret;
57822}
57823#endif
57824
57825#ifdef __LITTLE_ENDIAN__
57826__ai uint16_t vqsubh_u16(uint16_t __p0, uint16_t __p1) {
57827  uint16_t __ret;
57828  __ret = (uint16_t) __builtin_neon_vqsubh_u16(__p0, __p1);
57829  return __ret;
57830}
57831#else
57832__ai uint16_t vqsubh_u16(uint16_t __p0, uint16_t __p1) {
57833  uint16_t __ret;
57834  __ret = (uint16_t) __builtin_neon_vqsubh_u16(__p0, __p1);
57835  return __ret;
57836}
57837#endif
57838
57839#ifdef __LITTLE_ENDIAN__
57840__ai int8_t vqsubb_s8(int8_t __p0, int8_t __p1) {
57841  int8_t __ret;
57842  __ret = (int8_t) __builtin_neon_vqsubb_s8(__p0, __p1);
57843  return __ret;
57844}
57845#else
57846__ai int8_t vqsubb_s8(int8_t __p0, int8_t __p1) {
57847  int8_t __ret;
57848  __ret = (int8_t) __builtin_neon_vqsubb_s8(__p0, __p1);
57849  return __ret;
57850}
57851#endif
57852
57853#ifdef __LITTLE_ENDIAN__
57854__ai int32_t vqsubs_s32(int32_t __p0, int32_t __p1) {
57855  int32_t __ret;
57856  __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
57857  return __ret;
57858}
57859#else
57860__ai int32_t vqsubs_s32(int32_t __p0, int32_t __p1) {
57861  int32_t __ret;
57862  __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
57863  return __ret;
57864}
57865#endif
57866
57867#ifdef __LITTLE_ENDIAN__
57868__ai int64_t vqsubd_s64(int64_t __p0, int64_t __p1) {
57869  int64_t __ret;
57870  __ret = (int64_t) __builtin_neon_vqsubd_s64(__p0, __p1);
57871  return __ret;
57872}
57873#else
57874__ai int64_t vqsubd_s64(int64_t __p0, int64_t __p1) {
57875  int64_t __ret;
57876  __ret = (int64_t) __builtin_neon_vqsubd_s64(__p0, __p1);
57877  return __ret;
57878}
57879#endif
57880
57881#ifdef __LITTLE_ENDIAN__
57882__ai int16_t vqsubh_s16(int16_t __p0, int16_t __p1) {
57883  int16_t __ret;
57884  __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
57885  return __ret;
57886}
57887#else
57888__ai int16_t vqsubh_s16(int16_t __p0, int16_t __p1) {
57889  int16_t __ret;
57890  __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
57891  return __ret;
57892}
57893#endif
57894
57895#ifdef __LITTLE_ENDIAN__
57896__ai poly8x8_t vqtbl1_p8(poly8x16_t __p0, uint8x8_t __p1) {
57897  poly8x8_t __ret;
57898  __ret = (poly8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 4);
57899  return __ret;
57900}
57901#else
57902__ai poly8x8_t vqtbl1_p8(poly8x16_t __p0, uint8x8_t __p1) {
57903  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57904  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57905  poly8x8_t __ret;
57906  __ret = (poly8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 4);
57907  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57908  return __ret;
57909}
57910#endif
57911
57912#ifdef __LITTLE_ENDIAN__
57913__ai poly8x16_t vqtbl1q_p8(poly8x16_t __p0, uint8x16_t __p1) {
57914  poly8x16_t __ret;
57915  __ret = (poly8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
57916  return __ret;
57917}
57918#else
57919__ai poly8x16_t vqtbl1q_p8(poly8x16_t __p0, uint8x16_t __p1) {
57920  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57921  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57922  poly8x16_t __ret;
57923  __ret = (poly8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
57924  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57925  return __ret;
57926}
57927#endif
57928
57929#ifdef __LITTLE_ENDIAN__
57930__ai uint8x16_t vqtbl1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
57931  uint8x16_t __ret;
57932  __ret = (uint8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
57933  return __ret;
57934}
57935#else
57936__ai uint8x16_t vqtbl1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
57937  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57938  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57939  uint8x16_t __ret;
57940  __ret = (uint8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
57941  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57942  return __ret;
57943}
57944#endif
57945
57946#ifdef __LITTLE_ENDIAN__
57947__ai int8x16_t vqtbl1q_s8(int8x16_t __p0, int8x16_t __p1) {
57948  int8x16_t __ret;
57949  __ret = (int8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
57950  return __ret;
57951}
57952#else
57953__ai int8x16_t vqtbl1q_s8(int8x16_t __p0, int8x16_t __p1) {
57954  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57955  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57956  int8x16_t __ret;
57957  __ret = (int8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
57958  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57959  return __ret;
57960}
57961#endif
57962
57963#ifdef __LITTLE_ENDIAN__
57964__ai uint8x8_t vqtbl1_u8(uint8x16_t __p0, uint8x8_t __p1) {
57965  uint8x8_t __ret;
57966  __ret = (uint8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 16);
57967  return __ret;
57968}
57969#else
57970__ai uint8x8_t vqtbl1_u8(uint8x16_t __p0, uint8x8_t __p1) {
57971  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57972  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57973  uint8x8_t __ret;
57974  __ret = (uint8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 16);
57975  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57976  return __ret;
57977}
57978#endif
57979
57980#ifdef __LITTLE_ENDIAN__
57981__ai int8x8_t vqtbl1_s8(int8x16_t __p0, int8x8_t __p1) {
57982  int8x8_t __ret;
57983  __ret = (int8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 0);
57984  return __ret;
57985}
57986#else
57987__ai int8x8_t vqtbl1_s8(int8x16_t __p0, int8x8_t __p1) {
57988  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57989  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57990  int8x8_t __ret;
57991  __ret = (int8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 0);
57992  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57993  return __ret;
57994}
57995#endif
57996
57997#ifdef __LITTLE_ENDIAN__
57998__ai poly8x8_t vqtbl2_p8(poly8x16x2_t __p0, uint8x8_t __p1) {
57999  poly8x8_t __ret;
58000  __ret = (poly8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 4);
58001  return __ret;
58002}
58003#else
58004__ai poly8x8_t vqtbl2_p8(poly8x16x2_t __p0, uint8x8_t __p1) {
58005  poly8x16x2_t __rev0;
58006  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58007  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58008  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58009  poly8x8_t __ret;
58010  __ret = (poly8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 4);
58011  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58012  return __ret;
58013}
58014#endif
58015
58016#ifdef __LITTLE_ENDIAN__
58017__ai poly8x16_t vqtbl2q_p8(poly8x16x2_t __p0, uint8x16_t __p1) {
58018  poly8x16_t __ret;
58019  __ret = (poly8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 36);
58020  return __ret;
58021}
58022#else
58023__ai poly8x16_t vqtbl2q_p8(poly8x16x2_t __p0, uint8x16_t __p1) {
58024  poly8x16x2_t __rev0;
58025  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58026  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58027  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58028  poly8x16_t __ret;
58029  __ret = (poly8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 36);
58030  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58031  return __ret;
58032}
58033#endif
58034
58035#ifdef __LITTLE_ENDIAN__
58036__ai uint8x16_t vqtbl2q_u8(uint8x16x2_t __p0, uint8x16_t __p1) {
58037  uint8x16_t __ret;
58038  __ret = (uint8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 48);
58039  return __ret;
58040}
58041#else
58042__ai uint8x16_t vqtbl2q_u8(uint8x16x2_t __p0, uint8x16_t __p1) {
58043  uint8x16x2_t __rev0;
58044  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58045  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58046  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58047  uint8x16_t __ret;
58048  __ret = (uint8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 48);
58049  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58050  return __ret;
58051}
58052#endif
58053
58054#ifdef __LITTLE_ENDIAN__
58055__ai int8x16_t vqtbl2q_s8(int8x16x2_t __p0, int8x16_t __p1) {
58056  int8x16_t __ret;
58057  __ret = (int8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 32);
58058  return __ret;
58059}
58060#else
58061__ai int8x16_t vqtbl2q_s8(int8x16x2_t __p0, int8x16_t __p1) {
58062  int8x16x2_t __rev0;
58063  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58064  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58065  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58066  int8x16_t __ret;
58067  __ret = (int8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 32);
58068  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58069  return __ret;
58070}
58071#endif
58072
58073#ifdef __LITTLE_ENDIAN__
58074__ai uint8x8_t vqtbl2_u8(uint8x16x2_t __p0, uint8x8_t __p1) {
58075  uint8x8_t __ret;
58076  __ret = (uint8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 16);
58077  return __ret;
58078}
58079#else
58080__ai uint8x8_t vqtbl2_u8(uint8x16x2_t __p0, uint8x8_t __p1) {
58081  uint8x16x2_t __rev0;
58082  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58083  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58084  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58085  uint8x8_t __ret;
58086  __ret = (uint8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 16);
58087  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58088  return __ret;
58089}
58090#endif
58091
58092#ifdef __LITTLE_ENDIAN__
58093__ai int8x8_t vqtbl2_s8(int8x16x2_t __p0, int8x8_t __p1) {
58094  int8x8_t __ret;
58095  __ret = (int8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 0);
58096  return __ret;
58097}
58098#else
58099__ai int8x8_t vqtbl2_s8(int8x16x2_t __p0, int8x8_t __p1) {
58100  int8x16x2_t __rev0;
58101  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58102  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58103  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58104  int8x8_t __ret;
58105  __ret = (int8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 0);
58106  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58107  return __ret;
58108}
58109#endif
58110
58111#ifdef __LITTLE_ENDIAN__
58112__ai poly8x8_t vqtbl3_p8(poly8x16x3_t __p0, uint8x8_t __p1) {
58113  poly8x8_t __ret;
58114  __ret = (poly8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 4);
58115  return __ret;
58116}
58117#else
58118__ai poly8x8_t vqtbl3_p8(poly8x16x3_t __p0, uint8x8_t __p1) {
58119  poly8x16x3_t __rev0;
58120  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58121  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58122  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58123  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58124  poly8x8_t __ret;
58125  __ret = (poly8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 4);
58126  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58127  return __ret;
58128}
58129#endif
58130
58131#ifdef __LITTLE_ENDIAN__
58132__ai poly8x16_t vqtbl3q_p8(poly8x16x3_t __p0, uint8x16_t __p1) {
58133  poly8x16_t __ret;
58134  __ret = (poly8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 36);
58135  return __ret;
58136}
58137#else
58138__ai poly8x16_t vqtbl3q_p8(poly8x16x3_t __p0, uint8x16_t __p1) {
58139  poly8x16x3_t __rev0;
58140  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58141  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58142  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58143  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58144  poly8x16_t __ret;
58145  __ret = (poly8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 36);
58146  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58147  return __ret;
58148}
58149#endif
58150
58151#ifdef __LITTLE_ENDIAN__
58152__ai uint8x16_t vqtbl3q_u8(uint8x16x3_t __p0, uint8x16_t __p1) {
58153  uint8x16_t __ret;
58154  __ret = (uint8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 48);
58155  return __ret;
58156}
58157#else
58158__ai uint8x16_t vqtbl3q_u8(uint8x16x3_t __p0, uint8x16_t __p1) {
58159  uint8x16x3_t __rev0;
58160  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58161  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58162  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58163  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58164  uint8x16_t __ret;
58165  __ret = (uint8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 48);
58166  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58167  return __ret;
58168}
58169#endif
58170
58171#ifdef __LITTLE_ENDIAN__
58172__ai int8x16_t vqtbl3q_s8(int8x16x3_t __p0, int8x16_t __p1) {
58173  int8x16_t __ret;
58174  __ret = (int8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 32);
58175  return __ret;
58176}
58177#else
58178__ai int8x16_t vqtbl3q_s8(int8x16x3_t __p0, int8x16_t __p1) {
58179  int8x16x3_t __rev0;
58180  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58181  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58182  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58183  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58184  int8x16_t __ret;
58185  __ret = (int8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 32);
58186  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58187  return __ret;
58188}
58189#endif
58190
58191#ifdef __LITTLE_ENDIAN__
58192__ai uint8x8_t vqtbl3_u8(uint8x16x3_t __p0, uint8x8_t __p1) {
58193  uint8x8_t __ret;
58194  __ret = (uint8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 16);
58195  return __ret;
58196}
58197#else
58198__ai uint8x8_t vqtbl3_u8(uint8x16x3_t __p0, uint8x8_t __p1) {
58199  uint8x16x3_t __rev0;
58200  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58201  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58202  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58203  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58204  uint8x8_t __ret;
58205  __ret = (uint8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 16);
58206  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58207  return __ret;
58208}
58209#endif
58210
58211#ifdef __LITTLE_ENDIAN__
58212__ai int8x8_t vqtbl3_s8(int8x16x3_t __p0, int8x8_t __p1) {
58213  int8x8_t __ret;
58214  __ret = (int8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 0);
58215  return __ret;
58216}
58217#else
58218__ai int8x8_t vqtbl3_s8(int8x16x3_t __p0, int8x8_t __p1) {
58219  int8x16x3_t __rev0;
58220  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58221  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58222  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58223  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58224  int8x8_t __ret;
58225  __ret = (int8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 0);
58226  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58227  return __ret;
58228}
58229#endif
58230
58231#ifdef __LITTLE_ENDIAN__
58232__ai poly8x8_t vqtbl4_p8(poly8x16x4_t __p0, uint8x8_t __p1) {
58233  poly8x8_t __ret;
58234  __ret = (poly8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 4);
58235  return __ret;
58236}
58237#else
58238__ai poly8x8_t vqtbl4_p8(poly8x16x4_t __p0, uint8x8_t __p1) {
58239  poly8x16x4_t __rev0;
58240  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58241  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58242  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58243  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58244  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58245  poly8x8_t __ret;
58246  __ret = (poly8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 4);
58247  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58248  return __ret;
58249}
58250#endif
58251
58252#ifdef __LITTLE_ENDIAN__
58253__ai poly8x16_t vqtbl4q_p8(poly8x16x4_t __p0, uint8x16_t __p1) {
58254  poly8x16_t __ret;
58255  __ret = (poly8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 36);
58256  return __ret;
58257}
58258#else
58259__ai poly8x16_t vqtbl4q_p8(poly8x16x4_t __p0, uint8x16_t __p1) {
58260  poly8x16x4_t __rev0;
58261  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58262  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58263  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58264  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58265  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58266  poly8x16_t __ret;
58267  __ret = (poly8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 36);
58268  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58269  return __ret;
58270}
58271#endif
58272
58273#ifdef __LITTLE_ENDIAN__
58274__ai uint8x16_t vqtbl4q_u8(uint8x16x4_t __p0, uint8x16_t __p1) {
58275  uint8x16_t __ret;
58276  __ret = (uint8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 48);
58277  return __ret;
58278}
58279#else
58280__ai uint8x16_t vqtbl4q_u8(uint8x16x4_t __p0, uint8x16_t __p1) {
58281  uint8x16x4_t __rev0;
58282  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58283  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58284  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58285  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58286  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58287  uint8x16_t __ret;
58288  __ret = (uint8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 48);
58289  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58290  return __ret;
58291}
58292#endif
58293
58294#ifdef __LITTLE_ENDIAN__
58295__ai int8x16_t vqtbl4q_s8(int8x16x4_t __p0, int8x16_t __p1) {
58296  int8x16_t __ret;
58297  __ret = (int8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 32);
58298  return __ret;
58299}
58300#else
58301__ai int8x16_t vqtbl4q_s8(int8x16x4_t __p0, int8x16_t __p1) {
58302  int8x16x4_t __rev0;
58303  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58304  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58305  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58306  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58307  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58308  int8x16_t __ret;
58309  __ret = (int8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 32);
58310  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58311  return __ret;
58312}
58313#endif
58314
58315#ifdef __LITTLE_ENDIAN__
58316__ai uint8x8_t vqtbl4_u8(uint8x16x4_t __p0, uint8x8_t __p1) {
58317  uint8x8_t __ret;
58318  __ret = (uint8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 16);
58319  return __ret;
58320}
58321#else
58322__ai uint8x8_t vqtbl4_u8(uint8x16x4_t __p0, uint8x8_t __p1) {
58323  uint8x16x4_t __rev0;
58324  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58325  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58326  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58327  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58328  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58329  uint8x8_t __ret;
58330  __ret = (uint8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 16);
58331  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58332  return __ret;
58333}
58334#endif
58335
58336#ifdef __LITTLE_ENDIAN__
58337__ai int8x8_t vqtbl4_s8(int8x16x4_t __p0, int8x8_t __p1) {
58338  int8x8_t __ret;
58339  __ret = (int8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 0);
58340  return __ret;
58341}
58342#else
58343__ai int8x8_t vqtbl4_s8(int8x16x4_t __p0, int8x8_t __p1) {
58344  int8x16x4_t __rev0;
58345  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58346  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58347  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58348  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58349  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58350  int8x8_t __ret;
58351  __ret = (int8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 0);
58352  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58353  return __ret;
58354}
58355#endif
58356
58357#ifdef __LITTLE_ENDIAN__
58358__ai poly8x8_t vqtbx1_p8(poly8x8_t __p0, poly8x16_t __p1, uint8x8_t __p2) {
58359  poly8x8_t __ret;
58360  __ret = (poly8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 4);
58361  return __ret;
58362}
58363#else
58364__ai poly8x8_t vqtbx1_p8(poly8x8_t __p0, poly8x16_t __p1, uint8x8_t __p2) {
58365  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58366  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58367  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58368  poly8x8_t __ret;
58369  __ret = (poly8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 4);
58370  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58371  return __ret;
58372}
58373#endif
58374
58375#ifdef __LITTLE_ENDIAN__
58376__ai poly8x16_t vqtbx1q_p8(poly8x16_t __p0, poly8x16_t __p1, uint8x16_t __p2) {
58377  poly8x16_t __ret;
58378  __ret = (poly8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 36);
58379  return __ret;
58380}
58381#else
58382__ai poly8x16_t vqtbx1q_p8(poly8x16_t __p0, poly8x16_t __p1, uint8x16_t __p2) {
58383  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58384  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58385  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58386  poly8x16_t __ret;
58387  __ret = (poly8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 36);
58388  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58389  return __ret;
58390}
58391#endif
58392
58393#ifdef __LITTLE_ENDIAN__
58394__ai uint8x16_t vqtbx1q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
58395  uint8x16_t __ret;
58396  __ret = (uint8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
58397  return __ret;
58398}
58399#else
58400__ai uint8x16_t vqtbx1q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
58401  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58402  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58403  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58404  uint8x16_t __ret;
58405  __ret = (uint8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
58406  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58407  return __ret;
58408}
58409#endif
58410
58411#ifdef __LITTLE_ENDIAN__
58412__ai int8x16_t vqtbx1q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
58413  int8x16_t __ret;
58414  __ret = (int8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
58415  return __ret;
58416}
58417#else
58418__ai int8x16_t vqtbx1q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
58419  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58420  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58421  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58422  int8x16_t __ret;
58423  __ret = (int8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
58424  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58425  return __ret;
58426}
58427#endif
58428
58429#ifdef __LITTLE_ENDIAN__
58430__ai uint8x8_t vqtbx1_u8(uint8x8_t __p0, uint8x16_t __p1, uint8x8_t __p2) {
58431  uint8x8_t __ret;
58432  __ret = (uint8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 16);
58433  return __ret;
58434}
58435#else
58436__ai uint8x8_t vqtbx1_u8(uint8x8_t __p0, uint8x16_t __p1, uint8x8_t __p2) {
58437  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58438  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58439  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58440  uint8x8_t __ret;
58441  __ret = (uint8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 16);
58442  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58443  return __ret;
58444}
58445#endif
58446
58447#ifdef __LITTLE_ENDIAN__
58448__ai int8x8_t vqtbx1_s8(int8x8_t __p0, int8x16_t __p1, int8x8_t __p2) {
58449  int8x8_t __ret;
58450  __ret = (int8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 0);
58451  return __ret;
58452}
58453#else
58454__ai int8x8_t vqtbx1_s8(int8x8_t __p0, int8x16_t __p1, int8x8_t __p2) {
58455  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58456  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58457  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58458  int8x8_t __ret;
58459  __ret = (int8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 0);
58460  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58461  return __ret;
58462}
58463#endif
58464
58465#ifdef __LITTLE_ENDIAN__
58466__ai poly8x8_t vqtbx2_p8(poly8x8_t __p0, poly8x16x2_t __p1, uint8x8_t __p2) {
58467  poly8x8_t __ret;
58468  __ret = (poly8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 4);
58469  return __ret;
58470}
58471#else
58472__ai poly8x8_t vqtbx2_p8(poly8x8_t __p0, poly8x16x2_t __p1, uint8x8_t __p2) {
58473  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58474  poly8x16x2_t __rev1;
58475  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58476  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58477  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58478  poly8x8_t __ret;
58479  __ret = (poly8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 4);
58480  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58481  return __ret;
58482}
58483#endif
58484
58485#ifdef __LITTLE_ENDIAN__
58486__ai poly8x16_t vqtbx2q_p8(poly8x16_t __p0, poly8x16x2_t __p1, uint8x16_t __p2) {
58487  poly8x16_t __ret;
58488  __ret = (poly8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 36);
58489  return __ret;
58490}
58491#else
58492__ai poly8x16_t vqtbx2q_p8(poly8x16_t __p0, poly8x16x2_t __p1, uint8x16_t __p2) {
58493  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58494  poly8x16x2_t __rev1;
58495  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58496  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58497  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58498  poly8x16_t __ret;
58499  __ret = (poly8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 36);
58500  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58501  return __ret;
58502}
58503#endif
58504
58505#ifdef __LITTLE_ENDIAN__
58506__ai uint8x16_t vqtbx2q_u8(uint8x16_t __p0, uint8x16x2_t __p1, uint8x16_t __p2) {
58507  uint8x16_t __ret;
58508  __ret = (uint8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 48);
58509  return __ret;
58510}
58511#else
58512__ai uint8x16_t vqtbx2q_u8(uint8x16_t __p0, uint8x16x2_t __p1, uint8x16_t __p2) {
58513  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58514  uint8x16x2_t __rev1;
58515  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58516  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58517  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58518  uint8x16_t __ret;
58519  __ret = (uint8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 48);
58520  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58521  return __ret;
58522}
58523#endif
58524
58525#ifdef __LITTLE_ENDIAN__
58526__ai int8x16_t vqtbx2q_s8(int8x16_t __p0, int8x16x2_t __p1, int8x16_t __p2) {
58527  int8x16_t __ret;
58528  __ret = (int8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 32);
58529  return __ret;
58530}
58531#else
58532__ai int8x16_t vqtbx2q_s8(int8x16_t __p0, int8x16x2_t __p1, int8x16_t __p2) {
58533  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58534  int8x16x2_t __rev1;
58535  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58536  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58537  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58538  int8x16_t __ret;
58539  __ret = (int8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 32);
58540  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58541  return __ret;
58542}
58543#endif
58544
58545#ifdef __LITTLE_ENDIAN__
58546__ai uint8x8_t vqtbx2_u8(uint8x8_t __p0, uint8x16x2_t __p1, uint8x8_t __p2) {
58547  uint8x8_t __ret;
58548  __ret = (uint8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 16);
58549  return __ret;
58550}
58551#else
58552__ai uint8x8_t vqtbx2_u8(uint8x8_t __p0, uint8x16x2_t __p1, uint8x8_t __p2) {
58553  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58554  uint8x16x2_t __rev1;
58555  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58556  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58557  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58558  uint8x8_t __ret;
58559  __ret = (uint8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 16);
58560  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58561  return __ret;
58562}
58563#endif
58564
58565#ifdef __LITTLE_ENDIAN__
58566__ai int8x8_t vqtbx2_s8(int8x8_t __p0, int8x16x2_t __p1, int8x8_t __p2) {
58567  int8x8_t __ret;
58568  __ret = (int8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 0);
58569  return __ret;
58570}
58571#else
58572__ai int8x8_t vqtbx2_s8(int8x8_t __p0, int8x16x2_t __p1, int8x8_t __p2) {
58573  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58574  int8x16x2_t __rev1;
58575  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58576  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58577  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58578  int8x8_t __ret;
58579  __ret = (int8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 0);
58580  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58581  return __ret;
58582}
58583#endif
58584
58585#ifdef __LITTLE_ENDIAN__
58586__ai poly8x8_t vqtbx3_p8(poly8x8_t __p0, poly8x16x3_t __p1, uint8x8_t __p2) {
58587  poly8x8_t __ret;
58588  __ret = (poly8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 4);
58589  return __ret;
58590}
58591#else
58592__ai poly8x8_t vqtbx3_p8(poly8x8_t __p0, poly8x16x3_t __p1, uint8x8_t __p2) {
58593  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58594  poly8x16x3_t __rev1;
58595  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58596  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58597  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58598  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58599  poly8x8_t __ret;
58600  __ret = (poly8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 4);
58601  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58602  return __ret;
58603}
58604#endif
58605
58606#ifdef __LITTLE_ENDIAN__
58607__ai poly8x16_t vqtbx3q_p8(poly8x16_t __p0, poly8x16x3_t __p1, uint8x16_t __p2) {
58608  poly8x16_t __ret;
58609  __ret = (poly8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 36);
58610  return __ret;
58611}
58612#else
58613__ai poly8x16_t vqtbx3q_p8(poly8x16_t __p0, poly8x16x3_t __p1, uint8x16_t __p2) {
58614  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58615  poly8x16x3_t __rev1;
58616  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58617  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58618  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58619  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58620  poly8x16_t __ret;
58621  __ret = (poly8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 36);
58622  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58623  return __ret;
58624}
58625#endif
58626
58627#ifdef __LITTLE_ENDIAN__
58628__ai uint8x16_t vqtbx3q_u8(uint8x16_t __p0, uint8x16x3_t __p1, uint8x16_t __p2) {
58629  uint8x16_t __ret;
58630  __ret = (uint8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 48);
58631  return __ret;
58632}
58633#else
58634__ai uint8x16_t vqtbx3q_u8(uint8x16_t __p0, uint8x16x3_t __p1, uint8x16_t __p2) {
58635  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58636  uint8x16x3_t __rev1;
58637  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58638  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58639  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58640  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58641  uint8x16_t __ret;
58642  __ret = (uint8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 48);
58643  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58644  return __ret;
58645}
58646#endif
58647
58648#ifdef __LITTLE_ENDIAN__
58649__ai int8x16_t vqtbx3q_s8(int8x16_t __p0, int8x16x3_t __p1, int8x16_t __p2) {
58650  int8x16_t __ret;
58651  __ret = (int8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 32);
58652  return __ret;
58653}
58654#else
58655__ai int8x16_t vqtbx3q_s8(int8x16_t __p0, int8x16x3_t __p1, int8x16_t __p2) {
58656  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58657  int8x16x3_t __rev1;
58658  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58659  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58660  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58661  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58662  int8x16_t __ret;
58663  __ret = (int8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 32);
58664  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58665  return __ret;
58666}
58667#endif
58668
58669#ifdef __LITTLE_ENDIAN__
58670__ai uint8x8_t vqtbx3_u8(uint8x8_t __p0, uint8x16x3_t __p1, uint8x8_t __p2) {
58671  uint8x8_t __ret;
58672  __ret = (uint8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 16);
58673  return __ret;
58674}
58675#else
58676__ai uint8x8_t vqtbx3_u8(uint8x8_t __p0, uint8x16x3_t __p1, uint8x8_t __p2) {
58677  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58678  uint8x16x3_t __rev1;
58679  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58680  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58681  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58682  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58683  uint8x8_t __ret;
58684  __ret = (uint8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 16);
58685  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58686  return __ret;
58687}
58688#endif
58689
58690#ifdef __LITTLE_ENDIAN__
58691__ai int8x8_t vqtbx3_s8(int8x8_t __p0, int8x16x3_t __p1, int8x8_t __p2) {
58692  int8x8_t __ret;
58693  __ret = (int8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 0);
58694  return __ret;
58695}
58696#else
58697__ai int8x8_t vqtbx3_s8(int8x8_t __p0, int8x16x3_t __p1, int8x8_t __p2) {
58698  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58699  int8x16x3_t __rev1;
58700  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58701  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58702  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58703  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58704  int8x8_t __ret;
58705  __ret = (int8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 0);
58706  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58707  return __ret;
58708}
58709#endif
58710
58711#ifdef __LITTLE_ENDIAN__
58712__ai poly8x8_t vqtbx4_p8(poly8x8_t __p0, poly8x16x4_t __p1, uint8x8_t __p2) {
58713  poly8x8_t __ret;
58714  __ret = (poly8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 4);
58715  return __ret;
58716}
58717#else
58718__ai poly8x8_t vqtbx4_p8(poly8x8_t __p0, poly8x16x4_t __p1, uint8x8_t __p2) {
58719  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58720  poly8x16x4_t __rev1;
58721  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58722  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58723  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58724  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58725  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58726  poly8x8_t __ret;
58727  __ret = (poly8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 4);
58728  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58729  return __ret;
58730}
58731#endif
58732
58733#ifdef __LITTLE_ENDIAN__
58734__ai poly8x16_t vqtbx4q_p8(poly8x16_t __p0, poly8x16x4_t __p1, uint8x16_t __p2) {
58735  poly8x16_t __ret;
58736  __ret = (poly8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 36);
58737  return __ret;
58738}
58739#else
58740__ai poly8x16_t vqtbx4q_p8(poly8x16_t __p0, poly8x16x4_t __p1, uint8x16_t __p2) {
58741  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58742  poly8x16x4_t __rev1;
58743  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58744  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58745  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58746  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58747  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58748  poly8x16_t __ret;
58749  __ret = (poly8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 36);
58750  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58751  return __ret;
58752}
58753#endif
58754
58755#ifdef __LITTLE_ENDIAN__
58756__ai uint8x16_t vqtbx4q_u8(uint8x16_t __p0, uint8x16x4_t __p1, uint8x16_t __p2) {
58757  uint8x16_t __ret;
58758  __ret = (uint8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 48);
58759  return __ret;
58760}
58761#else
58762__ai uint8x16_t vqtbx4q_u8(uint8x16_t __p0, uint8x16x4_t __p1, uint8x16_t __p2) {
58763  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58764  uint8x16x4_t __rev1;
58765  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58766  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58767  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58768  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58769  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58770  uint8x16_t __ret;
58771  __ret = (uint8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 48);
58772  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58773  return __ret;
58774}
58775#endif
58776
58777#ifdef __LITTLE_ENDIAN__
58778__ai int8x16_t vqtbx4q_s8(int8x16_t __p0, int8x16x4_t __p1, int8x16_t __p2) {
58779  int8x16_t __ret;
58780  __ret = (int8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 32);
58781  return __ret;
58782}
58783#else
58784__ai int8x16_t vqtbx4q_s8(int8x16_t __p0, int8x16x4_t __p1, int8x16_t __p2) {
58785  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58786  int8x16x4_t __rev1;
58787  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58788  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58789  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58790  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58791  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58792  int8x16_t __ret;
58793  __ret = (int8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 32);
58794  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58795  return __ret;
58796}
58797#endif
58798
58799#ifdef __LITTLE_ENDIAN__
58800__ai uint8x8_t vqtbx4_u8(uint8x8_t __p0, uint8x16x4_t __p1, uint8x8_t __p2) {
58801  uint8x8_t __ret;
58802  __ret = (uint8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 16);
58803  return __ret;
58804}
58805#else
58806__ai uint8x8_t vqtbx4_u8(uint8x8_t __p0, uint8x16x4_t __p1, uint8x8_t __p2) {
58807  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58808  uint8x16x4_t __rev1;
58809  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58810  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58811  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58812  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58813  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58814  uint8x8_t __ret;
58815  __ret = (uint8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 16);
58816  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58817  return __ret;
58818}
58819#endif
58820
58821#ifdef __LITTLE_ENDIAN__
58822__ai int8x8_t vqtbx4_s8(int8x8_t __p0, int8x16x4_t __p1, int8x8_t __p2) {
58823  int8x8_t __ret;
58824  __ret = (int8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 0);
58825  return __ret;
58826}
58827#else
58828__ai int8x8_t vqtbx4_s8(int8x8_t __p0, int8x16x4_t __p1, int8x8_t __p2) {
58829  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58830  int8x16x4_t __rev1;
58831  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58832  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58833  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58834  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58835  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58836  int8x8_t __ret;
58837  __ret = (int8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 0);
58838  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58839  return __ret;
58840}
58841#endif
58842
58843#ifdef __LITTLE_ENDIAN__
58844__ai uint16x8_t vraddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
58845  uint16x8_t __ret;
58846  __ret = vcombine_u16(__p0, vraddhn_u32(__p1, __p2));
58847  return __ret;
58848}
58849#else
58850__ai uint16x8_t vraddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
58851  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
58852  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
58853  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
58854  uint16x8_t __ret;
58855  __ret = __noswap_vcombine_u16(__rev0, __noswap_vraddhn_u32(__rev1, __rev2));
58856  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58857  return __ret;
58858}
58859#endif
58860
58861#ifdef __LITTLE_ENDIAN__
58862__ai uint32x4_t vraddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
58863  uint32x4_t __ret;
58864  __ret = vcombine_u32(__p0, vraddhn_u64(__p1, __p2));
58865  return __ret;
58866}
58867#else
58868__ai uint32x4_t vraddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
58869  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
58870  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
58871  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
58872  uint32x4_t __ret;
58873  __ret = __noswap_vcombine_u32(__rev0, __noswap_vraddhn_u64(__rev1, __rev2));
58874  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
58875  return __ret;
58876}
58877#endif
58878
58879#ifdef __LITTLE_ENDIAN__
58880__ai uint8x16_t vraddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
58881  uint8x16_t __ret;
58882  __ret = vcombine_u8(__p0, vraddhn_u16(__p1, __p2));
58883  return __ret;
58884}
58885#else
58886__ai uint8x16_t vraddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
58887  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58888  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58889  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58890  uint8x16_t __ret;
58891  __ret = __noswap_vcombine_u8(__rev0, __noswap_vraddhn_u16(__rev1, __rev2));
58892  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58893  return __ret;
58894}
58895#endif
58896
58897#ifdef __LITTLE_ENDIAN__
58898__ai int16x8_t vraddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
58899  int16x8_t __ret;
58900  __ret = vcombine_s16(__p0, vraddhn_s32(__p1, __p2));
58901  return __ret;
58902}
58903#else
58904__ai int16x8_t vraddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
58905  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
58906  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
58907  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
58908  int16x8_t __ret;
58909  __ret = __noswap_vcombine_s16(__rev0, __noswap_vraddhn_s32(__rev1, __rev2));
58910  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58911  return __ret;
58912}
58913#endif
58914
58915#ifdef __LITTLE_ENDIAN__
58916__ai int32x4_t vraddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
58917  int32x4_t __ret;
58918  __ret = vcombine_s32(__p0, vraddhn_s64(__p1, __p2));
58919  return __ret;
58920}
58921#else
58922__ai int32x4_t vraddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
58923  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
58924  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
58925  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
58926  int32x4_t __ret;
58927  __ret = __noswap_vcombine_s32(__rev0, __noswap_vraddhn_s64(__rev1, __rev2));
58928  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
58929  return __ret;
58930}
58931#endif
58932
58933#ifdef __LITTLE_ENDIAN__
58934__ai int8x16_t vraddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
58935  int8x16_t __ret;
58936  __ret = vcombine_s8(__p0, vraddhn_s16(__p1, __p2));
58937  return __ret;
58938}
58939#else
58940__ai int8x16_t vraddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
58941  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58942  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58943  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
58944  int8x16_t __ret;
58945  __ret = __noswap_vcombine_s8(__rev0, __noswap_vraddhn_s16(__rev1, __rev2));
58946  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58947  return __ret;
58948}
58949#endif
58950
58951#ifdef __LITTLE_ENDIAN__
58952__ai poly8x8_t vrbit_p8(poly8x8_t __p0) {
58953  poly8x8_t __ret;
58954  __ret = (poly8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 4);
58955  return __ret;
58956}
58957#else
58958__ai poly8x8_t vrbit_p8(poly8x8_t __p0) {
58959  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
58960  poly8x8_t __ret;
58961  __ret = (poly8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 4);
58962  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58963  return __ret;
58964}
58965#endif
58966
58967#ifdef __LITTLE_ENDIAN__
58968__ai poly8x16_t vrbitq_p8(poly8x16_t __p0) {
58969  poly8x16_t __ret;
58970  __ret = (poly8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 36);
58971  return __ret;
58972}
58973#else
58974__ai poly8x16_t vrbitq_p8(poly8x16_t __p0) {
58975  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58976  poly8x16_t __ret;
58977  __ret = (poly8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 36);
58978  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58979  return __ret;
58980}
58981#endif
58982
58983#ifdef __LITTLE_ENDIAN__
58984__ai uint8x16_t vrbitq_u8(uint8x16_t __p0) {
58985  uint8x16_t __ret;
58986  __ret = (uint8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 48);
58987  return __ret;
58988}
58989#else
58990__ai uint8x16_t vrbitq_u8(uint8x16_t __p0) {
58991  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58992  uint8x16_t __ret;
58993  __ret = (uint8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 48);
58994  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58995  return __ret;
58996}
58997#endif
58998
58999#ifdef __LITTLE_ENDIAN__
59000__ai int8x16_t vrbitq_s8(int8x16_t __p0) {
59001  int8x16_t __ret;
59002  __ret = (int8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 32);
59003  return __ret;
59004}
59005#else
59006__ai int8x16_t vrbitq_s8(int8x16_t __p0) {
59007  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59008  int8x16_t __ret;
59009  __ret = (int8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 32);
59010  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59011  return __ret;
59012}
59013#endif
59014
59015#ifdef __LITTLE_ENDIAN__
59016__ai uint8x8_t vrbit_u8(uint8x8_t __p0) {
59017  uint8x8_t __ret;
59018  __ret = (uint8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 16);
59019  return __ret;
59020}
59021#else
59022__ai uint8x8_t vrbit_u8(uint8x8_t __p0) {
59023  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59024  uint8x8_t __ret;
59025  __ret = (uint8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 16);
59026  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59027  return __ret;
59028}
59029#endif
59030
59031#ifdef __LITTLE_ENDIAN__
59032__ai int8x8_t vrbit_s8(int8x8_t __p0) {
59033  int8x8_t __ret;
59034  __ret = (int8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 0);
59035  return __ret;
59036}
59037#else
59038__ai int8x8_t vrbit_s8(int8x8_t __p0) {
59039  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59040  int8x8_t __ret;
59041  __ret = (int8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 0);
59042  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59043  return __ret;
59044}
59045#endif
59046
59047#ifdef __LITTLE_ENDIAN__
59048__ai float64x2_t vrecpeq_f64(float64x2_t __p0) {
59049  float64x2_t __ret;
59050  __ret = (float64x2_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 42);
59051  return __ret;
59052}
59053#else
59054__ai float64x2_t vrecpeq_f64(float64x2_t __p0) {
59055  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59056  float64x2_t __ret;
59057  __ret = (float64x2_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 42);
59058  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59059  return __ret;
59060}
59061#endif
59062
59063#ifdef __LITTLE_ENDIAN__
59064__ai float64x1_t vrecpe_f64(float64x1_t __p0) {
59065  float64x1_t __ret;
59066  __ret = (float64x1_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 10);
59067  return __ret;
59068}
59069#else
59070__ai float64x1_t vrecpe_f64(float64x1_t __p0) {
59071  float64x1_t __ret;
59072  __ret = (float64x1_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 10);
59073  return __ret;
59074}
59075#endif
59076
59077#ifdef __LITTLE_ENDIAN__
59078__ai float64_t vrecped_f64(float64_t __p0) {
59079  float64_t __ret;
59080  __ret = (float64_t) __builtin_neon_vrecped_f64(__p0);
59081  return __ret;
59082}
59083#else
59084__ai float64_t vrecped_f64(float64_t __p0) {
59085  float64_t __ret;
59086  __ret = (float64_t) __builtin_neon_vrecped_f64(__p0);
59087  return __ret;
59088}
59089#endif
59090
59091#ifdef __LITTLE_ENDIAN__
59092__ai float32_t vrecpes_f32(float32_t __p0) {
59093  float32_t __ret;
59094  __ret = (float32_t) __builtin_neon_vrecpes_f32(__p0);
59095  return __ret;
59096}
59097#else
59098__ai float32_t vrecpes_f32(float32_t __p0) {
59099  float32_t __ret;
59100  __ret = (float32_t) __builtin_neon_vrecpes_f32(__p0);
59101  return __ret;
59102}
59103#endif
59104
59105#ifdef __LITTLE_ENDIAN__
59106__ai float64x2_t vrecpsq_f64(float64x2_t __p0, float64x2_t __p1) {
59107  float64x2_t __ret;
59108  __ret = (float64x2_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
59109  return __ret;
59110}
59111#else
59112__ai float64x2_t vrecpsq_f64(float64x2_t __p0, float64x2_t __p1) {
59113  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59114  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59115  float64x2_t __ret;
59116  __ret = (float64x2_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
59117  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59118  return __ret;
59119}
59120#endif
59121
59122#ifdef __LITTLE_ENDIAN__
59123__ai float64x1_t vrecps_f64(float64x1_t __p0, float64x1_t __p1) {
59124  float64x1_t __ret;
59125  __ret = (float64x1_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
59126  return __ret;
59127}
59128#else
59129__ai float64x1_t vrecps_f64(float64x1_t __p0, float64x1_t __p1) {
59130  float64x1_t __ret;
59131  __ret = (float64x1_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
59132  return __ret;
59133}
59134#endif
59135
59136#ifdef __LITTLE_ENDIAN__
59137__ai float64_t vrecpsd_f64(float64_t __p0, float64_t __p1) {
59138  float64_t __ret;
59139  __ret = (float64_t) __builtin_neon_vrecpsd_f64(__p0, __p1);
59140  return __ret;
59141}
59142#else
59143__ai float64_t vrecpsd_f64(float64_t __p0, float64_t __p1) {
59144  float64_t __ret;
59145  __ret = (float64_t) __builtin_neon_vrecpsd_f64(__p0, __p1);
59146  return __ret;
59147}
59148#endif
59149
59150#ifdef __LITTLE_ENDIAN__
59151__ai float32_t vrecpss_f32(float32_t __p0, float32_t __p1) {
59152  float32_t __ret;
59153  __ret = (float32_t) __builtin_neon_vrecpss_f32(__p0, __p1);
59154  return __ret;
59155}
59156#else
59157__ai float32_t vrecpss_f32(float32_t __p0, float32_t __p1) {
59158  float32_t __ret;
59159  __ret = (float32_t) __builtin_neon_vrecpss_f32(__p0, __p1);
59160  return __ret;
59161}
59162#endif
59163
59164#ifdef __LITTLE_ENDIAN__
59165__ai float64_t vrecpxd_f64(float64_t __p0) {
59166  float64_t __ret;
59167  __ret = (float64_t) __builtin_neon_vrecpxd_f64(__p0);
59168  return __ret;
59169}
59170#else
59171__ai float64_t vrecpxd_f64(float64_t __p0) {
59172  float64_t __ret;
59173  __ret = (float64_t) __builtin_neon_vrecpxd_f64(__p0);
59174  return __ret;
59175}
59176#endif
59177
59178#ifdef __LITTLE_ENDIAN__
59179__ai float32_t vrecpxs_f32(float32_t __p0) {
59180  float32_t __ret;
59181  __ret = (float32_t) __builtin_neon_vrecpxs_f32(__p0);
59182  return __ret;
59183}
59184#else
59185__ai float32_t vrecpxs_f32(float32_t __p0) {
59186  float32_t __ret;
59187  __ret = (float32_t) __builtin_neon_vrecpxs_f32(__p0);
59188  return __ret;
59189}
59190#endif
59191
59192#ifdef __LITTLE_ENDIAN__
59193__ai float64x2_t vrndq_f64(float64x2_t __p0) {
59194  float64x2_t __ret;
59195  __ret = (float64x2_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 42);
59196  return __ret;
59197}
59198#else
59199__ai float64x2_t vrndq_f64(float64x2_t __p0) {
59200  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59201  float64x2_t __ret;
59202  __ret = (float64x2_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 42);
59203  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59204  return __ret;
59205}
59206#endif
59207
59208#ifdef __LITTLE_ENDIAN__
59209__ai float32x4_t vrndq_f32(float32x4_t __p0) {
59210  float32x4_t __ret;
59211  __ret = (float32x4_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 41);
59212  return __ret;
59213}
59214#else
59215__ai float32x4_t vrndq_f32(float32x4_t __p0) {
59216  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59217  float32x4_t __ret;
59218  __ret = (float32x4_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 41);
59219  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59220  return __ret;
59221}
59222#endif
59223
59224#ifdef __LITTLE_ENDIAN__
59225__ai float64x1_t vrnd_f64(float64x1_t __p0) {
59226  float64x1_t __ret;
59227  __ret = (float64x1_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 10);
59228  return __ret;
59229}
59230#else
59231__ai float64x1_t vrnd_f64(float64x1_t __p0) {
59232  float64x1_t __ret;
59233  __ret = (float64x1_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 10);
59234  return __ret;
59235}
59236#endif
59237
59238#ifdef __LITTLE_ENDIAN__
59239__ai float32x2_t vrnd_f32(float32x2_t __p0) {
59240  float32x2_t __ret;
59241  __ret = (float32x2_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 9);
59242  return __ret;
59243}
59244#else
59245__ai float32x2_t vrnd_f32(float32x2_t __p0) {
59246  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59247  float32x2_t __ret;
59248  __ret = (float32x2_t) __builtin_neon_vrnd_v((int8x8_t)__rev0, 9);
59249  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59250  return __ret;
59251}
59252#endif
59253
59254#ifdef __LITTLE_ENDIAN__
59255__ai float64x2_t vrndaq_f64(float64x2_t __p0) {
59256  float64x2_t __ret;
59257  __ret = (float64x2_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 42);
59258  return __ret;
59259}
59260#else
59261__ai float64x2_t vrndaq_f64(float64x2_t __p0) {
59262  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59263  float64x2_t __ret;
59264  __ret = (float64x2_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 42);
59265  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59266  return __ret;
59267}
59268#endif
59269
59270#ifdef __LITTLE_ENDIAN__
59271__ai float32x4_t vrndaq_f32(float32x4_t __p0) {
59272  float32x4_t __ret;
59273  __ret = (float32x4_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 41);
59274  return __ret;
59275}
59276#else
59277__ai float32x4_t vrndaq_f32(float32x4_t __p0) {
59278  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59279  float32x4_t __ret;
59280  __ret = (float32x4_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 41);
59281  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59282  return __ret;
59283}
59284#endif
59285
59286#ifdef __LITTLE_ENDIAN__
59287__ai float64x1_t vrnda_f64(float64x1_t __p0) {
59288  float64x1_t __ret;
59289  __ret = (float64x1_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 10);
59290  return __ret;
59291}
59292#else
59293__ai float64x1_t vrnda_f64(float64x1_t __p0) {
59294  float64x1_t __ret;
59295  __ret = (float64x1_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 10);
59296  return __ret;
59297}
59298#endif
59299
59300#ifdef __LITTLE_ENDIAN__
59301__ai float32x2_t vrnda_f32(float32x2_t __p0) {
59302  float32x2_t __ret;
59303  __ret = (float32x2_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 9);
59304  return __ret;
59305}
59306#else
59307__ai float32x2_t vrnda_f32(float32x2_t __p0) {
59308  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59309  float32x2_t __ret;
59310  __ret = (float32x2_t) __builtin_neon_vrnda_v((int8x8_t)__rev0, 9);
59311  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59312  return __ret;
59313}
59314#endif
59315
59316#ifdef __LITTLE_ENDIAN__
59317__ai float64x2_t vrndiq_f64(float64x2_t __p0) {
59318  float64x2_t __ret;
59319  __ret = (float64x2_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 42);
59320  return __ret;
59321}
59322#else
59323__ai float64x2_t vrndiq_f64(float64x2_t __p0) {
59324  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59325  float64x2_t __ret;
59326  __ret = (float64x2_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 42);
59327  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59328  return __ret;
59329}
59330#endif
59331
59332#ifdef __LITTLE_ENDIAN__
59333__ai float32x4_t vrndiq_f32(float32x4_t __p0) {
59334  float32x4_t __ret;
59335  __ret = (float32x4_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 41);
59336  return __ret;
59337}
59338#else
59339__ai float32x4_t vrndiq_f32(float32x4_t __p0) {
59340  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59341  float32x4_t __ret;
59342  __ret = (float32x4_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 41);
59343  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59344  return __ret;
59345}
59346#endif
59347
59348#ifdef __LITTLE_ENDIAN__
59349__ai float64x1_t vrndi_f64(float64x1_t __p0) {
59350  float64x1_t __ret;
59351  __ret = (float64x1_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 10);
59352  return __ret;
59353}
59354#else
59355__ai float64x1_t vrndi_f64(float64x1_t __p0) {
59356  float64x1_t __ret;
59357  __ret = (float64x1_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 10);
59358  return __ret;
59359}
59360#endif
59361
59362#ifdef __LITTLE_ENDIAN__
59363__ai float32x2_t vrndi_f32(float32x2_t __p0) {
59364  float32x2_t __ret;
59365  __ret = (float32x2_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 9);
59366  return __ret;
59367}
59368#else
59369__ai float32x2_t vrndi_f32(float32x2_t __p0) {
59370  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59371  float32x2_t __ret;
59372  __ret = (float32x2_t) __builtin_neon_vrndi_v((int8x8_t)__rev0, 9);
59373  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59374  return __ret;
59375}
59376#endif
59377
59378#ifdef __LITTLE_ENDIAN__
59379__ai float64x2_t vrndmq_f64(float64x2_t __p0) {
59380  float64x2_t __ret;
59381  __ret = (float64x2_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 42);
59382  return __ret;
59383}
59384#else
59385__ai float64x2_t vrndmq_f64(float64x2_t __p0) {
59386  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59387  float64x2_t __ret;
59388  __ret = (float64x2_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 42);
59389  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59390  return __ret;
59391}
59392#endif
59393
59394#ifdef __LITTLE_ENDIAN__
59395__ai float32x4_t vrndmq_f32(float32x4_t __p0) {
59396  float32x4_t __ret;
59397  __ret = (float32x4_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 41);
59398  return __ret;
59399}
59400#else
59401__ai float32x4_t vrndmq_f32(float32x4_t __p0) {
59402  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59403  float32x4_t __ret;
59404  __ret = (float32x4_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 41);
59405  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59406  return __ret;
59407}
59408#endif
59409
59410#ifdef __LITTLE_ENDIAN__
59411__ai float64x1_t vrndm_f64(float64x1_t __p0) {
59412  float64x1_t __ret;
59413  __ret = (float64x1_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 10);
59414  return __ret;
59415}
59416#else
59417__ai float64x1_t vrndm_f64(float64x1_t __p0) {
59418  float64x1_t __ret;
59419  __ret = (float64x1_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 10);
59420  return __ret;
59421}
59422#endif
59423
59424#ifdef __LITTLE_ENDIAN__
59425__ai float32x2_t vrndm_f32(float32x2_t __p0) {
59426  float32x2_t __ret;
59427  __ret = (float32x2_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 9);
59428  return __ret;
59429}
59430#else
59431__ai float32x2_t vrndm_f32(float32x2_t __p0) {
59432  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59433  float32x2_t __ret;
59434  __ret = (float32x2_t) __builtin_neon_vrndm_v((int8x8_t)__rev0, 9);
59435  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59436  return __ret;
59437}
59438#endif
59439
59440#ifdef __LITTLE_ENDIAN__
59441__ai float64x2_t vrndnq_f64(float64x2_t __p0) {
59442  float64x2_t __ret;
59443  __ret = (float64x2_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 42);
59444  return __ret;
59445}
59446#else
59447__ai float64x2_t vrndnq_f64(float64x2_t __p0) {
59448  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59449  float64x2_t __ret;
59450  __ret = (float64x2_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 42);
59451  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59452  return __ret;
59453}
59454#endif
59455
59456#ifdef __LITTLE_ENDIAN__
59457__ai float32x4_t vrndnq_f32(float32x4_t __p0) {
59458  float32x4_t __ret;
59459  __ret = (float32x4_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 41);
59460  return __ret;
59461}
59462#else
59463__ai float32x4_t vrndnq_f32(float32x4_t __p0) {
59464  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59465  float32x4_t __ret;
59466  __ret = (float32x4_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 41);
59467  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59468  return __ret;
59469}
59470#endif
59471
59472#ifdef __LITTLE_ENDIAN__
59473__ai float64x1_t vrndn_f64(float64x1_t __p0) {
59474  float64x1_t __ret;
59475  __ret = (float64x1_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 10);
59476  return __ret;
59477}
59478#else
59479__ai float64x1_t vrndn_f64(float64x1_t __p0) {
59480  float64x1_t __ret;
59481  __ret = (float64x1_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 10);
59482  return __ret;
59483}
59484#endif
59485
59486#ifdef __LITTLE_ENDIAN__
59487__ai float32x2_t vrndn_f32(float32x2_t __p0) {
59488  float32x2_t __ret;
59489  __ret = (float32x2_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 9);
59490  return __ret;
59491}
59492#else
59493__ai float32x2_t vrndn_f32(float32x2_t __p0) {
59494  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59495  float32x2_t __ret;
59496  __ret = (float32x2_t) __builtin_neon_vrndn_v((int8x8_t)__rev0, 9);
59497  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59498  return __ret;
59499}
59500#endif
59501
59502#ifdef __LITTLE_ENDIAN__
59503__ai float64x2_t vrndpq_f64(float64x2_t __p0) {
59504  float64x2_t __ret;
59505  __ret = (float64x2_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 42);
59506  return __ret;
59507}
59508#else
59509__ai float64x2_t vrndpq_f64(float64x2_t __p0) {
59510  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59511  float64x2_t __ret;
59512  __ret = (float64x2_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 42);
59513  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59514  return __ret;
59515}
59516#endif
59517
59518#ifdef __LITTLE_ENDIAN__
59519__ai float32x4_t vrndpq_f32(float32x4_t __p0) {
59520  float32x4_t __ret;
59521  __ret = (float32x4_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 41);
59522  return __ret;
59523}
59524#else
59525__ai float32x4_t vrndpq_f32(float32x4_t __p0) {
59526  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59527  float32x4_t __ret;
59528  __ret = (float32x4_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 41);
59529  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59530  return __ret;
59531}
59532#endif
59533
59534#ifdef __LITTLE_ENDIAN__
59535__ai float64x1_t vrndp_f64(float64x1_t __p0) {
59536  float64x1_t __ret;
59537  __ret = (float64x1_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 10);
59538  return __ret;
59539}
59540#else
59541__ai float64x1_t vrndp_f64(float64x1_t __p0) {
59542  float64x1_t __ret;
59543  __ret = (float64x1_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 10);
59544  return __ret;
59545}
59546#endif
59547
59548#ifdef __LITTLE_ENDIAN__
59549__ai float32x2_t vrndp_f32(float32x2_t __p0) {
59550  float32x2_t __ret;
59551  __ret = (float32x2_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 9);
59552  return __ret;
59553}
59554#else
59555__ai float32x2_t vrndp_f32(float32x2_t __p0) {
59556  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59557  float32x2_t __ret;
59558  __ret = (float32x2_t) __builtin_neon_vrndp_v((int8x8_t)__rev0, 9);
59559  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59560  return __ret;
59561}
59562#endif
59563
59564#ifdef __LITTLE_ENDIAN__
59565__ai float64x2_t vrndxq_f64(float64x2_t __p0) {
59566  float64x2_t __ret;
59567  __ret = (float64x2_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 42);
59568  return __ret;
59569}
59570#else
59571__ai float64x2_t vrndxq_f64(float64x2_t __p0) {
59572  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59573  float64x2_t __ret;
59574  __ret = (float64x2_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 42);
59575  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59576  return __ret;
59577}
59578#endif
59579
59580#ifdef __LITTLE_ENDIAN__
59581__ai float32x4_t vrndxq_f32(float32x4_t __p0) {
59582  float32x4_t __ret;
59583  __ret = (float32x4_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 41);
59584  return __ret;
59585}
59586#else
59587__ai float32x4_t vrndxq_f32(float32x4_t __p0) {
59588  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59589  float32x4_t __ret;
59590  __ret = (float32x4_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 41);
59591  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59592  return __ret;
59593}
59594#endif
59595
59596#ifdef __LITTLE_ENDIAN__
59597__ai float64x1_t vrndx_f64(float64x1_t __p0) {
59598  float64x1_t __ret;
59599  __ret = (float64x1_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 10);
59600  return __ret;
59601}
59602#else
59603__ai float64x1_t vrndx_f64(float64x1_t __p0) {
59604  float64x1_t __ret;
59605  __ret = (float64x1_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 10);
59606  return __ret;
59607}
59608#endif
59609
59610#ifdef __LITTLE_ENDIAN__
59611__ai float32x2_t vrndx_f32(float32x2_t __p0) {
59612  float32x2_t __ret;
59613  __ret = (float32x2_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 9);
59614  return __ret;
59615}
59616#else
59617__ai float32x2_t vrndx_f32(float32x2_t __p0) {
59618  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59619  float32x2_t __ret;
59620  __ret = (float32x2_t) __builtin_neon_vrndx_v((int8x8_t)__rev0, 9);
59621  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59622  return __ret;
59623}
59624#endif
59625
59626#ifdef __LITTLE_ENDIAN__
59627__ai uint64_t vrshld_u64(uint64_t __p0, uint64_t __p1) {
59628  uint64_t __ret;
59629  __ret = (uint64_t) __builtin_neon_vrshld_u64(__p0, __p1);
59630  return __ret;
59631}
59632#else
59633__ai uint64_t vrshld_u64(uint64_t __p0, uint64_t __p1) {
59634  uint64_t __ret;
59635  __ret = (uint64_t) __builtin_neon_vrshld_u64(__p0, __p1);
59636  return __ret;
59637}
59638#endif
59639
59640#ifdef __LITTLE_ENDIAN__
59641__ai int64_t vrshld_s64(int64_t __p0, int64_t __p1) {
59642  int64_t __ret;
59643  __ret = (int64_t) __builtin_neon_vrshld_s64(__p0, __p1);
59644  return __ret;
59645}
59646#else
59647__ai int64_t vrshld_s64(int64_t __p0, int64_t __p1) {
59648  int64_t __ret;
59649  __ret = (int64_t) __builtin_neon_vrshld_s64(__p0, __p1);
59650  return __ret;
59651}
59652#endif
59653
59654#ifdef __LITTLE_ENDIAN__
59655#define vrshrd_n_u64(__p0, __p1) __extension__ ({ \
59656  uint64_t __s0 = __p0; \
59657  uint64_t __ret; \
59658  __ret = (uint64_t) __builtin_neon_vrshrd_n_u64(__s0, __p1); \
59659  __ret; \
59660})
59661#else
59662#define vrshrd_n_u64(__p0, __p1) __extension__ ({ \
59663  uint64_t __s0 = __p0; \
59664  uint64_t __ret; \
59665  __ret = (uint64_t) __builtin_neon_vrshrd_n_u64(__s0, __p1); \
59666  __ret; \
59667})
59668#endif
59669
59670#ifdef __LITTLE_ENDIAN__
59671#define vrshrd_n_s64(__p0, __p1) __extension__ ({ \
59672  int64_t __s0 = __p0; \
59673  int64_t __ret; \
59674  __ret = (int64_t) __builtin_neon_vrshrd_n_s64(__s0, __p1); \
59675  __ret; \
59676})
59677#else
59678#define vrshrd_n_s64(__p0, __p1) __extension__ ({ \
59679  int64_t __s0 = __p0; \
59680  int64_t __ret; \
59681  __ret = (int64_t) __builtin_neon_vrshrd_n_s64(__s0, __p1); \
59682  __ret; \
59683})
59684#endif
59685
59686#ifdef __LITTLE_ENDIAN__
59687#define vrshrn_high_n_u32(__p0_210, __p1_210, __p2_210) __extension__ ({ \
59688  uint16x4_t __s0_210 = __p0_210; \
59689  uint32x4_t __s1_210 = __p1_210; \
59690  uint16x8_t __ret_210; \
59691  __ret_210 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_210), (uint16x4_t)(vrshrn_n_u32(__s1_210, __p2_210)))); \
59692  __ret_210; \
59693})
59694#else
59695#define vrshrn_high_n_u32(__p0_211, __p1_211, __p2_211) __extension__ ({ \
59696  uint16x4_t __s0_211 = __p0_211; \
59697  uint32x4_t __s1_211 = __p1_211; \
59698  uint16x4_t __rev0_211;  __rev0_211 = __builtin_shufflevector(__s0_211, __s0_211, 3, 2, 1, 0); \
59699  uint32x4_t __rev1_211;  __rev1_211 = __builtin_shufflevector(__s1_211, __s1_211, 3, 2, 1, 0); \
59700  uint16x8_t __ret_211; \
59701  __ret_211 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_211), (uint16x4_t)(__noswap_vrshrn_n_u32(__rev1_211, __p2_211)))); \
59702  __ret_211 = __builtin_shufflevector(__ret_211, __ret_211, 7, 6, 5, 4, 3, 2, 1, 0); \
59703  __ret_211; \
59704})
59705#endif
59706
59707#ifdef __LITTLE_ENDIAN__
59708#define vrshrn_high_n_u64(__p0_212, __p1_212, __p2_212) __extension__ ({ \
59709  uint32x2_t __s0_212 = __p0_212; \
59710  uint64x2_t __s1_212 = __p1_212; \
59711  uint32x4_t __ret_212; \
59712  __ret_212 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_212), (uint32x2_t)(vrshrn_n_u64(__s1_212, __p2_212)))); \
59713  __ret_212; \
59714})
59715#else
59716#define vrshrn_high_n_u64(__p0_213, __p1_213, __p2_213) __extension__ ({ \
59717  uint32x2_t __s0_213 = __p0_213; \
59718  uint64x2_t __s1_213 = __p1_213; \
59719  uint32x2_t __rev0_213;  __rev0_213 = __builtin_shufflevector(__s0_213, __s0_213, 1, 0); \
59720  uint64x2_t __rev1_213;  __rev1_213 = __builtin_shufflevector(__s1_213, __s1_213, 1, 0); \
59721  uint32x4_t __ret_213; \
59722  __ret_213 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_213), (uint32x2_t)(__noswap_vrshrn_n_u64(__rev1_213, __p2_213)))); \
59723  __ret_213 = __builtin_shufflevector(__ret_213, __ret_213, 3, 2, 1, 0); \
59724  __ret_213; \
59725})
59726#endif
59727
59728#ifdef __LITTLE_ENDIAN__
59729#define vrshrn_high_n_u16(__p0_214, __p1_214, __p2_214) __extension__ ({ \
59730  uint8x8_t __s0_214 = __p0_214; \
59731  uint16x8_t __s1_214 = __p1_214; \
59732  uint8x16_t __ret_214; \
59733  __ret_214 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_214), (uint8x8_t)(vrshrn_n_u16(__s1_214, __p2_214)))); \
59734  __ret_214; \
59735})
59736#else
59737#define vrshrn_high_n_u16(__p0_215, __p1_215, __p2_215) __extension__ ({ \
59738  uint8x8_t __s0_215 = __p0_215; \
59739  uint16x8_t __s1_215 = __p1_215; \
59740  uint8x8_t __rev0_215;  __rev0_215 = __builtin_shufflevector(__s0_215, __s0_215, 7, 6, 5, 4, 3, 2, 1, 0); \
59741  uint16x8_t __rev1_215;  __rev1_215 = __builtin_shufflevector(__s1_215, __s1_215, 7, 6, 5, 4, 3, 2, 1, 0); \
59742  uint8x16_t __ret_215; \
59743  __ret_215 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_215), (uint8x8_t)(__noswap_vrshrn_n_u16(__rev1_215, __p2_215)))); \
59744  __ret_215 = __builtin_shufflevector(__ret_215, __ret_215, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59745  __ret_215; \
59746})
59747#endif
59748
59749#ifdef __LITTLE_ENDIAN__
59750#define vrshrn_high_n_s32(__p0_216, __p1_216, __p2_216) __extension__ ({ \
59751  int16x4_t __s0_216 = __p0_216; \
59752  int32x4_t __s1_216 = __p1_216; \
59753  int16x8_t __ret_216; \
59754  __ret_216 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_216), (int16x4_t)(vrshrn_n_s32(__s1_216, __p2_216)))); \
59755  __ret_216; \
59756})
59757#else
59758#define vrshrn_high_n_s32(__p0_217, __p1_217, __p2_217) __extension__ ({ \
59759  int16x4_t __s0_217 = __p0_217; \
59760  int32x4_t __s1_217 = __p1_217; \
59761  int16x4_t __rev0_217;  __rev0_217 = __builtin_shufflevector(__s0_217, __s0_217, 3, 2, 1, 0); \
59762  int32x4_t __rev1_217;  __rev1_217 = __builtin_shufflevector(__s1_217, __s1_217, 3, 2, 1, 0); \
59763  int16x8_t __ret_217; \
59764  __ret_217 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_217), (int16x4_t)(__noswap_vrshrn_n_s32(__rev1_217, __p2_217)))); \
59765  __ret_217 = __builtin_shufflevector(__ret_217, __ret_217, 7, 6, 5, 4, 3, 2, 1, 0); \
59766  __ret_217; \
59767})
59768#endif
59769
59770#ifdef __LITTLE_ENDIAN__
59771#define vrshrn_high_n_s64(__p0_218, __p1_218, __p2_218) __extension__ ({ \
59772  int32x2_t __s0_218 = __p0_218; \
59773  int64x2_t __s1_218 = __p1_218; \
59774  int32x4_t __ret_218; \
59775  __ret_218 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_218), (int32x2_t)(vrshrn_n_s64(__s1_218, __p2_218)))); \
59776  __ret_218; \
59777})
59778#else
59779#define vrshrn_high_n_s64(__p0_219, __p1_219, __p2_219) __extension__ ({ \
59780  int32x2_t __s0_219 = __p0_219; \
59781  int64x2_t __s1_219 = __p1_219; \
59782  int32x2_t __rev0_219;  __rev0_219 = __builtin_shufflevector(__s0_219, __s0_219, 1, 0); \
59783  int64x2_t __rev1_219;  __rev1_219 = __builtin_shufflevector(__s1_219, __s1_219, 1, 0); \
59784  int32x4_t __ret_219; \
59785  __ret_219 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_219), (int32x2_t)(__noswap_vrshrn_n_s64(__rev1_219, __p2_219)))); \
59786  __ret_219 = __builtin_shufflevector(__ret_219, __ret_219, 3, 2, 1, 0); \
59787  __ret_219; \
59788})
59789#endif
59790
59791#ifdef __LITTLE_ENDIAN__
59792#define vrshrn_high_n_s16(__p0_220, __p1_220, __p2_220) __extension__ ({ \
59793  int8x8_t __s0_220 = __p0_220; \
59794  int16x8_t __s1_220 = __p1_220; \
59795  int8x16_t __ret_220; \
59796  __ret_220 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_220), (int8x8_t)(vrshrn_n_s16(__s1_220, __p2_220)))); \
59797  __ret_220; \
59798})
59799#else
59800#define vrshrn_high_n_s16(__p0_221, __p1_221, __p2_221) __extension__ ({ \
59801  int8x8_t __s0_221 = __p0_221; \
59802  int16x8_t __s1_221 = __p1_221; \
59803  int8x8_t __rev0_221;  __rev0_221 = __builtin_shufflevector(__s0_221, __s0_221, 7, 6, 5, 4, 3, 2, 1, 0); \
59804  int16x8_t __rev1_221;  __rev1_221 = __builtin_shufflevector(__s1_221, __s1_221, 7, 6, 5, 4, 3, 2, 1, 0); \
59805  int8x16_t __ret_221; \
59806  __ret_221 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_221), (int8x8_t)(__noswap_vrshrn_n_s16(__rev1_221, __p2_221)))); \
59807  __ret_221 = __builtin_shufflevector(__ret_221, __ret_221, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
59808  __ret_221; \
59809})
59810#endif
59811
59812#ifdef __LITTLE_ENDIAN__
59813__ai float64x2_t vrsqrteq_f64(float64x2_t __p0) {
59814  float64x2_t __ret;
59815  __ret = (float64x2_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 42);
59816  return __ret;
59817}
59818#else
59819__ai float64x2_t vrsqrteq_f64(float64x2_t __p0) {
59820  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59821  float64x2_t __ret;
59822  __ret = (float64x2_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 42);
59823  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59824  return __ret;
59825}
59826#endif
59827
59828#ifdef __LITTLE_ENDIAN__
59829__ai float64x1_t vrsqrte_f64(float64x1_t __p0) {
59830  float64x1_t __ret;
59831  __ret = (float64x1_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 10);
59832  return __ret;
59833}
59834#else
59835__ai float64x1_t vrsqrte_f64(float64x1_t __p0) {
59836  float64x1_t __ret;
59837  __ret = (float64x1_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 10);
59838  return __ret;
59839}
59840#endif
59841
59842#ifdef __LITTLE_ENDIAN__
59843__ai float64_t vrsqrted_f64(float64_t __p0) {
59844  float64_t __ret;
59845  __ret = (float64_t) __builtin_neon_vrsqrted_f64(__p0);
59846  return __ret;
59847}
59848#else
59849__ai float64_t vrsqrted_f64(float64_t __p0) {
59850  float64_t __ret;
59851  __ret = (float64_t) __builtin_neon_vrsqrted_f64(__p0);
59852  return __ret;
59853}
59854#endif
59855
59856#ifdef __LITTLE_ENDIAN__
59857__ai float32_t vrsqrtes_f32(float32_t __p0) {
59858  float32_t __ret;
59859  __ret = (float32_t) __builtin_neon_vrsqrtes_f32(__p0);
59860  return __ret;
59861}
59862#else
59863__ai float32_t vrsqrtes_f32(float32_t __p0) {
59864  float32_t __ret;
59865  __ret = (float32_t) __builtin_neon_vrsqrtes_f32(__p0);
59866  return __ret;
59867}
59868#endif
59869
59870#ifdef __LITTLE_ENDIAN__
59871__ai float64x2_t vrsqrtsq_f64(float64x2_t __p0, float64x2_t __p1) {
59872  float64x2_t __ret;
59873  __ret = (float64x2_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
59874  return __ret;
59875}
59876#else
59877__ai float64x2_t vrsqrtsq_f64(float64x2_t __p0, float64x2_t __p1) {
59878  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59879  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59880  float64x2_t __ret;
59881  __ret = (float64x2_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
59882  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
59883  return __ret;
59884}
59885#endif
59886
59887#ifdef __LITTLE_ENDIAN__
59888__ai float64x1_t vrsqrts_f64(float64x1_t __p0, float64x1_t __p1) {
59889  float64x1_t __ret;
59890  __ret = (float64x1_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
59891  return __ret;
59892}
59893#else
59894__ai float64x1_t vrsqrts_f64(float64x1_t __p0, float64x1_t __p1) {
59895  float64x1_t __ret;
59896  __ret = (float64x1_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
59897  return __ret;
59898}
59899#endif
59900
59901#ifdef __LITTLE_ENDIAN__
59902__ai float64_t vrsqrtsd_f64(float64_t __p0, float64_t __p1) {
59903  float64_t __ret;
59904  __ret = (float64_t) __builtin_neon_vrsqrtsd_f64(__p0, __p1);
59905  return __ret;
59906}
59907#else
59908__ai float64_t vrsqrtsd_f64(float64_t __p0, float64_t __p1) {
59909  float64_t __ret;
59910  __ret = (float64_t) __builtin_neon_vrsqrtsd_f64(__p0, __p1);
59911  return __ret;
59912}
59913#endif
59914
59915#ifdef __LITTLE_ENDIAN__
59916__ai float32_t vrsqrtss_f32(float32_t __p0, float32_t __p1) {
59917  float32_t __ret;
59918  __ret = (float32_t) __builtin_neon_vrsqrtss_f32(__p0, __p1);
59919  return __ret;
59920}
59921#else
59922__ai float32_t vrsqrtss_f32(float32_t __p0, float32_t __p1) {
59923  float32_t __ret;
59924  __ret = (float32_t) __builtin_neon_vrsqrtss_f32(__p0, __p1);
59925  return __ret;
59926}
59927#endif
59928
59929#ifdef __LITTLE_ENDIAN__
59930#define vrsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
59931  uint64_t __s0 = __p0; \
59932  uint64_t __s1 = __p1; \
59933  uint64_t __ret; \
59934  __ret = (uint64_t) __builtin_neon_vrsrad_n_u64(__s0, __s1, __p2); \
59935  __ret; \
59936})
59937#else
59938#define vrsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
59939  uint64_t __s0 = __p0; \
59940  uint64_t __s1 = __p1; \
59941  uint64_t __ret; \
59942  __ret = (uint64_t) __builtin_neon_vrsrad_n_u64(__s0, __s1, __p2); \
59943  __ret; \
59944})
59945#endif
59946
59947#ifdef __LITTLE_ENDIAN__
59948#define vrsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
59949  int64_t __s0 = __p0; \
59950  int64_t __s1 = __p1; \
59951  int64_t __ret; \
59952  __ret = (int64_t) __builtin_neon_vrsrad_n_s64(__s0, __s1, __p2); \
59953  __ret; \
59954})
59955#else
59956#define vrsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
59957  int64_t __s0 = __p0; \
59958  int64_t __s1 = __p1; \
59959  int64_t __ret; \
59960  __ret = (int64_t) __builtin_neon_vrsrad_n_s64(__s0, __s1, __p2); \
59961  __ret; \
59962})
59963#endif
59964
59965#ifdef __LITTLE_ENDIAN__
59966__ai uint16x8_t vrsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
59967  uint16x8_t __ret;
59968  __ret = vcombine_u16(__p0, vrsubhn_u32(__p1, __p2));
59969  return __ret;
59970}
59971#else
59972__ai uint16x8_t vrsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
59973  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59974  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59975  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
59976  uint16x8_t __ret;
59977  __ret = __noswap_vcombine_u16(__rev0, __noswap_vrsubhn_u32(__rev1, __rev2));
59978  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59979  return __ret;
59980}
59981#endif
59982
59983#ifdef __LITTLE_ENDIAN__
59984__ai uint32x4_t vrsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
59985  uint32x4_t __ret;
59986  __ret = vcombine_u32(__p0, vrsubhn_u64(__p1, __p2));
59987  return __ret;
59988}
59989#else
59990__ai uint32x4_t vrsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
59991  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59992  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59993  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
59994  uint32x4_t __ret;
59995  __ret = __noswap_vcombine_u32(__rev0, __noswap_vrsubhn_u64(__rev1, __rev2));
59996  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59997  return __ret;
59998}
59999#endif
60000
60001#ifdef __LITTLE_ENDIAN__
60002__ai uint8x16_t vrsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
60003  uint8x16_t __ret;
60004  __ret = vcombine_u8(__p0, vrsubhn_u16(__p1, __p2));
60005  return __ret;
60006}
60007#else
60008__ai uint8x16_t vrsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
60009  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60010  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60011  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
60012  uint8x16_t __ret;
60013  __ret = __noswap_vcombine_u8(__rev0, __noswap_vrsubhn_u16(__rev1, __rev2));
60014  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60015  return __ret;
60016}
60017#endif
60018
60019#ifdef __LITTLE_ENDIAN__
60020__ai int16x8_t vrsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
60021  int16x8_t __ret;
60022  __ret = vcombine_s16(__p0, vrsubhn_s32(__p1, __p2));
60023  return __ret;
60024}
60025#else
60026__ai int16x8_t vrsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
60027  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60028  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60029  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
60030  int16x8_t __ret;
60031  __ret = __noswap_vcombine_s16(__rev0, __noswap_vrsubhn_s32(__rev1, __rev2));
60032  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60033  return __ret;
60034}
60035#endif
60036
60037#ifdef __LITTLE_ENDIAN__
60038__ai int32x4_t vrsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
60039  int32x4_t __ret;
60040  __ret = vcombine_s32(__p0, vrsubhn_s64(__p1, __p2));
60041  return __ret;
60042}
60043#else
60044__ai int32x4_t vrsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
60045  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60046  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60047  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
60048  int32x4_t __ret;
60049  __ret = __noswap_vcombine_s32(__rev0, __noswap_vrsubhn_s64(__rev1, __rev2));
60050  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60051  return __ret;
60052}
60053#endif
60054
60055#ifdef __LITTLE_ENDIAN__
60056__ai int8x16_t vrsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
60057  int8x16_t __ret;
60058  __ret = vcombine_s8(__p0, vrsubhn_s16(__p1, __p2));
60059  return __ret;
60060}
60061#else
60062__ai int8x16_t vrsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
60063  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60064  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60065  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
60066  int8x16_t __ret;
60067  __ret = __noswap_vcombine_s8(__rev0, __noswap_vrsubhn_s16(__rev1, __rev2));
60068  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60069  return __ret;
60070}
60071#endif
60072
60073#ifdef __LITTLE_ENDIAN__
60074#define vset_lane_f16(__p0_222, __p1_222, __p2_222) __extension__ ({ \
60075  float16_t __s0_222 = __p0_222; \
60076  float16x4_t __s1_222 = __p1_222; \
60077  float16x4_t __ret_222; \
60078float16_t __reint_222 = __s0_222; \
60079float16x4_t __reint1_222 = __s1_222; \
60080int16x4_t __reint2_222 = vset_lane_s16(*(int16_t *) &__reint_222, *(int16x4_t *) &__reint1_222, __p2_222); \
60081  __ret_222 = *(float16x4_t *) &__reint2_222; \
60082  __ret_222; \
60083})
60084#else
60085#define vset_lane_f16(__p0_223, __p1_223, __p2_223) __extension__ ({ \
60086  float16_t __s0_223 = __p0_223; \
60087  float16x4_t __s1_223 = __p1_223; \
60088  float16x4_t __rev1_223;  __rev1_223 = __builtin_shufflevector(__s1_223, __s1_223, 3, 2, 1, 0); \
60089  float16x4_t __ret_223; \
60090float16_t __reint_223 = __s0_223; \
60091float16x4_t __reint1_223 = __rev1_223; \
60092int16x4_t __reint2_223 = __noswap_vset_lane_s16(*(int16_t *) &__reint_223, *(int16x4_t *) &__reint1_223, __p2_223); \
60093  __ret_223 = *(float16x4_t *) &__reint2_223; \
60094  __ret_223 = __builtin_shufflevector(__ret_223, __ret_223, 3, 2, 1, 0); \
60095  __ret_223; \
60096})
60097#endif
60098
60099#ifdef __LITTLE_ENDIAN__
60100#define vsetq_lane_f16(__p0_224, __p1_224, __p2_224) __extension__ ({ \
60101  float16_t __s0_224 = __p0_224; \
60102  float16x8_t __s1_224 = __p1_224; \
60103  float16x8_t __ret_224; \
60104float16_t __reint_224 = __s0_224; \
60105float16x8_t __reint1_224 = __s1_224; \
60106int16x8_t __reint2_224 = vsetq_lane_s16(*(int16_t *) &__reint_224, *(int16x8_t *) &__reint1_224, __p2_224); \
60107  __ret_224 = *(float16x8_t *) &__reint2_224; \
60108  __ret_224; \
60109})
60110#else
60111#define vsetq_lane_f16(__p0_225, __p1_225, __p2_225) __extension__ ({ \
60112  float16_t __s0_225 = __p0_225; \
60113  float16x8_t __s1_225 = __p1_225; \
60114  float16x8_t __rev1_225;  __rev1_225 = __builtin_shufflevector(__s1_225, __s1_225, 7, 6, 5, 4, 3, 2, 1, 0); \
60115  float16x8_t __ret_225; \
60116float16_t __reint_225 = __s0_225; \
60117float16x8_t __reint1_225 = __rev1_225; \
60118int16x8_t __reint2_225 = __noswap_vsetq_lane_s16(*(int16_t *) &__reint_225, *(int16x8_t *) &__reint1_225, __p2_225); \
60119  __ret_225 = *(float16x8_t *) &__reint2_225; \
60120  __ret_225 = __builtin_shufflevector(__ret_225, __ret_225, 7, 6, 5, 4, 3, 2, 1, 0); \
60121  __ret_225; \
60122})
60123#endif
60124
60125#ifdef __LITTLE_ENDIAN__
60126#define vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60127  poly64_t __s0 = __p0; \
60128  poly64x1_t __s1 = __p1; \
60129  poly64x1_t __ret; \
60130  __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
60131  __ret; \
60132})
60133#else
60134#define vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60135  poly64_t __s0 = __p0; \
60136  poly64x1_t __s1 = __p1; \
60137  poly64x1_t __ret; \
60138  __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
60139  __ret; \
60140})
60141#define __noswap_vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60142  poly64_t __s0 = __p0; \
60143  poly64x1_t __s1 = __p1; \
60144  poly64x1_t __ret; \
60145  __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
60146  __ret; \
60147})
60148#endif
60149
60150#ifdef __LITTLE_ENDIAN__
60151#define vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60152  poly64_t __s0 = __p0; \
60153  poly64x2_t __s1 = __p1; \
60154  poly64x2_t __ret; \
60155  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
60156  __ret; \
60157})
60158#else
60159#define vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60160  poly64_t __s0 = __p0; \
60161  poly64x2_t __s1 = __p1; \
60162  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60163  poly64x2_t __ret; \
60164  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
60165  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
60166  __ret; \
60167})
60168#define __noswap_vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60169  poly64_t __s0 = __p0; \
60170  poly64x2_t __s1 = __p1; \
60171  poly64x2_t __ret; \
60172  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
60173  __ret; \
60174})
60175#endif
60176
60177#ifdef __LITTLE_ENDIAN__
60178#define vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60179  float64_t __s0 = __p0; \
60180  float64x2_t __s1 = __p1; \
60181  float64x2_t __ret; \
60182  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__s1, __p2); \
60183  __ret; \
60184})
60185#else
60186#define vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60187  float64_t __s0 = __p0; \
60188  float64x2_t __s1 = __p1; \
60189  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60190  float64x2_t __ret; \
60191  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__rev1, __p2); \
60192  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
60193  __ret; \
60194})
60195#define __noswap_vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60196  float64_t __s0 = __p0; \
60197  float64x2_t __s1 = __p1; \
60198  float64x2_t __ret; \
60199  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__s1, __p2); \
60200  __ret; \
60201})
60202#endif
60203
60204#ifdef __LITTLE_ENDIAN__
60205#define vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60206  float64_t __s0 = __p0; \
60207  float64x1_t __s1 = __p1; \
60208  float64x1_t __ret; \
60209  __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
60210  __ret; \
60211})
60212#else
60213#define vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60214  float64_t __s0 = __p0; \
60215  float64x1_t __s1 = __p1; \
60216  float64x1_t __ret; \
60217  __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
60218  __ret; \
60219})
60220#define __noswap_vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60221  float64_t __s0 = __p0; \
60222  float64x1_t __s1 = __p1; \
60223  float64x1_t __ret; \
60224  __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
60225  __ret; \
60226})
60227#endif
60228
60229#ifdef __LITTLE_ENDIAN__
60230__ai uint64_t vshld_u64(uint64_t __p0, uint64_t __p1) {
60231  uint64_t __ret;
60232  __ret = (uint64_t) __builtin_neon_vshld_u64(__p0, __p1);
60233  return __ret;
60234}
60235#else
60236__ai uint64_t vshld_u64(uint64_t __p0, uint64_t __p1) {
60237  uint64_t __ret;
60238  __ret = (uint64_t) __builtin_neon_vshld_u64(__p0, __p1);
60239  return __ret;
60240}
60241#endif
60242
60243#ifdef __LITTLE_ENDIAN__
60244__ai int64_t vshld_s64(int64_t __p0, int64_t __p1) {
60245  int64_t __ret;
60246  __ret = (int64_t) __builtin_neon_vshld_s64(__p0, __p1);
60247  return __ret;
60248}
60249#else
60250__ai int64_t vshld_s64(int64_t __p0, int64_t __p1) {
60251  int64_t __ret;
60252  __ret = (int64_t) __builtin_neon_vshld_s64(__p0, __p1);
60253  return __ret;
60254}
60255#endif
60256
60257#ifdef __LITTLE_ENDIAN__
60258#define vshld_n_u64(__p0, __p1) __extension__ ({ \
60259  uint64_t __s0 = __p0; \
60260  uint64_t __ret; \
60261  __ret = (uint64_t) __builtin_neon_vshld_n_u64(__s0, __p1); \
60262  __ret; \
60263})
60264#else
60265#define vshld_n_u64(__p0, __p1) __extension__ ({ \
60266  uint64_t __s0 = __p0; \
60267  uint64_t __ret; \
60268  __ret = (uint64_t) __builtin_neon_vshld_n_u64(__s0, __p1); \
60269  __ret; \
60270})
60271#endif
60272
60273#ifdef __LITTLE_ENDIAN__
60274#define vshld_n_s64(__p0, __p1) __extension__ ({ \
60275  int64_t __s0 = __p0; \
60276  int64_t __ret; \
60277  __ret = (int64_t) __builtin_neon_vshld_n_s64(__s0, __p1); \
60278  __ret; \
60279})
60280#else
60281#define vshld_n_s64(__p0, __p1) __extension__ ({ \
60282  int64_t __s0 = __p0; \
60283  int64_t __ret; \
60284  __ret = (int64_t) __builtin_neon_vshld_n_s64(__s0, __p1); \
60285  __ret; \
60286})
60287#endif
60288
60289#ifdef __LITTLE_ENDIAN__
60290#define vshll_high_n_u8(__p0_226, __p1_226) __extension__ ({ \
60291  uint8x16_t __s0_226 = __p0_226; \
60292  uint16x8_t __ret_226; \
60293  __ret_226 = (uint16x8_t)(vshll_n_u8(vget_high_u8(__s0_226), __p1_226)); \
60294  __ret_226; \
60295})
60296#else
60297#define vshll_high_n_u8(__p0_227, __p1_227) __extension__ ({ \
60298  uint8x16_t __s0_227 = __p0_227; \
60299  uint8x16_t __rev0_227;  __rev0_227 = __builtin_shufflevector(__s0_227, __s0_227, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60300  uint16x8_t __ret_227; \
60301  __ret_227 = (uint16x8_t)(__noswap_vshll_n_u8(__noswap_vget_high_u8(__rev0_227), __p1_227)); \
60302  __ret_227 = __builtin_shufflevector(__ret_227, __ret_227, 7, 6, 5, 4, 3, 2, 1, 0); \
60303  __ret_227; \
60304})
60305#endif
60306
60307#ifdef __LITTLE_ENDIAN__
60308#define vshll_high_n_u32(__p0_228, __p1_228) __extension__ ({ \
60309  uint32x4_t __s0_228 = __p0_228; \
60310  uint64x2_t __ret_228; \
60311  __ret_228 = (uint64x2_t)(vshll_n_u32(vget_high_u32(__s0_228), __p1_228)); \
60312  __ret_228; \
60313})
60314#else
60315#define vshll_high_n_u32(__p0_229, __p1_229) __extension__ ({ \
60316  uint32x4_t __s0_229 = __p0_229; \
60317  uint32x4_t __rev0_229;  __rev0_229 = __builtin_shufflevector(__s0_229, __s0_229, 3, 2, 1, 0); \
60318  uint64x2_t __ret_229; \
60319  __ret_229 = (uint64x2_t)(__noswap_vshll_n_u32(__noswap_vget_high_u32(__rev0_229), __p1_229)); \
60320  __ret_229 = __builtin_shufflevector(__ret_229, __ret_229, 1, 0); \
60321  __ret_229; \
60322})
60323#endif
60324
60325#ifdef __LITTLE_ENDIAN__
60326#define vshll_high_n_u16(__p0_230, __p1_230) __extension__ ({ \
60327  uint16x8_t __s0_230 = __p0_230; \
60328  uint32x4_t __ret_230; \
60329  __ret_230 = (uint32x4_t)(vshll_n_u16(vget_high_u16(__s0_230), __p1_230)); \
60330  __ret_230; \
60331})
60332#else
60333#define vshll_high_n_u16(__p0_231, __p1_231) __extension__ ({ \
60334  uint16x8_t __s0_231 = __p0_231; \
60335  uint16x8_t __rev0_231;  __rev0_231 = __builtin_shufflevector(__s0_231, __s0_231, 7, 6, 5, 4, 3, 2, 1, 0); \
60336  uint32x4_t __ret_231; \
60337  __ret_231 = (uint32x4_t)(__noswap_vshll_n_u16(__noswap_vget_high_u16(__rev0_231), __p1_231)); \
60338  __ret_231 = __builtin_shufflevector(__ret_231, __ret_231, 3, 2, 1, 0); \
60339  __ret_231; \
60340})
60341#endif
60342
60343#ifdef __LITTLE_ENDIAN__
60344#define vshll_high_n_s8(__p0_232, __p1_232) __extension__ ({ \
60345  int8x16_t __s0_232 = __p0_232; \
60346  int16x8_t __ret_232; \
60347  __ret_232 = (int16x8_t)(vshll_n_s8(vget_high_s8(__s0_232), __p1_232)); \
60348  __ret_232; \
60349})
60350#else
60351#define vshll_high_n_s8(__p0_233, __p1_233) __extension__ ({ \
60352  int8x16_t __s0_233 = __p0_233; \
60353  int8x16_t __rev0_233;  __rev0_233 = __builtin_shufflevector(__s0_233, __s0_233, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60354  int16x8_t __ret_233; \
60355  __ret_233 = (int16x8_t)(__noswap_vshll_n_s8(__noswap_vget_high_s8(__rev0_233), __p1_233)); \
60356  __ret_233 = __builtin_shufflevector(__ret_233, __ret_233, 7, 6, 5, 4, 3, 2, 1, 0); \
60357  __ret_233; \
60358})
60359#endif
60360
60361#ifdef __LITTLE_ENDIAN__
60362#define vshll_high_n_s32(__p0_234, __p1_234) __extension__ ({ \
60363  int32x4_t __s0_234 = __p0_234; \
60364  int64x2_t __ret_234; \
60365  __ret_234 = (int64x2_t)(vshll_n_s32(vget_high_s32(__s0_234), __p1_234)); \
60366  __ret_234; \
60367})
60368#else
60369#define vshll_high_n_s32(__p0_235, __p1_235) __extension__ ({ \
60370  int32x4_t __s0_235 = __p0_235; \
60371  int32x4_t __rev0_235;  __rev0_235 = __builtin_shufflevector(__s0_235, __s0_235, 3, 2, 1, 0); \
60372  int64x2_t __ret_235; \
60373  __ret_235 = (int64x2_t)(__noswap_vshll_n_s32(__noswap_vget_high_s32(__rev0_235), __p1_235)); \
60374  __ret_235 = __builtin_shufflevector(__ret_235, __ret_235, 1, 0); \
60375  __ret_235; \
60376})
60377#endif
60378
60379#ifdef __LITTLE_ENDIAN__
60380#define vshll_high_n_s16(__p0_236, __p1_236) __extension__ ({ \
60381  int16x8_t __s0_236 = __p0_236; \
60382  int32x4_t __ret_236; \
60383  __ret_236 = (int32x4_t)(vshll_n_s16(vget_high_s16(__s0_236), __p1_236)); \
60384  __ret_236; \
60385})
60386#else
60387#define vshll_high_n_s16(__p0_237, __p1_237) __extension__ ({ \
60388  int16x8_t __s0_237 = __p0_237; \
60389  int16x8_t __rev0_237;  __rev0_237 = __builtin_shufflevector(__s0_237, __s0_237, 7, 6, 5, 4, 3, 2, 1, 0); \
60390  int32x4_t __ret_237; \
60391  __ret_237 = (int32x4_t)(__noswap_vshll_n_s16(__noswap_vget_high_s16(__rev0_237), __p1_237)); \
60392  __ret_237 = __builtin_shufflevector(__ret_237, __ret_237, 3, 2, 1, 0); \
60393  __ret_237; \
60394})
60395#endif
60396
60397#ifdef __LITTLE_ENDIAN__
60398#define vshrd_n_u64(__p0, __p1) __extension__ ({ \
60399  uint64_t __s0 = __p0; \
60400  uint64_t __ret; \
60401  __ret = (uint64_t) __builtin_neon_vshrd_n_u64(__s0, __p1); \
60402  __ret; \
60403})
60404#else
60405#define vshrd_n_u64(__p0, __p1) __extension__ ({ \
60406  uint64_t __s0 = __p0; \
60407  uint64_t __ret; \
60408  __ret = (uint64_t) __builtin_neon_vshrd_n_u64(__s0, __p1); \
60409  __ret; \
60410})
60411#endif
60412
60413#ifdef __LITTLE_ENDIAN__
60414#define vshrd_n_s64(__p0, __p1) __extension__ ({ \
60415  int64_t __s0 = __p0; \
60416  int64_t __ret; \
60417  __ret = (int64_t) __builtin_neon_vshrd_n_s64(__s0, __p1); \
60418  __ret; \
60419})
60420#else
60421#define vshrd_n_s64(__p0, __p1) __extension__ ({ \
60422  int64_t __s0 = __p0; \
60423  int64_t __ret; \
60424  __ret = (int64_t) __builtin_neon_vshrd_n_s64(__s0, __p1); \
60425  __ret; \
60426})
60427#endif
60428
60429#ifdef __LITTLE_ENDIAN__
60430#define vshrn_high_n_u32(__p0_238, __p1_238, __p2_238) __extension__ ({ \
60431  uint16x4_t __s0_238 = __p0_238; \
60432  uint32x4_t __s1_238 = __p1_238; \
60433  uint16x8_t __ret_238; \
60434  __ret_238 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_238), (uint16x4_t)(vshrn_n_u32(__s1_238, __p2_238)))); \
60435  __ret_238; \
60436})
60437#else
60438#define vshrn_high_n_u32(__p0_239, __p1_239, __p2_239) __extension__ ({ \
60439  uint16x4_t __s0_239 = __p0_239; \
60440  uint32x4_t __s1_239 = __p1_239; \
60441  uint16x4_t __rev0_239;  __rev0_239 = __builtin_shufflevector(__s0_239, __s0_239, 3, 2, 1, 0); \
60442  uint32x4_t __rev1_239;  __rev1_239 = __builtin_shufflevector(__s1_239, __s1_239, 3, 2, 1, 0); \
60443  uint16x8_t __ret_239; \
60444  __ret_239 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_239), (uint16x4_t)(__noswap_vshrn_n_u32(__rev1_239, __p2_239)))); \
60445  __ret_239 = __builtin_shufflevector(__ret_239, __ret_239, 7, 6, 5, 4, 3, 2, 1, 0); \
60446  __ret_239; \
60447})
60448#endif
60449
60450#ifdef __LITTLE_ENDIAN__
60451#define vshrn_high_n_u64(__p0_240, __p1_240, __p2_240) __extension__ ({ \
60452  uint32x2_t __s0_240 = __p0_240; \
60453  uint64x2_t __s1_240 = __p1_240; \
60454  uint32x4_t __ret_240; \
60455  __ret_240 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_240), (uint32x2_t)(vshrn_n_u64(__s1_240, __p2_240)))); \
60456  __ret_240; \
60457})
60458#else
60459#define vshrn_high_n_u64(__p0_241, __p1_241, __p2_241) __extension__ ({ \
60460  uint32x2_t __s0_241 = __p0_241; \
60461  uint64x2_t __s1_241 = __p1_241; \
60462  uint32x2_t __rev0_241;  __rev0_241 = __builtin_shufflevector(__s0_241, __s0_241, 1, 0); \
60463  uint64x2_t __rev1_241;  __rev1_241 = __builtin_shufflevector(__s1_241, __s1_241, 1, 0); \
60464  uint32x4_t __ret_241; \
60465  __ret_241 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_241), (uint32x2_t)(__noswap_vshrn_n_u64(__rev1_241, __p2_241)))); \
60466  __ret_241 = __builtin_shufflevector(__ret_241, __ret_241, 3, 2, 1, 0); \
60467  __ret_241; \
60468})
60469#endif
60470
60471#ifdef __LITTLE_ENDIAN__
60472#define vshrn_high_n_u16(__p0_242, __p1_242, __p2_242) __extension__ ({ \
60473  uint8x8_t __s0_242 = __p0_242; \
60474  uint16x8_t __s1_242 = __p1_242; \
60475  uint8x16_t __ret_242; \
60476  __ret_242 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_242), (uint8x8_t)(vshrn_n_u16(__s1_242, __p2_242)))); \
60477  __ret_242; \
60478})
60479#else
60480#define vshrn_high_n_u16(__p0_243, __p1_243, __p2_243) __extension__ ({ \
60481  uint8x8_t __s0_243 = __p0_243; \
60482  uint16x8_t __s1_243 = __p1_243; \
60483  uint8x8_t __rev0_243;  __rev0_243 = __builtin_shufflevector(__s0_243, __s0_243, 7, 6, 5, 4, 3, 2, 1, 0); \
60484  uint16x8_t __rev1_243;  __rev1_243 = __builtin_shufflevector(__s1_243, __s1_243, 7, 6, 5, 4, 3, 2, 1, 0); \
60485  uint8x16_t __ret_243; \
60486  __ret_243 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_243), (uint8x8_t)(__noswap_vshrn_n_u16(__rev1_243, __p2_243)))); \
60487  __ret_243 = __builtin_shufflevector(__ret_243, __ret_243, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60488  __ret_243; \
60489})
60490#endif
60491
60492#ifdef __LITTLE_ENDIAN__
60493#define vshrn_high_n_s32(__p0_244, __p1_244, __p2_244) __extension__ ({ \
60494  int16x4_t __s0_244 = __p0_244; \
60495  int32x4_t __s1_244 = __p1_244; \
60496  int16x8_t __ret_244; \
60497  __ret_244 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_244), (int16x4_t)(vshrn_n_s32(__s1_244, __p2_244)))); \
60498  __ret_244; \
60499})
60500#else
60501#define vshrn_high_n_s32(__p0_245, __p1_245, __p2_245) __extension__ ({ \
60502  int16x4_t __s0_245 = __p0_245; \
60503  int32x4_t __s1_245 = __p1_245; \
60504  int16x4_t __rev0_245;  __rev0_245 = __builtin_shufflevector(__s0_245, __s0_245, 3, 2, 1, 0); \
60505  int32x4_t __rev1_245;  __rev1_245 = __builtin_shufflevector(__s1_245, __s1_245, 3, 2, 1, 0); \
60506  int16x8_t __ret_245; \
60507  __ret_245 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_245), (int16x4_t)(__noswap_vshrn_n_s32(__rev1_245, __p2_245)))); \
60508  __ret_245 = __builtin_shufflevector(__ret_245, __ret_245, 7, 6, 5, 4, 3, 2, 1, 0); \
60509  __ret_245; \
60510})
60511#endif
60512
60513#ifdef __LITTLE_ENDIAN__
60514#define vshrn_high_n_s64(__p0_246, __p1_246, __p2_246) __extension__ ({ \
60515  int32x2_t __s0_246 = __p0_246; \
60516  int64x2_t __s1_246 = __p1_246; \
60517  int32x4_t __ret_246; \
60518  __ret_246 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_246), (int32x2_t)(vshrn_n_s64(__s1_246, __p2_246)))); \
60519  __ret_246; \
60520})
60521#else
60522#define vshrn_high_n_s64(__p0_247, __p1_247, __p2_247) __extension__ ({ \
60523  int32x2_t __s0_247 = __p0_247; \
60524  int64x2_t __s1_247 = __p1_247; \
60525  int32x2_t __rev0_247;  __rev0_247 = __builtin_shufflevector(__s0_247, __s0_247, 1, 0); \
60526  int64x2_t __rev1_247;  __rev1_247 = __builtin_shufflevector(__s1_247, __s1_247, 1, 0); \
60527  int32x4_t __ret_247; \
60528  __ret_247 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_247), (int32x2_t)(__noswap_vshrn_n_s64(__rev1_247, __p2_247)))); \
60529  __ret_247 = __builtin_shufflevector(__ret_247, __ret_247, 3, 2, 1, 0); \
60530  __ret_247; \
60531})
60532#endif
60533
60534#ifdef __LITTLE_ENDIAN__
60535#define vshrn_high_n_s16(__p0_248, __p1_248, __p2_248) __extension__ ({ \
60536  int8x8_t __s0_248 = __p0_248; \
60537  int16x8_t __s1_248 = __p1_248; \
60538  int8x16_t __ret_248; \
60539  __ret_248 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_248), (int8x8_t)(vshrn_n_s16(__s1_248, __p2_248)))); \
60540  __ret_248; \
60541})
60542#else
60543#define vshrn_high_n_s16(__p0_249, __p1_249, __p2_249) __extension__ ({ \
60544  int8x8_t __s0_249 = __p0_249; \
60545  int16x8_t __s1_249 = __p1_249; \
60546  int8x8_t __rev0_249;  __rev0_249 = __builtin_shufflevector(__s0_249, __s0_249, 7, 6, 5, 4, 3, 2, 1, 0); \
60547  int16x8_t __rev1_249;  __rev1_249 = __builtin_shufflevector(__s1_249, __s1_249, 7, 6, 5, 4, 3, 2, 1, 0); \
60548  int8x16_t __ret_249; \
60549  __ret_249 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_249), (int8x8_t)(__noswap_vshrn_n_s16(__rev1_249, __p2_249)))); \
60550  __ret_249 = __builtin_shufflevector(__ret_249, __ret_249, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60551  __ret_249; \
60552})
60553#endif
60554
60555#ifdef __LITTLE_ENDIAN__
60556#define vslid_n_u64(__p0, __p1, __p2) __extension__ ({ \
60557  uint64_t __s0 = __p0; \
60558  uint64_t __s1 = __p1; \
60559  uint64_t __ret; \
60560  __ret = (uint64_t) __builtin_neon_vslid_n_u64(__s0, __s1, __p2); \
60561  __ret; \
60562})
60563#else
60564#define vslid_n_u64(__p0, __p1, __p2) __extension__ ({ \
60565  uint64_t __s0 = __p0; \
60566  uint64_t __s1 = __p1; \
60567  uint64_t __ret; \
60568  __ret = (uint64_t) __builtin_neon_vslid_n_u64(__s0, __s1, __p2); \
60569  __ret; \
60570})
60571#endif
60572
60573#ifdef __LITTLE_ENDIAN__
60574#define vslid_n_s64(__p0, __p1, __p2) __extension__ ({ \
60575  int64_t __s0 = __p0; \
60576  int64_t __s1 = __p1; \
60577  int64_t __ret; \
60578  __ret = (int64_t) __builtin_neon_vslid_n_s64(__s0, __s1, __p2); \
60579  __ret; \
60580})
60581#else
60582#define vslid_n_s64(__p0, __p1, __p2) __extension__ ({ \
60583  int64_t __s0 = __p0; \
60584  int64_t __s1 = __p1; \
60585  int64_t __ret; \
60586  __ret = (int64_t) __builtin_neon_vslid_n_s64(__s0, __s1, __p2); \
60587  __ret; \
60588})
60589#endif
60590
60591#ifdef __LITTLE_ENDIAN__
60592#define vsli_n_p64(__p0, __p1, __p2) __extension__ ({ \
60593  poly64x1_t __s0 = __p0; \
60594  poly64x1_t __s1 = __p1; \
60595  poly64x1_t __ret; \
60596  __ret = (poly64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
60597  __ret; \
60598})
60599#else
60600#define vsli_n_p64(__p0, __p1, __p2) __extension__ ({ \
60601  poly64x1_t __s0 = __p0; \
60602  poly64x1_t __s1 = __p1; \
60603  poly64x1_t __ret; \
60604  __ret = (poly64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
60605  __ret; \
60606})
60607#endif
60608
60609#ifdef __LITTLE_ENDIAN__
60610#define vsliq_n_p64(__p0, __p1, __p2) __extension__ ({ \
60611  poly64x2_t __s0 = __p0; \
60612  poly64x2_t __s1 = __p1; \
60613  poly64x2_t __ret; \
60614  __ret = (poly64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
60615  __ret; \
60616})
60617#else
60618#define vsliq_n_p64(__p0, __p1, __p2) __extension__ ({ \
60619  poly64x2_t __s0 = __p0; \
60620  poly64x2_t __s1 = __p1; \
60621  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
60622  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60623  poly64x2_t __ret; \
60624  __ret = (poly64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
60625  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
60626  __ret; \
60627})
60628#endif
60629
60630#ifdef __LITTLE_ENDIAN__
60631__ai uint8_t vsqaddb_u8(uint8_t __p0, uint8_t __p1) {
60632  uint8_t __ret;
60633  __ret = (uint8_t) __builtin_neon_vsqaddb_u8(__p0, __p1);
60634  return __ret;
60635}
60636#else
60637__ai uint8_t vsqaddb_u8(uint8_t __p0, uint8_t __p1) {
60638  uint8_t __ret;
60639  __ret = (uint8_t) __builtin_neon_vsqaddb_u8(__p0, __p1);
60640  return __ret;
60641}
60642#endif
60643
60644#ifdef __LITTLE_ENDIAN__
60645__ai uint32_t vsqadds_u32(uint32_t __p0, uint32_t __p1) {
60646  uint32_t __ret;
60647  __ret = (uint32_t) __builtin_neon_vsqadds_u32(__p0, __p1);
60648  return __ret;
60649}
60650#else
60651__ai uint32_t vsqadds_u32(uint32_t __p0, uint32_t __p1) {
60652  uint32_t __ret;
60653  __ret = (uint32_t) __builtin_neon_vsqadds_u32(__p0, __p1);
60654  return __ret;
60655}
60656#endif
60657
60658#ifdef __LITTLE_ENDIAN__
60659__ai uint64_t vsqaddd_u64(uint64_t __p0, uint64_t __p1) {
60660  uint64_t __ret;
60661  __ret = (uint64_t) __builtin_neon_vsqaddd_u64(__p0, __p1);
60662  return __ret;
60663}
60664#else
60665__ai uint64_t vsqaddd_u64(uint64_t __p0, uint64_t __p1) {
60666  uint64_t __ret;
60667  __ret = (uint64_t) __builtin_neon_vsqaddd_u64(__p0, __p1);
60668  return __ret;
60669}
60670#endif
60671
60672#ifdef __LITTLE_ENDIAN__
60673__ai uint16_t vsqaddh_u16(uint16_t __p0, uint16_t __p1) {
60674  uint16_t __ret;
60675  __ret = (uint16_t) __builtin_neon_vsqaddh_u16(__p0, __p1);
60676  return __ret;
60677}
60678#else
60679__ai uint16_t vsqaddh_u16(uint16_t __p0, uint16_t __p1) {
60680  uint16_t __ret;
60681  __ret = (uint16_t) __builtin_neon_vsqaddh_u16(__p0, __p1);
60682  return __ret;
60683}
60684#endif
60685
60686#ifdef __LITTLE_ENDIAN__
60687__ai uint8x16_t vsqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
60688  uint8x16_t __ret;
60689  __ret = (uint8x16_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
60690  return __ret;
60691}
60692#else
60693__ai uint8x16_t vsqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
60694  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60695  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60696  uint8x16_t __ret;
60697  __ret = (uint8x16_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
60698  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60699  return __ret;
60700}
60701#endif
60702
60703#ifdef __LITTLE_ENDIAN__
60704__ai uint32x4_t vsqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
60705  uint32x4_t __ret;
60706  __ret = (uint32x4_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
60707  return __ret;
60708}
60709#else
60710__ai uint32x4_t vsqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
60711  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60712  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60713  uint32x4_t __ret;
60714  __ret = (uint32x4_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
60715  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60716  return __ret;
60717}
60718#endif
60719
60720#ifdef __LITTLE_ENDIAN__
60721__ai uint64x2_t vsqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
60722  uint64x2_t __ret;
60723  __ret = (uint64x2_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
60724  return __ret;
60725}
60726#else
60727__ai uint64x2_t vsqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
60728  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60729  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60730  uint64x2_t __ret;
60731  __ret = (uint64x2_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
60732  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60733  return __ret;
60734}
60735#endif
60736
60737#ifdef __LITTLE_ENDIAN__
60738__ai uint16x8_t vsqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
60739  uint16x8_t __ret;
60740  __ret = (uint16x8_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
60741  return __ret;
60742}
60743#else
60744__ai uint16x8_t vsqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
60745  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60746  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60747  uint16x8_t __ret;
60748  __ret = (uint16x8_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
60749  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60750  return __ret;
60751}
60752#endif
60753
60754#ifdef __LITTLE_ENDIAN__
60755__ai uint8x8_t vsqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
60756  uint8x8_t __ret;
60757  __ret = (uint8x8_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
60758  return __ret;
60759}
60760#else
60761__ai uint8x8_t vsqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
60762  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60763  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60764  uint8x8_t __ret;
60765  __ret = (uint8x8_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
60766  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60767  return __ret;
60768}
60769#endif
60770
60771#ifdef __LITTLE_ENDIAN__
60772__ai uint32x2_t vsqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
60773  uint32x2_t __ret;
60774  __ret = (uint32x2_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
60775  return __ret;
60776}
60777#else
60778__ai uint32x2_t vsqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
60779  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60780  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60781  uint32x2_t __ret;
60782  __ret = (uint32x2_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
60783  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60784  return __ret;
60785}
60786#endif
60787
60788#ifdef __LITTLE_ENDIAN__
60789__ai uint64x1_t vsqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
60790  uint64x1_t __ret;
60791  __ret = (uint64x1_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
60792  return __ret;
60793}
60794#else
60795__ai uint64x1_t vsqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
60796  uint64x1_t __ret;
60797  __ret = (uint64x1_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
60798  return __ret;
60799}
60800#endif
60801
60802#ifdef __LITTLE_ENDIAN__
60803__ai uint16x4_t vsqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
60804  uint16x4_t __ret;
60805  __ret = (uint16x4_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
60806  return __ret;
60807}
60808#else
60809__ai uint16x4_t vsqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
60810  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60811  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60812  uint16x4_t __ret;
60813  __ret = (uint16x4_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
60814  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60815  return __ret;
60816}
60817#endif
60818
60819#ifdef __LITTLE_ENDIAN__
60820__ai float64x2_t vsqrtq_f64(float64x2_t __p0) {
60821  float64x2_t __ret;
60822  __ret = (float64x2_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 42);
60823  return __ret;
60824}
60825#else
60826__ai float64x2_t vsqrtq_f64(float64x2_t __p0) {
60827  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60828  float64x2_t __ret;
60829  __ret = (float64x2_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 42);
60830  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60831  return __ret;
60832}
60833#endif
60834
60835#ifdef __LITTLE_ENDIAN__
60836__ai float32x4_t vsqrtq_f32(float32x4_t __p0) {
60837  float32x4_t __ret;
60838  __ret = (float32x4_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 41);
60839  return __ret;
60840}
60841#else
60842__ai float32x4_t vsqrtq_f32(float32x4_t __p0) {
60843  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60844  float32x4_t __ret;
60845  __ret = (float32x4_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 41);
60846  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60847  return __ret;
60848}
60849#endif
60850
60851#ifdef __LITTLE_ENDIAN__
60852__ai float64x1_t vsqrt_f64(float64x1_t __p0) {
60853  float64x1_t __ret;
60854  __ret = (float64x1_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 10);
60855  return __ret;
60856}
60857#else
60858__ai float64x1_t vsqrt_f64(float64x1_t __p0) {
60859  float64x1_t __ret;
60860  __ret = (float64x1_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 10);
60861  return __ret;
60862}
60863#endif
60864
60865#ifdef __LITTLE_ENDIAN__
60866__ai float32x2_t vsqrt_f32(float32x2_t __p0) {
60867  float32x2_t __ret;
60868  __ret = (float32x2_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 9);
60869  return __ret;
60870}
60871#else
60872__ai float32x2_t vsqrt_f32(float32x2_t __p0) {
60873  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60874  float32x2_t __ret;
60875  __ret = (float32x2_t) __builtin_neon_vsqrt_v((int8x8_t)__rev0, 9);
60876  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60877  return __ret;
60878}
60879#endif
60880
60881#ifdef __LITTLE_ENDIAN__
60882#define vsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
60883  uint64_t __s0 = __p0; \
60884  uint64_t __s1 = __p1; \
60885  uint64_t __ret; \
60886  __ret = (uint64_t) __builtin_neon_vsrad_n_u64(__s0, __s1, __p2); \
60887  __ret; \
60888})
60889#else
60890#define vsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
60891  uint64_t __s0 = __p0; \
60892  uint64_t __s1 = __p1; \
60893  uint64_t __ret; \
60894  __ret = (uint64_t) __builtin_neon_vsrad_n_u64(__s0, __s1, __p2); \
60895  __ret; \
60896})
60897#endif
60898
60899#ifdef __LITTLE_ENDIAN__
60900#define vsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
60901  int64_t __s0 = __p0; \
60902  int64_t __s1 = __p1; \
60903  int64_t __ret; \
60904  __ret = (int64_t) __builtin_neon_vsrad_n_s64(__s0, __s1, __p2); \
60905  __ret; \
60906})
60907#else
60908#define vsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
60909  int64_t __s0 = __p0; \
60910  int64_t __s1 = __p1; \
60911  int64_t __ret; \
60912  __ret = (int64_t) __builtin_neon_vsrad_n_s64(__s0, __s1, __p2); \
60913  __ret; \
60914})
60915#endif
60916
60917#ifdef __LITTLE_ENDIAN__
60918#define vsrid_n_u64(__p0, __p1, __p2) __extension__ ({ \
60919  uint64_t __s0 = __p0; \
60920  uint64_t __s1 = __p1; \
60921  uint64_t __ret; \
60922  __ret = (uint64_t) __builtin_neon_vsrid_n_u64(__s0, __s1, __p2); \
60923  __ret; \
60924})
60925#else
60926#define vsrid_n_u64(__p0, __p1, __p2) __extension__ ({ \
60927  uint64_t __s0 = __p0; \
60928  uint64_t __s1 = __p1; \
60929  uint64_t __ret; \
60930  __ret = (uint64_t) __builtin_neon_vsrid_n_u64(__s0, __s1, __p2); \
60931  __ret; \
60932})
60933#endif
60934
60935#ifdef __LITTLE_ENDIAN__
60936#define vsrid_n_s64(__p0, __p1, __p2) __extension__ ({ \
60937  int64_t __s0 = __p0; \
60938  int64_t __s1 = __p1; \
60939  int64_t __ret; \
60940  __ret = (int64_t) __builtin_neon_vsrid_n_s64(__s0, __s1, __p2); \
60941  __ret; \
60942})
60943#else
60944#define vsrid_n_s64(__p0, __p1, __p2) __extension__ ({ \
60945  int64_t __s0 = __p0; \
60946  int64_t __s1 = __p1; \
60947  int64_t __ret; \
60948  __ret = (int64_t) __builtin_neon_vsrid_n_s64(__s0, __s1, __p2); \
60949  __ret; \
60950})
60951#endif
60952
60953#ifdef __LITTLE_ENDIAN__
60954#define vsri_n_p64(__p0, __p1, __p2) __extension__ ({ \
60955  poly64x1_t __s0 = __p0; \
60956  poly64x1_t __s1 = __p1; \
60957  poly64x1_t __ret; \
60958  __ret = (poly64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
60959  __ret; \
60960})
60961#else
60962#define vsri_n_p64(__p0, __p1, __p2) __extension__ ({ \
60963  poly64x1_t __s0 = __p0; \
60964  poly64x1_t __s1 = __p1; \
60965  poly64x1_t __ret; \
60966  __ret = (poly64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
60967  __ret; \
60968})
60969#endif
60970
60971#ifdef __LITTLE_ENDIAN__
60972#define vsriq_n_p64(__p0, __p1, __p2) __extension__ ({ \
60973  poly64x2_t __s0 = __p0; \
60974  poly64x2_t __s1 = __p1; \
60975  poly64x2_t __ret; \
60976  __ret = (poly64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
60977  __ret; \
60978})
60979#else
60980#define vsriq_n_p64(__p0, __p1, __p2) __extension__ ({ \
60981  poly64x2_t __s0 = __p0; \
60982  poly64x2_t __s1 = __p1; \
60983  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
60984  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60985  poly64x2_t __ret; \
60986  __ret = (poly64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
60987  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
60988  __ret; \
60989})
60990#endif
60991
60992#ifdef __LITTLE_ENDIAN__
60993#define vst1_p64(__p0, __p1) __extension__ ({ \
60994  poly64x1_t __s1 = __p1; \
60995  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 6); \
60996})
60997#else
60998#define vst1_p64(__p0, __p1) __extension__ ({ \
60999  poly64x1_t __s1 = __p1; \
61000  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 6); \
61001})
61002#endif
61003
61004#ifdef __LITTLE_ENDIAN__
61005#define vst1q_p64(__p0, __p1) __extension__ ({ \
61006  poly64x2_t __s1 = __p1; \
61007  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 38); \
61008})
61009#else
61010#define vst1q_p64(__p0, __p1) __extension__ ({ \
61011  poly64x2_t __s1 = __p1; \
61012  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61013  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 38); \
61014})
61015#endif
61016
61017#ifdef __LITTLE_ENDIAN__
61018#define vst1q_f64(__p0, __p1) __extension__ ({ \
61019  float64x2_t __s1 = __p1; \
61020  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 42); \
61021})
61022#else
61023#define vst1q_f64(__p0, __p1) __extension__ ({ \
61024  float64x2_t __s1 = __p1; \
61025  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61026  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 42); \
61027})
61028#endif
61029
61030#ifdef __LITTLE_ENDIAN__
61031#define vst1_f64(__p0, __p1) __extension__ ({ \
61032  float64x1_t __s1 = __p1; \
61033  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 10); \
61034})
61035#else
61036#define vst1_f64(__p0, __p1) __extension__ ({ \
61037  float64x1_t __s1 = __p1; \
61038  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 10); \
61039})
61040#endif
61041
61042#ifdef __LITTLE_ENDIAN__
61043#define vst1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61044  poly64x1_t __s1 = __p1; \
61045  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
61046})
61047#else
61048#define vst1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61049  poly64x1_t __s1 = __p1; \
61050  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
61051})
61052#endif
61053
61054#ifdef __LITTLE_ENDIAN__
61055#define vst1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61056  poly64x2_t __s1 = __p1; \
61057  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 38); \
61058})
61059#else
61060#define vst1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61061  poly64x2_t __s1 = __p1; \
61062  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61063  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 38); \
61064})
61065#endif
61066
61067#ifdef __LITTLE_ENDIAN__
61068#define vst1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61069  float64x2_t __s1 = __p1; \
61070  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 42); \
61071})
61072#else
61073#define vst1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61074  float64x2_t __s1 = __p1; \
61075  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61076  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 42); \
61077})
61078#endif
61079
61080#ifdef __LITTLE_ENDIAN__
61081#define vst1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61082  float64x1_t __s1 = __p1; \
61083  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
61084})
61085#else
61086#define vst1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61087  float64x1_t __s1 = __p1; \
61088  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
61089})
61090#endif
61091
61092#ifdef __LITTLE_ENDIAN__
61093#define vst1_p8_x2(__p0, __p1) __extension__ ({ \
61094  poly8x8x2_t __s1 = __p1; \
61095  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 4); \
61096})
61097#else
61098#define vst1_p8_x2(__p0, __p1) __extension__ ({ \
61099  poly8x8x2_t __s1 = __p1; \
61100  poly8x8x2_t __rev1; \
61101  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61102  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61103  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 4); \
61104})
61105#endif
61106
61107#ifdef __LITTLE_ENDIAN__
61108#define vst1_p64_x2(__p0, __p1) __extension__ ({ \
61109  poly64x1x2_t __s1 = __p1; \
61110  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
61111})
61112#else
61113#define vst1_p64_x2(__p0, __p1) __extension__ ({ \
61114  poly64x1x2_t __s1 = __p1; \
61115  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
61116})
61117#endif
61118
61119#ifdef __LITTLE_ENDIAN__
61120#define vst1_p16_x2(__p0, __p1) __extension__ ({ \
61121  poly16x4x2_t __s1 = __p1; \
61122  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 5); \
61123})
61124#else
61125#define vst1_p16_x2(__p0, __p1) __extension__ ({ \
61126  poly16x4x2_t __s1 = __p1; \
61127  poly16x4x2_t __rev1; \
61128  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61129  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61130  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 5); \
61131})
61132#endif
61133
61134#ifdef __LITTLE_ENDIAN__
61135#define vst1q_p8_x2(__p0, __p1) __extension__ ({ \
61136  poly8x16x2_t __s1 = __p1; \
61137  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 36); \
61138})
61139#else
61140#define vst1q_p8_x2(__p0, __p1) __extension__ ({ \
61141  poly8x16x2_t __s1 = __p1; \
61142  poly8x16x2_t __rev1; \
61143  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61144  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61145  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 36); \
61146})
61147#endif
61148
61149#ifdef __LITTLE_ENDIAN__
61150#define vst1q_p64_x2(__p0, __p1) __extension__ ({ \
61151  poly64x2x2_t __s1 = __p1; \
61152  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 38); \
61153})
61154#else
61155#define vst1q_p64_x2(__p0, __p1) __extension__ ({ \
61156  poly64x2x2_t __s1 = __p1; \
61157  poly64x2x2_t __rev1; \
61158  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61159  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61160  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 38); \
61161})
61162#endif
61163
61164#ifdef __LITTLE_ENDIAN__
61165#define vst1q_p16_x2(__p0, __p1) __extension__ ({ \
61166  poly16x8x2_t __s1 = __p1; \
61167  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 37); \
61168})
61169#else
61170#define vst1q_p16_x2(__p0, __p1) __extension__ ({ \
61171  poly16x8x2_t __s1 = __p1; \
61172  poly16x8x2_t __rev1; \
61173  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61174  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61175  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 37); \
61176})
61177#endif
61178
61179#ifdef __LITTLE_ENDIAN__
61180#define vst1q_u8_x2(__p0, __p1) __extension__ ({ \
61181  uint8x16x2_t __s1 = __p1; \
61182  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 48); \
61183})
61184#else
61185#define vst1q_u8_x2(__p0, __p1) __extension__ ({ \
61186  uint8x16x2_t __s1 = __p1; \
61187  uint8x16x2_t __rev1; \
61188  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61189  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61190  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 48); \
61191})
61192#endif
61193
61194#ifdef __LITTLE_ENDIAN__
61195#define vst1q_u32_x2(__p0, __p1) __extension__ ({ \
61196  uint32x4x2_t __s1 = __p1; \
61197  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 50); \
61198})
61199#else
61200#define vst1q_u32_x2(__p0, __p1) __extension__ ({ \
61201  uint32x4x2_t __s1 = __p1; \
61202  uint32x4x2_t __rev1; \
61203  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61204  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61205  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 50); \
61206})
61207#endif
61208
61209#ifdef __LITTLE_ENDIAN__
61210#define vst1q_u64_x2(__p0, __p1) __extension__ ({ \
61211  uint64x2x2_t __s1 = __p1; \
61212  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 51); \
61213})
61214#else
61215#define vst1q_u64_x2(__p0, __p1) __extension__ ({ \
61216  uint64x2x2_t __s1 = __p1; \
61217  uint64x2x2_t __rev1; \
61218  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61219  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61220  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 51); \
61221})
61222#endif
61223
61224#ifdef __LITTLE_ENDIAN__
61225#define vst1q_u16_x2(__p0, __p1) __extension__ ({ \
61226  uint16x8x2_t __s1 = __p1; \
61227  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 49); \
61228})
61229#else
61230#define vst1q_u16_x2(__p0, __p1) __extension__ ({ \
61231  uint16x8x2_t __s1 = __p1; \
61232  uint16x8x2_t __rev1; \
61233  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61234  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61235  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 49); \
61236})
61237#endif
61238
61239#ifdef __LITTLE_ENDIAN__
61240#define vst1q_s8_x2(__p0, __p1) __extension__ ({ \
61241  int8x16x2_t __s1 = __p1; \
61242  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 32); \
61243})
61244#else
61245#define vst1q_s8_x2(__p0, __p1) __extension__ ({ \
61246  int8x16x2_t __s1 = __p1; \
61247  int8x16x2_t __rev1; \
61248  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61249  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61250  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 32); \
61251})
61252#endif
61253
61254#ifdef __LITTLE_ENDIAN__
61255#define vst1q_f64_x2(__p0, __p1) __extension__ ({ \
61256  float64x2x2_t __s1 = __p1; \
61257  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 42); \
61258})
61259#else
61260#define vst1q_f64_x2(__p0, __p1) __extension__ ({ \
61261  float64x2x2_t __s1 = __p1; \
61262  float64x2x2_t __rev1; \
61263  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61264  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61265  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 42); \
61266})
61267#endif
61268
61269#ifdef __LITTLE_ENDIAN__
61270#define vst1q_f32_x2(__p0, __p1) __extension__ ({ \
61271  float32x4x2_t __s1 = __p1; \
61272  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 41); \
61273})
61274#else
61275#define vst1q_f32_x2(__p0, __p1) __extension__ ({ \
61276  float32x4x2_t __s1 = __p1; \
61277  float32x4x2_t __rev1; \
61278  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61279  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61280  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 41); \
61281})
61282#endif
61283
61284#ifdef __LITTLE_ENDIAN__
61285#define vst1q_f16_x2(__p0, __p1) __extension__ ({ \
61286  float16x8x2_t __s1 = __p1; \
61287  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 40); \
61288})
61289#else
61290#define vst1q_f16_x2(__p0, __p1) __extension__ ({ \
61291  float16x8x2_t __s1 = __p1; \
61292  float16x8x2_t __rev1; \
61293  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61294  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61295  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 40); \
61296})
61297#endif
61298
61299#ifdef __LITTLE_ENDIAN__
61300#define vst1q_s32_x2(__p0, __p1) __extension__ ({ \
61301  int32x4x2_t __s1 = __p1; \
61302  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 34); \
61303})
61304#else
61305#define vst1q_s32_x2(__p0, __p1) __extension__ ({ \
61306  int32x4x2_t __s1 = __p1; \
61307  int32x4x2_t __rev1; \
61308  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61309  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61310  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 34); \
61311})
61312#endif
61313
61314#ifdef __LITTLE_ENDIAN__
61315#define vst1q_s64_x2(__p0, __p1) __extension__ ({ \
61316  int64x2x2_t __s1 = __p1; \
61317  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 35); \
61318})
61319#else
61320#define vst1q_s64_x2(__p0, __p1) __extension__ ({ \
61321  int64x2x2_t __s1 = __p1; \
61322  int64x2x2_t __rev1; \
61323  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61324  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61325  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 35); \
61326})
61327#endif
61328
61329#ifdef __LITTLE_ENDIAN__
61330#define vst1q_s16_x2(__p0, __p1) __extension__ ({ \
61331  int16x8x2_t __s1 = __p1; \
61332  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 33); \
61333})
61334#else
61335#define vst1q_s16_x2(__p0, __p1) __extension__ ({ \
61336  int16x8x2_t __s1 = __p1; \
61337  int16x8x2_t __rev1; \
61338  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61339  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61340  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 33); \
61341})
61342#endif
61343
61344#ifdef __LITTLE_ENDIAN__
61345#define vst1_u8_x2(__p0, __p1) __extension__ ({ \
61346  uint8x8x2_t __s1 = __p1; \
61347  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 16); \
61348})
61349#else
61350#define vst1_u8_x2(__p0, __p1) __extension__ ({ \
61351  uint8x8x2_t __s1 = __p1; \
61352  uint8x8x2_t __rev1; \
61353  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61354  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61355  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 16); \
61356})
61357#endif
61358
61359#ifdef __LITTLE_ENDIAN__
61360#define vst1_u32_x2(__p0, __p1) __extension__ ({ \
61361  uint32x2x2_t __s1 = __p1; \
61362  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 18); \
61363})
61364#else
61365#define vst1_u32_x2(__p0, __p1) __extension__ ({ \
61366  uint32x2x2_t __s1 = __p1; \
61367  uint32x2x2_t __rev1; \
61368  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61369  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61370  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 18); \
61371})
61372#endif
61373
61374#ifdef __LITTLE_ENDIAN__
61375#define vst1_u64_x2(__p0, __p1) __extension__ ({ \
61376  uint64x1x2_t __s1 = __p1; \
61377  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
61378})
61379#else
61380#define vst1_u64_x2(__p0, __p1) __extension__ ({ \
61381  uint64x1x2_t __s1 = __p1; \
61382  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
61383})
61384#endif
61385
61386#ifdef __LITTLE_ENDIAN__
61387#define vst1_u16_x2(__p0, __p1) __extension__ ({ \
61388  uint16x4x2_t __s1 = __p1; \
61389  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 17); \
61390})
61391#else
61392#define vst1_u16_x2(__p0, __p1) __extension__ ({ \
61393  uint16x4x2_t __s1 = __p1; \
61394  uint16x4x2_t __rev1; \
61395  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61396  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61397  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 17); \
61398})
61399#endif
61400
61401#ifdef __LITTLE_ENDIAN__
61402#define vst1_s8_x2(__p0, __p1) __extension__ ({ \
61403  int8x8x2_t __s1 = __p1; \
61404  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 0); \
61405})
61406#else
61407#define vst1_s8_x2(__p0, __p1) __extension__ ({ \
61408  int8x8x2_t __s1 = __p1; \
61409  int8x8x2_t __rev1; \
61410  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61411  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61412  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 0); \
61413})
61414#endif
61415
61416#ifdef __LITTLE_ENDIAN__
61417#define vst1_f64_x2(__p0, __p1) __extension__ ({ \
61418  float64x1x2_t __s1 = __p1; \
61419  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 10); \
61420})
61421#else
61422#define vst1_f64_x2(__p0, __p1) __extension__ ({ \
61423  float64x1x2_t __s1 = __p1; \
61424  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 10); \
61425})
61426#endif
61427
61428#ifdef __LITTLE_ENDIAN__
61429#define vst1_f32_x2(__p0, __p1) __extension__ ({ \
61430  float32x2x2_t __s1 = __p1; \
61431  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 9); \
61432})
61433#else
61434#define vst1_f32_x2(__p0, __p1) __extension__ ({ \
61435  float32x2x2_t __s1 = __p1; \
61436  float32x2x2_t __rev1; \
61437  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61438  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61439  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 9); \
61440})
61441#endif
61442
61443#ifdef __LITTLE_ENDIAN__
61444#define vst1_f16_x2(__p0, __p1) __extension__ ({ \
61445  float16x4x2_t __s1 = __p1; \
61446  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 8); \
61447})
61448#else
61449#define vst1_f16_x2(__p0, __p1) __extension__ ({ \
61450  float16x4x2_t __s1 = __p1; \
61451  float16x4x2_t __rev1; \
61452  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61453  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61454  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 8); \
61455})
61456#endif
61457
61458#ifdef __LITTLE_ENDIAN__
61459#define vst1_s32_x2(__p0, __p1) __extension__ ({ \
61460  int32x2x2_t __s1 = __p1; \
61461  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 2); \
61462})
61463#else
61464#define vst1_s32_x2(__p0, __p1) __extension__ ({ \
61465  int32x2x2_t __s1 = __p1; \
61466  int32x2x2_t __rev1; \
61467  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61468  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61469  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 2); \
61470})
61471#endif
61472
61473#ifdef __LITTLE_ENDIAN__
61474#define vst1_s64_x2(__p0, __p1) __extension__ ({ \
61475  int64x1x2_t __s1 = __p1; \
61476  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 3); \
61477})
61478#else
61479#define vst1_s64_x2(__p0, __p1) __extension__ ({ \
61480  int64x1x2_t __s1 = __p1; \
61481  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 3); \
61482})
61483#endif
61484
61485#ifdef __LITTLE_ENDIAN__
61486#define vst1_s16_x2(__p0, __p1) __extension__ ({ \
61487  int16x4x2_t __s1 = __p1; \
61488  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 1); \
61489})
61490#else
61491#define vst1_s16_x2(__p0, __p1) __extension__ ({ \
61492  int16x4x2_t __s1 = __p1; \
61493  int16x4x2_t __rev1; \
61494  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61495  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61496  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 1); \
61497})
61498#endif
61499
61500#ifdef __LITTLE_ENDIAN__
61501#define vst1_p8_x3(__p0, __p1) __extension__ ({ \
61502  poly8x8x3_t __s1 = __p1; \
61503  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 4); \
61504})
61505#else
61506#define vst1_p8_x3(__p0, __p1) __extension__ ({ \
61507  poly8x8x3_t __s1 = __p1; \
61508  poly8x8x3_t __rev1; \
61509  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61510  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61511  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
61512  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 4); \
61513})
61514#endif
61515
61516#ifdef __LITTLE_ENDIAN__
61517#define vst1_p64_x3(__p0, __p1) __extension__ ({ \
61518  poly64x1x3_t __s1 = __p1; \
61519  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
61520})
61521#else
61522#define vst1_p64_x3(__p0, __p1) __extension__ ({ \
61523  poly64x1x3_t __s1 = __p1; \
61524  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
61525})
61526#endif
61527
61528#ifdef __LITTLE_ENDIAN__
61529#define vst1_p16_x3(__p0, __p1) __extension__ ({ \
61530  poly16x4x3_t __s1 = __p1; \
61531  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 5); \
61532})
61533#else
61534#define vst1_p16_x3(__p0, __p1) __extension__ ({ \
61535  poly16x4x3_t __s1 = __p1; \
61536  poly16x4x3_t __rev1; \
61537  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61538  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61539  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
61540  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 5); \
61541})
61542#endif
61543
61544#ifdef __LITTLE_ENDIAN__
61545#define vst1q_p8_x3(__p0, __p1) __extension__ ({ \
61546  poly8x16x3_t __s1 = __p1; \
61547  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 36); \
61548})
61549#else
61550#define vst1q_p8_x3(__p0, __p1) __extension__ ({ \
61551  poly8x16x3_t __s1 = __p1; \
61552  poly8x16x3_t __rev1; \
61553  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61554  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61555  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61556  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 36); \
61557})
61558#endif
61559
61560#ifdef __LITTLE_ENDIAN__
61561#define vst1q_p64_x3(__p0, __p1) __extension__ ({ \
61562  poly64x2x3_t __s1 = __p1; \
61563  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 38); \
61564})
61565#else
61566#define vst1q_p64_x3(__p0, __p1) __extension__ ({ \
61567  poly64x2x3_t __s1 = __p1; \
61568  poly64x2x3_t __rev1; \
61569  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61570  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61571  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61572  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 38); \
61573})
61574#endif
61575
61576#ifdef __LITTLE_ENDIAN__
61577#define vst1q_p16_x3(__p0, __p1) __extension__ ({ \
61578  poly16x8x3_t __s1 = __p1; \
61579  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 37); \
61580})
61581#else
61582#define vst1q_p16_x3(__p0, __p1) __extension__ ({ \
61583  poly16x8x3_t __s1 = __p1; \
61584  poly16x8x3_t __rev1; \
61585  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61586  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61587  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
61588  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 37); \
61589})
61590#endif
61591
61592#ifdef __LITTLE_ENDIAN__
61593#define vst1q_u8_x3(__p0, __p1) __extension__ ({ \
61594  uint8x16x3_t __s1 = __p1; \
61595  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 48); \
61596})
61597#else
61598#define vst1q_u8_x3(__p0, __p1) __extension__ ({ \
61599  uint8x16x3_t __s1 = __p1; \
61600  uint8x16x3_t __rev1; \
61601  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61602  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61603  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61604  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 48); \
61605})
61606#endif
61607
61608#ifdef __LITTLE_ENDIAN__
61609#define vst1q_u32_x3(__p0, __p1) __extension__ ({ \
61610  uint32x4x3_t __s1 = __p1; \
61611  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 50); \
61612})
61613#else
61614#define vst1q_u32_x3(__p0, __p1) __extension__ ({ \
61615  uint32x4x3_t __s1 = __p1; \
61616  uint32x4x3_t __rev1; \
61617  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61618  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61619  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
61620  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 50); \
61621})
61622#endif
61623
61624#ifdef __LITTLE_ENDIAN__
61625#define vst1q_u64_x3(__p0, __p1) __extension__ ({ \
61626  uint64x2x3_t __s1 = __p1; \
61627  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 51); \
61628})
61629#else
61630#define vst1q_u64_x3(__p0, __p1) __extension__ ({ \
61631  uint64x2x3_t __s1 = __p1; \
61632  uint64x2x3_t __rev1; \
61633  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61634  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61635  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61636  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 51); \
61637})
61638#endif
61639
61640#ifdef __LITTLE_ENDIAN__
61641#define vst1q_u16_x3(__p0, __p1) __extension__ ({ \
61642  uint16x8x3_t __s1 = __p1; \
61643  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 49); \
61644})
61645#else
61646#define vst1q_u16_x3(__p0, __p1) __extension__ ({ \
61647  uint16x8x3_t __s1 = __p1; \
61648  uint16x8x3_t __rev1; \
61649  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61650  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61651  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
61652  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 49); \
61653})
61654#endif
61655
61656#ifdef __LITTLE_ENDIAN__
61657#define vst1q_s8_x3(__p0, __p1) __extension__ ({ \
61658  int8x16x3_t __s1 = __p1; \
61659  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 32); \
61660})
61661#else
61662#define vst1q_s8_x3(__p0, __p1) __extension__ ({ \
61663  int8x16x3_t __s1 = __p1; \
61664  int8x16x3_t __rev1; \
61665  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61666  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61667  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61668  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 32); \
61669})
61670#endif
61671
61672#ifdef __LITTLE_ENDIAN__
61673#define vst1q_f64_x3(__p0, __p1) __extension__ ({ \
61674  float64x2x3_t __s1 = __p1; \
61675  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 42); \
61676})
61677#else
61678#define vst1q_f64_x3(__p0, __p1) __extension__ ({ \
61679  float64x2x3_t __s1 = __p1; \
61680  float64x2x3_t __rev1; \
61681  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61682  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61683  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61684  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 42); \
61685})
61686#endif
61687
61688#ifdef __LITTLE_ENDIAN__
61689#define vst1q_f32_x3(__p0, __p1) __extension__ ({ \
61690  float32x4x3_t __s1 = __p1; \
61691  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 41); \
61692})
61693#else
61694#define vst1q_f32_x3(__p0, __p1) __extension__ ({ \
61695  float32x4x3_t __s1 = __p1; \
61696  float32x4x3_t __rev1; \
61697  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61698  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61699  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
61700  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 41); \
61701})
61702#endif
61703
61704#ifdef __LITTLE_ENDIAN__
61705#define vst1q_f16_x3(__p0, __p1) __extension__ ({ \
61706  float16x8x3_t __s1 = __p1; \
61707  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 40); \
61708})
61709#else
61710#define vst1q_f16_x3(__p0, __p1) __extension__ ({ \
61711  float16x8x3_t __s1 = __p1; \
61712  float16x8x3_t __rev1; \
61713  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61714  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61715  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
61716  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 40); \
61717})
61718#endif
61719
61720#ifdef __LITTLE_ENDIAN__
61721#define vst1q_s32_x3(__p0, __p1) __extension__ ({ \
61722  int32x4x3_t __s1 = __p1; \
61723  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 34); \
61724})
61725#else
61726#define vst1q_s32_x3(__p0, __p1) __extension__ ({ \
61727  int32x4x3_t __s1 = __p1; \
61728  int32x4x3_t __rev1; \
61729  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61730  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61731  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
61732  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 34); \
61733})
61734#endif
61735
61736#ifdef __LITTLE_ENDIAN__
61737#define vst1q_s64_x3(__p0, __p1) __extension__ ({ \
61738  int64x2x3_t __s1 = __p1; \
61739  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 35); \
61740})
61741#else
61742#define vst1q_s64_x3(__p0, __p1) __extension__ ({ \
61743  int64x2x3_t __s1 = __p1; \
61744  int64x2x3_t __rev1; \
61745  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61746  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61747  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61748  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 35); \
61749})
61750#endif
61751
61752#ifdef __LITTLE_ENDIAN__
61753#define vst1q_s16_x3(__p0, __p1) __extension__ ({ \
61754  int16x8x3_t __s1 = __p1; \
61755  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 33); \
61756})
61757#else
61758#define vst1q_s16_x3(__p0, __p1) __extension__ ({ \
61759  int16x8x3_t __s1 = __p1; \
61760  int16x8x3_t __rev1; \
61761  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61762  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61763  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
61764  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 33); \
61765})
61766#endif
61767
61768#ifdef __LITTLE_ENDIAN__
61769#define vst1_u8_x3(__p0, __p1) __extension__ ({ \
61770  uint8x8x3_t __s1 = __p1; \
61771  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 16); \
61772})
61773#else
61774#define vst1_u8_x3(__p0, __p1) __extension__ ({ \
61775  uint8x8x3_t __s1 = __p1; \
61776  uint8x8x3_t __rev1; \
61777  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61778  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61779  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
61780  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 16); \
61781})
61782#endif
61783
61784#ifdef __LITTLE_ENDIAN__
61785#define vst1_u32_x3(__p0, __p1) __extension__ ({ \
61786  uint32x2x3_t __s1 = __p1; \
61787  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 18); \
61788})
61789#else
61790#define vst1_u32_x3(__p0, __p1) __extension__ ({ \
61791  uint32x2x3_t __s1 = __p1; \
61792  uint32x2x3_t __rev1; \
61793  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61794  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61795  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61796  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 18); \
61797})
61798#endif
61799
61800#ifdef __LITTLE_ENDIAN__
61801#define vst1_u64_x3(__p0, __p1) __extension__ ({ \
61802  uint64x1x3_t __s1 = __p1; \
61803  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
61804})
61805#else
61806#define vst1_u64_x3(__p0, __p1) __extension__ ({ \
61807  uint64x1x3_t __s1 = __p1; \
61808  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
61809})
61810#endif
61811
61812#ifdef __LITTLE_ENDIAN__
61813#define vst1_u16_x3(__p0, __p1) __extension__ ({ \
61814  uint16x4x3_t __s1 = __p1; \
61815  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 17); \
61816})
61817#else
61818#define vst1_u16_x3(__p0, __p1) __extension__ ({ \
61819  uint16x4x3_t __s1 = __p1; \
61820  uint16x4x3_t __rev1; \
61821  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61822  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61823  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
61824  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 17); \
61825})
61826#endif
61827
61828#ifdef __LITTLE_ENDIAN__
61829#define vst1_s8_x3(__p0, __p1) __extension__ ({ \
61830  int8x8x3_t __s1 = __p1; \
61831  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 0); \
61832})
61833#else
61834#define vst1_s8_x3(__p0, __p1) __extension__ ({ \
61835  int8x8x3_t __s1 = __p1; \
61836  int8x8x3_t __rev1; \
61837  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61838  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61839  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
61840  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 0); \
61841})
61842#endif
61843
61844#ifdef __LITTLE_ENDIAN__
61845#define vst1_f64_x3(__p0, __p1) __extension__ ({ \
61846  float64x1x3_t __s1 = __p1; \
61847  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
61848})
61849#else
61850#define vst1_f64_x3(__p0, __p1) __extension__ ({ \
61851  float64x1x3_t __s1 = __p1; \
61852  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
61853})
61854#endif
61855
61856#ifdef __LITTLE_ENDIAN__
61857#define vst1_f32_x3(__p0, __p1) __extension__ ({ \
61858  float32x2x3_t __s1 = __p1; \
61859  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 9); \
61860})
61861#else
61862#define vst1_f32_x3(__p0, __p1) __extension__ ({ \
61863  float32x2x3_t __s1 = __p1; \
61864  float32x2x3_t __rev1; \
61865  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61866  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61867  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61868  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 9); \
61869})
61870#endif
61871
61872#ifdef __LITTLE_ENDIAN__
61873#define vst1_f16_x3(__p0, __p1) __extension__ ({ \
61874  float16x4x3_t __s1 = __p1; \
61875  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 8); \
61876})
61877#else
61878#define vst1_f16_x3(__p0, __p1) __extension__ ({ \
61879  float16x4x3_t __s1 = __p1; \
61880  float16x4x3_t __rev1; \
61881  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61882  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61883  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
61884  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 8); \
61885})
61886#endif
61887
61888#ifdef __LITTLE_ENDIAN__
61889#define vst1_s32_x3(__p0, __p1) __extension__ ({ \
61890  int32x2x3_t __s1 = __p1; \
61891  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 2); \
61892})
61893#else
61894#define vst1_s32_x3(__p0, __p1) __extension__ ({ \
61895  int32x2x3_t __s1 = __p1; \
61896  int32x2x3_t __rev1; \
61897  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61898  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61899  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
61900  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 2); \
61901})
61902#endif
61903
61904#ifdef __LITTLE_ENDIAN__
61905#define vst1_s64_x3(__p0, __p1) __extension__ ({ \
61906  int64x1x3_t __s1 = __p1; \
61907  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
61908})
61909#else
61910#define vst1_s64_x3(__p0, __p1) __extension__ ({ \
61911  int64x1x3_t __s1 = __p1; \
61912  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
61913})
61914#endif
61915
61916#ifdef __LITTLE_ENDIAN__
61917#define vst1_s16_x3(__p0, __p1) __extension__ ({ \
61918  int16x4x3_t __s1 = __p1; \
61919  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 1); \
61920})
61921#else
61922#define vst1_s16_x3(__p0, __p1) __extension__ ({ \
61923  int16x4x3_t __s1 = __p1; \
61924  int16x4x3_t __rev1; \
61925  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61926  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61927  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
61928  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 1); \
61929})
61930#endif
61931
61932#ifdef __LITTLE_ENDIAN__
61933#define vst1_p8_x4(__p0, __p1) __extension__ ({ \
61934  poly8x8x4_t __s1 = __p1; \
61935  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 4); \
61936})
61937#else
61938#define vst1_p8_x4(__p0, __p1) __extension__ ({ \
61939  poly8x8x4_t __s1 = __p1; \
61940  poly8x8x4_t __rev1; \
61941  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61942  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61943  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
61944  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
61945  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 4); \
61946})
61947#endif
61948
61949#ifdef __LITTLE_ENDIAN__
61950#define vst1_p64_x4(__p0, __p1) __extension__ ({ \
61951  poly64x1x4_t __s1 = __p1; \
61952  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
61953})
61954#else
61955#define vst1_p64_x4(__p0, __p1) __extension__ ({ \
61956  poly64x1x4_t __s1 = __p1; \
61957  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
61958})
61959#endif
61960
61961#ifdef __LITTLE_ENDIAN__
61962#define vst1_p16_x4(__p0, __p1) __extension__ ({ \
61963  poly16x4x4_t __s1 = __p1; \
61964  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 5); \
61965})
61966#else
61967#define vst1_p16_x4(__p0, __p1) __extension__ ({ \
61968  poly16x4x4_t __s1 = __p1; \
61969  poly16x4x4_t __rev1; \
61970  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61971  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61972  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
61973  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
61974  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 5); \
61975})
61976#endif
61977
61978#ifdef __LITTLE_ENDIAN__
61979#define vst1q_p8_x4(__p0, __p1) __extension__ ({ \
61980  poly8x16x4_t __s1 = __p1; \
61981  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 36); \
61982})
61983#else
61984#define vst1q_p8_x4(__p0, __p1) __extension__ ({ \
61985  poly8x16x4_t __s1 = __p1; \
61986  poly8x16x4_t __rev1; \
61987  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61988  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61989  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61990  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61991  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 36); \
61992})
61993#endif
61994
61995#ifdef __LITTLE_ENDIAN__
61996#define vst1q_p64_x4(__p0, __p1) __extension__ ({ \
61997  poly64x2x4_t __s1 = __p1; \
61998  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 38); \
61999})
62000#else
62001#define vst1q_p64_x4(__p0, __p1) __extension__ ({ \
62002  poly64x2x4_t __s1 = __p1; \
62003  poly64x2x4_t __rev1; \
62004  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62005  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62006  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62007  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62008  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 38); \
62009})
62010#endif
62011
62012#ifdef __LITTLE_ENDIAN__
62013#define vst1q_p16_x4(__p0, __p1) __extension__ ({ \
62014  poly16x8x4_t __s1 = __p1; \
62015  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 37); \
62016})
62017#else
62018#define vst1q_p16_x4(__p0, __p1) __extension__ ({ \
62019  poly16x8x4_t __s1 = __p1; \
62020  poly16x8x4_t __rev1; \
62021  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62022  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62023  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62024  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62025  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 37); \
62026})
62027#endif
62028
62029#ifdef __LITTLE_ENDIAN__
62030#define vst1q_u8_x4(__p0, __p1) __extension__ ({ \
62031  uint8x16x4_t __s1 = __p1; \
62032  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 48); \
62033})
62034#else
62035#define vst1q_u8_x4(__p0, __p1) __extension__ ({ \
62036  uint8x16x4_t __s1 = __p1; \
62037  uint8x16x4_t __rev1; \
62038  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62039  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62040  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62041  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62042  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 48); \
62043})
62044#endif
62045
62046#ifdef __LITTLE_ENDIAN__
62047#define vst1q_u32_x4(__p0, __p1) __extension__ ({ \
62048  uint32x4x4_t __s1 = __p1; \
62049  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 50); \
62050})
62051#else
62052#define vst1q_u32_x4(__p0, __p1) __extension__ ({ \
62053  uint32x4x4_t __s1 = __p1; \
62054  uint32x4x4_t __rev1; \
62055  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62056  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62057  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62058  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62059  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 50); \
62060})
62061#endif
62062
62063#ifdef __LITTLE_ENDIAN__
62064#define vst1q_u64_x4(__p0, __p1) __extension__ ({ \
62065  uint64x2x4_t __s1 = __p1; \
62066  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 51); \
62067})
62068#else
62069#define vst1q_u64_x4(__p0, __p1) __extension__ ({ \
62070  uint64x2x4_t __s1 = __p1; \
62071  uint64x2x4_t __rev1; \
62072  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62073  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62074  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62075  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62076  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 51); \
62077})
62078#endif
62079
62080#ifdef __LITTLE_ENDIAN__
62081#define vst1q_u16_x4(__p0, __p1) __extension__ ({ \
62082  uint16x8x4_t __s1 = __p1; \
62083  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 49); \
62084})
62085#else
62086#define vst1q_u16_x4(__p0, __p1) __extension__ ({ \
62087  uint16x8x4_t __s1 = __p1; \
62088  uint16x8x4_t __rev1; \
62089  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62090  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62091  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62092  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62093  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 49); \
62094})
62095#endif
62096
62097#ifdef __LITTLE_ENDIAN__
62098#define vst1q_s8_x4(__p0, __p1) __extension__ ({ \
62099  int8x16x4_t __s1 = __p1; \
62100  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 32); \
62101})
62102#else
62103#define vst1q_s8_x4(__p0, __p1) __extension__ ({ \
62104  int8x16x4_t __s1 = __p1; \
62105  int8x16x4_t __rev1; \
62106  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62107  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62108  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62109  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62110  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 32); \
62111})
62112#endif
62113
62114#ifdef __LITTLE_ENDIAN__
62115#define vst1q_f64_x4(__p0, __p1) __extension__ ({ \
62116  float64x2x4_t __s1 = __p1; \
62117  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 42); \
62118})
62119#else
62120#define vst1q_f64_x4(__p0, __p1) __extension__ ({ \
62121  float64x2x4_t __s1 = __p1; \
62122  float64x2x4_t __rev1; \
62123  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62124  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62125  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62126  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62127  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 42); \
62128})
62129#endif
62130
62131#ifdef __LITTLE_ENDIAN__
62132#define vst1q_f32_x4(__p0, __p1) __extension__ ({ \
62133  float32x4x4_t __s1 = __p1; \
62134  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 41); \
62135})
62136#else
62137#define vst1q_f32_x4(__p0, __p1) __extension__ ({ \
62138  float32x4x4_t __s1 = __p1; \
62139  float32x4x4_t __rev1; \
62140  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62141  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62142  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62143  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62144  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 41); \
62145})
62146#endif
62147
62148#ifdef __LITTLE_ENDIAN__
62149#define vst1q_f16_x4(__p0, __p1) __extension__ ({ \
62150  float16x8x4_t __s1 = __p1; \
62151  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 40); \
62152})
62153#else
62154#define vst1q_f16_x4(__p0, __p1) __extension__ ({ \
62155  float16x8x4_t __s1 = __p1; \
62156  float16x8x4_t __rev1; \
62157  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62158  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62159  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62160  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62161  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 40); \
62162})
62163#endif
62164
62165#ifdef __LITTLE_ENDIAN__
62166#define vst1q_s32_x4(__p0, __p1) __extension__ ({ \
62167  int32x4x4_t __s1 = __p1; \
62168  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 34); \
62169})
62170#else
62171#define vst1q_s32_x4(__p0, __p1) __extension__ ({ \
62172  int32x4x4_t __s1 = __p1; \
62173  int32x4x4_t __rev1; \
62174  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62175  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62176  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62177  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62178  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 34); \
62179})
62180#endif
62181
62182#ifdef __LITTLE_ENDIAN__
62183#define vst1q_s64_x4(__p0, __p1) __extension__ ({ \
62184  int64x2x4_t __s1 = __p1; \
62185  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 35); \
62186})
62187#else
62188#define vst1q_s64_x4(__p0, __p1) __extension__ ({ \
62189  int64x2x4_t __s1 = __p1; \
62190  int64x2x4_t __rev1; \
62191  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62192  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62193  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62194  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62195  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 35); \
62196})
62197#endif
62198
62199#ifdef __LITTLE_ENDIAN__
62200#define vst1q_s16_x4(__p0, __p1) __extension__ ({ \
62201  int16x8x4_t __s1 = __p1; \
62202  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 33); \
62203})
62204#else
62205#define vst1q_s16_x4(__p0, __p1) __extension__ ({ \
62206  int16x8x4_t __s1 = __p1; \
62207  int16x8x4_t __rev1; \
62208  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62209  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62210  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62211  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62212  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 33); \
62213})
62214#endif
62215
62216#ifdef __LITTLE_ENDIAN__
62217#define vst1_u8_x4(__p0, __p1) __extension__ ({ \
62218  uint8x8x4_t __s1 = __p1; \
62219  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 16); \
62220})
62221#else
62222#define vst1_u8_x4(__p0, __p1) __extension__ ({ \
62223  uint8x8x4_t __s1 = __p1; \
62224  uint8x8x4_t __rev1; \
62225  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62226  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62227  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62228  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62229  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 16); \
62230})
62231#endif
62232
62233#ifdef __LITTLE_ENDIAN__
62234#define vst1_u32_x4(__p0, __p1) __extension__ ({ \
62235  uint32x2x4_t __s1 = __p1; \
62236  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 18); \
62237})
62238#else
62239#define vst1_u32_x4(__p0, __p1) __extension__ ({ \
62240  uint32x2x4_t __s1 = __p1; \
62241  uint32x2x4_t __rev1; \
62242  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62243  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62244  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62245  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62246  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 18); \
62247})
62248#endif
62249
62250#ifdef __LITTLE_ENDIAN__
62251#define vst1_u64_x4(__p0, __p1) __extension__ ({ \
62252  uint64x1x4_t __s1 = __p1; \
62253  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
62254})
62255#else
62256#define vst1_u64_x4(__p0, __p1) __extension__ ({ \
62257  uint64x1x4_t __s1 = __p1; \
62258  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
62259})
62260#endif
62261
62262#ifdef __LITTLE_ENDIAN__
62263#define vst1_u16_x4(__p0, __p1) __extension__ ({ \
62264  uint16x4x4_t __s1 = __p1; \
62265  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 17); \
62266})
62267#else
62268#define vst1_u16_x4(__p0, __p1) __extension__ ({ \
62269  uint16x4x4_t __s1 = __p1; \
62270  uint16x4x4_t __rev1; \
62271  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62272  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62273  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62274  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62275  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 17); \
62276})
62277#endif
62278
62279#ifdef __LITTLE_ENDIAN__
62280#define vst1_s8_x4(__p0, __p1) __extension__ ({ \
62281  int8x8x4_t __s1 = __p1; \
62282  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 0); \
62283})
62284#else
62285#define vst1_s8_x4(__p0, __p1) __extension__ ({ \
62286  int8x8x4_t __s1 = __p1; \
62287  int8x8x4_t __rev1; \
62288  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62289  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62290  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62291  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62292  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 0); \
62293})
62294#endif
62295
62296#ifdef __LITTLE_ENDIAN__
62297#define vst1_f64_x4(__p0, __p1) __extension__ ({ \
62298  float64x1x4_t __s1 = __p1; \
62299  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
62300})
62301#else
62302#define vst1_f64_x4(__p0, __p1) __extension__ ({ \
62303  float64x1x4_t __s1 = __p1; \
62304  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
62305})
62306#endif
62307
62308#ifdef __LITTLE_ENDIAN__
62309#define vst1_f32_x4(__p0, __p1) __extension__ ({ \
62310  float32x2x4_t __s1 = __p1; \
62311  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 9); \
62312})
62313#else
62314#define vst1_f32_x4(__p0, __p1) __extension__ ({ \
62315  float32x2x4_t __s1 = __p1; \
62316  float32x2x4_t __rev1; \
62317  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62318  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62319  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62320  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62321  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 9); \
62322})
62323#endif
62324
62325#ifdef __LITTLE_ENDIAN__
62326#define vst1_f16_x4(__p0, __p1) __extension__ ({ \
62327  float16x4x4_t __s1 = __p1; \
62328  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 8); \
62329})
62330#else
62331#define vst1_f16_x4(__p0, __p1) __extension__ ({ \
62332  float16x4x4_t __s1 = __p1; \
62333  float16x4x4_t __rev1; \
62334  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62335  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62336  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62337  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62338  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 8); \
62339})
62340#endif
62341
62342#ifdef __LITTLE_ENDIAN__
62343#define vst1_s32_x4(__p0, __p1) __extension__ ({ \
62344  int32x2x4_t __s1 = __p1; \
62345  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 2); \
62346})
62347#else
62348#define vst1_s32_x4(__p0, __p1) __extension__ ({ \
62349  int32x2x4_t __s1 = __p1; \
62350  int32x2x4_t __rev1; \
62351  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62352  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62353  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62354  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62355  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 2); \
62356})
62357#endif
62358
62359#ifdef __LITTLE_ENDIAN__
62360#define vst1_s64_x4(__p0, __p1) __extension__ ({ \
62361  int64x1x4_t __s1 = __p1; \
62362  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
62363})
62364#else
62365#define vst1_s64_x4(__p0, __p1) __extension__ ({ \
62366  int64x1x4_t __s1 = __p1; \
62367  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
62368})
62369#endif
62370
62371#ifdef __LITTLE_ENDIAN__
62372#define vst1_s16_x4(__p0, __p1) __extension__ ({ \
62373  int16x4x4_t __s1 = __p1; \
62374  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 1); \
62375})
62376#else
62377#define vst1_s16_x4(__p0, __p1) __extension__ ({ \
62378  int16x4x4_t __s1 = __p1; \
62379  int16x4x4_t __rev1; \
62380  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62381  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62382  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62383  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62384  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 1); \
62385})
62386#endif
62387
62388#ifdef __LITTLE_ENDIAN__
62389#define vst2_p64(__p0, __p1) __extension__ ({ \
62390  poly64x1x2_t __s1 = __p1; \
62391  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
62392})
62393#else
62394#define vst2_p64(__p0, __p1) __extension__ ({ \
62395  poly64x1x2_t __s1 = __p1; \
62396  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
62397})
62398#endif
62399
62400#ifdef __LITTLE_ENDIAN__
62401#define vst2q_p64(__p0, __p1) __extension__ ({ \
62402  poly64x2x2_t __s1 = __p1; \
62403  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 38); \
62404})
62405#else
62406#define vst2q_p64(__p0, __p1) __extension__ ({ \
62407  poly64x2x2_t __s1 = __p1; \
62408  poly64x2x2_t __rev1; \
62409  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62410  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62411  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 38); \
62412})
62413#endif
62414
62415#ifdef __LITTLE_ENDIAN__
62416#define vst2q_u64(__p0, __p1) __extension__ ({ \
62417  uint64x2x2_t __s1 = __p1; \
62418  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 51); \
62419})
62420#else
62421#define vst2q_u64(__p0, __p1) __extension__ ({ \
62422  uint64x2x2_t __s1 = __p1; \
62423  uint64x2x2_t __rev1; \
62424  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62425  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62426  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 51); \
62427})
62428#endif
62429
62430#ifdef __LITTLE_ENDIAN__
62431#define vst2q_f64(__p0, __p1) __extension__ ({ \
62432  float64x2x2_t __s1 = __p1; \
62433  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 42); \
62434})
62435#else
62436#define vst2q_f64(__p0, __p1) __extension__ ({ \
62437  float64x2x2_t __s1 = __p1; \
62438  float64x2x2_t __rev1; \
62439  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62440  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62441  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 42); \
62442})
62443#endif
62444
62445#ifdef __LITTLE_ENDIAN__
62446#define vst2q_s64(__p0, __p1) __extension__ ({ \
62447  int64x2x2_t __s1 = __p1; \
62448  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 35); \
62449})
62450#else
62451#define vst2q_s64(__p0, __p1) __extension__ ({ \
62452  int64x2x2_t __s1 = __p1; \
62453  int64x2x2_t __rev1; \
62454  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62455  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62456  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 35); \
62457})
62458#endif
62459
62460#ifdef __LITTLE_ENDIAN__
62461#define vst2_f64(__p0, __p1) __extension__ ({ \
62462  float64x1x2_t __s1 = __p1; \
62463  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 10); \
62464})
62465#else
62466#define vst2_f64(__p0, __p1) __extension__ ({ \
62467  float64x1x2_t __s1 = __p1; \
62468  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 10); \
62469})
62470#endif
62471
62472#ifdef __LITTLE_ENDIAN__
62473#define vst2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
62474  poly64x1x2_t __s1 = __p1; \
62475  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
62476})
62477#else
62478#define vst2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
62479  poly64x1x2_t __s1 = __p1; \
62480  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
62481})
62482#endif
62483
62484#ifdef __LITTLE_ENDIAN__
62485#define vst2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
62486  poly8x16x2_t __s1 = __p1; \
62487  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 36); \
62488})
62489#else
62490#define vst2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
62491  poly8x16x2_t __s1 = __p1; \
62492  poly8x16x2_t __rev1; \
62493  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62494  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62495  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 36); \
62496})
62497#endif
62498
62499#ifdef __LITTLE_ENDIAN__
62500#define vst2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
62501  poly64x2x2_t __s1 = __p1; \
62502  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 38); \
62503})
62504#else
62505#define vst2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
62506  poly64x2x2_t __s1 = __p1; \
62507  poly64x2x2_t __rev1; \
62508  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62509  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62510  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 38); \
62511})
62512#endif
62513
62514#ifdef __LITTLE_ENDIAN__
62515#define vst2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
62516  uint8x16x2_t __s1 = __p1; \
62517  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 48); \
62518})
62519#else
62520#define vst2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
62521  uint8x16x2_t __s1 = __p1; \
62522  uint8x16x2_t __rev1; \
62523  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62524  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62525  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 48); \
62526})
62527#endif
62528
62529#ifdef __LITTLE_ENDIAN__
62530#define vst2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
62531  uint64x2x2_t __s1 = __p1; \
62532  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 51); \
62533})
62534#else
62535#define vst2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
62536  uint64x2x2_t __s1 = __p1; \
62537  uint64x2x2_t __rev1; \
62538  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62539  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62540  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 51); \
62541})
62542#endif
62543
62544#ifdef __LITTLE_ENDIAN__
62545#define vst2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
62546  int8x16x2_t __s1 = __p1; \
62547  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 32); \
62548})
62549#else
62550#define vst2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
62551  int8x16x2_t __s1 = __p1; \
62552  int8x16x2_t __rev1; \
62553  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62554  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62555  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 32); \
62556})
62557#endif
62558
62559#ifdef __LITTLE_ENDIAN__
62560#define vst2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
62561  float64x2x2_t __s1 = __p1; \
62562  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 42); \
62563})
62564#else
62565#define vst2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
62566  float64x2x2_t __s1 = __p1; \
62567  float64x2x2_t __rev1; \
62568  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62569  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62570  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 42); \
62571})
62572#endif
62573
62574#ifdef __LITTLE_ENDIAN__
62575#define vst2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
62576  int64x2x2_t __s1 = __p1; \
62577  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 35); \
62578})
62579#else
62580#define vst2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
62581  int64x2x2_t __s1 = __p1; \
62582  int64x2x2_t __rev1; \
62583  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62584  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62585  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 35); \
62586})
62587#endif
62588
62589#ifdef __LITTLE_ENDIAN__
62590#define vst2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
62591  uint64x1x2_t __s1 = __p1; \
62592  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
62593})
62594#else
62595#define vst2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
62596  uint64x1x2_t __s1 = __p1; \
62597  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
62598})
62599#endif
62600
62601#ifdef __LITTLE_ENDIAN__
62602#define vst2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
62603  float64x1x2_t __s1 = __p1; \
62604  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 10); \
62605})
62606#else
62607#define vst2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
62608  float64x1x2_t __s1 = __p1; \
62609  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 10); \
62610})
62611#endif
62612
62613#ifdef __LITTLE_ENDIAN__
62614#define vst2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
62615  int64x1x2_t __s1 = __p1; \
62616  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 3); \
62617})
62618#else
62619#define vst2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
62620  int64x1x2_t __s1 = __p1; \
62621  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 3); \
62622})
62623#endif
62624
62625#ifdef __LITTLE_ENDIAN__
62626#define vst3_p64(__p0, __p1) __extension__ ({ \
62627  poly64x1x3_t __s1 = __p1; \
62628  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
62629})
62630#else
62631#define vst3_p64(__p0, __p1) __extension__ ({ \
62632  poly64x1x3_t __s1 = __p1; \
62633  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
62634})
62635#endif
62636
62637#ifdef __LITTLE_ENDIAN__
62638#define vst3q_p64(__p0, __p1) __extension__ ({ \
62639  poly64x2x3_t __s1 = __p1; \
62640  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 38); \
62641})
62642#else
62643#define vst3q_p64(__p0, __p1) __extension__ ({ \
62644  poly64x2x3_t __s1 = __p1; \
62645  poly64x2x3_t __rev1; \
62646  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62647  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62648  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62649  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 38); \
62650})
62651#endif
62652
62653#ifdef __LITTLE_ENDIAN__
62654#define vst3q_u64(__p0, __p1) __extension__ ({ \
62655  uint64x2x3_t __s1 = __p1; \
62656  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 51); \
62657})
62658#else
62659#define vst3q_u64(__p0, __p1) __extension__ ({ \
62660  uint64x2x3_t __s1 = __p1; \
62661  uint64x2x3_t __rev1; \
62662  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62663  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62664  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62665  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 51); \
62666})
62667#endif
62668
62669#ifdef __LITTLE_ENDIAN__
62670#define vst3q_f64(__p0, __p1) __extension__ ({ \
62671  float64x2x3_t __s1 = __p1; \
62672  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 42); \
62673})
62674#else
62675#define vst3q_f64(__p0, __p1) __extension__ ({ \
62676  float64x2x3_t __s1 = __p1; \
62677  float64x2x3_t __rev1; \
62678  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62679  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62680  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62681  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 42); \
62682})
62683#endif
62684
62685#ifdef __LITTLE_ENDIAN__
62686#define vst3q_s64(__p0, __p1) __extension__ ({ \
62687  int64x2x3_t __s1 = __p1; \
62688  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 35); \
62689})
62690#else
62691#define vst3q_s64(__p0, __p1) __extension__ ({ \
62692  int64x2x3_t __s1 = __p1; \
62693  int64x2x3_t __rev1; \
62694  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62695  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62696  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62697  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 35); \
62698})
62699#endif
62700
62701#ifdef __LITTLE_ENDIAN__
62702#define vst3_f64(__p0, __p1) __extension__ ({ \
62703  float64x1x3_t __s1 = __p1; \
62704  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
62705})
62706#else
62707#define vst3_f64(__p0, __p1) __extension__ ({ \
62708  float64x1x3_t __s1 = __p1; \
62709  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
62710})
62711#endif
62712
62713#ifdef __LITTLE_ENDIAN__
62714#define vst3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
62715  poly64x1x3_t __s1 = __p1; \
62716  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
62717})
62718#else
62719#define vst3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
62720  poly64x1x3_t __s1 = __p1; \
62721  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
62722})
62723#endif
62724
62725#ifdef __LITTLE_ENDIAN__
62726#define vst3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
62727  poly8x16x3_t __s1 = __p1; \
62728  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 36); \
62729})
62730#else
62731#define vst3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
62732  poly8x16x3_t __s1 = __p1; \
62733  poly8x16x3_t __rev1; \
62734  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62735  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62736  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62737  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 36); \
62738})
62739#endif
62740
62741#ifdef __LITTLE_ENDIAN__
62742#define vst3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
62743  poly64x2x3_t __s1 = __p1; \
62744  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 38); \
62745})
62746#else
62747#define vst3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
62748  poly64x2x3_t __s1 = __p1; \
62749  poly64x2x3_t __rev1; \
62750  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62751  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62752  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62753  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 38); \
62754})
62755#endif
62756
62757#ifdef __LITTLE_ENDIAN__
62758#define vst3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
62759  uint8x16x3_t __s1 = __p1; \
62760  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 48); \
62761})
62762#else
62763#define vst3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
62764  uint8x16x3_t __s1 = __p1; \
62765  uint8x16x3_t __rev1; \
62766  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62767  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62768  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62769  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 48); \
62770})
62771#endif
62772
62773#ifdef __LITTLE_ENDIAN__
62774#define vst3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
62775  uint64x2x3_t __s1 = __p1; \
62776  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 51); \
62777})
62778#else
62779#define vst3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
62780  uint64x2x3_t __s1 = __p1; \
62781  uint64x2x3_t __rev1; \
62782  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62783  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62784  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62785  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 51); \
62786})
62787#endif
62788
62789#ifdef __LITTLE_ENDIAN__
62790#define vst3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
62791  int8x16x3_t __s1 = __p1; \
62792  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 32); \
62793})
62794#else
62795#define vst3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
62796  int8x16x3_t __s1 = __p1; \
62797  int8x16x3_t __rev1; \
62798  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62799  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62800  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62801  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 32); \
62802})
62803#endif
62804
62805#ifdef __LITTLE_ENDIAN__
62806#define vst3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
62807  float64x2x3_t __s1 = __p1; \
62808  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 42); \
62809})
62810#else
62811#define vst3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
62812  float64x2x3_t __s1 = __p1; \
62813  float64x2x3_t __rev1; \
62814  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62815  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62816  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62817  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 42); \
62818})
62819#endif
62820
62821#ifdef __LITTLE_ENDIAN__
62822#define vst3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
62823  int64x2x3_t __s1 = __p1; \
62824  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 35); \
62825})
62826#else
62827#define vst3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
62828  int64x2x3_t __s1 = __p1; \
62829  int64x2x3_t __rev1; \
62830  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62831  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62832  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62833  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 35); \
62834})
62835#endif
62836
62837#ifdef __LITTLE_ENDIAN__
62838#define vst3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
62839  uint64x1x3_t __s1 = __p1; \
62840  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
62841})
62842#else
62843#define vst3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
62844  uint64x1x3_t __s1 = __p1; \
62845  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
62846})
62847#endif
62848
62849#ifdef __LITTLE_ENDIAN__
62850#define vst3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
62851  float64x1x3_t __s1 = __p1; \
62852  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
62853})
62854#else
62855#define vst3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
62856  float64x1x3_t __s1 = __p1; \
62857  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
62858})
62859#endif
62860
62861#ifdef __LITTLE_ENDIAN__
62862#define vst3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
62863  int64x1x3_t __s1 = __p1; \
62864  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
62865})
62866#else
62867#define vst3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
62868  int64x1x3_t __s1 = __p1; \
62869  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
62870})
62871#endif
62872
62873#ifdef __LITTLE_ENDIAN__
62874#define vst4_p64(__p0, __p1) __extension__ ({ \
62875  poly64x1x4_t __s1 = __p1; \
62876  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
62877})
62878#else
62879#define vst4_p64(__p0, __p1) __extension__ ({ \
62880  poly64x1x4_t __s1 = __p1; \
62881  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
62882})
62883#endif
62884
62885#ifdef __LITTLE_ENDIAN__
62886#define vst4q_p64(__p0, __p1) __extension__ ({ \
62887  poly64x2x4_t __s1 = __p1; \
62888  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 38); \
62889})
62890#else
62891#define vst4q_p64(__p0, __p1) __extension__ ({ \
62892  poly64x2x4_t __s1 = __p1; \
62893  poly64x2x4_t __rev1; \
62894  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62895  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62896  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62897  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62898  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 38); \
62899})
62900#endif
62901
62902#ifdef __LITTLE_ENDIAN__
62903#define vst4q_u64(__p0, __p1) __extension__ ({ \
62904  uint64x2x4_t __s1 = __p1; \
62905  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 51); \
62906})
62907#else
62908#define vst4q_u64(__p0, __p1) __extension__ ({ \
62909  uint64x2x4_t __s1 = __p1; \
62910  uint64x2x4_t __rev1; \
62911  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62912  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62913  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62914  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62915  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 51); \
62916})
62917#endif
62918
62919#ifdef __LITTLE_ENDIAN__
62920#define vst4q_f64(__p0, __p1) __extension__ ({ \
62921  float64x2x4_t __s1 = __p1; \
62922  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 42); \
62923})
62924#else
62925#define vst4q_f64(__p0, __p1) __extension__ ({ \
62926  float64x2x4_t __s1 = __p1; \
62927  float64x2x4_t __rev1; \
62928  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62929  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62930  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62931  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62932  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 42); \
62933})
62934#endif
62935
62936#ifdef __LITTLE_ENDIAN__
62937#define vst4q_s64(__p0, __p1) __extension__ ({ \
62938  int64x2x4_t __s1 = __p1; \
62939  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 35); \
62940})
62941#else
62942#define vst4q_s64(__p0, __p1) __extension__ ({ \
62943  int64x2x4_t __s1 = __p1; \
62944  int64x2x4_t __rev1; \
62945  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62946  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62947  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62948  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62949  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 35); \
62950})
62951#endif
62952
62953#ifdef __LITTLE_ENDIAN__
62954#define vst4_f64(__p0, __p1) __extension__ ({ \
62955  float64x1x4_t __s1 = __p1; \
62956  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
62957})
62958#else
62959#define vst4_f64(__p0, __p1) __extension__ ({ \
62960  float64x1x4_t __s1 = __p1; \
62961  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
62962})
62963#endif
62964
62965#ifdef __LITTLE_ENDIAN__
62966#define vst4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
62967  poly64x1x4_t __s1 = __p1; \
62968  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
62969})
62970#else
62971#define vst4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
62972  poly64x1x4_t __s1 = __p1; \
62973  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
62974})
62975#endif
62976
62977#ifdef __LITTLE_ENDIAN__
62978#define vst4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
62979  poly8x16x4_t __s1 = __p1; \
62980  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 36); \
62981})
62982#else
62983#define vst4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
62984  poly8x16x4_t __s1 = __p1; \
62985  poly8x16x4_t __rev1; \
62986  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62987  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62988  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62989  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62990  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 36); \
62991})
62992#endif
62993
62994#ifdef __LITTLE_ENDIAN__
62995#define vst4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
62996  poly64x2x4_t __s1 = __p1; \
62997  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 38); \
62998})
62999#else
63000#define vst4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63001  poly64x2x4_t __s1 = __p1; \
63002  poly64x2x4_t __rev1; \
63003  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63004  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63005  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63006  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63007  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 38); \
63008})
63009#endif
63010
63011#ifdef __LITTLE_ENDIAN__
63012#define vst4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63013  uint8x16x4_t __s1 = __p1; \
63014  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 48); \
63015})
63016#else
63017#define vst4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63018  uint8x16x4_t __s1 = __p1; \
63019  uint8x16x4_t __rev1; \
63020  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63021  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63022  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63023  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63024  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 48); \
63025})
63026#endif
63027
63028#ifdef __LITTLE_ENDIAN__
63029#define vst4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63030  uint64x2x4_t __s1 = __p1; \
63031  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 51); \
63032})
63033#else
63034#define vst4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63035  uint64x2x4_t __s1 = __p1; \
63036  uint64x2x4_t __rev1; \
63037  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63038  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63039  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63040  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63041  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 51); \
63042})
63043#endif
63044
63045#ifdef __LITTLE_ENDIAN__
63046#define vst4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63047  int8x16x4_t __s1 = __p1; \
63048  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 32); \
63049})
63050#else
63051#define vst4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63052  int8x16x4_t __s1 = __p1; \
63053  int8x16x4_t __rev1; \
63054  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63055  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63056  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63057  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63058  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 32); \
63059})
63060#endif
63061
63062#ifdef __LITTLE_ENDIAN__
63063#define vst4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63064  float64x2x4_t __s1 = __p1; \
63065  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 42); \
63066})
63067#else
63068#define vst4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63069  float64x2x4_t __s1 = __p1; \
63070  float64x2x4_t __rev1; \
63071  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63072  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63073  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63074  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63075  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 42); \
63076})
63077#endif
63078
63079#ifdef __LITTLE_ENDIAN__
63080#define vst4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63081  int64x2x4_t __s1 = __p1; \
63082  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 35); \
63083})
63084#else
63085#define vst4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63086  int64x2x4_t __s1 = __p1; \
63087  int64x2x4_t __rev1; \
63088  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63089  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63090  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63091  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63092  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 35); \
63093})
63094#endif
63095
63096#ifdef __LITTLE_ENDIAN__
63097#define vst4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63098  uint64x1x4_t __s1 = __p1; \
63099  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
63100})
63101#else
63102#define vst4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63103  uint64x1x4_t __s1 = __p1; \
63104  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
63105})
63106#endif
63107
63108#ifdef __LITTLE_ENDIAN__
63109#define vst4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63110  float64x1x4_t __s1 = __p1; \
63111  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
63112})
63113#else
63114#define vst4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63115  float64x1x4_t __s1 = __p1; \
63116  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
63117})
63118#endif
63119
63120#ifdef __LITTLE_ENDIAN__
63121#define vst4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63122  int64x1x4_t __s1 = __p1; \
63123  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
63124})
63125#else
63126#define vst4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63127  int64x1x4_t __s1 = __p1; \
63128  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
63129})
63130#endif
63131
63132#ifdef __LITTLE_ENDIAN__
63133#define vstrq_p128(__p0, __p1) __extension__ ({ \
63134  poly128_t __s1 = __p1; \
63135  __builtin_neon_vstrq_p128(__p0, __s1); \
63136})
63137#else
63138#define vstrq_p128(__p0, __p1) __extension__ ({ \
63139  poly128_t __s1 = __p1; \
63140  __builtin_neon_vstrq_p128(__p0, __s1); \
63141})
63142#endif
63143
63144#ifdef __LITTLE_ENDIAN__
63145__ai uint64_t vsubd_u64(uint64_t __p0, uint64_t __p1) {
63146  uint64_t __ret;
63147  __ret = (uint64_t) __builtin_neon_vsubd_u64(__p0, __p1);
63148  return __ret;
63149}
63150#else
63151__ai uint64_t vsubd_u64(uint64_t __p0, uint64_t __p1) {
63152  uint64_t __ret;
63153  __ret = (uint64_t) __builtin_neon_vsubd_u64(__p0, __p1);
63154  return __ret;
63155}
63156#endif
63157
63158#ifdef __LITTLE_ENDIAN__
63159__ai int64_t vsubd_s64(int64_t __p0, int64_t __p1) {
63160  int64_t __ret;
63161  __ret = (int64_t) __builtin_neon_vsubd_s64(__p0, __p1);
63162  return __ret;
63163}
63164#else
63165__ai int64_t vsubd_s64(int64_t __p0, int64_t __p1) {
63166  int64_t __ret;
63167  __ret = (int64_t) __builtin_neon_vsubd_s64(__p0, __p1);
63168  return __ret;
63169}
63170#endif
63171
63172#ifdef __LITTLE_ENDIAN__
63173__ai float64x2_t vsubq_f64(float64x2_t __p0, float64x2_t __p1) {
63174  float64x2_t __ret;
63175  __ret = __p0 - __p1;
63176  return __ret;
63177}
63178#else
63179__ai float64x2_t vsubq_f64(float64x2_t __p0, float64x2_t __p1) {
63180  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63181  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63182  float64x2_t __ret;
63183  __ret = __rev0 - __rev1;
63184  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63185  return __ret;
63186}
63187#endif
63188
63189#ifdef __LITTLE_ENDIAN__
63190__ai float64x1_t vsub_f64(float64x1_t __p0, float64x1_t __p1) {
63191  float64x1_t __ret;
63192  __ret = __p0 - __p1;
63193  return __ret;
63194}
63195#else
63196__ai float64x1_t vsub_f64(float64x1_t __p0, float64x1_t __p1) {
63197  float64x1_t __ret;
63198  __ret = __p0 - __p1;
63199  return __ret;
63200}
63201#endif
63202
63203#ifdef __LITTLE_ENDIAN__
63204__ai uint16x8_t vsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
63205  uint16x8_t __ret;
63206  __ret = vcombine_u16(__p0, vsubhn_u32(__p1, __p2));
63207  return __ret;
63208}
63209#else
63210__ai uint16x8_t vsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
63211  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63212  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63213  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
63214  uint16x8_t __ret;
63215  __ret = __noswap_vcombine_u16(__rev0, __noswap_vsubhn_u32(__rev1, __rev2));
63216  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63217  return __ret;
63218}
63219#endif
63220
63221#ifdef __LITTLE_ENDIAN__
63222__ai uint32x4_t vsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
63223  uint32x4_t __ret;
63224  __ret = vcombine_u32(__p0, vsubhn_u64(__p1, __p2));
63225  return __ret;
63226}
63227#else
63228__ai uint32x4_t vsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
63229  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63230  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63231  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
63232  uint32x4_t __ret;
63233  __ret = __noswap_vcombine_u32(__rev0, __noswap_vsubhn_u64(__rev1, __rev2));
63234  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63235  return __ret;
63236}
63237#endif
63238
63239#ifdef __LITTLE_ENDIAN__
63240__ai uint8x16_t vsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
63241  uint8x16_t __ret;
63242  __ret = vcombine_u8(__p0, vsubhn_u16(__p1, __p2));
63243  return __ret;
63244}
63245#else
63246__ai uint8x16_t vsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
63247  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63248  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63249  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
63250  uint8x16_t __ret;
63251  __ret = __noswap_vcombine_u8(__rev0, __noswap_vsubhn_u16(__rev1, __rev2));
63252  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63253  return __ret;
63254}
63255#endif
63256
63257#ifdef __LITTLE_ENDIAN__
63258__ai int16x8_t vsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
63259  int16x8_t __ret;
63260  __ret = vcombine_s16(__p0, vsubhn_s32(__p1, __p2));
63261  return __ret;
63262}
63263#else
63264__ai int16x8_t vsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
63265  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63266  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63267  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
63268  int16x8_t __ret;
63269  __ret = __noswap_vcombine_s16(__rev0, __noswap_vsubhn_s32(__rev1, __rev2));
63270  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63271  return __ret;
63272}
63273#endif
63274
63275#ifdef __LITTLE_ENDIAN__
63276__ai int32x4_t vsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
63277  int32x4_t __ret;
63278  __ret = vcombine_s32(__p0, vsubhn_s64(__p1, __p2));
63279  return __ret;
63280}
63281#else
63282__ai int32x4_t vsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
63283  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63284  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63285  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
63286  int32x4_t __ret;
63287  __ret = __noswap_vcombine_s32(__rev0, __noswap_vsubhn_s64(__rev1, __rev2));
63288  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63289  return __ret;
63290}
63291#endif
63292
63293#ifdef __LITTLE_ENDIAN__
63294__ai int8x16_t vsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
63295  int8x16_t __ret;
63296  __ret = vcombine_s8(__p0, vsubhn_s16(__p1, __p2));
63297  return __ret;
63298}
63299#else
63300__ai int8x16_t vsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
63301  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63302  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63303  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
63304  int8x16_t __ret;
63305  __ret = __noswap_vcombine_s8(__rev0, __noswap_vsubhn_s16(__rev1, __rev2));
63306  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63307  return __ret;
63308}
63309#endif
63310
63311#ifdef __LITTLE_ENDIAN__
63312__ai uint16x8_t vsubl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
63313  uint16x8_t __ret;
63314  __ret = vmovl_high_u8(__p0) - vmovl_high_u8(__p1);
63315  return __ret;
63316}
63317#else
63318__ai uint16x8_t vsubl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
63319  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63320  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63321  uint16x8_t __ret;
63322  __ret = __noswap_vmovl_high_u8(__rev0) - __noswap_vmovl_high_u8(__rev1);
63323  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63324  return __ret;
63325}
63326#endif
63327
63328#ifdef __LITTLE_ENDIAN__
63329__ai uint64x2_t vsubl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
63330  uint64x2_t __ret;
63331  __ret = vmovl_high_u32(__p0) - vmovl_high_u32(__p1);
63332  return __ret;
63333}
63334#else
63335__ai uint64x2_t vsubl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
63336  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63337  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63338  uint64x2_t __ret;
63339  __ret = __noswap_vmovl_high_u32(__rev0) - __noswap_vmovl_high_u32(__rev1);
63340  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63341  return __ret;
63342}
63343#endif
63344
63345#ifdef __LITTLE_ENDIAN__
63346__ai uint32x4_t vsubl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
63347  uint32x4_t __ret;
63348  __ret = vmovl_high_u16(__p0) - vmovl_high_u16(__p1);
63349  return __ret;
63350}
63351#else
63352__ai uint32x4_t vsubl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
63353  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63354  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63355  uint32x4_t __ret;
63356  __ret = __noswap_vmovl_high_u16(__rev0) - __noswap_vmovl_high_u16(__rev1);
63357  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63358  return __ret;
63359}
63360#endif
63361
63362#ifdef __LITTLE_ENDIAN__
63363__ai int16x8_t vsubl_high_s8(int8x16_t __p0, int8x16_t __p1) {
63364  int16x8_t __ret;
63365  __ret = vmovl_high_s8(__p0) - vmovl_high_s8(__p1);
63366  return __ret;
63367}
63368#else
63369__ai int16x8_t vsubl_high_s8(int8x16_t __p0, int8x16_t __p1) {
63370  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63371  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63372  int16x8_t __ret;
63373  __ret = __noswap_vmovl_high_s8(__rev0) - __noswap_vmovl_high_s8(__rev1);
63374  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63375  return __ret;
63376}
63377#endif
63378
63379#ifdef __LITTLE_ENDIAN__
63380__ai int64x2_t vsubl_high_s32(int32x4_t __p0, int32x4_t __p1) {
63381  int64x2_t __ret;
63382  __ret = vmovl_high_s32(__p0) - vmovl_high_s32(__p1);
63383  return __ret;
63384}
63385#else
63386__ai int64x2_t vsubl_high_s32(int32x4_t __p0, int32x4_t __p1) {
63387  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63388  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63389  int64x2_t __ret;
63390  __ret = __noswap_vmovl_high_s32(__rev0) - __noswap_vmovl_high_s32(__rev1);
63391  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63392  return __ret;
63393}
63394#endif
63395
63396#ifdef __LITTLE_ENDIAN__
63397__ai int32x4_t vsubl_high_s16(int16x8_t __p0, int16x8_t __p1) {
63398  int32x4_t __ret;
63399  __ret = vmovl_high_s16(__p0) - vmovl_high_s16(__p1);
63400  return __ret;
63401}
63402#else
63403__ai int32x4_t vsubl_high_s16(int16x8_t __p0, int16x8_t __p1) {
63404  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63405  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63406  int32x4_t __ret;
63407  __ret = __noswap_vmovl_high_s16(__rev0) - __noswap_vmovl_high_s16(__rev1);
63408  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63409  return __ret;
63410}
63411#endif
63412
63413#ifdef __LITTLE_ENDIAN__
63414__ai uint16x8_t vsubw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
63415  uint16x8_t __ret;
63416  __ret = __p0 - vmovl_high_u8(__p1);
63417  return __ret;
63418}
63419#else
63420__ai uint16x8_t vsubw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
63421  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63422  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63423  uint16x8_t __ret;
63424  __ret = __rev0 - __noswap_vmovl_high_u8(__rev1);
63425  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63426  return __ret;
63427}
63428#endif
63429
63430#ifdef __LITTLE_ENDIAN__
63431__ai uint64x2_t vsubw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
63432  uint64x2_t __ret;
63433  __ret = __p0 - vmovl_high_u32(__p1);
63434  return __ret;
63435}
63436#else
63437__ai uint64x2_t vsubw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
63438  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63439  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63440  uint64x2_t __ret;
63441  __ret = __rev0 - __noswap_vmovl_high_u32(__rev1);
63442  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63443  return __ret;
63444}
63445#endif
63446
63447#ifdef __LITTLE_ENDIAN__
63448__ai uint32x4_t vsubw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
63449  uint32x4_t __ret;
63450  __ret = __p0 - vmovl_high_u16(__p1);
63451  return __ret;
63452}
63453#else
63454__ai uint32x4_t vsubw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
63455  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63456  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63457  uint32x4_t __ret;
63458  __ret = __rev0 - __noswap_vmovl_high_u16(__rev1);
63459  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63460  return __ret;
63461}
63462#endif
63463
63464#ifdef __LITTLE_ENDIAN__
63465__ai int16x8_t vsubw_high_s8(int16x8_t __p0, int8x16_t __p1) {
63466  int16x8_t __ret;
63467  __ret = __p0 - vmovl_high_s8(__p1);
63468  return __ret;
63469}
63470#else
63471__ai int16x8_t vsubw_high_s8(int16x8_t __p0, int8x16_t __p1) {
63472  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63473  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63474  int16x8_t __ret;
63475  __ret = __rev0 - __noswap_vmovl_high_s8(__rev1);
63476  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63477  return __ret;
63478}
63479#endif
63480
63481#ifdef __LITTLE_ENDIAN__
63482__ai int64x2_t vsubw_high_s32(int64x2_t __p0, int32x4_t __p1) {
63483  int64x2_t __ret;
63484  __ret = __p0 - vmovl_high_s32(__p1);
63485  return __ret;
63486}
63487#else
63488__ai int64x2_t vsubw_high_s32(int64x2_t __p0, int32x4_t __p1) {
63489  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63490  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63491  int64x2_t __ret;
63492  __ret = __rev0 - __noswap_vmovl_high_s32(__rev1);
63493  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63494  return __ret;
63495}
63496#endif
63497
63498#ifdef __LITTLE_ENDIAN__
63499__ai int32x4_t vsubw_high_s16(int32x4_t __p0, int16x8_t __p1) {
63500  int32x4_t __ret;
63501  __ret = __p0 - vmovl_high_s16(__p1);
63502  return __ret;
63503}
63504#else
63505__ai int32x4_t vsubw_high_s16(int32x4_t __p0, int16x8_t __p1) {
63506  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63507  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63508  int32x4_t __ret;
63509  __ret = __rev0 - __noswap_vmovl_high_s16(__rev1);
63510  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63511  return __ret;
63512}
63513#endif
63514
63515#ifdef __LITTLE_ENDIAN__
63516__ai poly8x8_t vtrn1_p8(poly8x8_t __p0, poly8x8_t __p1) {
63517  poly8x8_t __ret;
63518  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
63519  return __ret;
63520}
63521#else
63522__ai poly8x8_t vtrn1_p8(poly8x8_t __p0, poly8x8_t __p1) {
63523  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63524  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63525  poly8x8_t __ret;
63526  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
63527  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63528  return __ret;
63529}
63530#endif
63531
63532#ifdef __LITTLE_ENDIAN__
63533__ai poly16x4_t vtrn1_p16(poly16x4_t __p0, poly16x4_t __p1) {
63534  poly16x4_t __ret;
63535  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
63536  return __ret;
63537}
63538#else
63539__ai poly16x4_t vtrn1_p16(poly16x4_t __p0, poly16x4_t __p1) {
63540  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63541  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63542  poly16x4_t __ret;
63543  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
63544  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63545  return __ret;
63546}
63547#endif
63548
63549#ifdef __LITTLE_ENDIAN__
63550__ai poly8x16_t vtrn1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
63551  poly8x16_t __ret;
63552  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
63553  return __ret;
63554}
63555#else
63556__ai poly8x16_t vtrn1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
63557  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63558  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63559  poly8x16_t __ret;
63560  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
63561  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63562  return __ret;
63563}
63564#endif
63565
63566#ifdef __LITTLE_ENDIAN__
63567__ai poly64x2_t vtrn1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
63568  poly64x2_t __ret;
63569  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
63570  return __ret;
63571}
63572#else
63573__ai poly64x2_t vtrn1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
63574  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63575  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63576  poly64x2_t __ret;
63577  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
63578  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63579  return __ret;
63580}
63581#endif
63582
63583#ifdef __LITTLE_ENDIAN__
63584__ai poly16x8_t vtrn1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
63585  poly16x8_t __ret;
63586  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
63587  return __ret;
63588}
63589#else
63590__ai poly16x8_t vtrn1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
63591  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63592  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63593  poly16x8_t __ret;
63594  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
63595  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63596  return __ret;
63597}
63598#endif
63599
63600#ifdef __LITTLE_ENDIAN__
63601__ai uint8x16_t vtrn1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
63602  uint8x16_t __ret;
63603  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
63604  return __ret;
63605}
63606#else
63607__ai uint8x16_t vtrn1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
63608  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63609  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63610  uint8x16_t __ret;
63611  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
63612  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63613  return __ret;
63614}
63615#endif
63616
63617#ifdef __LITTLE_ENDIAN__
63618__ai uint32x4_t vtrn1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
63619  uint32x4_t __ret;
63620  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
63621  return __ret;
63622}
63623#else
63624__ai uint32x4_t vtrn1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
63625  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63626  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63627  uint32x4_t __ret;
63628  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
63629  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63630  return __ret;
63631}
63632#endif
63633
63634#ifdef __LITTLE_ENDIAN__
63635__ai uint64x2_t vtrn1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
63636  uint64x2_t __ret;
63637  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
63638  return __ret;
63639}
63640#else
63641__ai uint64x2_t vtrn1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
63642  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63643  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63644  uint64x2_t __ret;
63645  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
63646  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63647  return __ret;
63648}
63649#endif
63650
63651#ifdef __LITTLE_ENDIAN__
63652__ai uint16x8_t vtrn1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
63653  uint16x8_t __ret;
63654  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
63655  return __ret;
63656}
63657#else
63658__ai uint16x8_t vtrn1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
63659  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63660  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63661  uint16x8_t __ret;
63662  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
63663  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63664  return __ret;
63665}
63666#endif
63667
63668#ifdef __LITTLE_ENDIAN__
63669__ai int8x16_t vtrn1q_s8(int8x16_t __p0, int8x16_t __p1) {
63670  int8x16_t __ret;
63671  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
63672  return __ret;
63673}
63674#else
63675__ai int8x16_t vtrn1q_s8(int8x16_t __p0, int8x16_t __p1) {
63676  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63677  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63678  int8x16_t __ret;
63679  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
63680  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63681  return __ret;
63682}
63683#endif
63684
63685#ifdef __LITTLE_ENDIAN__
63686__ai float64x2_t vtrn1q_f64(float64x2_t __p0, float64x2_t __p1) {
63687  float64x2_t __ret;
63688  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
63689  return __ret;
63690}
63691#else
63692__ai float64x2_t vtrn1q_f64(float64x2_t __p0, float64x2_t __p1) {
63693  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63694  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63695  float64x2_t __ret;
63696  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
63697  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63698  return __ret;
63699}
63700#endif
63701
63702#ifdef __LITTLE_ENDIAN__
63703__ai float32x4_t vtrn1q_f32(float32x4_t __p0, float32x4_t __p1) {
63704  float32x4_t __ret;
63705  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
63706  return __ret;
63707}
63708#else
63709__ai float32x4_t vtrn1q_f32(float32x4_t __p0, float32x4_t __p1) {
63710  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63711  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63712  float32x4_t __ret;
63713  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
63714  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63715  return __ret;
63716}
63717#endif
63718
63719#ifdef __LITTLE_ENDIAN__
63720__ai int32x4_t vtrn1q_s32(int32x4_t __p0, int32x4_t __p1) {
63721  int32x4_t __ret;
63722  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
63723  return __ret;
63724}
63725#else
63726__ai int32x4_t vtrn1q_s32(int32x4_t __p0, int32x4_t __p1) {
63727  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63728  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63729  int32x4_t __ret;
63730  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
63731  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63732  return __ret;
63733}
63734#endif
63735
63736#ifdef __LITTLE_ENDIAN__
63737__ai int64x2_t vtrn1q_s64(int64x2_t __p0, int64x2_t __p1) {
63738  int64x2_t __ret;
63739  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
63740  return __ret;
63741}
63742#else
63743__ai int64x2_t vtrn1q_s64(int64x2_t __p0, int64x2_t __p1) {
63744  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63745  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63746  int64x2_t __ret;
63747  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
63748  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63749  return __ret;
63750}
63751#endif
63752
63753#ifdef __LITTLE_ENDIAN__
63754__ai int16x8_t vtrn1q_s16(int16x8_t __p0, int16x8_t __p1) {
63755  int16x8_t __ret;
63756  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
63757  return __ret;
63758}
63759#else
63760__ai int16x8_t vtrn1q_s16(int16x8_t __p0, int16x8_t __p1) {
63761  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63762  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63763  int16x8_t __ret;
63764  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
63765  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63766  return __ret;
63767}
63768#endif
63769
63770#ifdef __LITTLE_ENDIAN__
63771__ai uint8x8_t vtrn1_u8(uint8x8_t __p0, uint8x8_t __p1) {
63772  uint8x8_t __ret;
63773  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
63774  return __ret;
63775}
63776#else
63777__ai uint8x8_t vtrn1_u8(uint8x8_t __p0, uint8x8_t __p1) {
63778  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63779  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63780  uint8x8_t __ret;
63781  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
63782  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63783  return __ret;
63784}
63785#endif
63786
63787#ifdef __LITTLE_ENDIAN__
63788__ai uint32x2_t vtrn1_u32(uint32x2_t __p0, uint32x2_t __p1) {
63789  uint32x2_t __ret;
63790  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
63791  return __ret;
63792}
63793#else
63794__ai uint32x2_t vtrn1_u32(uint32x2_t __p0, uint32x2_t __p1) {
63795  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63796  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63797  uint32x2_t __ret;
63798  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
63799  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63800  return __ret;
63801}
63802#endif
63803
63804#ifdef __LITTLE_ENDIAN__
63805__ai uint16x4_t vtrn1_u16(uint16x4_t __p0, uint16x4_t __p1) {
63806  uint16x4_t __ret;
63807  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
63808  return __ret;
63809}
63810#else
63811__ai uint16x4_t vtrn1_u16(uint16x4_t __p0, uint16x4_t __p1) {
63812  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63813  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63814  uint16x4_t __ret;
63815  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
63816  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63817  return __ret;
63818}
63819#endif
63820
63821#ifdef __LITTLE_ENDIAN__
63822__ai int8x8_t vtrn1_s8(int8x8_t __p0, int8x8_t __p1) {
63823  int8x8_t __ret;
63824  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
63825  return __ret;
63826}
63827#else
63828__ai int8x8_t vtrn1_s8(int8x8_t __p0, int8x8_t __p1) {
63829  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63830  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63831  int8x8_t __ret;
63832  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
63833  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63834  return __ret;
63835}
63836#endif
63837
63838#ifdef __LITTLE_ENDIAN__
63839__ai float32x2_t vtrn1_f32(float32x2_t __p0, float32x2_t __p1) {
63840  float32x2_t __ret;
63841  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
63842  return __ret;
63843}
63844#else
63845__ai float32x2_t vtrn1_f32(float32x2_t __p0, float32x2_t __p1) {
63846  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63847  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63848  float32x2_t __ret;
63849  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
63850  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63851  return __ret;
63852}
63853#endif
63854
63855#ifdef __LITTLE_ENDIAN__
63856__ai int32x2_t vtrn1_s32(int32x2_t __p0, int32x2_t __p1) {
63857  int32x2_t __ret;
63858  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
63859  return __ret;
63860}
63861#else
63862__ai int32x2_t vtrn1_s32(int32x2_t __p0, int32x2_t __p1) {
63863  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63864  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63865  int32x2_t __ret;
63866  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
63867  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63868  return __ret;
63869}
63870#endif
63871
63872#ifdef __LITTLE_ENDIAN__
63873__ai int16x4_t vtrn1_s16(int16x4_t __p0, int16x4_t __p1) {
63874  int16x4_t __ret;
63875  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
63876  return __ret;
63877}
63878#else
63879__ai int16x4_t vtrn1_s16(int16x4_t __p0, int16x4_t __p1) {
63880  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63881  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63882  int16x4_t __ret;
63883  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
63884  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63885  return __ret;
63886}
63887#endif
63888
63889#ifdef __LITTLE_ENDIAN__
63890__ai poly8x8_t vtrn2_p8(poly8x8_t __p0, poly8x8_t __p1) {
63891  poly8x8_t __ret;
63892  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
63893  return __ret;
63894}
63895#else
63896__ai poly8x8_t vtrn2_p8(poly8x8_t __p0, poly8x8_t __p1) {
63897  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63898  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63899  poly8x8_t __ret;
63900  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
63901  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63902  return __ret;
63903}
63904#endif
63905
63906#ifdef __LITTLE_ENDIAN__
63907__ai poly16x4_t vtrn2_p16(poly16x4_t __p0, poly16x4_t __p1) {
63908  poly16x4_t __ret;
63909  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
63910  return __ret;
63911}
63912#else
63913__ai poly16x4_t vtrn2_p16(poly16x4_t __p0, poly16x4_t __p1) {
63914  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63915  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63916  poly16x4_t __ret;
63917  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
63918  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63919  return __ret;
63920}
63921#endif
63922
63923#ifdef __LITTLE_ENDIAN__
63924__ai poly8x16_t vtrn2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
63925  poly8x16_t __ret;
63926  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
63927  return __ret;
63928}
63929#else
63930__ai poly8x16_t vtrn2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
63931  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63932  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63933  poly8x16_t __ret;
63934  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
63935  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63936  return __ret;
63937}
63938#endif
63939
63940#ifdef __LITTLE_ENDIAN__
63941__ai poly64x2_t vtrn2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
63942  poly64x2_t __ret;
63943  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
63944  return __ret;
63945}
63946#else
63947__ai poly64x2_t vtrn2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
63948  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63949  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63950  poly64x2_t __ret;
63951  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
63952  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63953  return __ret;
63954}
63955#endif
63956
63957#ifdef __LITTLE_ENDIAN__
63958__ai poly16x8_t vtrn2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
63959  poly16x8_t __ret;
63960  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
63961  return __ret;
63962}
63963#else
63964__ai poly16x8_t vtrn2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
63965  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63966  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63967  poly16x8_t __ret;
63968  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
63969  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63970  return __ret;
63971}
63972#endif
63973
63974#ifdef __LITTLE_ENDIAN__
63975__ai uint8x16_t vtrn2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
63976  uint8x16_t __ret;
63977  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
63978  return __ret;
63979}
63980#else
63981__ai uint8x16_t vtrn2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
63982  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63983  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63984  uint8x16_t __ret;
63985  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
63986  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63987  return __ret;
63988}
63989#endif
63990
63991#ifdef __LITTLE_ENDIAN__
63992__ai uint32x4_t vtrn2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
63993  uint32x4_t __ret;
63994  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
63995  return __ret;
63996}
63997#else
63998__ai uint32x4_t vtrn2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
63999  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64000  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64001  uint32x4_t __ret;
64002  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64003  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64004  return __ret;
64005}
64006#endif
64007
64008#ifdef __LITTLE_ENDIAN__
64009__ai uint64x2_t vtrn2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64010  uint64x2_t __ret;
64011  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64012  return __ret;
64013}
64014#else
64015__ai uint64x2_t vtrn2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64016  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64017  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64018  uint64x2_t __ret;
64019  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64020  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64021  return __ret;
64022}
64023#endif
64024
64025#ifdef __LITTLE_ENDIAN__
64026__ai uint16x8_t vtrn2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64027  uint16x8_t __ret;
64028  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64029  return __ret;
64030}
64031#else
64032__ai uint16x8_t vtrn2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64033  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64034  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64035  uint16x8_t __ret;
64036  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64037  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64038  return __ret;
64039}
64040#endif
64041
64042#ifdef __LITTLE_ENDIAN__
64043__ai int8x16_t vtrn2q_s8(int8x16_t __p0, int8x16_t __p1) {
64044  int8x16_t __ret;
64045  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64046  return __ret;
64047}
64048#else
64049__ai int8x16_t vtrn2q_s8(int8x16_t __p0, int8x16_t __p1) {
64050  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64051  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64052  int8x16_t __ret;
64053  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64054  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64055  return __ret;
64056}
64057#endif
64058
64059#ifdef __LITTLE_ENDIAN__
64060__ai float64x2_t vtrn2q_f64(float64x2_t __p0, float64x2_t __p1) {
64061  float64x2_t __ret;
64062  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64063  return __ret;
64064}
64065#else
64066__ai float64x2_t vtrn2q_f64(float64x2_t __p0, float64x2_t __p1) {
64067  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64068  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64069  float64x2_t __ret;
64070  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64071  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64072  return __ret;
64073}
64074#endif
64075
64076#ifdef __LITTLE_ENDIAN__
64077__ai float32x4_t vtrn2q_f32(float32x4_t __p0, float32x4_t __p1) {
64078  float32x4_t __ret;
64079  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64080  return __ret;
64081}
64082#else
64083__ai float32x4_t vtrn2q_f32(float32x4_t __p0, float32x4_t __p1) {
64084  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64085  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64086  float32x4_t __ret;
64087  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64088  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64089  return __ret;
64090}
64091#endif
64092
64093#ifdef __LITTLE_ENDIAN__
64094__ai int32x4_t vtrn2q_s32(int32x4_t __p0, int32x4_t __p1) {
64095  int32x4_t __ret;
64096  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64097  return __ret;
64098}
64099#else
64100__ai int32x4_t vtrn2q_s32(int32x4_t __p0, int32x4_t __p1) {
64101  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64102  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64103  int32x4_t __ret;
64104  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64105  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64106  return __ret;
64107}
64108#endif
64109
64110#ifdef __LITTLE_ENDIAN__
64111__ai int64x2_t vtrn2q_s64(int64x2_t __p0, int64x2_t __p1) {
64112  int64x2_t __ret;
64113  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64114  return __ret;
64115}
64116#else
64117__ai int64x2_t vtrn2q_s64(int64x2_t __p0, int64x2_t __p1) {
64118  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64119  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64120  int64x2_t __ret;
64121  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64122  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64123  return __ret;
64124}
64125#endif
64126
64127#ifdef __LITTLE_ENDIAN__
64128__ai int16x8_t vtrn2q_s16(int16x8_t __p0, int16x8_t __p1) {
64129  int16x8_t __ret;
64130  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64131  return __ret;
64132}
64133#else
64134__ai int16x8_t vtrn2q_s16(int16x8_t __p0, int16x8_t __p1) {
64135  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64136  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64137  int16x8_t __ret;
64138  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64139  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64140  return __ret;
64141}
64142#endif
64143
64144#ifdef __LITTLE_ENDIAN__
64145__ai uint8x8_t vtrn2_u8(uint8x8_t __p0, uint8x8_t __p1) {
64146  uint8x8_t __ret;
64147  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64148  return __ret;
64149}
64150#else
64151__ai uint8x8_t vtrn2_u8(uint8x8_t __p0, uint8x8_t __p1) {
64152  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64153  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64154  uint8x8_t __ret;
64155  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64156  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64157  return __ret;
64158}
64159#endif
64160
64161#ifdef __LITTLE_ENDIAN__
64162__ai uint32x2_t vtrn2_u32(uint32x2_t __p0, uint32x2_t __p1) {
64163  uint32x2_t __ret;
64164  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64165  return __ret;
64166}
64167#else
64168__ai uint32x2_t vtrn2_u32(uint32x2_t __p0, uint32x2_t __p1) {
64169  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64170  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64171  uint32x2_t __ret;
64172  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64173  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64174  return __ret;
64175}
64176#endif
64177
64178#ifdef __LITTLE_ENDIAN__
64179__ai uint16x4_t vtrn2_u16(uint16x4_t __p0, uint16x4_t __p1) {
64180  uint16x4_t __ret;
64181  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64182  return __ret;
64183}
64184#else
64185__ai uint16x4_t vtrn2_u16(uint16x4_t __p0, uint16x4_t __p1) {
64186  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64187  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64188  uint16x4_t __ret;
64189  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64190  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64191  return __ret;
64192}
64193#endif
64194
64195#ifdef __LITTLE_ENDIAN__
64196__ai int8x8_t vtrn2_s8(int8x8_t __p0, int8x8_t __p1) {
64197  int8x8_t __ret;
64198  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64199  return __ret;
64200}
64201#else
64202__ai int8x8_t vtrn2_s8(int8x8_t __p0, int8x8_t __p1) {
64203  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64204  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64205  int8x8_t __ret;
64206  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64207  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64208  return __ret;
64209}
64210#endif
64211
64212#ifdef __LITTLE_ENDIAN__
64213__ai float32x2_t vtrn2_f32(float32x2_t __p0, float32x2_t __p1) {
64214  float32x2_t __ret;
64215  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64216  return __ret;
64217}
64218#else
64219__ai float32x2_t vtrn2_f32(float32x2_t __p0, float32x2_t __p1) {
64220  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64221  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64222  float32x2_t __ret;
64223  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64224  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64225  return __ret;
64226}
64227#endif
64228
64229#ifdef __LITTLE_ENDIAN__
64230__ai int32x2_t vtrn2_s32(int32x2_t __p0, int32x2_t __p1) {
64231  int32x2_t __ret;
64232  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64233  return __ret;
64234}
64235#else
64236__ai int32x2_t vtrn2_s32(int32x2_t __p0, int32x2_t __p1) {
64237  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64238  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64239  int32x2_t __ret;
64240  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64241  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64242  return __ret;
64243}
64244#endif
64245
64246#ifdef __LITTLE_ENDIAN__
64247__ai int16x4_t vtrn2_s16(int16x4_t __p0, int16x4_t __p1) {
64248  int16x4_t __ret;
64249  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64250  return __ret;
64251}
64252#else
64253__ai int16x4_t vtrn2_s16(int16x4_t __p0, int16x4_t __p1) {
64254  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64255  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64256  int16x4_t __ret;
64257  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64258  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64259  return __ret;
64260}
64261#endif
64262
64263#ifdef __LITTLE_ENDIAN__
64264__ai uint64x1_t vtst_p64(poly64x1_t __p0, poly64x1_t __p1) {
64265  uint64x1_t __ret;
64266  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64267  return __ret;
64268}
64269#else
64270__ai uint64x1_t vtst_p64(poly64x1_t __p0, poly64x1_t __p1) {
64271  uint64x1_t __ret;
64272  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64273  return __ret;
64274}
64275#endif
64276
64277#ifdef __LITTLE_ENDIAN__
64278__ai uint64x2_t vtstq_p64(poly64x2_t __p0, poly64x2_t __p1) {
64279  uint64x2_t __ret;
64280  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
64281  return __ret;
64282}
64283#else
64284__ai uint64x2_t vtstq_p64(poly64x2_t __p0, poly64x2_t __p1) {
64285  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64286  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64287  uint64x2_t __ret;
64288  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
64289  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64290  return __ret;
64291}
64292#endif
64293
64294#ifdef __LITTLE_ENDIAN__
64295__ai uint64x2_t vtstq_u64(uint64x2_t __p0, uint64x2_t __p1) {
64296  uint64x2_t __ret;
64297  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
64298  return __ret;
64299}
64300#else
64301__ai uint64x2_t vtstq_u64(uint64x2_t __p0, uint64x2_t __p1) {
64302  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64303  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64304  uint64x2_t __ret;
64305  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
64306  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64307  return __ret;
64308}
64309#endif
64310
64311#ifdef __LITTLE_ENDIAN__
64312__ai uint64x2_t vtstq_s64(int64x2_t __p0, int64x2_t __p1) {
64313  uint64x2_t __ret;
64314  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
64315  return __ret;
64316}
64317#else
64318__ai uint64x2_t vtstq_s64(int64x2_t __p0, int64x2_t __p1) {
64319  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64320  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64321  uint64x2_t __ret;
64322  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
64323  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64324  return __ret;
64325}
64326#endif
64327
64328#ifdef __LITTLE_ENDIAN__
64329__ai uint64x1_t vtst_u64(uint64x1_t __p0, uint64x1_t __p1) {
64330  uint64x1_t __ret;
64331  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64332  return __ret;
64333}
64334#else
64335__ai uint64x1_t vtst_u64(uint64x1_t __p0, uint64x1_t __p1) {
64336  uint64x1_t __ret;
64337  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64338  return __ret;
64339}
64340#endif
64341
64342#ifdef __LITTLE_ENDIAN__
64343__ai uint64x1_t vtst_s64(int64x1_t __p0, int64x1_t __p1) {
64344  uint64x1_t __ret;
64345  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64346  return __ret;
64347}
64348#else
64349__ai uint64x1_t vtst_s64(int64x1_t __p0, int64x1_t __p1) {
64350  uint64x1_t __ret;
64351  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64352  return __ret;
64353}
64354#endif
64355
64356#ifdef __LITTLE_ENDIAN__
64357__ai uint64_t vtstd_u64(uint64_t __p0, uint64_t __p1) {
64358  uint64_t __ret;
64359  __ret = (uint64_t) __builtin_neon_vtstd_u64(__p0, __p1);
64360  return __ret;
64361}
64362#else
64363__ai uint64_t vtstd_u64(uint64_t __p0, uint64_t __p1) {
64364  uint64_t __ret;
64365  __ret = (uint64_t) __builtin_neon_vtstd_u64(__p0, __p1);
64366  return __ret;
64367}
64368#endif
64369
64370#ifdef __LITTLE_ENDIAN__
64371__ai int64_t vtstd_s64(int64_t __p0, int64_t __p1) {
64372  int64_t __ret;
64373  __ret = (int64_t) __builtin_neon_vtstd_s64(__p0, __p1);
64374  return __ret;
64375}
64376#else
64377__ai int64_t vtstd_s64(int64_t __p0, int64_t __p1) {
64378  int64_t __ret;
64379  __ret = (int64_t) __builtin_neon_vtstd_s64(__p0, __p1);
64380  return __ret;
64381}
64382#endif
64383
64384#ifdef __LITTLE_ENDIAN__
64385__ai int8_t vuqaddb_s8(int8_t __p0, int8_t __p1) {
64386  int8_t __ret;
64387  __ret = (int8_t) __builtin_neon_vuqaddb_s8(__p0, __p1);
64388  return __ret;
64389}
64390#else
64391__ai int8_t vuqaddb_s8(int8_t __p0, int8_t __p1) {
64392  int8_t __ret;
64393  __ret = (int8_t) __builtin_neon_vuqaddb_s8(__p0, __p1);
64394  return __ret;
64395}
64396#endif
64397
64398#ifdef __LITTLE_ENDIAN__
64399__ai int32_t vuqadds_s32(int32_t __p0, int32_t __p1) {
64400  int32_t __ret;
64401  __ret = (int32_t) __builtin_neon_vuqadds_s32(__p0, __p1);
64402  return __ret;
64403}
64404#else
64405__ai int32_t vuqadds_s32(int32_t __p0, int32_t __p1) {
64406  int32_t __ret;
64407  __ret = (int32_t) __builtin_neon_vuqadds_s32(__p0, __p1);
64408  return __ret;
64409}
64410#endif
64411
64412#ifdef __LITTLE_ENDIAN__
64413__ai int64_t vuqaddd_s64(int64_t __p0, int64_t __p1) {
64414  int64_t __ret;
64415  __ret = (int64_t) __builtin_neon_vuqaddd_s64(__p0, __p1);
64416  return __ret;
64417}
64418#else
64419__ai int64_t vuqaddd_s64(int64_t __p0, int64_t __p1) {
64420  int64_t __ret;
64421  __ret = (int64_t) __builtin_neon_vuqaddd_s64(__p0, __p1);
64422  return __ret;
64423}
64424#endif
64425
64426#ifdef __LITTLE_ENDIAN__
64427__ai int16_t vuqaddh_s16(int16_t __p0, int16_t __p1) {
64428  int16_t __ret;
64429  __ret = (int16_t) __builtin_neon_vuqaddh_s16(__p0, __p1);
64430  return __ret;
64431}
64432#else
64433__ai int16_t vuqaddh_s16(int16_t __p0, int16_t __p1) {
64434  int16_t __ret;
64435  __ret = (int16_t) __builtin_neon_vuqaddh_s16(__p0, __p1);
64436  return __ret;
64437}
64438#endif
64439
64440#ifdef __LITTLE_ENDIAN__
64441__ai int8x16_t vuqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
64442  int8x16_t __ret;
64443  __ret = (int8x16_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
64444  return __ret;
64445}
64446#else
64447__ai int8x16_t vuqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
64448  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64449  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64450  int8x16_t __ret;
64451  __ret = (int8x16_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
64452  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64453  return __ret;
64454}
64455#endif
64456
64457#ifdef __LITTLE_ENDIAN__
64458__ai int32x4_t vuqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
64459  int32x4_t __ret;
64460  __ret = (int32x4_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
64461  return __ret;
64462}
64463#else
64464__ai int32x4_t vuqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
64465  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64466  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64467  int32x4_t __ret;
64468  __ret = (int32x4_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
64469  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64470  return __ret;
64471}
64472#endif
64473
64474#ifdef __LITTLE_ENDIAN__
64475__ai int64x2_t vuqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
64476  int64x2_t __ret;
64477  __ret = (int64x2_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
64478  return __ret;
64479}
64480#else
64481__ai int64x2_t vuqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
64482  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64483  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64484  int64x2_t __ret;
64485  __ret = (int64x2_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
64486  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64487  return __ret;
64488}
64489#endif
64490
64491#ifdef __LITTLE_ENDIAN__
64492__ai int16x8_t vuqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
64493  int16x8_t __ret;
64494  __ret = (int16x8_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
64495  return __ret;
64496}
64497#else
64498__ai int16x8_t vuqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
64499  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64500  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64501  int16x8_t __ret;
64502  __ret = (int16x8_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
64503  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64504  return __ret;
64505}
64506#endif
64507
64508#ifdef __LITTLE_ENDIAN__
64509__ai int8x8_t vuqadd_s8(int8x8_t __p0, int8x8_t __p1) {
64510  int8x8_t __ret;
64511  __ret = (int8x8_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
64512  return __ret;
64513}
64514#else
64515__ai int8x8_t vuqadd_s8(int8x8_t __p0, int8x8_t __p1) {
64516  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64517  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64518  int8x8_t __ret;
64519  __ret = (int8x8_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
64520  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64521  return __ret;
64522}
64523#endif
64524
64525#ifdef __LITTLE_ENDIAN__
64526__ai int32x2_t vuqadd_s32(int32x2_t __p0, int32x2_t __p1) {
64527  int32x2_t __ret;
64528  __ret = (int32x2_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
64529  return __ret;
64530}
64531#else
64532__ai int32x2_t vuqadd_s32(int32x2_t __p0, int32x2_t __p1) {
64533  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64534  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64535  int32x2_t __ret;
64536  __ret = (int32x2_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
64537  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64538  return __ret;
64539}
64540#endif
64541
64542#ifdef __LITTLE_ENDIAN__
64543__ai int64x1_t vuqadd_s64(int64x1_t __p0, int64x1_t __p1) {
64544  int64x1_t __ret;
64545  __ret = (int64x1_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
64546  return __ret;
64547}
64548#else
64549__ai int64x1_t vuqadd_s64(int64x1_t __p0, int64x1_t __p1) {
64550  int64x1_t __ret;
64551  __ret = (int64x1_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
64552  return __ret;
64553}
64554#endif
64555
64556#ifdef __LITTLE_ENDIAN__
64557__ai int16x4_t vuqadd_s16(int16x4_t __p0, int16x4_t __p1) {
64558  int16x4_t __ret;
64559  __ret = (int16x4_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
64560  return __ret;
64561}
64562#else
64563__ai int16x4_t vuqadd_s16(int16x4_t __p0, int16x4_t __p1) {
64564  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64565  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64566  int16x4_t __ret;
64567  __ret = (int16x4_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
64568  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64569  return __ret;
64570}
64571#endif
64572
64573#ifdef __LITTLE_ENDIAN__
64574__ai poly8x8_t vuzp1_p8(poly8x8_t __p0, poly8x8_t __p1) {
64575  poly8x8_t __ret;
64576  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
64577  return __ret;
64578}
64579#else
64580__ai poly8x8_t vuzp1_p8(poly8x8_t __p0, poly8x8_t __p1) {
64581  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64582  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64583  poly8x8_t __ret;
64584  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
64585  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64586  return __ret;
64587}
64588#endif
64589
64590#ifdef __LITTLE_ENDIAN__
64591__ai poly16x4_t vuzp1_p16(poly16x4_t __p0, poly16x4_t __p1) {
64592  poly16x4_t __ret;
64593  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
64594  return __ret;
64595}
64596#else
64597__ai poly16x4_t vuzp1_p16(poly16x4_t __p0, poly16x4_t __p1) {
64598  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64599  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64600  poly16x4_t __ret;
64601  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
64602  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64603  return __ret;
64604}
64605#endif
64606
64607#ifdef __LITTLE_ENDIAN__
64608__ai poly8x16_t vuzp1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64609  poly8x16_t __ret;
64610  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
64611  return __ret;
64612}
64613#else
64614__ai poly8x16_t vuzp1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64615  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64616  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64617  poly8x16_t __ret;
64618  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
64619  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64620  return __ret;
64621}
64622#endif
64623
64624#ifdef __LITTLE_ENDIAN__
64625__ai poly64x2_t vuzp1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
64626  poly64x2_t __ret;
64627  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64628  return __ret;
64629}
64630#else
64631__ai poly64x2_t vuzp1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
64632  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64633  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64634  poly64x2_t __ret;
64635  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64636  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64637  return __ret;
64638}
64639#endif
64640
64641#ifdef __LITTLE_ENDIAN__
64642__ai poly16x8_t vuzp1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
64643  poly16x8_t __ret;
64644  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
64645  return __ret;
64646}
64647#else
64648__ai poly16x8_t vuzp1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
64649  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64650  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64651  poly16x8_t __ret;
64652  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
64653  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64654  return __ret;
64655}
64656#endif
64657
64658#ifdef __LITTLE_ENDIAN__
64659__ai uint8x16_t vuzp1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
64660  uint8x16_t __ret;
64661  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
64662  return __ret;
64663}
64664#else
64665__ai uint8x16_t vuzp1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
64666  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64667  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64668  uint8x16_t __ret;
64669  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
64670  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64671  return __ret;
64672}
64673#endif
64674
64675#ifdef __LITTLE_ENDIAN__
64676__ai uint32x4_t vuzp1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
64677  uint32x4_t __ret;
64678  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
64679  return __ret;
64680}
64681#else
64682__ai uint32x4_t vuzp1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
64683  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64684  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64685  uint32x4_t __ret;
64686  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
64687  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64688  return __ret;
64689}
64690#endif
64691
64692#ifdef __LITTLE_ENDIAN__
64693__ai uint64x2_t vuzp1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64694  uint64x2_t __ret;
64695  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64696  return __ret;
64697}
64698#else
64699__ai uint64x2_t vuzp1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64700  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64701  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64702  uint64x2_t __ret;
64703  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64704  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64705  return __ret;
64706}
64707#endif
64708
64709#ifdef __LITTLE_ENDIAN__
64710__ai uint16x8_t vuzp1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64711  uint16x8_t __ret;
64712  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
64713  return __ret;
64714}
64715#else
64716__ai uint16x8_t vuzp1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64717  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64718  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64719  uint16x8_t __ret;
64720  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
64721  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64722  return __ret;
64723}
64724#endif
64725
64726#ifdef __LITTLE_ENDIAN__
64727__ai int8x16_t vuzp1q_s8(int8x16_t __p0, int8x16_t __p1) {
64728  int8x16_t __ret;
64729  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
64730  return __ret;
64731}
64732#else
64733__ai int8x16_t vuzp1q_s8(int8x16_t __p0, int8x16_t __p1) {
64734  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64735  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64736  int8x16_t __ret;
64737  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
64738  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64739  return __ret;
64740}
64741#endif
64742
64743#ifdef __LITTLE_ENDIAN__
64744__ai float64x2_t vuzp1q_f64(float64x2_t __p0, float64x2_t __p1) {
64745  float64x2_t __ret;
64746  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64747  return __ret;
64748}
64749#else
64750__ai float64x2_t vuzp1q_f64(float64x2_t __p0, float64x2_t __p1) {
64751  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64752  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64753  float64x2_t __ret;
64754  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64755  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64756  return __ret;
64757}
64758#endif
64759
64760#ifdef __LITTLE_ENDIAN__
64761__ai float32x4_t vuzp1q_f32(float32x4_t __p0, float32x4_t __p1) {
64762  float32x4_t __ret;
64763  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
64764  return __ret;
64765}
64766#else
64767__ai float32x4_t vuzp1q_f32(float32x4_t __p0, float32x4_t __p1) {
64768  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64769  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64770  float32x4_t __ret;
64771  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
64772  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64773  return __ret;
64774}
64775#endif
64776
64777#ifdef __LITTLE_ENDIAN__
64778__ai int32x4_t vuzp1q_s32(int32x4_t __p0, int32x4_t __p1) {
64779  int32x4_t __ret;
64780  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
64781  return __ret;
64782}
64783#else
64784__ai int32x4_t vuzp1q_s32(int32x4_t __p0, int32x4_t __p1) {
64785  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64786  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64787  int32x4_t __ret;
64788  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
64789  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64790  return __ret;
64791}
64792#endif
64793
64794#ifdef __LITTLE_ENDIAN__
64795__ai int64x2_t vuzp1q_s64(int64x2_t __p0, int64x2_t __p1) {
64796  int64x2_t __ret;
64797  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64798  return __ret;
64799}
64800#else
64801__ai int64x2_t vuzp1q_s64(int64x2_t __p0, int64x2_t __p1) {
64802  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64803  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64804  int64x2_t __ret;
64805  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64806  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64807  return __ret;
64808}
64809#endif
64810
64811#ifdef __LITTLE_ENDIAN__
64812__ai int16x8_t vuzp1q_s16(int16x8_t __p0, int16x8_t __p1) {
64813  int16x8_t __ret;
64814  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
64815  return __ret;
64816}
64817#else
64818__ai int16x8_t vuzp1q_s16(int16x8_t __p0, int16x8_t __p1) {
64819  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64820  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64821  int16x8_t __ret;
64822  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
64823  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64824  return __ret;
64825}
64826#endif
64827
64828#ifdef __LITTLE_ENDIAN__
64829__ai uint8x8_t vuzp1_u8(uint8x8_t __p0, uint8x8_t __p1) {
64830  uint8x8_t __ret;
64831  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
64832  return __ret;
64833}
64834#else
64835__ai uint8x8_t vuzp1_u8(uint8x8_t __p0, uint8x8_t __p1) {
64836  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64837  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64838  uint8x8_t __ret;
64839  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
64840  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64841  return __ret;
64842}
64843#endif
64844
64845#ifdef __LITTLE_ENDIAN__
64846__ai uint32x2_t vuzp1_u32(uint32x2_t __p0, uint32x2_t __p1) {
64847  uint32x2_t __ret;
64848  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64849  return __ret;
64850}
64851#else
64852__ai uint32x2_t vuzp1_u32(uint32x2_t __p0, uint32x2_t __p1) {
64853  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64854  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64855  uint32x2_t __ret;
64856  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64857  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64858  return __ret;
64859}
64860#endif
64861
64862#ifdef __LITTLE_ENDIAN__
64863__ai uint16x4_t vuzp1_u16(uint16x4_t __p0, uint16x4_t __p1) {
64864  uint16x4_t __ret;
64865  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
64866  return __ret;
64867}
64868#else
64869__ai uint16x4_t vuzp1_u16(uint16x4_t __p0, uint16x4_t __p1) {
64870  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64871  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64872  uint16x4_t __ret;
64873  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
64874  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64875  return __ret;
64876}
64877#endif
64878
64879#ifdef __LITTLE_ENDIAN__
64880__ai int8x8_t vuzp1_s8(int8x8_t __p0, int8x8_t __p1) {
64881  int8x8_t __ret;
64882  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
64883  return __ret;
64884}
64885#else
64886__ai int8x8_t vuzp1_s8(int8x8_t __p0, int8x8_t __p1) {
64887  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64888  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64889  int8x8_t __ret;
64890  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
64891  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64892  return __ret;
64893}
64894#endif
64895
64896#ifdef __LITTLE_ENDIAN__
64897__ai float32x2_t vuzp1_f32(float32x2_t __p0, float32x2_t __p1) {
64898  float32x2_t __ret;
64899  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64900  return __ret;
64901}
64902#else
64903__ai float32x2_t vuzp1_f32(float32x2_t __p0, float32x2_t __p1) {
64904  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64905  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64906  float32x2_t __ret;
64907  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64908  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64909  return __ret;
64910}
64911#endif
64912
64913#ifdef __LITTLE_ENDIAN__
64914__ai int32x2_t vuzp1_s32(int32x2_t __p0, int32x2_t __p1) {
64915  int32x2_t __ret;
64916  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64917  return __ret;
64918}
64919#else
64920__ai int32x2_t vuzp1_s32(int32x2_t __p0, int32x2_t __p1) {
64921  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64922  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64923  int32x2_t __ret;
64924  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64925  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64926  return __ret;
64927}
64928#endif
64929
64930#ifdef __LITTLE_ENDIAN__
64931__ai int16x4_t vuzp1_s16(int16x4_t __p0, int16x4_t __p1) {
64932  int16x4_t __ret;
64933  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
64934  return __ret;
64935}
64936#else
64937__ai int16x4_t vuzp1_s16(int16x4_t __p0, int16x4_t __p1) {
64938  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64939  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64940  int16x4_t __ret;
64941  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
64942  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64943  return __ret;
64944}
64945#endif
64946
64947#ifdef __LITTLE_ENDIAN__
64948__ai poly8x8_t vuzp2_p8(poly8x8_t __p0, poly8x8_t __p1) {
64949  poly8x8_t __ret;
64950  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
64951  return __ret;
64952}
64953#else
64954__ai poly8x8_t vuzp2_p8(poly8x8_t __p0, poly8x8_t __p1) {
64955  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64956  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64957  poly8x8_t __ret;
64958  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
64959  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64960  return __ret;
64961}
64962#endif
64963
64964#ifdef __LITTLE_ENDIAN__
64965__ai poly16x4_t vuzp2_p16(poly16x4_t __p0, poly16x4_t __p1) {
64966  poly16x4_t __ret;
64967  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
64968  return __ret;
64969}
64970#else
64971__ai poly16x4_t vuzp2_p16(poly16x4_t __p0, poly16x4_t __p1) {
64972  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64973  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64974  poly16x4_t __ret;
64975  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
64976  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64977  return __ret;
64978}
64979#endif
64980
64981#ifdef __LITTLE_ENDIAN__
64982__ai poly8x16_t vuzp2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64983  poly8x16_t __ret;
64984  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
64985  return __ret;
64986}
64987#else
64988__ai poly8x16_t vuzp2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64989  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64990  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64991  poly8x16_t __ret;
64992  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
64993  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64994  return __ret;
64995}
64996#endif
64997
64998#ifdef __LITTLE_ENDIAN__
64999__ai poly64x2_t vuzp2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65000  poly64x2_t __ret;
65001  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65002  return __ret;
65003}
65004#else
65005__ai poly64x2_t vuzp2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65006  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65007  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65008  poly64x2_t __ret;
65009  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65010  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65011  return __ret;
65012}
65013#endif
65014
65015#ifdef __LITTLE_ENDIAN__
65016__ai poly16x8_t vuzp2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65017  poly16x8_t __ret;
65018  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65019  return __ret;
65020}
65021#else
65022__ai poly16x8_t vuzp2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65023  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65024  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65025  poly16x8_t __ret;
65026  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65027  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65028  return __ret;
65029}
65030#endif
65031
65032#ifdef __LITTLE_ENDIAN__
65033__ai uint8x16_t vuzp2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65034  uint8x16_t __ret;
65035  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65036  return __ret;
65037}
65038#else
65039__ai uint8x16_t vuzp2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65040  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65041  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65042  uint8x16_t __ret;
65043  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65044  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65045  return __ret;
65046}
65047#endif
65048
65049#ifdef __LITTLE_ENDIAN__
65050__ai uint32x4_t vuzp2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65051  uint32x4_t __ret;
65052  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65053  return __ret;
65054}
65055#else
65056__ai uint32x4_t vuzp2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65057  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65058  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65059  uint32x4_t __ret;
65060  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65061  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65062  return __ret;
65063}
65064#endif
65065
65066#ifdef __LITTLE_ENDIAN__
65067__ai uint64x2_t vuzp2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65068  uint64x2_t __ret;
65069  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65070  return __ret;
65071}
65072#else
65073__ai uint64x2_t vuzp2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65074  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65075  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65076  uint64x2_t __ret;
65077  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65078  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65079  return __ret;
65080}
65081#endif
65082
65083#ifdef __LITTLE_ENDIAN__
65084__ai uint16x8_t vuzp2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65085  uint16x8_t __ret;
65086  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65087  return __ret;
65088}
65089#else
65090__ai uint16x8_t vuzp2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65091  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65092  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65093  uint16x8_t __ret;
65094  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65095  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65096  return __ret;
65097}
65098#endif
65099
65100#ifdef __LITTLE_ENDIAN__
65101__ai int8x16_t vuzp2q_s8(int8x16_t __p0, int8x16_t __p1) {
65102  int8x16_t __ret;
65103  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65104  return __ret;
65105}
65106#else
65107__ai int8x16_t vuzp2q_s8(int8x16_t __p0, int8x16_t __p1) {
65108  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65109  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65110  int8x16_t __ret;
65111  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65112  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65113  return __ret;
65114}
65115#endif
65116
65117#ifdef __LITTLE_ENDIAN__
65118__ai float64x2_t vuzp2q_f64(float64x2_t __p0, float64x2_t __p1) {
65119  float64x2_t __ret;
65120  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65121  return __ret;
65122}
65123#else
65124__ai float64x2_t vuzp2q_f64(float64x2_t __p0, float64x2_t __p1) {
65125  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65126  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65127  float64x2_t __ret;
65128  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65129  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65130  return __ret;
65131}
65132#endif
65133
65134#ifdef __LITTLE_ENDIAN__
65135__ai float32x4_t vuzp2q_f32(float32x4_t __p0, float32x4_t __p1) {
65136  float32x4_t __ret;
65137  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65138  return __ret;
65139}
65140#else
65141__ai float32x4_t vuzp2q_f32(float32x4_t __p0, float32x4_t __p1) {
65142  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65143  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65144  float32x4_t __ret;
65145  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65146  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65147  return __ret;
65148}
65149#endif
65150
65151#ifdef __LITTLE_ENDIAN__
65152__ai int32x4_t vuzp2q_s32(int32x4_t __p0, int32x4_t __p1) {
65153  int32x4_t __ret;
65154  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65155  return __ret;
65156}
65157#else
65158__ai int32x4_t vuzp2q_s32(int32x4_t __p0, int32x4_t __p1) {
65159  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65160  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65161  int32x4_t __ret;
65162  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65163  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65164  return __ret;
65165}
65166#endif
65167
65168#ifdef __LITTLE_ENDIAN__
65169__ai int64x2_t vuzp2q_s64(int64x2_t __p0, int64x2_t __p1) {
65170  int64x2_t __ret;
65171  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65172  return __ret;
65173}
65174#else
65175__ai int64x2_t vuzp2q_s64(int64x2_t __p0, int64x2_t __p1) {
65176  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65177  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65178  int64x2_t __ret;
65179  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65180  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65181  return __ret;
65182}
65183#endif
65184
65185#ifdef __LITTLE_ENDIAN__
65186__ai int16x8_t vuzp2q_s16(int16x8_t __p0, int16x8_t __p1) {
65187  int16x8_t __ret;
65188  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65189  return __ret;
65190}
65191#else
65192__ai int16x8_t vuzp2q_s16(int16x8_t __p0, int16x8_t __p1) {
65193  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65194  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65195  int16x8_t __ret;
65196  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65197  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65198  return __ret;
65199}
65200#endif
65201
65202#ifdef __LITTLE_ENDIAN__
65203__ai uint8x8_t vuzp2_u8(uint8x8_t __p0, uint8x8_t __p1) {
65204  uint8x8_t __ret;
65205  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65206  return __ret;
65207}
65208#else
65209__ai uint8x8_t vuzp2_u8(uint8x8_t __p0, uint8x8_t __p1) {
65210  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65211  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65212  uint8x8_t __ret;
65213  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65214  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65215  return __ret;
65216}
65217#endif
65218
65219#ifdef __LITTLE_ENDIAN__
65220__ai uint32x2_t vuzp2_u32(uint32x2_t __p0, uint32x2_t __p1) {
65221  uint32x2_t __ret;
65222  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65223  return __ret;
65224}
65225#else
65226__ai uint32x2_t vuzp2_u32(uint32x2_t __p0, uint32x2_t __p1) {
65227  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65228  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65229  uint32x2_t __ret;
65230  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65231  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65232  return __ret;
65233}
65234#endif
65235
65236#ifdef __LITTLE_ENDIAN__
65237__ai uint16x4_t vuzp2_u16(uint16x4_t __p0, uint16x4_t __p1) {
65238  uint16x4_t __ret;
65239  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65240  return __ret;
65241}
65242#else
65243__ai uint16x4_t vuzp2_u16(uint16x4_t __p0, uint16x4_t __p1) {
65244  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65245  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65246  uint16x4_t __ret;
65247  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65248  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65249  return __ret;
65250}
65251#endif
65252
65253#ifdef __LITTLE_ENDIAN__
65254__ai int8x8_t vuzp2_s8(int8x8_t __p0, int8x8_t __p1) {
65255  int8x8_t __ret;
65256  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65257  return __ret;
65258}
65259#else
65260__ai int8x8_t vuzp2_s8(int8x8_t __p0, int8x8_t __p1) {
65261  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65262  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65263  int8x8_t __ret;
65264  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65265  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65266  return __ret;
65267}
65268#endif
65269
65270#ifdef __LITTLE_ENDIAN__
65271__ai float32x2_t vuzp2_f32(float32x2_t __p0, float32x2_t __p1) {
65272  float32x2_t __ret;
65273  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65274  return __ret;
65275}
65276#else
65277__ai float32x2_t vuzp2_f32(float32x2_t __p0, float32x2_t __p1) {
65278  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65279  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65280  float32x2_t __ret;
65281  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65282  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65283  return __ret;
65284}
65285#endif
65286
65287#ifdef __LITTLE_ENDIAN__
65288__ai int32x2_t vuzp2_s32(int32x2_t __p0, int32x2_t __p1) {
65289  int32x2_t __ret;
65290  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65291  return __ret;
65292}
65293#else
65294__ai int32x2_t vuzp2_s32(int32x2_t __p0, int32x2_t __p1) {
65295  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65296  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65297  int32x2_t __ret;
65298  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65299  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65300  return __ret;
65301}
65302#endif
65303
65304#ifdef __LITTLE_ENDIAN__
65305__ai int16x4_t vuzp2_s16(int16x4_t __p0, int16x4_t __p1) {
65306  int16x4_t __ret;
65307  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65308  return __ret;
65309}
65310#else
65311__ai int16x4_t vuzp2_s16(int16x4_t __p0, int16x4_t __p1) {
65312  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65313  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65314  int16x4_t __ret;
65315  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65316  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65317  return __ret;
65318}
65319#endif
65320
65321#ifdef __LITTLE_ENDIAN__
65322__ai poly8x8_t vzip1_p8(poly8x8_t __p0, poly8x8_t __p1) {
65323  poly8x8_t __ret;
65324  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65325  return __ret;
65326}
65327#else
65328__ai poly8x8_t vzip1_p8(poly8x8_t __p0, poly8x8_t __p1) {
65329  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65330  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65331  poly8x8_t __ret;
65332  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
65333  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65334  return __ret;
65335}
65336#endif
65337
65338#ifdef __LITTLE_ENDIAN__
65339__ai poly16x4_t vzip1_p16(poly16x4_t __p0, poly16x4_t __p1) {
65340  poly16x4_t __ret;
65341  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
65342  return __ret;
65343}
65344#else
65345__ai poly16x4_t vzip1_p16(poly16x4_t __p0, poly16x4_t __p1) {
65346  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65347  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65348  poly16x4_t __ret;
65349  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
65350  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65351  return __ret;
65352}
65353#endif
65354
65355#ifdef __LITTLE_ENDIAN__
65356__ai poly8x16_t vzip1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65357  poly8x16_t __ret;
65358  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65359  return __ret;
65360}
65361#else
65362__ai poly8x16_t vzip1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65363  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65364  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65365  poly8x16_t __ret;
65366  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65367  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65368  return __ret;
65369}
65370#endif
65371
65372#ifdef __LITTLE_ENDIAN__
65373__ai poly64x2_t vzip1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65374  poly64x2_t __ret;
65375  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65376  return __ret;
65377}
65378#else
65379__ai poly64x2_t vzip1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65380  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65381  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65382  poly64x2_t __ret;
65383  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65384  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65385  return __ret;
65386}
65387#endif
65388
65389#ifdef __LITTLE_ENDIAN__
65390__ai poly16x8_t vzip1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65391  poly16x8_t __ret;
65392  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65393  return __ret;
65394}
65395#else
65396__ai poly16x8_t vzip1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65397  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65398  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65399  poly16x8_t __ret;
65400  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
65401  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65402  return __ret;
65403}
65404#endif
65405
65406#ifdef __LITTLE_ENDIAN__
65407__ai uint8x16_t vzip1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65408  uint8x16_t __ret;
65409  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65410  return __ret;
65411}
65412#else
65413__ai uint8x16_t vzip1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65414  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65415  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65416  uint8x16_t __ret;
65417  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65418  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65419  return __ret;
65420}
65421#endif
65422
65423#ifdef __LITTLE_ENDIAN__
65424__ai uint32x4_t vzip1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65425  uint32x4_t __ret;
65426  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
65427  return __ret;
65428}
65429#else
65430__ai uint32x4_t vzip1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65431  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65432  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65433  uint32x4_t __ret;
65434  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
65435  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65436  return __ret;
65437}
65438#endif
65439
65440#ifdef __LITTLE_ENDIAN__
65441__ai uint64x2_t vzip1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65442  uint64x2_t __ret;
65443  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65444  return __ret;
65445}
65446#else
65447__ai uint64x2_t vzip1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65448  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65449  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65450  uint64x2_t __ret;
65451  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65452  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65453  return __ret;
65454}
65455#endif
65456
65457#ifdef __LITTLE_ENDIAN__
65458__ai uint16x8_t vzip1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65459  uint16x8_t __ret;
65460  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65461  return __ret;
65462}
65463#else
65464__ai uint16x8_t vzip1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65465  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65466  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65467  uint16x8_t __ret;
65468  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
65469  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65470  return __ret;
65471}
65472#endif
65473
65474#ifdef __LITTLE_ENDIAN__
65475__ai int8x16_t vzip1q_s8(int8x16_t __p0, int8x16_t __p1) {
65476  int8x16_t __ret;
65477  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65478  return __ret;
65479}
65480#else
65481__ai int8x16_t vzip1q_s8(int8x16_t __p0, int8x16_t __p1) {
65482  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65483  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65484  int8x16_t __ret;
65485  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65486  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65487  return __ret;
65488}
65489#endif
65490
65491#ifdef __LITTLE_ENDIAN__
65492__ai float64x2_t vzip1q_f64(float64x2_t __p0, float64x2_t __p1) {
65493  float64x2_t __ret;
65494  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65495  return __ret;
65496}
65497#else
65498__ai float64x2_t vzip1q_f64(float64x2_t __p0, float64x2_t __p1) {
65499  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65500  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65501  float64x2_t __ret;
65502  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65503  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65504  return __ret;
65505}
65506#endif
65507
65508#ifdef __LITTLE_ENDIAN__
65509__ai float32x4_t vzip1q_f32(float32x4_t __p0, float32x4_t __p1) {
65510  float32x4_t __ret;
65511  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
65512  return __ret;
65513}
65514#else
65515__ai float32x4_t vzip1q_f32(float32x4_t __p0, float32x4_t __p1) {
65516  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65517  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65518  float32x4_t __ret;
65519  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
65520  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65521  return __ret;
65522}
65523#endif
65524
65525#ifdef __LITTLE_ENDIAN__
65526__ai int32x4_t vzip1q_s32(int32x4_t __p0, int32x4_t __p1) {
65527  int32x4_t __ret;
65528  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
65529  return __ret;
65530}
65531#else
65532__ai int32x4_t vzip1q_s32(int32x4_t __p0, int32x4_t __p1) {
65533  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65534  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65535  int32x4_t __ret;
65536  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
65537  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65538  return __ret;
65539}
65540#endif
65541
65542#ifdef __LITTLE_ENDIAN__
65543__ai int64x2_t vzip1q_s64(int64x2_t __p0, int64x2_t __p1) {
65544  int64x2_t __ret;
65545  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65546  return __ret;
65547}
65548#else
65549__ai int64x2_t vzip1q_s64(int64x2_t __p0, int64x2_t __p1) {
65550  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65551  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65552  int64x2_t __ret;
65553  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65554  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65555  return __ret;
65556}
65557#endif
65558
65559#ifdef __LITTLE_ENDIAN__
65560__ai int16x8_t vzip1q_s16(int16x8_t __p0, int16x8_t __p1) {
65561  int16x8_t __ret;
65562  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65563  return __ret;
65564}
65565#else
65566__ai int16x8_t vzip1q_s16(int16x8_t __p0, int16x8_t __p1) {
65567  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65568  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65569  int16x8_t __ret;
65570  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
65571  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65572  return __ret;
65573}
65574#endif
65575
65576#ifdef __LITTLE_ENDIAN__
65577__ai uint8x8_t vzip1_u8(uint8x8_t __p0, uint8x8_t __p1) {
65578  uint8x8_t __ret;
65579  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65580  return __ret;
65581}
65582#else
65583__ai uint8x8_t vzip1_u8(uint8x8_t __p0, uint8x8_t __p1) {
65584  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65585  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65586  uint8x8_t __ret;
65587  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
65588  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65589  return __ret;
65590}
65591#endif
65592
65593#ifdef __LITTLE_ENDIAN__
65594__ai uint32x2_t vzip1_u32(uint32x2_t __p0, uint32x2_t __p1) {
65595  uint32x2_t __ret;
65596  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65597  return __ret;
65598}
65599#else
65600__ai uint32x2_t vzip1_u32(uint32x2_t __p0, uint32x2_t __p1) {
65601  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65602  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65603  uint32x2_t __ret;
65604  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65605  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65606  return __ret;
65607}
65608#endif
65609
65610#ifdef __LITTLE_ENDIAN__
65611__ai uint16x4_t vzip1_u16(uint16x4_t __p0, uint16x4_t __p1) {
65612  uint16x4_t __ret;
65613  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
65614  return __ret;
65615}
65616#else
65617__ai uint16x4_t vzip1_u16(uint16x4_t __p0, uint16x4_t __p1) {
65618  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65619  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65620  uint16x4_t __ret;
65621  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
65622  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65623  return __ret;
65624}
65625#endif
65626
65627#ifdef __LITTLE_ENDIAN__
65628__ai int8x8_t vzip1_s8(int8x8_t __p0, int8x8_t __p1) {
65629  int8x8_t __ret;
65630  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65631  return __ret;
65632}
65633#else
65634__ai int8x8_t vzip1_s8(int8x8_t __p0, int8x8_t __p1) {
65635  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65636  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65637  int8x8_t __ret;
65638  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
65639  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65640  return __ret;
65641}
65642#endif
65643
65644#ifdef __LITTLE_ENDIAN__
65645__ai float32x2_t vzip1_f32(float32x2_t __p0, float32x2_t __p1) {
65646  float32x2_t __ret;
65647  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65648  return __ret;
65649}
65650#else
65651__ai float32x2_t vzip1_f32(float32x2_t __p0, float32x2_t __p1) {
65652  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65653  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65654  float32x2_t __ret;
65655  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65656  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65657  return __ret;
65658}
65659#endif
65660
65661#ifdef __LITTLE_ENDIAN__
65662__ai int32x2_t vzip1_s32(int32x2_t __p0, int32x2_t __p1) {
65663  int32x2_t __ret;
65664  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65665  return __ret;
65666}
65667#else
65668__ai int32x2_t vzip1_s32(int32x2_t __p0, int32x2_t __p1) {
65669  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65670  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65671  int32x2_t __ret;
65672  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65673  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65674  return __ret;
65675}
65676#endif
65677
65678#ifdef __LITTLE_ENDIAN__
65679__ai int16x4_t vzip1_s16(int16x4_t __p0, int16x4_t __p1) {
65680  int16x4_t __ret;
65681  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
65682  return __ret;
65683}
65684#else
65685__ai int16x4_t vzip1_s16(int16x4_t __p0, int16x4_t __p1) {
65686  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65687  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65688  int16x4_t __ret;
65689  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
65690  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65691  return __ret;
65692}
65693#endif
65694
65695#ifdef __LITTLE_ENDIAN__
65696__ai poly8x8_t vzip2_p8(poly8x8_t __p0, poly8x8_t __p1) {
65697  poly8x8_t __ret;
65698  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
65699  return __ret;
65700}
65701#else
65702__ai poly8x8_t vzip2_p8(poly8x8_t __p0, poly8x8_t __p1) {
65703  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65704  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65705  poly8x8_t __ret;
65706  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
65707  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65708  return __ret;
65709}
65710#endif
65711
65712#ifdef __LITTLE_ENDIAN__
65713__ai poly16x4_t vzip2_p16(poly16x4_t __p0, poly16x4_t __p1) {
65714  poly16x4_t __ret;
65715  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
65716  return __ret;
65717}
65718#else
65719__ai poly16x4_t vzip2_p16(poly16x4_t __p0, poly16x4_t __p1) {
65720  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65721  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65722  poly16x4_t __ret;
65723  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
65724  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65725  return __ret;
65726}
65727#endif
65728
65729#ifdef __LITTLE_ENDIAN__
65730__ai poly8x16_t vzip2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65731  poly8x16_t __ret;
65732  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
65733  return __ret;
65734}
65735#else
65736__ai poly8x16_t vzip2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65737  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65738  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65739  poly8x16_t __ret;
65740  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
65741  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65742  return __ret;
65743}
65744#endif
65745
65746#ifdef __LITTLE_ENDIAN__
65747__ai poly64x2_t vzip2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65748  poly64x2_t __ret;
65749  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65750  return __ret;
65751}
65752#else
65753__ai poly64x2_t vzip2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65754  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65755  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65756  poly64x2_t __ret;
65757  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65758  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65759  return __ret;
65760}
65761#endif
65762
65763#ifdef __LITTLE_ENDIAN__
65764__ai poly16x8_t vzip2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65765  poly16x8_t __ret;
65766  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
65767  return __ret;
65768}
65769#else
65770__ai poly16x8_t vzip2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65771  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65772  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65773  poly16x8_t __ret;
65774  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
65775  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65776  return __ret;
65777}
65778#endif
65779
65780#ifdef __LITTLE_ENDIAN__
65781__ai uint8x16_t vzip2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65782  uint8x16_t __ret;
65783  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
65784  return __ret;
65785}
65786#else
65787__ai uint8x16_t vzip2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65788  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65789  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65790  uint8x16_t __ret;
65791  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
65792  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65793  return __ret;
65794}
65795#endif
65796
65797#ifdef __LITTLE_ENDIAN__
65798__ai uint32x4_t vzip2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65799  uint32x4_t __ret;
65800  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
65801  return __ret;
65802}
65803#else
65804__ai uint32x4_t vzip2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65805  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65806  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65807  uint32x4_t __ret;
65808  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
65809  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65810  return __ret;
65811}
65812#endif
65813
65814#ifdef __LITTLE_ENDIAN__
65815__ai uint64x2_t vzip2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65816  uint64x2_t __ret;
65817  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65818  return __ret;
65819}
65820#else
65821__ai uint64x2_t vzip2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65822  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65823  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65824  uint64x2_t __ret;
65825  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65826  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65827  return __ret;
65828}
65829#endif
65830
65831#ifdef __LITTLE_ENDIAN__
65832__ai uint16x8_t vzip2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65833  uint16x8_t __ret;
65834  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
65835  return __ret;
65836}
65837#else
65838__ai uint16x8_t vzip2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65839  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65840  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65841  uint16x8_t __ret;
65842  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
65843  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65844  return __ret;
65845}
65846#endif
65847
65848#ifdef __LITTLE_ENDIAN__
65849__ai int8x16_t vzip2q_s8(int8x16_t __p0, int8x16_t __p1) {
65850  int8x16_t __ret;
65851  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
65852  return __ret;
65853}
65854#else
65855__ai int8x16_t vzip2q_s8(int8x16_t __p0, int8x16_t __p1) {
65856  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65857  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65858  int8x16_t __ret;
65859  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
65860  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65861  return __ret;
65862}
65863#endif
65864
65865#ifdef __LITTLE_ENDIAN__
65866__ai float64x2_t vzip2q_f64(float64x2_t __p0, float64x2_t __p1) {
65867  float64x2_t __ret;
65868  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65869  return __ret;
65870}
65871#else
65872__ai float64x2_t vzip2q_f64(float64x2_t __p0, float64x2_t __p1) {
65873  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65874  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65875  float64x2_t __ret;
65876  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65877  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65878  return __ret;
65879}
65880#endif
65881
65882#ifdef __LITTLE_ENDIAN__
65883__ai float32x4_t vzip2q_f32(float32x4_t __p0, float32x4_t __p1) {
65884  float32x4_t __ret;
65885  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
65886  return __ret;
65887}
65888#else
65889__ai float32x4_t vzip2q_f32(float32x4_t __p0, float32x4_t __p1) {
65890  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65891  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65892  float32x4_t __ret;
65893  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
65894  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65895  return __ret;
65896}
65897#endif
65898
65899#ifdef __LITTLE_ENDIAN__
65900__ai int32x4_t vzip2q_s32(int32x4_t __p0, int32x4_t __p1) {
65901  int32x4_t __ret;
65902  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
65903  return __ret;
65904}
65905#else
65906__ai int32x4_t vzip2q_s32(int32x4_t __p0, int32x4_t __p1) {
65907  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65908  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65909  int32x4_t __ret;
65910  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
65911  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65912  return __ret;
65913}
65914#endif
65915
65916#ifdef __LITTLE_ENDIAN__
65917__ai int64x2_t vzip2q_s64(int64x2_t __p0, int64x2_t __p1) {
65918  int64x2_t __ret;
65919  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65920  return __ret;
65921}
65922#else
65923__ai int64x2_t vzip2q_s64(int64x2_t __p0, int64x2_t __p1) {
65924  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65925  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65926  int64x2_t __ret;
65927  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65928  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65929  return __ret;
65930}
65931#endif
65932
65933#ifdef __LITTLE_ENDIAN__
65934__ai int16x8_t vzip2q_s16(int16x8_t __p0, int16x8_t __p1) {
65935  int16x8_t __ret;
65936  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
65937  return __ret;
65938}
65939#else
65940__ai int16x8_t vzip2q_s16(int16x8_t __p0, int16x8_t __p1) {
65941  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65942  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65943  int16x8_t __ret;
65944  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
65945  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65946  return __ret;
65947}
65948#endif
65949
65950#ifdef __LITTLE_ENDIAN__
65951__ai uint8x8_t vzip2_u8(uint8x8_t __p0, uint8x8_t __p1) {
65952  uint8x8_t __ret;
65953  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
65954  return __ret;
65955}
65956#else
65957__ai uint8x8_t vzip2_u8(uint8x8_t __p0, uint8x8_t __p1) {
65958  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65959  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65960  uint8x8_t __ret;
65961  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
65962  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65963  return __ret;
65964}
65965#endif
65966
65967#ifdef __LITTLE_ENDIAN__
65968__ai uint32x2_t vzip2_u32(uint32x2_t __p0, uint32x2_t __p1) {
65969  uint32x2_t __ret;
65970  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65971  return __ret;
65972}
65973#else
65974__ai uint32x2_t vzip2_u32(uint32x2_t __p0, uint32x2_t __p1) {
65975  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65976  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65977  uint32x2_t __ret;
65978  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65979  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65980  return __ret;
65981}
65982#endif
65983
65984#ifdef __LITTLE_ENDIAN__
65985__ai uint16x4_t vzip2_u16(uint16x4_t __p0, uint16x4_t __p1) {
65986  uint16x4_t __ret;
65987  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
65988  return __ret;
65989}
65990#else
65991__ai uint16x4_t vzip2_u16(uint16x4_t __p0, uint16x4_t __p1) {
65992  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65993  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65994  uint16x4_t __ret;
65995  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
65996  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65997  return __ret;
65998}
65999#endif
66000
66001#ifdef __LITTLE_ENDIAN__
66002__ai int8x8_t vzip2_s8(int8x8_t __p0, int8x8_t __p1) {
66003  int8x8_t __ret;
66004  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66005  return __ret;
66006}
66007#else
66008__ai int8x8_t vzip2_s8(int8x8_t __p0, int8x8_t __p1) {
66009  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66010  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66011  int8x8_t __ret;
66012  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66013  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66014  return __ret;
66015}
66016#endif
66017
66018#ifdef __LITTLE_ENDIAN__
66019__ai float32x2_t vzip2_f32(float32x2_t __p0, float32x2_t __p1) {
66020  float32x2_t __ret;
66021  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66022  return __ret;
66023}
66024#else
66025__ai float32x2_t vzip2_f32(float32x2_t __p0, float32x2_t __p1) {
66026  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66027  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66028  float32x2_t __ret;
66029  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66030  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66031  return __ret;
66032}
66033#endif
66034
66035#ifdef __LITTLE_ENDIAN__
66036__ai int32x2_t vzip2_s32(int32x2_t __p0, int32x2_t __p1) {
66037  int32x2_t __ret;
66038  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66039  return __ret;
66040}
66041#else
66042__ai int32x2_t vzip2_s32(int32x2_t __p0, int32x2_t __p1) {
66043  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66044  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66045  int32x2_t __ret;
66046  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66047  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66048  return __ret;
66049}
66050#endif
66051
66052#ifdef __LITTLE_ENDIAN__
66053__ai int16x4_t vzip2_s16(int16x4_t __p0, int16x4_t __p1) {
66054  int16x4_t __ret;
66055  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66056  return __ret;
66057}
66058#else
66059__ai int16x4_t vzip2_s16(int16x4_t __p0, int16x4_t __p1) {
66060  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66061  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66062  int16x4_t __ret;
66063  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66064  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66065  return __ret;
66066}
66067#endif
66068
66069#endif
66070#ifdef __LITTLE_ENDIAN__
66071__ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
66072  uint8x16_t __ret;
66073  __ret = __p0 + vabdq_u8(__p1, __p2);
66074  return __ret;
66075}
66076#else
66077__ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
66078  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66079  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66080  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66081  uint8x16_t __ret;
66082  __ret = __rev0 + __noswap_vabdq_u8(__rev1, __rev2);
66083  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66084  return __ret;
66085}
66086#endif
66087
66088#ifdef __LITTLE_ENDIAN__
66089__ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
66090  uint32x4_t __ret;
66091  __ret = __p0 + vabdq_u32(__p1, __p2);
66092  return __ret;
66093}
66094#else
66095__ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
66096  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66097  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66098  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66099  uint32x4_t __ret;
66100  __ret = __rev0 + __noswap_vabdq_u32(__rev1, __rev2);
66101  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66102  return __ret;
66103}
66104#endif
66105
66106#ifdef __LITTLE_ENDIAN__
66107__ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
66108  uint16x8_t __ret;
66109  __ret = __p0 + vabdq_u16(__p1, __p2);
66110  return __ret;
66111}
66112#else
66113__ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
66114  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66115  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66116  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66117  uint16x8_t __ret;
66118  __ret = __rev0 + __noswap_vabdq_u16(__rev1, __rev2);
66119  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66120  return __ret;
66121}
66122#endif
66123
66124#ifdef __LITTLE_ENDIAN__
66125__ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
66126  int8x16_t __ret;
66127  __ret = __p0 + vabdq_s8(__p1, __p2);
66128  return __ret;
66129}
66130#else
66131__ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
66132  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66133  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66134  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66135  int8x16_t __ret;
66136  __ret = __rev0 + __noswap_vabdq_s8(__rev1, __rev2);
66137  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66138  return __ret;
66139}
66140#endif
66141
66142#ifdef __LITTLE_ENDIAN__
66143__ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
66144  int32x4_t __ret;
66145  __ret = __p0 + vabdq_s32(__p1, __p2);
66146  return __ret;
66147}
66148#else
66149__ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
66150  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66151  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66152  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66153  int32x4_t __ret;
66154  __ret = __rev0 + __noswap_vabdq_s32(__rev1, __rev2);
66155  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66156  return __ret;
66157}
66158#endif
66159
66160#ifdef __LITTLE_ENDIAN__
66161__ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
66162  int16x8_t __ret;
66163  __ret = __p0 + vabdq_s16(__p1, __p2);
66164  return __ret;
66165}
66166#else
66167__ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
66168  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66169  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66170  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66171  int16x8_t __ret;
66172  __ret = __rev0 + __noswap_vabdq_s16(__rev1, __rev2);
66173  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66174  return __ret;
66175}
66176#endif
66177
66178#ifdef __LITTLE_ENDIAN__
66179__ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66180  uint8x8_t __ret;
66181  __ret = __p0 + vabd_u8(__p1, __p2);
66182  return __ret;
66183}
66184#else
66185__ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66186  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66187  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66188  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66189  uint8x8_t __ret;
66190  __ret = __rev0 + __noswap_vabd_u8(__rev1, __rev2);
66191  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66192  return __ret;
66193}
66194#endif
66195
66196#ifdef __LITTLE_ENDIAN__
66197__ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66198  uint32x2_t __ret;
66199  __ret = __p0 + vabd_u32(__p1, __p2);
66200  return __ret;
66201}
66202#else
66203__ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66204  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66205  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66206  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66207  uint32x2_t __ret;
66208  __ret = __rev0 + __noswap_vabd_u32(__rev1, __rev2);
66209  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66210  return __ret;
66211}
66212#endif
66213
66214#ifdef __LITTLE_ENDIAN__
66215__ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66216  uint16x4_t __ret;
66217  __ret = __p0 + vabd_u16(__p1, __p2);
66218  return __ret;
66219}
66220#else
66221__ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66222  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66223  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66224  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66225  uint16x4_t __ret;
66226  __ret = __rev0 + __noswap_vabd_u16(__rev1, __rev2);
66227  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66228  return __ret;
66229}
66230#endif
66231
66232#ifdef __LITTLE_ENDIAN__
66233__ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
66234  int8x8_t __ret;
66235  __ret = __p0 + vabd_s8(__p1, __p2);
66236  return __ret;
66237}
66238#else
66239__ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
66240  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66241  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66242  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66243  int8x8_t __ret;
66244  __ret = __rev0 + __noswap_vabd_s8(__rev1, __rev2);
66245  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66246  return __ret;
66247}
66248#endif
66249
66250#ifdef __LITTLE_ENDIAN__
66251__ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
66252  int32x2_t __ret;
66253  __ret = __p0 + vabd_s32(__p1, __p2);
66254  return __ret;
66255}
66256#else
66257__ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
66258  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66259  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66260  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66261  int32x2_t __ret;
66262  __ret = __rev0 + __noswap_vabd_s32(__rev1, __rev2);
66263  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66264  return __ret;
66265}
66266#endif
66267
66268#ifdef __LITTLE_ENDIAN__
66269__ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
66270  int16x4_t __ret;
66271  __ret = __p0 + vabd_s16(__p1, __p2);
66272  return __ret;
66273}
66274#else
66275__ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
66276  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66277  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66278  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66279  int16x4_t __ret;
66280  __ret = __rev0 + __noswap_vabd_s16(__rev1, __rev2);
66281  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66282  return __ret;
66283}
66284#endif
66285
66286#ifdef __LITTLE_ENDIAN__
66287__ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66288  uint16x8_t __ret;
66289  __ret = (uint16x8_t)(vmovl_u8((uint8x8_t)(vabd_u8(__p0, __p1))));
66290  return __ret;
66291}
66292#else
66293__ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66294  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66295  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66296  uint16x8_t __ret;
66297  __ret = (uint16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_u8(__rev0, __rev1))));
66298  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66299  return __ret;
66300}
66301__ai uint16x8_t __noswap_vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66302  uint16x8_t __ret;
66303  __ret = (uint16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_u8(__p0, __p1))));
66304  return __ret;
66305}
66306#endif
66307
66308#ifdef __LITTLE_ENDIAN__
66309__ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66310  uint64x2_t __ret;
66311  __ret = (uint64x2_t)(vmovl_u32((uint32x2_t)(vabd_u32(__p0, __p1))));
66312  return __ret;
66313}
66314#else
66315__ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66316  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66317  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66318  uint64x2_t __ret;
66319  __ret = (uint64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_u32(__rev0, __rev1))));
66320  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66321  return __ret;
66322}
66323__ai uint64x2_t __noswap_vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66324  uint64x2_t __ret;
66325  __ret = (uint64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_u32(__p0, __p1))));
66326  return __ret;
66327}
66328#endif
66329
66330#ifdef __LITTLE_ENDIAN__
66331__ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66332  uint32x4_t __ret;
66333  __ret = (uint32x4_t)(vmovl_u16((uint16x4_t)(vabd_u16(__p0, __p1))));
66334  return __ret;
66335}
66336#else
66337__ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66338  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66339  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66340  uint32x4_t __ret;
66341  __ret = (uint32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_u16(__rev0, __rev1))));
66342  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66343  return __ret;
66344}
66345__ai uint32x4_t __noswap_vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66346  uint32x4_t __ret;
66347  __ret = (uint32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_u16(__p0, __p1))));
66348  return __ret;
66349}
66350#endif
66351
66352#ifdef __LITTLE_ENDIAN__
66353__ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66354  int16x8_t __ret;
66355  __ret = (int16x8_t)(vmovl_u8((uint8x8_t)(vabd_s8(__p0, __p1))));
66356  return __ret;
66357}
66358#else
66359__ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66360  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66361  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66362  int16x8_t __ret;
66363  __ret = (int16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_s8(__rev0, __rev1))));
66364  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66365  return __ret;
66366}
66367__ai int16x8_t __noswap_vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66368  int16x8_t __ret;
66369  __ret = (int16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_s8(__p0, __p1))));
66370  return __ret;
66371}
66372#endif
66373
66374#ifdef __LITTLE_ENDIAN__
66375__ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66376  int64x2_t __ret;
66377  __ret = (int64x2_t)(vmovl_u32((uint32x2_t)(vabd_s32(__p0, __p1))));
66378  return __ret;
66379}
66380#else
66381__ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66382  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66383  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66384  int64x2_t __ret;
66385  __ret = (int64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_s32(__rev0, __rev1))));
66386  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66387  return __ret;
66388}
66389__ai int64x2_t __noswap_vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66390  int64x2_t __ret;
66391  __ret = (int64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_s32(__p0, __p1))));
66392  return __ret;
66393}
66394#endif
66395
66396#ifdef __LITTLE_ENDIAN__
66397__ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66398  int32x4_t __ret;
66399  __ret = (int32x4_t)(vmovl_u16((uint16x4_t)(vabd_s16(__p0, __p1))));
66400  return __ret;
66401}
66402#else
66403__ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66404  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66405  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66406  int32x4_t __ret;
66407  __ret = (int32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_s16(__rev0, __rev1))));
66408  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66409  return __ret;
66410}
66411__ai int32x4_t __noswap_vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66412  int32x4_t __ret;
66413  __ret = (int32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_s16(__p0, __p1))));
66414  return __ret;
66415}
66416#endif
66417
66418#ifdef __LITTLE_ENDIAN__
66419__ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66420  uint16x8_t __ret;
66421  __ret = vmovl_u8(__p0) + vmovl_u8(__p1);
66422  return __ret;
66423}
66424#else
66425__ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66426  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66427  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66428  uint16x8_t __ret;
66429  __ret = __noswap_vmovl_u8(__rev0) + __noswap_vmovl_u8(__rev1);
66430  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66431  return __ret;
66432}
66433#endif
66434
66435#ifdef __LITTLE_ENDIAN__
66436__ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66437  uint64x2_t __ret;
66438  __ret = vmovl_u32(__p0) + vmovl_u32(__p1);
66439  return __ret;
66440}
66441#else
66442__ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66443  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66444  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66445  uint64x2_t __ret;
66446  __ret = __noswap_vmovl_u32(__rev0) + __noswap_vmovl_u32(__rev1);
66447  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66448  return __ret;
66449}
66450#endif
66451
66452#ifdef __LITTLE_ENDIAN__
66453__ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66454  uint32x4_t __ret;
66455  __ret = vmovl_u16(__p0) + vmovl_u16(__p1);
66456  return __ret;
66457}
66458#else
66459__ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66460  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66461  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66462  uint32x4_t __ret;
66463  __ret = __noswap_vmovl_u16(__rev0) + __noswap_vmovl_u16(__rev1);
66464  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66465  return __ret;
66466}
66467#endif
66468
66469#ifdef __LITTLE_ENDIAN__
66470__ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1) {
66471  int16x8_t __ret;
66472  __ret = vmovl_s8(__p0) + vmovl_s8(__p1);
66473  return __ret;
66474}
66475#else
66476__ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1) {
66477  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66478  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66479  int16x8_t __ret;
66480  __ret = __noswap_vmovl_s8(__rev0) + __noswap_vmovl_s8(__rev1);
66481  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66482  return __ret;
66483}
66484#endif
66485
66486#ifdef __LITTLE_ENDIAN__
66487__ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1) {
66488  int64x2_t __ret;
66489  __ret = vmovl_s32(__p0) + vmovl_s32(__p1);
66490  return __ret;
66491}
66492#else
66493__ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1) {
66494  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66495  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66496  int64x2_t __ret;
66497  __ret = __noswap_vmovl_s32(__rev0) + __noswap_vmovl_s32(__rev1);
66498  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66499  return __ret;
66500}
66501#endif
66502
66503#ifdef __LITTLE_ENDIAN__
66504__ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1) {
66505  int32x4_t __ret;
66506  __ret = vmovl_s16(__p0) + vmovl_s16(__p1);
66507  return __ret;
66508}
66509#else
66510__ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1) {
66511  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66512  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66513  int32x4_t __ret;
66514  __ret = __noswap_vmovl_s16(__rev0) + __noswap_vmovl_s16(__rev1);
66515  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66516  return __ret;
66517}
66518#endif
66519
66520#ifdef __LITTLE_ENDIAN__
66521__ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1) {
66522  uint16x8_t __ret;
66523  __ret = __p0 + vmovl_u8(__p1);
66524  return __ret;
66525}
66526#else
66527__ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1) {
66528  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66529  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66530  uint16x8_t __ret;
66531  __ret = __rev0 + __noswap_vmovl_u8(__rev1);
66532  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66533  return __ret;
66534}
66535#endif
66536
66537#ifdef __LITTLE_ENDIAN__
66538__ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1) {
66539  uint64x2_t __ret;
66540  __ret = __p0 + vmovl_u32(__p1);
66541  return __ret;
66542}
66543#else
66544__ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1) {
66545  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66546  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66547  uint64x2_t __ret;
66548  __ret = __rev0 + __noswap_vmovl_u32(__rev1);
66549  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66550  return __ret;
66551}
66552#endif
66553
66554#ifdef __LITTLE_ENDIAN__
66555__ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1) {
66556  uint32x4_t __ret;
66557  __ret = __p0 + vmovl_u16(__p1);
66558  return __ret;
66559}
66560#else
66561__ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1) {
66562  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66563  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66564  uint32x4_t __ret;
66565  __ret = __rev0 + __noswap_vmovl_u16(__rev1);
66566  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66567  return __ret;
66568}
66569#endif
66570
66571#ifdef __LITTLE_ENDIAN__
66572__ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1) {
66573  int16x8_t __ret;
66574  __ret = __p0 + vmovl_s8(__p1);
66575  return __ret;
66576}
66577#else
66578__ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1) {
66579  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66580  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66581  int16x8_t __ret;
66582  __ret = __rev0 + __noswap_vmovl_s8(__rev1);
66583  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66584  return __ret;
66585}
66586#endif
66587
66588#ifdef __LITTLE_ENDIAN__
66589__ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1) {
66590  int64x2_t __ret;
66591  __ret = __p0 + vmovl_s32(__p1);
66592  return __ret;
66593}
66594#else
66595__ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1) {
66596  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66597  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66598  int64x2_t __ret;
66599  __ret = __rev0 + __noswap_vmovl_s32(__rev1);
66600  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66601  return __ret;
66602}
66603#endif
66604
66605#ifdef __LITTLE_ENDIAN__
66606__ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1) {
66607  int32x4_t __ret;
66608  __ret = __p0 + vmovl_s16(__p1);
66609  return __ret;
66610}
66611#else
66612__ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1) {
66613  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66614  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66615  int32x4_t __ret;
66616  __ret = __rev0 + __noswap_vmovl_s16(__rev1);
66617  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66618  return __ret;
66619}
66620#endif
66621
66622#ifdef __LITTLE_ENDIAN__
66623__ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66624  uint16x8_t __ret;
66625  __ret = __p0 + vmull_u8(__p1, __p2);
66626  return __ret;
66627}
66628#else
66629__ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66630  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66631  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66632  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66633  uint16x8_t __ret;
66634  __ret = __rev0 + __noswap_vmull_u8(__rev1, __rev2);
66635  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66636  return __ret;
66637}
66638__ai uint16x8_t __noswap_vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66639  uint16x8_t __ret;
66640  __ret = __p0 + __noswap_vmull_u8(__p1, __p2);
66641  return __ret;
66642}
66643#endif
66644
66645#ifdef __LITTLE_ENDIAN__
66646__ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66647  uint64x2_t __ret;
66648  __ret = __p0 + vmull_u32(__p1, __p2);
66649  return __ret;
66650}
66651#else
66652__ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66653  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66654  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66655  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66656  uint64x2_t __ret;
66657  __ret = __rev0 + __noswap_vmull_u32(__rev1, __rev2);
66658  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66659  return __ret;
66660}
66661__ai uint64x2_t __noswap_vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66662  uint64x2_t __ret;
66663  __ret = __p0 + __noswap_vmull_u32(__p1, __p2);
66664  return __ret;
66665}
66666#endif
66667
66668#ifdef __LITTLE_ENDIAN__
66669__ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66670  uint32x4_t __ret;
66671  __ret = __p0 + vmull_u16(__p1, __p2);
66672  return __ret;
66673}
66674#else
66675__ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66676  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66677  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66678  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66679  uint32x4_t __ret;
66680  __ret = __rev0 + __noswap_vmull_u16(__rev1, __rev2);
66681  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66682  return __ret;
66683}
66684__ai uint32x4_t __noswap_vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66685  uint32x4_t __ret;
66686  __ret = __p0 + __noswap_vmull_u16(__p1, __p2);
66687  return __ret;
66688}
66689#endif
66690
66691#ifdef __LITTLE_ENDIAN__
66692__ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
66693  int16x8_t __ret;
66694  __ret = __p0 + vmull_s8(__p1, __p2);
66695  return __ret;
66696}
66697#else
66698__ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
66699  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66700  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66701  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66702  int16x8_t __ret;
66703  __ret = __rev0 + __noswap_vmull_s8(__rev1, __rev2);
66704  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66705  return __ret;
66706}
66707__ai int16x8_t __noswap_vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
66708  int16x8_t __ret;
66709  __ret = __p0 + __noswap_vmull_s8(__p1, __p2);
66710  return __ret;
66711}
66712#endif
66713
66714#ifdef __LITTLE_ENDIAN__
66715__ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
66716  int64x2_t __ret;
66717  __ret = __p0 + vmull_s32(__p1, __p2);
66718  return __ret;
66719}
66720#else
66721__ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
66722  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66723  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66724  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66725  int64x2_t __ret;
66726  __ret = __rev0 + __noswap_vmull_s32(__rev1, __rev2);
66727  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66728  return __ret;
66729}
66730__ai int64x2_t __noswap_vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
66731  int64x2_t __ret;
66732  __ret = __p0 + __noswap_vmull_s32(__p1, __p2);
66733  return __ret;
66734}
66735#endif
66736
66737#ifdef __LITTLE_ENDIAN__
66738__ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
66739  int32x4_t __ret;
66740  __ret = __p0 + vmull_s16(__p1, __p2);
66741  return __ret;
66742}
66743#else
66744__ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
66745  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66746  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66747  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66748  int32x4_t __ret;
66749  __ret = __rev0 + __noswap_vmull_s16(__rev1, __rev2);
66750  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66751  return __ret;
66752}
66753__ai int32x4_t __noswap_vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
66754  int32x4_t __ret;
66755  __ret = __p0 + __noswap_vmull_s16(__p1, __p2);
66756  return __ret;
66757}
66758#endif
66759
66760#ifdef __LITTLE_ENDIAN__
66761#define vmlal_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
66762  uint64x2_t __s0 = __p0; \
66763  uint32x2_t __s1 = __p1; \
66764  uint32x2_t __s2 = __p2; \
66765  uint64x2_t __ret; \
66766  __ret = __s0 + vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
66767  __ret; \
66768})
66769#else
66770#define vmlal_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
66771  uint64x2_t __s0 = __p0; \
66772  uint32x2_t __s1 = __p1; \
66773  uint32x2_t __s2 = __p2; \
66774  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
66775  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
66776  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
66777  uint64x2_t __ret; \
66778  __ret = __rev0 + __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
66779  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
66780  __ret; \
66781})
66782#endif
66783
66784#ifdef __LITTLE_ENDIAN__
66785#define vmlal_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
66786  uint32x4_t __s0 = __p0; \
66787  uint16x4_t __s1 = __p1; \
66788  uint16x4_t __s2 = __p2; \
66789  uint32x4_t __ret; \
66790  __ret = __s0 + vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
66791  __ret; \
66792})
66793#else
66794#define vmlal_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
66795  uint32x4_t __s0 = __p0; \
66796  uint16x4_t __s1 = __p1; \
66797  uint16x4_t __s2 = __p2; \
66798  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
66799  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
66800  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
66801  uint32x4_t __ret; \
66802  __ret = __rev0 + __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
66803  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
66804  __ret; \
66805})
66806#endif
66807
66808#ifdef __LITTLE_ENDIAN__
66809#define vmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
66810  int64x2_t __s0 = __p0; \
66811  int32x2_t __s1 = __p1; \
66812  int32x2_t __s2 = __p2; \
66813  int64x2_t __ret; \
66814  __ret = __s0 + vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
66815  __ret; \
66816})
66817#else
66818#define vmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
66819  int64x2_t __s0 = __p0; \
66820  int32x2_t __s1 = __p1; \
66821  int32x2_t __s2 = __p2; \
66822  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
66823  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
66824  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
66825  int64x2_t __ret; \
66826  __ret = __rev0 + __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
66827  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
66828  __ret; \
66829})
66830#endif
66831
66832#ifdef __LITTLE_ENDIAN__
66833#define vmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
66834  int32x4_t __s0 = __p0; \
66835  int16x4_t __s1 = __p1; \
66836  int16x4_t __s2 = __p2; \
66837  int32x4_t __ret; \
66838  __ret = __s0 + vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
66839  __ret; \
66840})
66841#else
66842#define vmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
66843  int32x4_t __s0 = __p0; \
66844  int16x4_t __s1 = __p1; \
66845  int16x4_t __s2 = __p2; \
66846  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
66847  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
66848  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
66849  int32x4_t __ret; \
66850  __ret = __rev0 + __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
66851  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
66852  __ret; \
66853})
66854#endif
66855
66856#ifdef __LITTLE_ENDIAN__
66857__ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
66858  uint64x2_t __ret;
66859  __ret = __p0 + vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
66860  return __ret;
66861}
66862#else
66863__ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
66864  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66865  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66866  uint64x2_t __ret;
66867  __ret = __rev0 + __noswap_vmull_u32(__rev1, (uint32x2_t) {__p2, __p2});
66868  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66869  return __ret;
66870}
66871__ai uint64x2_t __noswap_vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
66872  uint64x2_t __ret;
66873  __ret = __p0 + __noswap_vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
66874  return __ret;
66875}
66876#endif
66877
66878#ifdef __LITTLE_ENDIAN__
66879__ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
66880  uint32x4_t __ret;
66881  __ret = __p0 + vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
66882  return __ret;
66883}
66884#else
66885__ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
66886  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66887  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66888  uint32x4_t __ret;
66889  __ret = __rev0 + __noswap_vmull_u16(__rev1, (uint16x4_t) {__p2, __p2, __p2, __p2});
66890  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66891  return __ret;
66892}
66893__ai uint32x4_t __noswap_vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
66894  uint32x4_t __ret;
66895  __ret = __p0 + __noswap_vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
66896  return __ret;
66897}
66898#endif
66899
66900#ifdef __LITTLE_ENDIAN__
66901__ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
66902  int64x2_t __ret;
66903  __ret = __p0 + vmull_s32(__p1, (int32x2_t) {__p2, __p2});
66904  return __ret;
66905}
66906#else
66907__ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
66908  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66909  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66910  int64x2_t __ret;
66911  __ret = __rev0 + __noswap_vmull_s32(__rev1, (int32x2_t) {__p2, __p2});
66912  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66913  return __ret;
66914}
66915__ai int64x2_t __noswap_vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
66916  int64x2_t __ret;
66917  __ret = __p0 + __noswap_vmull_s32(__p1, (int32x2_t) {__p2, __p2});
66918  return __ret;
66919}
66920#endif
66921
66922#ifdef __LITTLE_ENDIAN__
66923__ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
66924  int32x4_t __ret;
66925  __ret = __p0 + vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
66926  return __ret;
66927}
66928#else
66929__ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
66930  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66931  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66932  int32x4_t __ret;
66933  __ret = __rev0 + __noswap_vmull_s16(__rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
66934  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66935  return __ret;
66936}
66937__ai int32x4_t __noswap_vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
66938  int32x4_t __ret;
66939  __ret = __p0 + __noswap_vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
66940  return __ret;
66941}
66942#endif
66943
66944#ifdef __LITTLE_ENDIAN__
66945__ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66946  uint16x8_t __ret;
66947  __ret = __p0 - vmull_u8(__p1, __p2);
66948  return __ret;
66949}
66950#else
66951__ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66952  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66953  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66954  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66955  uint16x8_t __ret;
66956  __ret = __rev0 - __noswap_vmull_u8(__rev1, __rev2);
66957  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66958  return __ret;
66959}
66960__ai uint16x8_t __noswap_vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66961  uint16x8_t __ret;
66962  __ret = __p0 - __noswap_vmull_u8(__p1, __p2);
66963  return __ret;
66964}
66965#endif
66966
66967#ifdef __LITTLE_ENDIAN__
66968__ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66969  uint64x2_t __ret;
66970  __ret = __p0 - vmull_u32(__p1, __p2);
66971  return __ret;
66972}
66973#else
66974__ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66975  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66976  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66977  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66978  uint64x2_t __ret;
66979  __ret = __rev0 - __noswap_vmull_u32(__rev1, __rev2);
66980  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66981  return __ret;
66982}
66983__ai uint64x2_t __noswap_vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66984  uint64x2_t __ret;
66985  __ret = __p0 - __noswap_vmull_u32(__p1, __p2);
66986  return __ret;
66987}
66988#endif
66989
66990#ifdef __LITTLE_ENDIAN__
66991__ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66992  uint32x4_t __ret;
66993  __ret = __p0 - vmull_u16(__p1, __p2);
66994  return __ret;
66995}
66996#else
66997__ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66998  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66999  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67000  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67001  uint32x4_t __ret;
67002  __ret = __rev0 - __noswap_vmull_u16(__rev1, __rev2);
67003  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67004  return __ret;
67005}
67006__ai uint32x4_t __noswap_vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67007  uint32x4_t __ret;
67008  __ret = __p0 - __noswap_vmull_u16(__p1, __p2);
67009  return __ret;
67010}
67011#endif
67012
67013#ifdef __LITTLE_ENDIAN__
67014__ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67015  int16x8_t __ret;
67016  __ret = __p0 - vmull_s8(__p1, __p2);
67017  return __ret;
67018}
67019#else
67020__ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67021  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67022  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67023  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67024  int16x8_t __ret;
67025  __ret = __rev0 - __noswap_vmull_s8(__rev1, __rev2);
67026  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67027  return __ret;
67028}
67029__ai int16x8_t __noswap_vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67030  int16x8_t __ret;
67031  __ret = __p0 - __noswap_vmull_s8(__p1, __p2);
67032  return __ret;
67033}
67034#endif
67035
67036#ifdef __LITTLE_ENDIAN__
67037__ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67038  int64x2_t __ret;
67039  __ret = __p0 - vmull_s32(__p1, __p2);
67040  return __ret;
67041}
67042#else
67043__ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67044  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67045  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67046  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67047  int64x2_t __ret;
67048  __ret = __rev0 - __noswap_vmull_s32(__rev1, __rev2);
67049  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67050  return __ret;
67051}
67052__ai int64x2_t __noswap_vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67053  int64x2_t __ret;
67054  __ret = __p0 - __noswap_vmull_s32(__p1, __p2);
67055  return __ret;
67056}
67057#endif
67058
67059#ifdef __LITTLE_ENDIAN__
67060__ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67061  int32x4_t __ret;
67062  __ret = __p0 - vmull_s16(__p1, __p2);
67063  return __ret;
67064}
67065#else
67066__ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67067  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67068  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67069  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67070  int32x4_t __ret;
67071  __ret = __rev0 - __noswap_vmull_s16(__rev1, __rev2);
67072  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67073  return __ret;
67074}
67075__ai int32x4_t __noswap_vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67076  int32x4_t __ret;
67077  __ret = __p0 - __noswap_vmull_s16(__p1, __p2);
67078  return __ret;
67079}
67080#endif
67081
67082#ifdef __LITTLE_ENDIAN__
67083#define vmlsl_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
67084  uint64x2_t __s0 = __p0; \
67085  uint32x2_t __s1 = __p1; \
67086  uint32x2_t __s2 = __p2; \
67087  uint64x2_t __ret; \
67088  __ret = __s0 - vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
67089  __ret; \
67090})
67091#else
67092#define vmlsl_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
67093  uint64x2_t __s0 = __p0; \
67094  uint32x2_t __s1 = __p1; \
67095  uint32x2_t __s2 = __p2; \
67096  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
67097  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
67098  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
67099  uint64x2_t __ret; \
67100  __ret = __rev0 - __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
67101  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
67102  __ret; \
67103})
67104#endif
67105
67106#ifdef __LITTLE_ENDIAN__
67107#define vmlsl_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
67108  uint32x4_t __s0 = __p0; \
67109  uint16x4_t __s1 = __p1; \
67110  uint16x4_t __s2 = __p2; \
67111  uint32x4_t __ret; \
67112  __ret = __s0 - vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
67113  __ret; \
67114})
67115#else
67116#define vmlsl_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
67117  uint32x4_t __s0 = __p0; \
67118  uint16x4_t __s1 = __p1; \
67119  uint16x4_t __s2 = __p2; \
67120  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
67121  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
67122  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
67123  uint32x4_t __ret; \
67124  __ret = __rev0 - __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
67125  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
67126  __ret; \
67127})
67128#endif
67129
67130#ifdef __LITTLE_ENDIAN__
67131#define vmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
67132  int64x2_t __s0 = __p0; \
67133  int32x2_t __s1 = __p1; \
67134  int32x2_t __s2 = __p2; \
67135  int64x2_t __ret; \
67136  __ret = __s0 - vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
67137  __ret; \
67138})
67139#else
67140#define vmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
67141  int64x2_t __s0 = __p0; \
67142  int32x2_t __s1 = __p1; \
67143  int32x2_t __s2 = __p2; \
67144  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
67145  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
67146  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
67147  int64x2_t __ret; \
67148  __ret = __rev0 - __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
67149  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
67150  __ret; \
67151})
67152#endif
67153
67154#ifdef __LITTLE_ENDIAN__
67155#define vmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
67156  int32x4_t __s0 = __p0; \
67157  int16x4_t __s1 = __p1; \
67158  int16x4_t __s2 = __p2; \
67159  int32x4_t __ret; \
67160  __ret = __s0 - vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
67161  __ret; \
67162})
67163#else
67164#define vmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
67165  int32x4_t __s0 = __p0; \
67166  int16x4_t __s1 = __p1; \
67167  int16x4_t __s2 = __p2; \
67168  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
67169  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
67170  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
67171  int32x4_t __ret; \
67172  __ret = __rev0 - __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
67173  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
67174  __ret; \
67175})
67176#endif
67177
67178#ifdef __LITTLE_ENDIAN__
67179__ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67180  uint64x2_t __ret;
67181  __ret = __p0 - vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67182  return __ret;
67183}
67184#else
67185__ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67186  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67187  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67188  uint64x2_t __ret;
67189  __ret = __rev0 - __noswap_vmull_u32(__rev1, (uint32x2_t) {__p2, __p2});
67190  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67191  return __ret;
67192}
67193__ai uint64x2_t __noswap_vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67194  uint64x2_t __ret;
67195  __ret = __p0 - __noswap_vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67196  return __ret;
67197}
67198#endif
67199
67200#ifdef __LITTLE_ENDIAN__
67201__ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67202  uint32x4_t __ret;
67203  __ret = __p0 - vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67204  return __ret;
67205}
67206#else
67207__ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67208  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67209  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67210  uint32x4_t __ret;
67211  __ret = __rev0 - __noswap_vmull_u16(__rev1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67212  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67213  return __ret;
67214}
67215__ai uint32x4_t __noswap_vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67216  uint32x4_t __ret;
67217  __ret = __p0 - __noswap_vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67218  return __ret;
67219}
67220#endif
67221
67222#ifdef __LITTLE_ENDIAN__
67223__ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67224  int64x2_t __ret;
67225  __ret = __p0 - vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67226  return __ret;
67227}
67228#else
67229__ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67230  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67231  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67232  int64x2_t __ret;
67233  __ret = __rev0 - __noswap_vmull_s32(__rev1, (int32x2_t) {__p2, __p2});
67234  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67235  return __ret;
67236}
67237__ai int64x2_t __noswap_vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67238  int64x2_t __ret;
67239  __ret = __p0 - __noswap_vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67240  return __ret;
67241}
67242#endif
67243
67244#ifdef __LITTLE_ENDIAN__
67245__ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67246  int32x4_t __ret;
67247  __ret = __p0 - vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67248  return __ret;
67249}
67250#else
67251__ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67252  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67253  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67254  int32x4_t __ret;
67255  __ret = __rev0 - __noswap_vmull_s16(__rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
67256  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67257  return __ret;
67258}
67259__ai int32x4_t __noswap_vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67260  int32x4_t __ret;
67261  __ret = __p0 - __noswap_vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67262  return __ret;
67263}
67264#endif
67265
67266#if defined(__aarch64__)
67267#ifdef __LITTLE_ENDIAN__
67268__ai uint16x8_t vabdl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
67269  uint16x8_t __ret;
67270  __ret = vabdl_u8(vget_high_u8(__p0), vget_high_u8(__p1));
67271  return __ret;
67272}
67273#else
67274__ai uint16x8_t vabdl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
67275  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67276  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67277  uint16x8_t __ret;
67278  __ret = __noswap_vabdl_u8(__noswap_vget_high_u8(__rev0), __noswap_vget_high_u8(__rev1));
67279  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67280  return __ret;
67281}
67282#endif
67283
67284#ifdef __LITTLE_ENDIAN__
67285__ai uint64x2_t vabdl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
67286  uint64x2_t __ret;
67287  __ret = vabdl_u32(vget_high_u32(__p0), vget_high_u32(__p1));
67288  return __ret;
67289}
67290#else
67291__ai uint64x2_t vabdl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
67292  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67293  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67294  uint64x2_t __ret;
67295  __ret = __noswap_vabdl_u32(__noswap_vget_high_u32(__rev0), __noswap_vget_high_u32(__rev1));
67296  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67297  return __ret;
67298}
67299#endif
67300
67301#ifdef __LITTLE_ENDIAN__
67302__ai uint32x4_t vabdl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
67303  uint32x4_t __ret;
67304  __ret = vabdl_u16(vget_high_u16(__p0), vget_high_u16(__p1));
67305  return __ret;
67306}
67307#else
67308__ai uint32x4_t vabdl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
67309  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67310  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67311  uint32x4_t __ret;
67312  __ret = __noswap_vabdl_u16(__noswap_vget_high_u16(__rev0), __noswap_vget_high_u16(__rev1));
67313  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67314  return __ret;
67315}
67316#endif
67317
67318#ifdef __LITTLE_ENDIAN__
67319__ai int16x8_t vabdl_high_s8(int8x16_t __p0, int8x16_t __p1) {
67320  int16x8_t __ret;
67321  __ret = vabdl_s8(vget_high_s8(__p0), vget_high_s8(__p1));
67322  return __ret;
67323}
67324#else
67325__ai int16x8_t vabdl_high_s8(int8x16_t __p0, int8x16_t __p1) {
67326  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67327  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67328  int16x8_t __ret;
67329  __ret = __noswap_vabdl_s8(__noswap_vget_high_s8(__rev0), __noswap_vget_high_s8(__rev1));
67330  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67331  return __ret;
67332}
67333#endif
67334
67335#ifdef __LITTLE_ENDIAN__
67336__ai int64x2_t vabdl_high_s32(int32x4_t __p0, int32x4_t __p1) {
67337  int64x2_t __ret;
67338  __ret = vabdl_s32(vget_high_s32(__p0), vget_high_s32(__p1));
67339  return __ret;
67340}
67341#else
67342__ai int64x2_t vabdl_high_s32(int32x4_t __p0, int32x4_t __p1) {
67343  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67344  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67345  int64x2_t __ret;
67346  __ret = __noswap_vabdl_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
67347  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67348  return __ret;
67349}
67350#endif
67351
67352#ifdef __LITTLE_ENDIAN__
67353__ai int32x4_t vabdl_high_s16(int16x8_t __p0, int16x8_t __p1) {
67354  int32x4_t __ret;
67355  __ret = vabdl_s16(vget_high_s16(__p0), vget_high_s16(__p1));
67356  return __ret;
67357}
67358#else
67359__ai int32x4_t vabdl_high_s16(int16x8_t __p0, int16x8_t __p1) {
67360  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67361  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67362  int32x4_t __ret;
67363  __ret = __noswap_vabdl_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
67364  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67365  return __ret;
67366}
67367#endif
67368
67369#ifdef __LITTLE_ENDIAN__
67370__ai uint16x8_t vaddl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
67371  uint16x8_t __ret;
67372  __ret = vmovl_high_u8(__p0) + vmovl_high_u8(__p1);
67373  return __ret;
67374}
67375#else
67376__ai uint16x8_t vaddl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
67377  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67378  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67379  uint16x8_t __ret;
67380  __ret = __noswap_vmovl_high_u8(__rev0) + __noswap_vmovl_high_u8(__rev1);
67381  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67382  return __ret;
67383}
67384#endif
67385
67386#ifdef __LITTLE_ENDIAN__
67387__ai uint64x2_t vaddl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
67388  uint64x2_t __ret;
67389  __ret = vmovl_high_u32(__p0) + vmovl_high_u32(__p1);
67390  return __ret;
67391}
67392#else
67393__ai uint64x2_t vaddl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
67394  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67395  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67396  uint64x2_t __ret;
67397  __ret = __noswap_vmovl_high_u32(__rev0) + __noswap_vmovl_high_u32(__rev1);
67398  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67399  return __ret;
67400}
67401#endif
67402
67403#ifdef __LITTLE_ENDIAN__
67404__ai uint32x4_t vaddl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
67405  uint32x4_t __ret;
67406  __ret = vmovl_high_u16(__p0) + vmovl_high_u16(__p1);
67407  return __ret;
67408}
67409#else
67410__ai uint32x4_t vaddl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
67411  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67412  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67413  uint32x4_t __ret;
67414  __ret = __noswap_vmovl_high_u16(__rev0) + __noswap_vmovl_high_u16(__rev1);
67415  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67416  return __ret;
67417}
67418#endif
67419
67420#ifdef __LITTLE_ENDIAN__
67421__ai int16x8_t vaddl_high_s8(int8x16_t __p0, int8x16_t __p1) {
67422  int16x8_t __ret;
67423  __ret = vmovl_high_s8(__p0) + vmovl_high_s8(__p1);
67424  return __ret;
67425}
67426#else
67427__ai int16x8_t vaddl_high_s8(int8x16_t __p0, int8x16_t __p1) {
67428  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67429  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67430  int16x8_t __ret;
67431  __ret = __noswap_vmovl_high_s8(__rev0) + __noswap_vmovl_high_s8(__rev1);
67432  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67433  return __ret;
67434}
67435#endif
67436
67437#ifdef __LITTLE_ENDIAN__
67438__ai int64x2_t vaddl_high_s32(int32x4_t __p0, int32x4_t __p1) {
67439  int64x2_t __ret;
67440  __ret = vmovl_high_s32(__p0) + vmovl_high_s32(__p1);
67441  return __ret;
67442}
67443#else
67444__ai int64x2_t vaddl_high_s32(int32x4_t __p0, int32x4_t __p1) {
67445  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67446  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67447  int64x2_t __ret;
67448  __ret = __noswap_vmovl_high_s32(__rev0) + __noswap_vmovl_high_s32(__rev1);
67449  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67450  return __ret;
67451}
67452#endif
67453
67454#ifdef __LITTLE_ENDIAN__
67455__ai int32x4_t vaddl_high_s16(int16x8_t __p0, int16x8_t __p1) {
67456  int32x4_t __ret;
67457  __ret = vmovl_high_s16(__p0) + vmovl_high_s16(__p1);
67458  return __ret;
67459}
67460#else
67461__ai int32x4_t vaddl_high_s16(int16x8_t __p0, int16x8_t __p1) {
67462  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67463  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67464  int32x4_t __ret;
67465  __ret = __noswap_vmovl_high_s16(__rev0) + __noswap_vmovl_high_s16(__rev1);
67466  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67467  return __ret;
67468}
67469#endif
67470
67471#ifdef __LITTLE_ENDIAN__
67472__ai uint16x8_t vaddw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
67473  uint16x8_t __ret;
67474  __ret = __p0 + vmovl_high_u8(__p1);
67475  return __ret;
67476}
67477#else
67478__ai uint16x8_t vaddw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
67479  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67480  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67481  uint16x8_t __ret;
67482  __ret = __rev0 + __noswap_vmovl_high_u8(__rev1);
67483  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67484  return __ret;
67485}
67486#endif
67487
67488#ifdef __LITTLE_ENDIAN__
67489__ai uint64x2_t vaddw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
67490  uint64x2_t __ret;
67491  __ret = __p0 + vmovl_high_u32(__p1);
67492  return __ret;
67493}
67494#else
67495__ai uint64x2_t vaddw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
67496  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67497  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67498  uint64x2_t __ret;
67499  __ret = __rev0 + __noswap_vmovl_high_u32(__rev1);
67500  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67501  return __ret;
67502}
67503#endif
67504
67505#ifdef __LITTLE_ENDIAN__
67506__ai uint32x4_t vaddw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
67507  uint32x4_t __ret;
67508  __ret = __p0 + vmovl_high_u16(__p1);
67509  return __ret;
67510}
67511#else
67512__ai uint32x4_t vaddw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
67513  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67514  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67515  uint32x4_t __ret;
67516  __ret = __rev0 + __noswap_vmovl_high_u16(__rev1);
67517  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67518  return __ret;
67519}
67520#endif
67521
67522#ifdef __LITTLE_ENDIAN__
67523__ai int16x8_t vaddw_high_s8(int16x8_t __p0, int8x16_t __p1) {
67524  int16x8_t __ret;
67525  __ret = __p0 + vmovl_high_s8(__p1);
67526  return __ret;
67527}
67528#else
67529__ai int16x8_t vaddw_high_s8(int16x8_t __p0, int8x16_t __p1) {
67530  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67531  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67532  int16x8_t __ret;
67533  __ret = __rev0 + __noswap_vmovl_high_s8(__rev1);
67534  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67535  return __ret;
67536}
67537#endif
67538
67539#ifdef __LITTLE_ENDIAN__
67540__ai int64x2_t vaddw_high_s32(int64x2_t __p0, int32x4_t __p1) {
67541  int64x2_t __ret;
67542  __ret = __p0 + vmovl_high_s32(__p1);
67543  return __ret;
67544}
67545#else
67546__ai int64x2_t vaddw_high_s32(int64x2_t __p0, int32x4_t __p1) {
67547  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67548  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67549  int64x2_t __ret;
67550  __ret = __rev0 + __noswap_vmovl_high_s32(__rev1);
67551  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67552  return __ret;
67553}
67554#endif
67555
67556#ifdef __LITTLE_ENDIAN__
67557__ai int32x4_t vaddw_high_s16(int32x4_t __p0, int16x8_t __p1) {
67558  int32x4_t __ret;
67559  __ret = __p0 + vmovl_high_s16(__p1);
67560  return __ret;
67561}
67562#else
67563__ai int32x4_t vaddw_high_s16(int32x4_t __p0, int16x8_t __p1) {
67564  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67565  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67566  int32x4_t __ret;
67567  __ret = __rev0 + __noswap_vmovl_high_s16(__rev1);
67568  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67569  return __ret;
67570}
67571#endif
67572
67573#ifdef __LITTLE_ENDIAN__
67574#define vcopyq_lane_p64(__p0_250, __p1_250, __p2_250, __p3_250) __extension__ ({ \
67575  poly64x2_t __s0_250 = __p0_250; \
67576  poly64x1_t __s2_250 = __p2_250; \
67577  poly64x2_t __ret_250; \
67578  __ret_250 = vsetq_lane_p64(vget_lane_p64(__s2_250, __p3_250), __s0_250, __p1_250); \
67579  __ret_250; \
67580})
67581#else
67582#define vcopyq_lane_p64(__p0_251, __p1_251, __p2_251, __p3_251) __extension__ ({ \
67583  poly64x2_t __s0_251 = __p0_251; \
67584  poly64x1_t __s2_251 = __p2_251; \
67585  poly64x2_t __rev0_251;  __rev0_251 = __builtin_shufflevector(__s0_251, __s0_251, 1, 0); \
67586  poly64x2_t __ret_251; \
67587  __ret_251 = __noswap_vsetq_lane_p64(__noswap_vget_lane_p64(__s2_251, __p3_251), __rev0_251, __p1_251); \
67588  __ret_251 = __builtin_shufflevector(__ret_251, __ret_251, 1, 0); \
67589  __ret_251; \
67590})
67591#endif
67592
67593#ifdef __LITTLE_ENDIAN__
67594#define vcopyq_lane_f64(__p0_252, __p1_252, __p2_252, __p3_252) __extension__ ({ \
67595  float64x2_t __s0_252 = __p0_252; \
67596  float64x1_t __s2_252 = __p2_252; \
67597  float64x2_t __ret_252; \
67598  __ret_252 = vsetq_lane_f64(vget_lane_f64(__s2_252, __p3_252), __s0_252, __p1_252); \
67599  __ret_252; \
67600})
67601#else
67602#define vcopyq_lane_f64(__p0_253, __p1_253, __p2_253, __p3_253) __extension__ ({ \
67603  float64x2_t __s0_253 = __p0_253; \
67604  float64x1_t __s2_253 = __p2_253; \
67605  float64x2_t __rev0_253;  __rev0_253 = __builtin_shufflevector(__s0_253, __s0_253, 1, 0); \
67606  float64x2_t __ret_253; \
67607  __ret_253 = __noswap_vsetq_lane_f64(__noswap_vget_lane_f64(__s2_253, __p3_253), __rev0_253, __p1_253); \
67608  __ret_253 = __builtin_shufflevector(__ret_253, __ret_253, 1, 0); \
67609  __ret_253; \
67610})
67611#endif
67612
67613#ifdef __LITTLE_ENDIAN__
67614#define vcopy_lane_p64(__p0_254, __p1_254, __p2_254, __p3_254) __extension__ ({ \
67615  poly64x1_t __s0_254 = __p0_254; \
67616  poly64x1_t __s2_254 = __p2_254; \
67617  poly64x1_t __ret_254; \
67618  __ret_254 = vset_lane_p64(vget_lane_p64(__s2_254, __p3_254), __s0_254, __p1_254); \
67619  __ret_254; \
67620})
67621#else
67622#define vcopy_lane_p64(__p0_255, __p1_255, __p2_255, __p3_255) __extension__ ({ \
67623  poly64x1_t __s0_255 = __p0_255; \
67624  poly64x1_t __s2_255 = __p2_255; \
67625  poly64x1_t __ret_255; \
67626  __ret_255 = __noswap_vset_lane_p64(__noswap_vget_lane_p64(__s2_255, __p3_255), __s0_255, __p1_255); \
67627  __ret_255; \
67628})
67629#endif
67630
67631#ifdef __LITTLE_ENDIAN__
67632#define vcopy_lane_f64(__p0_256, __p1_256, __p2_256, __p3_256) __extension__ ({ \
67633  float64x1_t __s0_256 = __p0_256; \
67634  float64x1_t __s2_256 = __p2_256; \
67635  float64x1_t __ret_256; \
67636  __ret_256 = vset_lane_f64(vget_lane_f64(__s2_256, __p3_256), __s0_256, __p1_256); \
67637  __ret_256; \
67638})
67639#else
67640#define vcopy_lane_f64(__p0_257, __p1_257, __p2_257, __p3_257) __extension__ ({ \
67641  float64x1_t __s0_257 = __p0_257; \
67642  float64x1_t __s2_257 = __p2_257; \
67643  float64x1_t __ret_257; \
67644  __ret_257 = __noswap_vset_lane_f64(__noswap_vget_lane_f64(__s2_257, __p3_257), __s0_257, __p1_257); \
67645  __ret_257; \
67646})
67647#endif
67648
67649#ifdef __LITTLE_ENDIAN__
67650#define vcopyq_laneq_p64(__p0_258, __p1_258, __p2_258, __p3_258) __extension__ ({ \
67651  poly64x2_t __s0_258 = __p0_258; \
67652  poly64x2_t __s2_258 = __p2_258; \
67653  poly64x2_t __ret_258; \
67654  __ret_258 = vsetq_lane_p64(vgetq_lane_p64(__s2_258, __p3_258), __s0_258, __p1_258); \
67655  __ret_258; \
67656})
67657#else
67658#define vcopyq_laneq_p64(__p0_259, __p1_259, __p2_259, __p3_259) __extension__ ({ \
67659  poly64x2_t __s0_259 = __p0_259; \
67660  poly64x2_t __s2_259 = __p2_259; \
67661  poly64x2_t __rev0_259;  __rev0_259 = __builtin_shufflevector(__s0_259, __s0_259, 1, 0); \
67662  poly64x2_t __rev2_259;  __rev2_259 = __builtin_shufflevector(__s2_259, __s2_259, 1, 0); \
67663  poly64x2_t __ret_259; \
67664  __ret_259 = __noswap_vsetq_lane_p64(__noswap_vgetq_lane_p64(__rev2_259, __p3_259), __rev0_259, __p1_259); \
67665  __ret_259 = __builtin_shufflevector(__ret_259, __ret_259, 1, 0); \
67666  __ret_259; \
67667})
67668#endif
67669
67670#ifdef __LITTLE_ENDIAN__
67671#define vcopyq_laneq_f64(__p0_260, __p1_260, __p2_260, __p3_260) __extension__ ({ \
67672  float64x2_t __s0_260 = __p0_260; \
67673  float64x2_t __s2_260 = __p2_260; \
67674  float64x2_t __ret_260; \
67675  __ret_260 = vsetq_lane_f64(vgetq_lane_f64(__s2_260, __p3_260), __s0_260, __p1_260); \
67676  __ret_260; \
67677})
67678#else
67679#define vcopyq_laneq_f64(__p0_261, __p1_261, __p2_261, __p3_261) __extension__ ({ \
67680  float64x2_t __s0_261 = __p0_261; \
67681  float64x2_t __s2_261 = __p2_261; \
67682  float64x2_t __rev0_261;  __rev0_261 = __builtin_shufflevector(__s0_261, __s0_261, 1, 0); \
67683  float64x2_t __rev2_261;  __rev2_261 = __builtin_shufflevector(__s2_261, __s2_261, 1, 0); \
67684  float64x2_t __ret_261; \
67685  __ret_261 = __noswap_vsetq_lane_f64(__noswap_vgetq_lane_f64(__rev2_261, __p3_261), __rev0_261, __p1_261); \
67686  __ret_261 = __builtin_shufflevector(__ret_261, __ret_261, 1, 0); \
67687  __ret_261; \
67688})
67689#endif
67690
67691#ifdef __LITTLE_ENDIAN__
67692#define vcopy_laneq_p64(__p0_262, __p1_262, __p2_262, __p3_262) __extension__ ({ \
67693  poly64x1_t __s0_262 = __p0_262; \
67694  poly64x2_t __s2_262 = __p2_262; \
67695  poly64x1_t __ret_262; \
67696  __ret_262 = vset_lane_p64(vgetq_lane_p64(__s2_262, __p3_262), __s0_262, __p1_262); \
67697  __ret_262; \
67698})
67699#else
67700#define vcopy_laneq_p64(__p0_263, __p1_263, __p2_263, __p3_263) __extension__ ({ \
67701  poly64x1_t __s0_263 = __p0_263; \
67702  poly64x2_t __s2_263 = __p2_263; \
67703  poly64x2_t __rev2_263;  __rev2_263 = __builtin_shufflevector(__s2_263, __s2_263, 1, 0); \
67704  poly64x1_t __ret_263; \
67705  __ret_263 = __noswap_vset_lane_p64(__noswap_vgetq_lane_p64(__rev2_263, __p3_263), __s0_263, __p1_263); \
67706  __ret_263; \
67707})
67708#endif
67709
67710#ifdef __LITTLE_ENDIAN__
67711#define vcopy_laneq_f64(__p0_264, __p1_264, __p2_264, __p3_264) __extension__ ({ \
67712  float64x1_t __s0_264 = __p0_264; \
67713  float64x2_t __s2_264 = __p2_264; \
67714  float64x1_t __ret_264; \
67715  __ret_264 = vset_lane_f64(vgetq_lane_f64(__s2_264, __p3_264), __s0_264, __p1_264); \
67716  __ret_264; \
67717})
67718#else
67719#define vcopy_laneq_f64(__p0_265, __p1_265, __p2_265, __p3_265) __extension__ ({ \
67720  float64x1_t __s0_265 = __p0_265; \
67721  float64x2_t __s2_265 = __p2_265; \
67722  float64x2_t __rev2_265;  __rev2_265 = __builtin_shufflevector(__s2_265, __s2_265, 1, 0); \
67723  float64x1_t __ret_265; \
67724  __ret_265 = __noswap_vset_lane_f64(__noswap_vgetq_lane_f64(__rev2_265, __p3_265), __s0_265, __p1_265); \
67725  __ret_265; \
67726})
67727#endif
67728
67729#ifdef __LITTLE_ENDIAN__
67730__ai uint16x8_t vmlal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
67731  uint16x8_t __ret;
67732  __ret = vmlal_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
67733  return __ret;
67734}
67735#else
67736__ai uint16x8_t vmlal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
67737  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67738  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67739  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67740  uint16x8_t __ret;
67741  __ret = __noswap_vmlal_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
67742  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67743  return __ret;
67744}
67745#endif
67746
67747#ifdef __LITTLE_ENDIAN__
67748__ai uint64x2_t vmlal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
67749  uint64x2_t __ret;
67750  __ret = vmlal_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
67751  return __ret;
67752}
67753#else
67754__ai uint64x2_t vmlal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
67755  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67756  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67757  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67758  uint64x2_t __ret;
67759  __ret = __noswap_vmlal_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
67760  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67761  return __ret;
67762}
67763#endif
67764
67765#ifdef __LITTLE_ENDIAN__
67766__ai uint32x4_t vmlal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
67767  uint32x4_t __ret;
67768  __ret = vmlal_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
67769  return __ret;
67770}
67771#else
67772__ai uint32x4_t vmlal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
67773  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67774  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67775  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67776  uint32x4_t __ret;
67777  __ret = __noswap_vmlal_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
67778  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67779  return __ret;
67780}
67781#endif
67782
67783#ifdef __LITTLE_ENDIAN__
67784__ai int16x8_t vmlal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
67785  int16x8_t __ret;
67786  __ret = vmlal_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
67787  return __ret;
67788}
67789#else
67790__ai int16x8_t vmlal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
67791  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67792  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67793  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67794  int16x8_t __ret;
67795  __ret = __noswap_vmlal_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
67796  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67797  return __ret;
67798}
67799#endif
67800
67801#ifdef __LITTLE_ENDIAN__
67802__ai int64x2_t vmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
67803  int64x2_t __ret;
67804  __ret = vmlal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
67805  return __ret;
67806}
67807#else
67808__ai int64x2_t vmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
67809  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67810  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67811  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67812  int64x2_t __ret;
67813  __ret = __noswap_vmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
67814  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67815  return __ret;
67816}
67817#endif
67818
67819#ifdef __LITTLE_ENDIAN__
67820__ai int32x4_t vmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
67821  int32x4_t __ret;
67822  __ret = vmlal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
67823  return __ret;
67824}
67825#else
67826__ai int32x4_t vmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
67827  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67828  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67829  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67830  int32x4_t __ret;
67831  __ret = __noswap_vmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
67832  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67833  return __ret;
67834}
67835#endif
67836
67837#ifdef __LITTLE_ENDIAN__
67838__ai uint64x2_t vmlal_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
67839  uint64x2_t __ret;
67840  __ret = vmlal_n_u32(__p0, vget_high_u32(__p1), __p2);
67841  return __ret;
67842}
67843#else
67844__ai uint64x2_t vmlal_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
67845  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67846  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67847  uint64x2_t __ret;
67848  __ret = __noswap_vmlal_n_u32(__rev0, __noswap_vget_high_u32(__rev1), __p2);
67849  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67850  return __ret;
67851}
67852#endif
67853
67854#ifdef __LITTLE_ENDIAN__
67855__ai uint32x4_t vmlal_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
67856  uint32x4_t __ret;
67857  __ret = vmlal_n_u16(__p0, vget_high_u16(__p1), __p2);
67858  return __ret;
67859}
67860#else
67861__ai uint32x4_t vmlal_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
67862  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67863  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67864  uint32x4_t __ret;
67865  __ret = __noswap_vmlal_n_u16(__rev0, __noswap_vget_high_u16(__rev1), __p2);
67866  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67867  return __ret;
67868}
67869#endif
67870
67871#ifdef __LITTLE_ENDIAN__
67872__ai int64x2_t vmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
67873  int64x2_t __ret;
67874  __ret = vmlal_n_s32(__p0, vget_high_s32(__p1), __p2);
67875  return __ret;
67876}
67877#else
67878__ai int64x2_t vmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
67879  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67880  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67881  int64x2_t __ret;
67882  __ret = __noswap_vmlal_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
67883  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67884  return __ret;
67885}
67886#endif
67887
67888#ifdef __LITTLE_ENDIAN__
67889__ai int32x4_t vmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
67890  int32x4_t __ret;
67891  __ret = vmlal_n_s16(__p0, vget_high_s16(__p1), __p2);
67892  return __ret;
67893}
67894#else
67895__ai int32x4_t vmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
67896  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67897  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67898  int32x4_t __ret;
67899  __ret = __noswap_vmlal_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
67900  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67901  return __ret;
67902}
67903#endif
67904
67905#ifdef __LITTLE_ENDIAN__
67906__ai uint16x8_t vmlsl_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
67907  uint16x8_t __ret;
67908  __ret = vmlsl_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
67909  return __ret;
67910}
67911#else
67912__ai uint16x8_t vmlsl_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
67913  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67914  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67915  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67916  uint16x8_t __ret;
67917  __ret = __noswap_vmlsl_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
67918  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67919  return __ret;
67920}
67921#endif
67922
67923#ifdef __LITTLE_ENDIAN__
67924__ai uint64x2_t vmlsl_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
67925  uint64x2_t __ret;
67926  __ret = vmlsl_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
67927  return __ret;
67928}
67929#else
67930__ai uint64x2_t vmlsl_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
67931  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67932  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67933  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67934  uint64x2_t __ret;
67935  __ret = __noswap_vmlsl_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
67936  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67937  return __ret;
67938}
67939#endif
67940
67941#ifdef __LITTLE_ENDIAN__
67942__ai uint32x4_t vmlsl_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
67943  uint32x4_t __ret;
67944  __ret = vmlsl_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
67945  return __ret;
67946}
67947#else
67948__ai uint32x4_t vmlsl_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
67949  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67950  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67951  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67952  uint32x4_t __ret;
67953  __ret = __noswap_vmlsl_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
67954  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67955  return __ret;
67956}
67957#endif
67958
67959#ifdef __LITTLE_ENDIAN__
67960__ai int16x8_t vmlsl_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
67961  int16x8_t __ret;
67962  __ret = vmlsl_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
67963  return __ret;
67964}
67965#else
67966__ai int16x8_t vmlsl_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
67967  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67968  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67969  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
67970  int16x8_t __ret;
67971  __ret = __noswap_vmlsl_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
67972  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67973  return __ret;
67974}
67975#endif
67976
67977#ifdef __LITTLE_ENDIAN__
67978__ai int64x2_t vmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
67979  int64x2_t __ret;
67980  __ret = vmlsl_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
67981  return __ret;
67982}
67983#else
67984__ai int64x2_t vmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
67985  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67986  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67987  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67988  int64x2_t __ret;
67989  __ret = __noswap_vmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
67990  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67991  return __ret;
67992}
67993#endif
67994
67995#ifdef __LITTLE_ENDIAN__
67996__ai int32x4_t vmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
67997  int32x4_t __ret;
67998  __ret = vmlsl_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
67999  return __ret;
68000}
68001#else
68002__ai int32x4_t vmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68003  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68004  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68005  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68006  int32x4_t __ret;
68007  __ret = __noswap_vmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
68008  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68009  return __ret;
68010}
68011#endif
68012
68013#ifdef __LITTLE_ENDIAN__
68014__ai uint64x2_t vmlsl_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68015  uint64x2_t __ret;
68016  __ret = vmlsl_n_u32(__p0, vget_high_u32(__p1), __p2);
68017  return __ret;
68018}
68019#else
68020__ai uint64x2_t vmlsl_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68021  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68022  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68023  uint64x2_t __ret;
68024  __ret = __noswap_vmlsl_n_u32(__rev0, __noswap_vget_high_u32(__rev1), __p2);
68025  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68026  return __ret;
68027}
68028#endif
68029
68030#ifdef __LITTLE_ENDIAN__
68031__ai uint32x4_t vmlsl_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68032  uint32x4_t __ret;
68033  __ret = vmlsl_n_u16(__p0, vget_high_u16(__p1), __p2);
68034  return __ret;
68035}
68036#else
68037__ai uint32x4_t vmlsl_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68038  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68039  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68040  uint32x4_t __ret;
68041  __ret = __noswap_vmlsl_n_u16(__rev0, __noswap_vget_high_u16(__rev1), __p2);
68042  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68043  return __ret;
68044}
68045#endif
68046
68047#ifdef __LITTLE_ENDIAN__
68048__ai int64x2_t vmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68049  int64x2_t __ret;
68050  __ret = vmlsl_n_s32(__p0, vget_high_s32(__p1), __p2);
68051  return __ret;
68052}
68053#else
68054__ai int64x2_t vmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68055  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68056  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68057  int64x2_t __ret;
68058  __ret = __noswap_vmlsl_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
68059  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68060  return __ret;
68061}
68062#endif
68063
68064#ifdef __LITTLE_ENDIAN__
68065__ai int32x4_t vmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68066  int32x4_t __ret;
68067  __ret = vmlsl_n_s16(__p0, vget_high_s16(__p1), __p2);
68068  return __ret;
68069}
68070#else
68071__ai int32x4_t vmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68072  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68073  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68074  int32x4_t __ret;
68075  __ret = __noswap_vmlsl_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
68076  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68077  return __ret;
68078}
68079#endif
68080
68081#ifdef __LITTLE_ENDIAN__
68082#define vmulx_lane_f64(__p0_266, __p1_266, __p2_266) __extension__ ({ \
68083  float64x1_t __s0_266 = __p0_266; \
68084  float64x1_t __s1_266 = __p1_266; \
68085  float64x1_t __ret_266; \
68086  float64_t __x_266 = vget_lane_f64(__s0_266, 0); \
68087  float64_t __y_266 = vget_lane_f64(__s1_266, __p2_266); \
68088  float64_t __z_266 = vmulxd_f64(__x_266, __y_266); \
68089  __ret_266 = vset_lane_f64(__z_266, __s0_266, __p2_266); \
68090  __ret_266; \
68091})
68092#else
68093#define vmulx_lane_f64(__p0_267, __p1_267, __p2_267) __extension__ ({ \
68094  float64x1_t __s0_267 = __p0_267; \
68095  float64x1_t __s1_267 = __p1_267; \
68096  float64x1_t __ret_267; \
68097  float64_t __x_267 = __noswap_vget_lane_f64(__s0_267, 0); \
68098  float64_t __y_267 = __noswap_vget_lane_f64(__s1_267, __p2_267); \
68099  float64_t __z_267 = __noswap_vmulxd_f64(__x_267, __y_267); \
68100  __ret_267 = __noswap_vset_lane_f64(__z_267, __s0_267, __p2_267); \
68101  __ret_267; \
68102})
68103#endif
68104
68105#ifdef __LITTLE_ENDIAN__
68106#define vmulx_laneq_f64(__p0_268, __p1_268, __p2_268) __extension__ ({ \
68107  float64x1_t __s0_268 = __p0_268; \
68108  float64x2_t __s1_268 = __p1_268; \
68109  float64x1_t __ret_268; \
68110  float64_t __x_268 = vget_lane_f64(__s0_268, 0); \
68111  float64_t __y_268 = vgetq_lane_f64(__s1_268, __p2_268); \
68112  float64_t __z_268 = vmulxd_f64(__x_268, __y_268); \
68113  __ret_268 = vset_lane_f64(__z_268, __s0_268, 0); \
68114  __ret_268; \
68115})
68116#else
68117#define vmulx_laneq_f64(__p0_269, __p1_269, __p2_269) __extension__ ({ \
68118  float64x1_t __s0_269 = __p0_269; \
68119  float64x2_t __s1_269 = __p1_269; \
68120  float64x2_t __rev1_269;  __rev1_269 = __builtin_shufflevector(__s1_269, __s1_269, 1, 0); \
68121  float64x1_t __ret_269; \
68122  float64_t __x_269 = __noswap_vget_lane_f64(__s0_269, 0); \
68123  float64_t __y_269 = __noswap_vgetq_lane_f64(__rev1_269, __p2_269); \
68124  float64_t __z_269 = __noswap_vmulxd_f64(__x_269, __y_269); \
68125  __ret_269 = __noswap_vset_lane_f64(__z_269, __s0_269, 0); \
68126  __ret_269; \
68127})
68128#endif
68129
68130#endif
68131#ifdef __LITTLE_ENDIAN__
68132__ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
68133  uint16x8_t __ret;
68134  __ret = __p0 + vabdl_u8(__p1, __p2);
68135  return __ret;
68136}
68137#else
68138__ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
68139  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68140  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68141  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68142  uint16x8_t __ret;
68143  __ret = __rev0 + __noswap_vabdl_u8(__rev1, __rev2);
68144  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68145  return __ret;
68146}
68147__ai uint16x8_t __noswap_vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
68148  uint16x8_t __ret;
68149  __ret = __p0 + __noswap_vabdl_u8(__p1, __p2);
68150  return __ret;
68151}
68152#endif
68153
68154#ifdef __LITTLE_ENDIAN__
68155__ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
68156  uint64x2_t __ret;
68157  __ret = __p0 + vabdl_u32(__p1, __p2);
68158  return __ret;
68159}
68160#else
68161__ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
68162  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68163  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
68164  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
68165  uint64x2_t __ret;
68166  __ret = __rev0 + __noswap_vabdl_u32(__rev1, __rev2);
68167  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68168  return __ret;
68169}
68170__ai uint64x2_t __noswap_vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
68171  uint64x2_t __ret;
68172  __ret = __p0 + __noswap_vabdl_u32(__p1, __p2);
68173  return __ret;
68174}
68175#endif
68176
68177#ifdef __LITTLE_ENDIAN__
68178__ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
68179  uint32x4_t __ret;
68180  __ret = __p0 + vabdl_u16(__p1, __p2);
68181  return __ret;
68182}
68183#else
68184__ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
68185  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68186  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68187  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68188  uint32x4_t __ret;
68189  __ret = __rev0 + __noswap_vabdl_u16(__rev1, __rev2);
68190  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68191  return __ret;
68192}
68193__ai uint32x4_t __noswap_vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
68194  uint32x4_t __ret;
68195  __ret = __p0 + __noswap_vabdl_u16(__p1, __p2);
68196  return __ret;
68197}
68198#endif
68199
68200#ifdef __LITTLE_ENDIAN__
68201__ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
68202  int16x8_t __ret;
68203  __ret = __p0 + vabdl_s8(__p1, __p2);
68204  return __ret;
68205}
68206#else
68207__ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
68208  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68209  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68210  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68211  int16x8_t __ret;
68212  __ret = __rev0 + __noswap_vabdl_s8(__rev1, __rev2);
68213  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68214  return __ret;
68215}
68216__ai int16x8_t __noswap_vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
68217  int16x8_t __ret;
68218  __ret = __p0 + __noswap_vabdl_s8(__p1, __p2);
68219  return __ret;
68220}
68221#endif
68222
68223#ifdef __LITTLE_ENDIAN__
68224__ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
68225  int64x2_t __ret;
68226  __ret = __p0 + vabdl_s32(__p1, __p2);
68227  return __ret;
68228}
68229#else
68230__ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
68231  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68232  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
68233  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
68234  int64x2_t __ret;
68235  __ret = __rev0 + __noswap_vabdl_s32(__rev1, __rev2);
68236  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68237  return __ret;
68238}
68239__ai int64x2_t __noswap_vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
68240  int64x2_t __ret;
68241  __ret = __p0 + __noswap_vabdl_s32(__p1, __p2);
68242  return __ret;
68243}
68244#endif
68245
68246#ifdef __LITTLE_ENDIAN__
68247__ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
68248  int32x4_t __ret;
68249  __ret = __p0 + vabdl_s16(__p1, __p2);
68250  return __ret;
68251}
68252#else
68253__ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
68254  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68255  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68256  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68257  int32x4_t __ret;
68258  __ret = __rev0 + __noswap_vabdl_s16(__rev1, __rev2);
68259  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68260  return __ret;
68261}
68262__ai int32x4_t __noswap_vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
68263  int32x4_t __ret;
68264  __ret = __p0 + __noswap_vabdl_s16(__p1, __p2);
68265  return __ret;
68266}
68267#endif
68268
68269#if defined(__aarch64__)
68270#ifdef __LITTLE_ENDIAN__
68271__ai uint16x8_t vabal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68272  uint16x8_t __ret;
68273  __ret = vabal_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
68274  return __ret;
68275}
68276#else
68277__ai uint16x8_t vabal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68278  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68279  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68280  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68281  uint16x8_t __ret;
68282  __ret = __noswap_vabal_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
68283  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68284  return __ret;
68285}
68286#endif
68287
68288#ifdef __LITTLE_ENDIAN__
68289__ai uint64x2_t vabal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68290  uint64x2_t __ret;
68291  __ret = vabal_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
68292  return __ret;
68293}
68294#else
68295__ai uint64x2_t vabal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68296  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68297  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68298  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68299  uint64x2_t __ret;
68300  __ret = __noswap_vabal_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
68301  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68302  return __ret;
68303}
68304#endif
68305
68306#ifdef __LITTLE_ENDIAN__
68307__ai uint32x4_t vabal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68308  uint32x4_t __ret;
68309  __ret = vabal_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
68310  return __ret;
68311}
68312#else
68313__ai uint32x4_t vabal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68314  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68315  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68316  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68317  uint32x4_t __ret;
68318  __ret = __noswap_vabal_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
68319  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68320  return __ret;
68321}
68322#endif
68323
68324#ifdef __LITTLE_ENDIAN__
68325__ai int16x8_t vabal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68326  int16x8_t __ret;
68327  __ret = vabal_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
68328  return __ret;
68329}
68330#else
68331__ai int16x8_t vabal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68332  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68333  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68334  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68335  int16x8_t __ret;
68336  __ret = __noswap_vabal_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
68337  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68338  return __ret;
68339}
68340#endif
68341
68342#ifdef __LITTLE_ENDIAN__
68343__ai int64x2_t vabal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68344  int64x2_t __ret;
68345  __ret = vabal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
68346  return __ret;
68347}
68348#else
68349__ai int64x2_t vabal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68350  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68351  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68352  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68353  int64x2_t __ret;
68354  __ret = __noswap_vabal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
68355  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68356  return __ret;
68357}
68358#endif
68359
68360#ifdef __LITTLE_ENDIAN__
68361__ai int32x4_t vabal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68362  int32x4_t __ret;
68363  __ret = vabal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
68364  return __ret;
68365}
68366#else
68367__ai int32x4_t vabal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68368  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68369  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68370  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68371  int32x4_t __ret;
68372  __ret = __noswap_vabal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
68373  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68374  return __ret;
68375}
68376#endif
68377
68378#endif
68379
68380#undef __ai
68381
68382#endif /* __ARM_NEON_H */
68383