1/*===---- arm_neon.h - ARM Neon intrinsics ---------------------------------===
2 *
3 * Permission is hereby granted, free of charge, to any person obtaining a copy
4 * of this software and associated documentation files (the "Software"), to deal
5 * in the Software without restriction, including without limitation the rights
6 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
7 * copies of the Software, and to permit persons to whom the Software is
8 * furnished to do so, subject to the following conditions:
9 *
10 * The above copyright notice and this permission notice shall be included in
11 * all copies or substantial portions of the Software.
12 *
13 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
14 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
15 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
16 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
17 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
18 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
19 * THE SOFTWARE.
20 *
21 *===-----------------------------------------------------------------------===
22 */
23
24#ifndef __ARM_NEON_H
25#define __ARM_NEON_H
26
27#if !defined(__ARM_NEON)
28#error "NEON support not enabled"
29#endif
30
31#include <stdint.h>
32
33typedef float float32_t;
34typedef __fp16 float16_t;
35#ifdef __aarch64__
36typedef double float64_t;
37#endif
38
39#ifdef __aarch64__
40typedef uint8_t poly8_t;
41typedef uint16_t poly16_t;
42typedef uint64_t poly64_t;
43typedef __uint128_t poly128_t;
44#else
45typedef int8_t poly8_t;
46typedef int16_t poly16_t;
47#endif
48typedef __attribute__((neon_vector_type(8))) int8_t int8x8_t;
49typedef __attribute__((neon_vector_type(16))) int8_t int8x16_t;
50typedef __attribute__((neon_vector_type(4))) int16_t int16x4_t;
51typedef __attribute__((neon_vector_type(8))) int16_t int16x8_t;
52typedef __attribute__((neon_vector_type(2))) int32_t int32x2_t;
53typedef __attribute__((neon_vector_type(4))) int32_t int32x4_t;
54typedef __attribute__((neon_vector_type(1))) int64_t int64x1_t;
55typedef __attribute__((neon_vector_type(2))) int64_t int64x2_t;
56typedef __attribute__((neon_vector_type(8))) uint8_t uint8x8_t;
57typedef __attribute__((neon_vector_type(16))) uint8_t uint8x16_t;
58typedef __attribute__((neon_vector_type(4))) uint16_t uint16x4_t;
59typedef __attribute__((neon_vector_type(8))) uint16_t uint16x8_t;
60typedef __attribute__((neon_vector_type(2))) uint32_t uint32x2_t;
61typedef __attribute__((neon_vector_type(4))) uint32_t uint32x4_t;
62typedef __attribute__((neon_vector_type(1))) uint64_t uint64x1_t;
63typedef __attribute__((neon_vector_type(2))) uint64_t uint64x2_t;
64typedef __attribute__((neon_vector_type(4))) float16_t float16x4_t;
65typedef __attribute__((neon_vector_type(8))) float16_t float16x8_t;
66typedef __attribute__((neon_vector_type(2))) float32_t float32x2_t;
67typedef __attribute__((neon_vector_type(4))) float32_t float32x4_t;
68#ifdef __aarch64__
69typedef __attribute__((neon_vector_type(1))) float64_t float64x1_t;
70typedef __attribute__((neon_vector_type(2))) float64_t float64x2_t;
71#endif
72typedef __attribute__((neon_polyvector_type(8))) poly8_t poly8x8_t;
73typedef __attribute__((neon_polyvector_type(16))) poly8_t poly8x16_t;
74typedef __attribute__((neon_polyvector_type(4))) poly16_t poly16x4_t;
75typedef __attribute__((neon_polyvector_type(8))) poly16_t poly16x8_t;
76#ifdef __aarch64__
77typedef __attribute__((neon_polyvector_type(1))) poly64_t poly64x1_t;
78typedef __attribute__((neon_polyvector_type(2))) poly64_t poly64x2_t;
79#endif
80
81typedef struct int8x8x2_t {
82  int8x8_t val[2];
83} int8x8x2_t;
84
85typedef struct int8x16x2_t {
86  int8x16_t val[2];
87} int8x16x2_t;
88
89typedef struct int16x4x2_t {
90  int16x4_t val[2];
91} int16x4x2_t;
92
93typedef struct int16x8x2_t {
94  int16x8_t val[2];
95} int16x8x2_t;
96
97typedef struct int32x2x2_t {
98  int32x2_t val[2];
99} int32x2x2_t;
100
101typedef struct int32x4x2_t {
102  int32x4_t val[2];
103} int32x4x2_t;
104
105typedef struct int64x1x2_t {
106  int64x1_t val[2];
107} int64x1x2_t;
108
109typedef struct int64x2x2_t {
110  int64x2_t val[2];
111} int64x2x2_t;
112
113typedef struct uint8x8x2_t {
114  uint8x8_t val[2];
115} uint8x8x2_t;
116
117typedef struct uint8x16x2_t {
118  uint8x16_t val[2];
119} uint8x16x2_t;
120
121typedef struct uint16x4x2_t {
122  uint16x4_t val[2];
123} uint16x4x2_t;
124
125typedef struct uint16x8x2_t {
126  uint16x8_t val[2];
127} uint16x8x2_t;
128
129typedef struct uint32x2x2_t {
130  uint32x2_t val[2];
131} uint32x2x2_t;
132
133typedef struct uint32x4x2_t {
134  uint32x4_t val[2];
135} uint32x4x2_t;
136
137typedef struct uint64x1x2_t {
138  uint64x1_t val[2];
139} uint64x1x2_t;
140
141typedef struct uint64x2x2_t {
142  uint64x2_t val[2];
143} uint64x2x2_t;
144
145typedef struct float16x4x2_t {
146  float16x4_t val[2];
147} float16x4x2_t;
148
149typedef struct float16x8x2_t {
150  float16x8_t val[2];
151} float16x8x2_t;
152
153typedef struct float32x2x2_t {
154  float32x2_t val[2];
155} float32x2x2_t;
156
157typedef struct float32x4x2_t {
158  float32x4_t val[2];
159} float32x4x2_t;
160
161#ifdef __aarch64__
162typedef struct float64x1x2_t {
163  float64x1_t val[2];
164} float64x1x2_t;
165
166typedef struct float64x2x2_t {
167  float64x2_t val[2];
168} float64x2x2_t;
169
170#endif
171typedef struct poly8x8x2_t {
172  poly8x8_t val[2];
173} poly8x8x2_t;
174
175typedef struct poly8x16x2_t {
176  poly8x16_t val[2];
177} poly8x16x2_t;
178
179typedef struct poly16x4x2_t {
180  poly16x4_t val[2];
181} poly16x4x2_t;
182
183typedef struct poly16x8x2_t {
184  poly16x8_t val[2];
185} poly16x8x2_t;
186
187#ifdef __aarch64__
188typedef struct poly64x1x2_t {
189  poly64x1_t val[2];
190} poly64x1x2_t;
191
192typedef struct poly64x2x2_t {
193  poly64x2_t val[2];
194} poly64x2x2_t;
195
196#endif
197typedef struct int8x8x3_t {
198  int8x8_t val[3];
199} int8x8x3_t;
200
201typedef struct int8x16x3_t {
202  int8x16_t val[3];
203} int8x16x3_t;
204
205typedef struct int16x4x3_t {
206  int16x4_t val[3];
207} int16x4x3_t;
208
209typedef struct int16x8x3_t {
210  int16x8_t val[3];
211} int16x8x3_t;
212
213typedef struct int32x2x3_t {
214  int32x2_t val[3];
215} int32x2x3_t;
216
217typedef struct int32x4x3_t {
218  int32x4_t val[3];
219} int32x4x3_t;
220
221typedef struct int64x1x3_t {
222  int64x1_t val[3];
223} int64x1x3_t;
224
225typedef struct int64x2x3_t {
226  int64x2_t val[3];
227} int64x2x3_t;
228
229typedef struct uint8x8x3_t {
230  uint8x8_t val[3];
231} uint8x8x3_t;
232
233typedef struct uint8x16x3_t {
234  uint8x16_t val[3];
235} uint8x16x3_t;
236
237typedef struct uint16x4x3_t {
238  uint16x4_t val[3];
239} uint16x4x3_t;
240
241typedef struct uint16x8x3_t {
242  uint16x8_t val[3];
243} uint16x8x3_t;
244
245typedef struct uint32x2x3_t {
246  uint32x2_t val[3];
247} uint32x2x3_t;
248
249typedef struct uint32x4x3_t {
250  uint32x4_t val[3];
251} uint32x4x3_t;
252
253typedef struct uint64x1x3_t {
254  uint64x1_t val[3];
255} uint64x1x3_t;
256
257typedef struct uint64x2x3_t {
258  uint64x2_t val[3];
259} uint64x2x3_t;
260
261typedef struct float16x4x3_t {
262  float16x4_t val[3];
263} float16x4x3_t;
264
265typedef struct float16x8x3_t {
266  float16x8_t val[3];
267} float16x8x3_t;
268
269typedef struct float32x2x3_t {
270  float32x2_t val[3];
271} float32x2x3_t;
272
273typedef struct float32x4x3_t {
274  float32x4_t val[3];
275} float32x4x3_t;
276
277#ifdef __aarch64__
278typedef struct float64x1x3_t {
279  float64x1_t val[3];
280} float64x1x3_t;
281
282typedef struct float64x2x3_t {
283  float64x2_t val[3];
284} float64x2x3_t;
285
286#endif
287typedef struct poly8x8x3_t {
288  poly8x8_t val[3];
289} poly8x8x3_t;
290
291typedef struct poly8x16x3_t {
292  poly8x16_t val[3];
293} poly8x16x3_t;
294
295typedef struct poly16x4x3_t {
296  poly16x4_t val[3];
297} poly16x4x3_t;
298
299typedef struct poly16x8x3_t {
300  poly16x8_t val[3];
301} poly16x8x3_t;
302
303#ifdef __aarch64__
304typedef struct poly64x1x3_t {
305  poly64x1_t val[3];
306} poly64x1x3_t;
307
308typedef struct poly64x2x3_t {
309  poly64x2_t val[3];
310} poly64x2x3_t;
311
312#endif
313typedef struct int8x8x4_t {
314  int8x8_t val[4];
315} int8x8x4_t;
316
317typedef struct int8x16x4_t {
318  int8x16_t val[4];
319} int8x16x4_t;
320
321typedef struct int16x4x4_t {
322  int16x4_t val[4];
323} int16x4x4_t;
324
325typedef struct int16x8x4_t {
326  int16x8_t val[4];
327} int16x8x4_t;
328
329typedef struct int32x2x4_t {
330  int32x2_t val[4];
331} int32x2x4_t;
332
333typedef struct int32x4x4_t {
334  int32x4_t val[4];
335} int32x4x4_t;
336
337typedef struct int64x1x4_t {
338  int64x1_t val[4];
339} int64x1x4_t;
340
341typedef struct int64x2x4_t {
342  int64x2_t val[4];
343} int64x2x4_t;
344
345typedef struct uint8x8x4_t {
346  uint8x8_t val[4];
347} uint8x8x4_t;
348
349typedef struct uint8x16x4_t {
350  uint8x16_t val[4];
351} uint8x16x4_t;
352
353typedef struct uint16x4x4_t {
354  uint16x4_t val[4];
355} uint16x4x4_t;
356
357typedef struct uint16x8x4_t {
358  uint16x8_t val[4];
359} uint16x8x4_t;
360
361typedef struct uint32x2x4_t {
362  uint32x2_t val[4];
363} uint32x2x4_t;
364
365typedef struct uint32x4x4_t {
366  uint32x4_t val[4];
367} uint32x4x4_t;
368
369typedef struct uint64x1x4_t {
370  uint64x1_t val[4];
371} uint64x1x4_t;
372
373typedef struct uint64x2x4_t {
374  uint64x2_t val[4];
375} uint64x2x4_t;
376
377typedef struct float16x4x4_t {
378  float16x4_t val[4];
379} float16x4x4_t;
380
381typedef struct float16x8x4_t {
382  float16x8_t val[4];
383} float16x8x4_t;
384
385typedef struct float32x2x4_t {
386  float32x2_t val[4];
387} float32x2x4_t;
388
389typedef struct float32x4x4_t {
390  float32x4_t val[4];
391} float32x4x4_t;
392
393#ifdef __aarch64__
394typedef struct float64x1x4_t {
395  float64x1_t val[4];
396} float64x1x4_t;
397
398typedef struct float64x2x4_t {
399  float64x2_t val[4];
400} float64x2x4_t;
401
402#endif
403typedef struct poly8x8x4_t {
404  poly8x8_t val[4];
405} poly8x8x4_t;
406
407typedef struct poly8x16x4_t {
408  poly8x16_t val[4];
409} poly8x16x4_t;
410
411typedef struct poly16x4x4_t {
412  poly16x4_t val[4];
413} poly16x4x4_t;
414
415typedef struct poly16x8x4_t {
416  poly16x8_t val[4];
417} poly16x8x4_t;
418
419#ifdef __aarch64__
420typedef struct poly64x1x4_t {
421  poly64x1_t val[4];
422} poly64x1x4_t;
423
424typedef struct poly64x2x4_t {
425  poly64x2_t val[4];
426} poly64x2x4_t;
427
428#endif
429
430#define __ai static inline __attribute__((__always_inline__, __nodebug__))
431
432#ifdef __LITTLE_ENDIAN__
433__ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
434  uint8x16_t __ret;
435  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
436  return __ret;
437}
438#else
439__ai uint8x16_t vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
440  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
441  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
442  uint8x16_t __ret;
443  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
444  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
445  return __ret;
446}
447__ai uint8x16_t __noswap_vabdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
448  uint8x16_t __ret;
449  __ret = (uint8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
450  return __ret;
451}
452#endif
453
454#ifdef __LITTLE_ENDIAN__
455__ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
456  uint32x4_t __ret;
457  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
458  return __ret;
459}
460#else
461__ai uint32x4_t vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
462  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
463  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
464  uint32x4_t __ret;
465  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
466  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
467  return __ret;
468}
469__ai uint32x4_t __noswap_vabdq_u32(uint32x4_t __p0, uint32x4_t __p1) {
470  uint32x4_t __ret;
471  __ret = (uint32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
472  return __ret;
473}
474#endif
475
476#ifdef __LITTLE_ENDIAN__
477__ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
478  uint16x8_t __ret;
479  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
480  return __ret;
481}
482#else
483__ai uint16x8_t vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
484  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
485  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
486  uint16x8_t __ret;
487  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
488  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
489  return __ret;
490}
491__ai uint16x8_t __noswap_vabdq_u16(uint16x8_t __p0, uint16x8_t __p1) {
492  uint16x8_t __ret;
493  __ret = (uint16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
494  return __ret;
495}
496#endif
497
498#ifdef __LITTLE_ENDIAN__
499__ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
500  int8x16_t __ret;
501  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
502  return __ret;
503}
504#else
505__ai int8x16_t vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
506  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
507  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
508  int8x16_t __ret;
509  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
510  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
511  return __ret;
512}
513__ai int8x16_t __noswap_vabdq_s8(int8x16_t __p0, int8x16_t __p1) {
514  int8x16_t __ret;
515  __ret = (int8x16_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
516  return __ret;
517}
518#endif
519
520#ifdef __LITTLE_ENDIAN__
521__ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1) {
522  float32x4_t __ret;
523  __ret = (float32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
524  return __ret;
525}
526#else
527__ai float32x4_t vabdq_f32(float32x4_t __p0, float32x4_t __p1) {
528  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
529  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
530  float32x4_t __ret;
531  __ret = (float32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
532  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
533  return __ret;
534}
535#endif
536
537#ifdef __LITTLE_ENDIAN__
538__ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
539  int32x4_t __ret;
540  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
541  return __ret;
542}
543#else
544__ai int32x4_t vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
545  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
546  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
547  int32x4_t __ret;
548  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
549  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
550  return __ret;
551}
552__ai int32x4_t __noswap_vabdq_s32(int32x4_t __p0, int32x4_t __p1) {
553  int32x4_t __ret;
554  __ret = (int32x4_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
555  return __ret;
556}
557#endif
558
559#ifdef __LITTLE_ENDIAN__
560__ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
561  int16x8_t __ret;
562  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
563  return __ret;
564}
565#else
566__ai int16x8_t vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
567  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
568  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
569  int16x8_t __ret;
570  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
571  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
572  return __ret;
573}
574__ai int16x8_t __noswap_vabdq_s16(int16x8_t __p0, int16x8_t __p1) {
575  int16x8_t __ret;
576  __ret = (int16x8_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
577  return __ret;
578}
579#endif
580
581#ifdef __LITTLE_ENDIAN__
582__ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
583  uint8x8_t __ret;
584  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
585  return __ret;
586}
587#else
588__ai uint8x8_t vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
589  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
590  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
591  uint8x8_t __ret;
592  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
593  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
594  return __ret;
595}
596__ai uint8x8_t __noswap_vabd_u8(uint8x8_t __p0, uint8x8_t __p1) {
597  uint8x8_t __ret;
598  __ret = (uint8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
599  return __ret;
600}
601#endif
602
603#ifdef __LITTLE_ENDIAN__
604__ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
605  uint32x2_t __ret;
606  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
607  return __ret;
608}
609#else
610__ai uint32x2_t vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
611  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
612  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
613  uint32x2_t __ret;
614  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
615  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
616  return __ret;
617}
618__ai uint32x2_t __noswap_vabd_u32(uint32x2_t __p0, uint32x2_t __p1) {
619  uint32x2_t __ret;
620  __ret = (uint32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
621  return __ret;
622}
623#endif
624
625#ifdef __LITTLE_ENDIAN__
626__ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
627  uint16x4_t __ret;
628  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
629  return __ret;
630}
631#else
632__ai uint16x4_t vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
633  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
634  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
635  uint16x4_t __ret;
636  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
637  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
638  return __ret;
639}
640__ai uint16x4_t __noswap_vabd_u16(uint16x4_t __p0, uint16x4_t __p1) {
641  uint16x4_t __ret;
642  __ret = (uint16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
643  return __ret;
644}
645#endif
646
647#ifdef __LITTLE_ENDIAN__
648__ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1) {
649  int8x8_t __ret;
650  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
651  return __ret;
652}
653#else
654__ai int8x8_t vabd_s8(int8x8_t __p0, int8x8_t __p1) {
655  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
656  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
657  int8x8_t __ret;
658  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
659  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
660  return __ret;
661}
662__ai int8x8_t __noswap_vabd_s8(int8x8_t __p0, int8x8_t __p1) {
663  int8x8_t __ret;
664  __ret = (int8x8_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
665  return __ret;
666}
667#endif
668
669#ifdef __LITTLE_ENDIAN__
670__ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1) {
671  float32x2_t __ret;
672  __ret = (float32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
673  return __ret;
674}
675#else
676__ai float32x2_t vabd_f32(float32x2_t __p0, float32x2_t __p1) {
677  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
678  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
679  float32x2_t __ret;
680  __ret = (float32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
681  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
682  return __ret;
683}
684#endif
685
686#ifdef __LITTLE_ENDIAN__
687__ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1) {
688  int32x2_t __ret;
689  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
690  return __ret;
691}
692#else
693__ai int32x2_t vabd_s32(int32x2_t __p0, int32x2_t __p1) {
694  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
695  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
696  int32x2_t __ret;
697  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
698  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
699  return __ret;
700}
701__ai int32x2_t __noswap_vabd_s32(int32x2_t __p0, int32x2_t __p1) {
702  int32x2_t __ret;
703  __ret = (int32x2_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
704  return __ret;
705}
706#endif
707
708#ifdef __LITTLE_ENDIAN__
709__ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1) {
710  int16x4_t __ret;
711  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
712  return __ret;
713}
714#else
715__ai int16x4_t vabd_s16(int16x4_t __p0, int16x4_t __p1) {
716  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
717  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
718  int16x4_t __ret;
719  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
720  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
721  return __ret;
722}
723__ai int16x4_t __noswap_vabd_s16(int16x4_t __p0, int16x4_t __p1) {
724  int16x4_t __ret;
725  __ret = (int16x4_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
726  return __ret;
727}
728#endif
729
730#ifdef __LITTLE_ENDIAN__
731__ai int8x16_t vabsq_s8(int8x16_t __p0) {
732  int8x16_t __ret;
733  __ret = (int8x16_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 32);
734  return __ret;
735}
736#else
737__ai int8x16_t vabsq_s8(int8x16_t __p0) {
738  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
739  int8x16_t __ret;
740  __ret = (int8x16_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 32);
741  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
742  return __ret;
743}
744#endif
745
746#ifdef __LITTLE_ENDIAN__
747__ai float32x4_t vabsq_f32(float32x4_t __p0) {
748  float32x4_t __ret;
749  __ret = (float32x4_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 41);
750  return __ret;
751}
752#else
753__ai float32x4_t vabsq_f32(float32x4_t __p0) {
754  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
755  float32x4_t __ret;
756  __ret = (float32x4_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 41);
757  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
758  return __ret;
759}
760#endif
761
762#ifdef __LITTLE_ENDIAN__
763__ai int32x4_t vabsq_s32(int32x4_t __p0) {
764  int32x4_t __ret;
765  __ret = (int32x4_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 34);
766  return __ret;
767}
768#else
769__ai int32x4_t vabsq_s32(int32x4_t __p0) {
770  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
771  int32x4_t __ret;
772  __ret = (int32x4_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 34);
773  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
774  return __ret;
775}
776#endif
777
778#ifdef __LITTLE_ENDIAN__
779__ai int16x8_t vabsq_s16(int16x8_t __p0) {
780  int16x8_t __ret;
781  __ret = (int16x8_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 33);
782  return __ret;
783}
784#else
785__ai int16x8_t vabsq_s16(int16x8_t __p0) {
786  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
787  int16x8_t __ret;
788  __ret = (int16x8_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 33);
789  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
790  return __ret;
791}
792#endif
793
794#ifdef __LITTLE_ENDIAN__
795__ai int8x8_t vabs_s8(int8x8_t __p0) {
796  int8x8_t __ret;
797  __ret = (int8x8_t) __builtin_neon_vabs_v((int8x8_t)__p0, 0);
798  return __ret;
799}
800#else
801__ai int8x8_t vabs_s8(int8x8_t __p0) {
802  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
803  int8x8_t __ret;
804  __ret = (int8x8_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 0);
805  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
806  return __ret;
807}
808#endif
809
810#ifdef __LITTLE_ENDIAN__
811__ai float32x2_t vabs_f32(float32x2_t __p0) {
812  float32x2_t __ret;
813  __ret = (float32x2_t) __builtin_neon_vabs_v((int8x8_t)__p0, 9);
814  return __ret;
815}
816#else
817__ai float32x2_t vabs_f32(float32x2_t __p0) {
818  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
819  float32x2_t __ret;
820  __ret = (float32x2_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 9);
821  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
822  return __ret;
823}
824#endif
825
826#ifdef __LITTLE_ENDIAN__
827__ai int32x2_t vabs_s32(int32x2_t __p0) {
828  int32x2_t __ret;
829  __ret = (int32x2_t) __builtin_neon_vabs_v((int8x8_t)__p0, 2);
830  return __ret;
831}
832#else
833__ai int32x2_t vabs_s32(int32x2_t __p0) {
834  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
835  int32x2_t __ret;
836  __ret = (int32x2_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 2);
837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
838  return __ret;
839}
840#endif
841
842#ifdef __LITTLE_ENDIAN__
843__ai int16x4_t vabs_s16(int16x4_t __p0) {
844  int16x4_t __ret;
845  __ret = (int16x4_t) __builtin_neon_vabs_v((int8x8_t)__p0, 1);
846  return __ret;
847}
848#else
849__ai int16x4_t vabs_s16(int16x4_t __p0) {
850  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
851  int16x4_t __ret;
852  __ret = (int16x4_t) __builtin_neon_vabs_v((int8x8_t)__rev0, 1);
853  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
854  return __ret;
855}
856#endif
857
858#ifdef __LITTLE_ENDIAN__
859__ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
860  uint8x16_t __ret;
861  __ret = __p0 + __p1;
862  return __ret;
863}
864#else
865__ai uint8x16_t vaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
866  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
867  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
868  uint8x16_t __ret;
869  __ret = __rev0 + __rev1;
870  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
871  return __ret;
872}
873#endif
874
875#ifdef __LITTLE_ENDIAN__
876__ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
877  uint32x4_t __ret;
878  __ret = __p0 + __p1;
879  return __ret;
880}
881#else
882__ai uint32x4_t vaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
883  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
884  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
885  uint32x4_t __ret;
886  __ret = __rev0 + __rev1;
887  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
888  return __ret;
889}
890#endif
891
892#ifdef __LITTLE_ENDIAN__
893__ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
894  uint64x2_t __ret;
895  __ret = __p0 + __p1;
896  return __ret;
897}
898#else
899__ai uint64x2_t vaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
900  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
901  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
902  uint64x2_t __ret;
903  __ret = __rev0 + __rev1;
904  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
905  return __ret;
906}
907#endif
908
909#ifdef __LITTLE_ENDIAN__
910__ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
911  uint16x8_t __ret;
912  __ret = __p0 + __p1;
913  return __ret;
914}
915#else
916__ai uint16x8_t vaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
917  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
918  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
919  uint16x8_t __ret;
920  __ret = __rev0 + __rev1;
921  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
922  return __ret;
923}
924#endif
925
926#ifdef __LITTLE_ENDIAN__
927__ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1) {
928  int8x16_t __ret;
929  __ret = __p0 + __p1;
930  return __ret;
931}
932#else
933__ai int8x16_t vaddq_s8(int8x16_t __p0, int8x16_t __p1) {
934  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
935  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
936  int8x16_t __ret;
937  __ret = __rev0 + __rev1;
938  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
939  return __ret;
940}
941#endif
942
943#ifdef __LITTLE_ENDIAN__
944__ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1) {
945  float32x4_t __ret;
946  __ret = __p0 + __p1;
947  return __ret;
948}
949#else
950__ai float32x4_t vaddq_f32(float32x4_t __p0, float32x4_t __p1) {
951  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
952  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
953  float32x4_t __ret;
954  __ret = __rev0 + __rev1;
955  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
956  return __ret;
957}
958#endif
959
960#ifdef __LITTLE_ENDIAN__
961__ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1) {
962  int32x4_t __ret;
963  __ret = __p0 + __p1;
964  return __ret;
965}
966#else
967__ai int32x4_t vaddq_s32(int32x4_t __p0, int32x4_t __p1) {
968  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
969  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
970  int32x4_t __ret;
971  __ret = __rev0 + __rev1;
972  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
973  return __ret;
974}
975#endif
976
977#ifdef __LITTLE_ENDIAN__
978__ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1) {
979  int64x2_t __ret;
980  __ret = __p0 + __p1;
981  return __ret;
982}
983#else
984__ai int64x2_t vaddq_s64(int64x2_t __p0, int64x2_t __p1) {
985  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
986  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
987  int64x2_t __ret;
988  __ret = __rev0 + __rev1;
989  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
990  return __ret;
991}
992#endif
993
994#ifdef __LITTLE_ENDIAN__
995__ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1) {
996  int16x8_t __ret;
997  __ret = __p0 + __p1;
998  return __ret;
999}
1000#else
1001__ai int16x8_t vaddq_s16(int16x8_t __p0, int16x8_t __p1) {
1002  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1003  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1004  int16x8_t __ret;
1005  __ret = __rev0 + __rev1;
1006  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1007  return __ret;
1008}
1009#endif
1010
1011#ifdef __LITTLE_ENDIAN__
1012__ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1013  uint8x8_t __ret;
1014  __ret = __p0 + __p1;
1015  return __ret;
1016}
1017#else
1018__ai uint8x8_t vadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
1019  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1020  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1021  uint8x8_t __ret;
1022  __ret = __rev0 + __rev1;
1023  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1024  return __ret;
1025}
1026#endif
1027
1028#ifdef __LITTLE_ENDIAN__
1029__ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1030  uint32x2_t __ret;
1031  __ret = __p0 + __p1;
1032  return __ret;
1033}
1034#else
1035__ai uint32x2_t vadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
1036  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1037  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1038  uint32x2_t __ret;
1039  __ret = __rev0 + __rev1;
1040  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1041  return __ret;
1042}
1043#endif
1044
1045#ifdef __LITTLE_ENDIAN__
1046__ai uint64x1_t vadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
1047  uint64x1_t __ret;
1048  __ret = __p0 + __p1;
1049  return __ret;
1050}
1051#else
1052__ai uint64x1_t vadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
1053  uint64x1_t __ret;
1054  __ret = __p0 + __p1;
1055  return __ret;
1056}
1057#endif
1058
1059#ifdef __LITTLE_ENDIAN__
1060__ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
1061  uint16x4_t __ret;
1062  __ret = __p0 + __p1;
1063  return __ret;
1064}
1065#else
1066__ai uint16x4_t vadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
1067  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1068  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1069  uint16x4_t __ret;
1070  __ret = __rev0 + __rev1;
1071  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1072  return __ret;
1073}
1074#endif
1075
1076#ifdef __LITTLE_ENDIAN__
1077__ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1) {
1078  int8x8_t __ret;
1079  __ret = __p0 + __p1;
1080  return __ret;
1081}
1082#else
1083__ai int8x8_t vadd_s8(int8x8_t __p0, int8x8_t __p1) {
1084  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1085  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1086  int8x8_t __ret;
1087  __ret = __rev0 + __rev1;
1088  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1089  return __ret;
1090}
1091#endif
1092
1093#ifdef __LITTLE_ENDIAN__
1094__ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1) {
1095  float32x2_t __ret;
1096  __ret = __p0 + __p1;
1097  return __ret;
1098}
1099#else
1100__ai float32x2_t vadd_f32(float32x2_t __p0, float32x2_t __p1) {
1101  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1102  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1103  float32x2_t __ret;
1104  __ret = __rev0 + __rev1;
1105  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1106  return __ret;
1107}
1108#endif
1109
1110#ifdef __LITTLE_ENDIAN__
1111__ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1) {
1112  int32x2_t __ret;
1113  __ret = __p0 + __p1;
1114  return __ret;
1115}
1116#else
1117__ai int32x2_t vadd_s32(int32x2_t __p0, int32x2_t __p1) {
1118  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1119  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1120  int32x2_t __ret;
1121  __ret = __rev0 + __rev1;
1122  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1123  return __ret;
1124}
1125#endif
1126
1127#ifdef __LITTLE_ENDIAN__
1128__ai int64x1_t vadd_s64(int64x1_t __p0, int64x1_t __p1) {
1129  int64x1_t __ret;
1130  __ret = __p0 + __p1;
1131  return __ret;
1132}
1133#else
1134__ai int64x1_t vadd_s64(int64x1_t __p0, int64x1_t __p1) {
1135  int64x1_t __ret;
1136  __ret = __p0 + __p1;
1137  return __ret;
1138}
1139#endif
1140
1141#ifdef __LITTLE_ENDIAN__
1142__ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) {
1143  int16x4_t __ret;
1144  __ret = __p0 + __p1;
1145  return __ret;
1146}
1147#else
1148__ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) {
1149  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1150  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1151  int16x4_t __ret;
1152  __ret = __rev0 + __rev1;
1153  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1154  return __ret;
1155}
1156#endif
1157
1158#ifdef __LITTLE_ENDIAN__
1159__ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
1160  uint16x4_t __ret;
1161  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
1162  return __ret;
1163}
1164#else
1165__ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
1166  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1167  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1168  uint16x4_t __ret;
1169  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
1170  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1171  return __ret;
1172}
1173__ai uint16x4_t __noswap_vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
1174  uint16x4_t __ret;
1175  __ret = (uint16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
1176  return __ret;
1177}
1178#endif
1179
1180#ifdef __LITTLE_ENDIAN__
1181__ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
1182  uint32x2_t __ret;
1183  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
1184  return __ret;
1185}
1186#else
1187__ai uint32x2_t vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
1188  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1189  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1190  uint32x2_t __ret;
1191  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
1192  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1193  return __ret;
1194}
1195__ai uint32x2_t __noswap_vaddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
1196  uint32x2_t __ret;
1197  __ret = (uint32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
1198  return __ret;
1199}
1200#endif
1201
1202#ifdef __LITTLE_ENDIAN__
1203__ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
1204  uint8x8_t __ret;
1205  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
1206  return __ret;
1207}
1208#else
1209__ai uint8x8_t vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
1210  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1211  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1212  uint8x8_t __ret;
1213  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
1214  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1215  return __ret;
1216}
1217__ai uint8x8_t __noswap_vaddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
1218  uint8x8_t __ret;
1219  __ret = (uint8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
1220  return __ret;
1221}
1222#endif
1223
1224#ifdef __LITTLE_ENDIAN__
1225__ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
1226  int16x4_t __ret;
1227  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
1228  return __ret;
1229}
1230#else
1231__ai int16x4_t vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
1232  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1233  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1234  int16x4_t __ret;
1235  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
1236  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1237  return __ret;
1238}
1239__ai int16x4_t __noswap_vaddhn_s32(int32x4_t __p0, int32x4_t __p1) {
1240  int16x4_t __ret;
1241  __ret = (int16x4_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
1242  return __ret;
1243}
1244#endif
1245
1246#ifdef __LITTLE_ENDIAN__
1247__ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
1248  int32x2_t __ret;
1249  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
1250  return __ret;
1251}
1252#else
1253__ai int32x2_t vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
1254  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1255  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1256  int32x2_t __ret;
1257  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
1258  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1259  return __ret;
1260}
1261__ai int32x2_t __noswap_vaddhn_s64(int64x2_t __p0, int64x2_t __p1) {
1262  int32x2_t __ret;
1263  __ret = (int32x2_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
1264  return __ret;
1265}
1266#endif
1267
1268#ifdef __LITTLE_ENDIAN__
1269__ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
1270  int8x8_t __ret;
1271  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
1272  return __ret;
1273}
1274#else
1275__ai int8x8_t vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
1276  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1277  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1278  int8x8_t __ret;
1279  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
1280  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1281  return __ret;
1282}
1283__ai int8x8_t __noswap_vaddhn_s16(int16x8_t __p0, int16x8_t __p1) {
1284  int8x8_t __ret;
1285  __ret = (int8x8_t) __builtin_neon_vaddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
1286  return __ret;
1287}
1288#endif
1289
1290#ifdef __LITTLE_ENDIAN__
1291__ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1292  uint8x16_t __ret;
1293  __ret = __p0 & __p1;
1294  return __ret;
1295}
1296#else
1297__ai uint8x16_t vandq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1298  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1299  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1300  uint8x16_t __ret;
1301  __ret = __rev0 & __rev1;
1302  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1303  return __ret;
1304}
1305#endif
1306
1307#ifdef __LITTLE_ENDIAN__
1308__ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1309  uint32x4_t __ret;
1310  __ret = __p0 & __p1;
1311  return __ret;
1312}
1313#else
1314__ai uint32x4_t vandq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1315  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1316  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1317  uint32x4_t __ret;
1318  __ret = __rev0 & __rev1;
1319  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1320  return __ret;
1321}
1322#endif
1323
1324#ifdef __LITTLE_ENDIAN__
1325__ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1326  uint64x2_t __ret;
1327  __ret = __p0 & __p1;
1328  return __ret;
1329}
1330#else
1331__ai uint64x2_t vandq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1332  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1333  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1334  uint64x2_t __ret;
1335  __ret = __rev0 & __rev1;
1336  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1337  return __ret;
1338}
1339#endif
1340
1341#ifdef __LITTLE_ENDIAN__
1342__ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1343  uint16x8_t __ret;
1344  __ret = __p0 & __p1;
1345  return __ret;
1346}
1347#else
1348__ai uint16x8_t vandq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1349  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1350  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1351  uint16x8_t __ret;
1352  __ret = __rev0 & __rev1;
1353  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1354  return __ret;
1355}
1356#endif
1357
1358#ifdef __LITTLE_ENDIAN__
1359__ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1) {
1360  int8x16_t __ret;
1361  __ret = __p0 & __p1;
1362  return __ret;
1363}
1364#else
1365__ai int8x16_t vandq_s8(int8x16_t __p0, int8x16_t __p1) {
1366  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1367  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1368  int8x16_t __ret;
1369  __ret = __rev0 & __rev1;
1370  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1371  return __ret;
1372}
1373#endif
1374
1375#ifdef __LITTLE_ENDIAN__
1376__ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1) {
1377  int32x4_t __ret;
1378  __ret = __p0 & __p1;
1379  return __ret;
1380}
1381#else
1382__ai int32x4_t vandq_s32(int32x4_t __p0, int32x4_t __p1) {
1383  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1384  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1385  int32x4_t __ret;
1386  __ret = __rev0 & __rev1;
1387  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1388  return __ret;
1389}
1390#endif
1391
1392#ifdef __LITTLE_ENDIAN__
1393__ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1) {
1394  int64x2_t __ret;
1395  __ret = __p0 & __p1;
1396  return __ret;
1397}
1398#else
1399__ai int64x2_t vandq_s64(int64x2_t __p0, int64x2_t __p1) {
1400  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1401  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1402  int64x2_t __ret;
1403  __ret = __rev0 & __rev1;
1404  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1405  return __ret;
1406}
1407#endif
1408
1409#ifdef __LITTLE_ENDIAN__
1410__ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1) {
1411  int16x8_t __ret;
1412  __ret = __p0 & __p1;
1413  return __ret;
1414}
1415#else
1416__ai int16x8_t vandq_s16(int16x8_t __p0, int16x8_t __p1) {
1417  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1418  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1419  int16x8_t __ret;
1420  __ret = __rev0 & __rev1;
1421  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1422  return __ret;
1423}
1424#endif
1425
1426#ifdef __LITTLE_ENDIAN__
1427__ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1) {
1428  uint8x8_t __ret;
1429  __ret = __p0 & __p1;
1430  return __ret;
1431}
1432#else
1433__ai uint8x8_t vand_u8(uint8x8_t __p0, uint8x8_t __p1) {
1434  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1435  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1436  uint8x8_t __ret;
1437  __ret = __rev0 & __rev1;
1438  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1439  return __ret;
1440}
1441#endif
1442
1443#ifdef __LITTLE_ENDIAN__
1444__ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1) {
1445  uint32x2_t __ret;
1446  __ret = __p0 & __p1;
1447  return __ret;
1448}
1449#else
1450__ai uint32x2_t vand_u32(uint32x2_t __p0, uint32x2_t __p1) {
1451  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1452  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1453  uint32x2_t __ret;
1454  __ret = __rev0 & __rev1;
1455  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1456  return __ret;
1457}
1458#endif
1459
1460#ifdef __LITTLE_ENDIAN__
1461__ai uint64x1_t vand_u64(uint64x1_t __p0, uint64x1_t __p1) {
1462  uint64x1_t __ret;
1463  __ret = __p0 & __p1;
1464  return __ret;
1465}
1466#else
1467__ai uint64x1_t vand_u64(uint64x1_t __p0, uint64x1_t __p1) {
1468  uint64x1_t __ret;
1469  __ret = __p0 & __p1;
1470  return __ret;
1471}
1472#endif
1473
1474#ifdef __LITTLE_ENDIAN__
1475__ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1) {
1476  uint16x4_t __ret;
1477  __ret = __p0 & __p1;
1478  return __ret;
1479}
1480#else
1481__ai uint16x4_t vand_u16(uint16x4_t __p0, uint16x4_t __p1) {
1482  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1483  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1484  uint16x4_t __ret;
1485  __ret = __rev0 & __rev1;
1486  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1487  return __ret;
1488}
1489#endif
1490
1491#ifdef __LITTLE_ENDIAN__
1492__ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1) {
1493  int8x8_t __ret;
1494  __ret = __p0 & __p1;
1495  return __ret;
1496}
1497#else
1498__ai int8x8_t vand_s8(int8x8_t __p0, int8x8_t __p1) {
1499  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1500  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1501  int8x8_t __ret;
1502  __ret = __rev0 & __rev1;
1503  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1504  return __ret;
1505}
1506#endif
1507
1508#ifdef __LITTLE_ENDIAN__
1509__ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1) {
1510  int32x2_t __ret;
1511  __ret = __p0 & __p1;
1512  return __ret;
1513}
1514#else
1515__ai int32x2_t vand_s32(int32x2_t __p0, int32x2_t __p1) {
1516  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1517  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1518  int32x2_t __ret;
1519  __ret = __rev0 & __rev1;
1520  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1521  return __ret;
1522}
1523#endif
1524
1525#ifdef __LITTLE_ENDIAN__
1526__ai int64x1_t vand_s64(int64x1_t __p0, int64x1_t __p1) {
1527  int64x1_t __ret;
1528  __ret = __p0 & __p1;
1529  return __ret;
1530}
1531#else
1532__ai int64x1_t vand_s64(int64x1_t __p0, int64x1_t __p1) {
1533  int64x1_t __ret;
1534  __ret = __p0 & __p1;
1535  return __ret;
1536}
1537#endif
1538
1539#ifdef __LITTLE_ENDIAN__
1540__ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1) {
1541  int16x4_t __ret;
1542  __ret = __p0 & __p1;
1543  return __ret;
1544}
1545#else
1546__ai int16x4_t vand_s16(int16x4_t __p0, int16x4_t __p1) {
1547  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1548  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1549  int16x4_t __ret;
1550  __ret = __rev0 & __rev1;
1551  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1552  return __ret;
1553}
1554#endif
1555
1556#ifdef __LITTLE_ENDIAN__
1557__ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1558  uint8x16_t __ret;
1559  __ret = __p0 & ~__p1;
1560  return __ret;
1561}
1562#else
1563__ai uint8x16_t vbicq_u8(uint8x16_t __p0, uint8x16_t __p1) {
1564  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1565  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1566  uint8x16_t __ret;
1567  __ret = __rev0 & ~__rev1;
1568  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1569  return __ret;
1570}
1571#endif
1572
1573#ifdef __LITTLE_ENDIAN__
1574__ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1575  uint32x4_t __ret;
1576  __ret = __p0 & ~__p1;
1577  return __ret;
1578}
1579#else
1580__ai uint32x4_t vbicq_u32(uint32x4_t __p0, uint32x4_t __p1) {
1581  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1582  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1583  uint32x4_t __ret;
1584  __ret = __rev0 & ~__rev1;
1585  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1586  return __ret;
1587}
1588#endif
1589
1590#ifdef __LITTLE_ENDIAN__
1591__ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1592  uint64x2_t __ret;
1593  __ret = __p0 & ~__p1;
1594  return __ret;
1595}
1596#else
1597__ai uint64x2_t vbicq_u64(uint64x2_t __p0, uint64x2_t __p1) {
1598  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1599  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1600  uint64x2_t __ret;
1601  __ret = __rev0 & ~__rev1;
1602  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1603  return __ret;
1604}
1605#endif
1606
1607#ifdef __LITTLE_ENDIAN__
1608__ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1609  uint16x8_t __ret;
1610  __ret = __p0 & ~__p1;
1611  return __ret;
1612}
1613#else
1614__ai uint16x8_t vbicq_u16(uint16x8_t __p0, uint16x8_t __p1) {
1615  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1616  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1617  uint16x8_t __ret;
1618  __ret = __rev0 & ~__rev1;
1619  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1620  return __ret;
1621}
1622#endif
1623
1624#ifdef __LITTLE_ENDIAN__
1625__ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1) {
1626  int8x16_t __ret;
1627  __ret = __p0 & ~__p1;
1628  return __ret;
1629}
1630#else
1631__ai int8x16_t vbicq_s8(int8x16_t __p0, int8x16_t __p1) {
1632  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1633  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1634  int8x16_t __ret;
1635  __ret = __rev0 & ~__rev1;
1636  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1637  return __ret;
1638}
1639#endif
1640
1641#ifdef __LITTLE_ENDIAN__
1642__ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1) {
1643  int32x4_t __ret;
1644  __ret = __p0 & ~__p1;
1645  return __ret;
1646}
1647#else
1648__ai int32x4_t vbicq_s32(int32x4_t __p0, int32x4_t __p1) {
1649  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1650  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1651  int32x4_t __ret;
1652  __ret = __rev0 & ~__rev1;
1653  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1654  return __ret;
1655}
1656#endif
1657
1658#ifdef __LITTLE_ENDIAN__
1659__ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1) {
1660  int64x2_t __ret;
1661  __ret = __p0 & ~__p1;
1662  return __ret;
1663}
1664#else
1665__ai int64x2_t vbicq_s64(int64x2_t __p0, int64x2_t __p1) {
1666  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1667  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1668  int64x2_t __ret;
1669  __ret = __rev0 & ~__rev1;
1670  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1671  return __ret;
1672}
1673#endif
1674
1675#ifdef __LITTLE_ENDIAN__
1676__ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1) {
1677  int16x8_t __ret;
1678  __ret = __p0 & ~__p1;
1679  return __ret;
1680}
1681#else
1682__ai int16x8_t vbicq_s16(int16x8_t __p0, int16x8_t __p1) {
1683  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1684  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1685  int16x8_t __ret;
1686  __ret = __rev0 & ~__rev1;
1687  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1688  return __ret;
1689}
1690#endif
1691
1692#ifdef __LITTLE_ENDIAN__
1693__ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1) {
1694  uint8x8_t __ret;
1695  __ret = __p0 & ~__p1;
1696  return __ret;
1697}
1698#else
1699__ai uint8x8_t vbic_u8(uint8x8_t __p0, uint8x8_t __p1) {
1700  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1701  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1702  uint8x8_t __ret;
1703  __ret = __rev0 & ~__rev1;
1704  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1705  return __ret;
1706}
1707#endif
1708
1709#ifdef __LITTLE_ENDIAN__
1710__ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1) {
1711  uint32x2_t __ret;
1712  __ret = __p0 & ~__p1;
1713  return __ret;
1714}
1715#else
1716__ai uint32x2_t vbic_u32(uint32x2_t __p0, uint32x2_t __p1) {
1717  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1718  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1719  uint32x2_t __ret;
1720  __ret = __rev0 & ~__rev1;
1721  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1722  return __ret;
1723}
1724#endif
1725
1726#ifdef __LITTLE_ENDIAN__
1727__ai uint64x1_t vbic_u64(uint64x1_t __p0, uint64x1_t __p1) {
1728  uint64x1_t __ret;
1729  __ret = __p0 & ~__p1;
1730  return __ret;
1731}
1732#else
1733__ai uint64x1_t vbic_u64(uint64x1_t __p0, uint64x1_t __p1) {
1734  uint64x1_t __ret;
1735  __ret = __p0 & ~__p1;
1736  return __ret;
1737}
1738#endif
1739
1740#ifdef __LITTLE_ENDIAN__
1741__ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1) {
1742  uint16x4_t __ret;
1743  __ret = __p0 & ~__p1;
1744  return __ret;
1745}
1746#else
1747__ai uint16x4_t vbic_u16(uint16x4_t __p0, uint16x4_t __p1) {
1748  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1749  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1750  uint16x4_t __ret;
1751  __ret = __rev0 & ~__rev1;
1752  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1753  return __ret;
1754}
1755#endif
1756
1757#ifdef __LITTLE_ENDIAN__
1758__ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1) {
1759  int8x8_t __ret;
1760  __ret = __p0 & ~__p1;
1761  return __ret;
1762}
1763#else
1764__ai int8x8_t vbic_s8(int8x8_t __p0, int8x8_t __p1) {
1765  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1766  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1767  int8x8_t __ret;
1768  __ret = __rev0 & ~__rev1;
1769  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1770  return __ret;
1771}
1772#endif
1773
1774#ifdef __LITTLE_ENDIAN__
1775__ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1) {
1776  int32x2_t __ret;
1777  __ret = __p0 & ~__p1;
1778  return __ret;
1779}
1780#else
1781__ai int32x2_t vbic_s32(int32x2_t __p0, int32x2_t __p1) {
1782  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1783  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1784  int32x2_t __ret;
1785  __ret = __rev0 & ~__rev1;
1786  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1787  return __ret;
1788}
1789#endif
1790
1791#ifdef __LITTLE_ENDIAN__
1792__ai int64x1_t vbic_s64(int64x1_t __p0, int64x1_t __p1) {
1793  int64x1_t __ret;
1794  __ret = __p0 & ~__p1;
1795  return __ret;
1796}
1797#else
1798__ai int64x1_t vbic_s64(int64x1_t __p0, int64x1_t __p1) {
1799  int64x1_t __ret;
1800  __ret = __p0 & ~__p1;
1801  return __ret;
1802}
1803#endif
1804
1805#ifdef __LITTLE_ENDIAN__
1806__ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1) {
1807  int16x4_t __ret;
1808  __ret = __p0 & ~__p1;
1809  return __ret;
1810}
1811#else
1812__ai int16x4_t vbic_s16(int16x4_t __p0, int16x4_t __p1) {
1813  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1814  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1815  int16x4_t __ret;
1816  __ret = __rev0 & ~__rev1;
1817  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1818  return __ret;
1819}
1820#endif
1821
1822#ifdef __LITTLE_ENDIAN__
1823__ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2) {
1824  poly8x8_t __ret;
1825  __ret = (poly8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 4);
1826  return __ret;
1827}
1828#else
1829__ai poly8x8_t vbsl_p8(uint8x8_t __p0, poly8x8_t __p1, poly8x8_t __p2) {
1830  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1831  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1832  poly8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
1833  poly8x8_t __ret;
1834  __ret = (poly8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 4);
1835  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1836  return __ret;
1837}
1838#endif
1839
1840#ifdef __LITTLE_ENDIAN__
1841__ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2) {
1842  poly16x4_t __ret;
1843  __ret = (poly16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 5);
1844  return __ret;
1845}
1846#else
1847__ai poly16x4_t vbsl_p16(uint16x4_t __p0, poly16x4_t __p1, poly16x4_t __p2) {
1848  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1849  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1850  poly16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
1851  poly16x4_t __ret;
1852  __ret = (poly16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 5);
1853  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1854  return __ret;
1855}
1856#endif
1857
1858#ifdef __LITTLE_ENDIAN__
1859__ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2) {
1860  poly8x16_t __ret;
1861  __ret = (poly8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 36);
1862  return __ret;
1863}
1864#else
1865__ai poly8x16_t vbslq_p8(uint8x16_t __p0, poly8x16_t __p1, poly8x16_t __p2) {
1866  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1867  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1868  poly8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1869  poly8x16_t __ret;
1870  __ret = (poly8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 36);
1871  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1872  return __ret;
1873}
1874#endif
1875
1876#ifdef __LITTLE_ENDIAN__
1877__ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2) {
1878  poly16x8_t __ret;
1879  __ret = (poly16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 37);
1880  return __ret;
1881}
1882#else
1883__ai poly16x8_t vbslq_p16(uint16x8_t __p0, poly16x8_t __p1, poly16x8_t __p2) {
1884  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1885  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1886  poly16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
1887  poly16x8_t __ret;
1888  __ret = (poly16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 37);
1889  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1890  return __ret;
1891}
1892#endif
1893
1894#ifdef __LITTLE_ENDIAN__
1895__ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
1896  uint8x16_t __ret;
1897  __ret = (uint8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
1898  return __ret;
1899}
1900#else
1901__ai uint8x16_t vbslq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
1902  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1903  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1904  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1905  uint8x16_t __ret;
1906  __ret = (uint8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
1907  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1908  return __ret;
1909}
1910#endif
1911
1912#ifdef __LITTLE_ENDIAN__
1913__ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
1914  uint32x4_t __ret;
1915  __ret = (uint32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
1916  return __ret;
1917}
1918#else
1919__ai uint32x4_t vbslq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
1920  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1921  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1922  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
1923  uint32x4_t __ret;
1924  __ret = (uint32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
1925  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1926  return __ret;
1927}
1928#endif
1929
1930#ifdef __LITTLE_ENDIAN__
1931__ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
1932  uint64x2_t __ret;
1933  __ret = (uint64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51);
1934  return __ret;
1935}
1936#else
1937__ai uint64x2_t vbslq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
1938  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
1939  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
1940  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
1941  uint64x2_t __ret;
1942  __ret = (uint64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51);
1943  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
1944  return __ret;
1945}
1946#endif
1947
1948#ifdef __LITTLE_ENDIAN__
1949__ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
1950  uint16x8_t __ret;
1951  __ret = (uint16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 49);
1952  return __ret;
1953}
1954#else
1955__ai uint16x8_t vbslq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
1956  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
1957  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
1958  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
1959  uint16x8_t __ret;
1960  __ret = (uint16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 49);
1961  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
1962  return __ret;
1963}
1964#endif
1965
1966#ifdef __LITTLE_ENDIAN__
1967__ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
1968  int8x16_t __ret;
1969  __ret = (int8x16_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
1970  return __ret;
1971}
1972#else
1973__ai int8x16_t vbslq_s8(uint8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
1974  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1975  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1976  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1977  int8x16_t __ret;
1978  __ret = (int8x16_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
1979  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
1980  return __ret;
1981}
1982#endif
1983
1984#ifdef __LITTLE_ENDIAN__
1985__ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
1986  float32x4_t __ret;
1987  __ret = (float32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
1988  return __ret;
1989}
1990#else
1991__ai float32x4_t vbslq_f32(uint32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
1992  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
1993  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
1994  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
1995  float32x4_t __ret;
1996  __ret = (float32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
1997  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
1998  return __ret;
1999}
2000#endif
2001
2002#ifdef __LITTLE_ENDIAN__
2003__ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
2004  int32x4_t __ret;
2005  __ret = (int32x4_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34);
2006  return __ret;
2007}
2008#else
2009__ai int32x4_t vbslq_s32(uint32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
2010  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2011  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2012  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2013  int32x4_t __ret;
2014  __ret = (int32x4_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34);
2015  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2016  return __ret;
2017}
2018#endif
2019
2020#ifdef __LITTLE_ENDIAN__
2021__ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
2022  int64x2_t __ret;
2023  __ret = (int64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 35);
2024  return __ret;
2025}
2026#else
2027__ai int64x2_t vbslq_s64(uint64x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
2028  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2029  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2030  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2031  int64x2_t __ret;
2032  __ret = (int64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 35);
2033  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2034  return __ret;
2035}
2036#endif
2037
2038#ifdef __LITTLE_ENDIAN__
2039__ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
2040  int16x8_t __ret;
2041  __ret = (int16x8_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33);
2042  return __ret;
2043}
2044#else
2045__ai int16x8_t vbslq_s16(uint16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
2046  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2047  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2048  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2049  int16x8_t __ret;
2050  __ret = (int16x8_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33);
2051  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2052  return __ret;
2053}
2054#endif
2055
2056#ifdef __LITTLE_ENDIAN__
2057__ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
2058  uint8x8_t __ret;
2059  __ret = (uint8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 16);
2060  return __ret;
2061}
2062#else
2063__ai uint8x8_t vbsl_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
2064  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2065  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2066  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2067  uint8x8_t __ret;
2068  __ret = (uint8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 16);
2069  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2070  return __ret;
2071}
2072#endif
2073
2074#ifdef __LITTLE_ENDIAN__
2075__ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
2076  uint32x2_t __ret;
2077  __ret = (uint32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 18);
2078  return __ret;
2079}
2080#else
2081__ai uint32x2_t vbsl_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
2082  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2083  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2084  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2085  uint32x2_t __ret;
2086  __ret = (uint32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 18);
2087  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2088  return __ret;
2089}
2090#endif
2091
2092#ifdef __LITTLE_ENDIAN__
2093__ai uint64x1_t vbsl_u64(uint64x1_t __p0, uint64x1_t __p1, uint64x1_t __p2) {
2094  uint64x1_t __ret;
2095  __ret = (uint64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 19);
2096  return __ret;
2097}
2098#else
2099__ai uint64x1_t vbsl_u64(uint64x1_t __p0, uint64x1_t __p1, uint64x1_t __p2) {
2100  uint64x1_t __ret;
2101  __ret = (uint64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 19);
2102  return __ret;
2103}
2104#endif
2105
2106#ifdef __LITTLE_ENDIAN__
2107__ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
2108  uint16x4_t __ret;
2109  __ret = (uint16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 17);
2110  return __ret;
2111}
2112#else
2113__ai uint16x4_t vbsl_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
2114  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2115  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2116  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2117  uint16x4_t __ret;
2118  __ret = (uint16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 17);
2119  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2120  return __ret;
2121}
2122#endif
2123
2124#ifdef __LITTLE_ENDIAN__
2125__ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
2126  int8x8_t __ret;
2127  __ret = (int8x8_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 0);
2128  return __ret;
2129}
2130#else
2131__ai int8x8_t vbsl_s8(uint8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
2132  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2133  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2134  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
2135  int8x8_t __ret;
2136  __ret = (int8x8_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 0);
2137  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2138  return __ret;
2139}
2140#endif
2141
2142#ifdef __LITTLE_ENDIAN__
2143__ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
2144  float32x2_t __ret;
2145  __ret = (float32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
2146  return __ret;
2147}
2148#else
2149__ai float32x2_t vbsl_f32(uint32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
2150  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2151  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2152  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2153  float32x2_t __ret;
2154  __ret = (float32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
2155  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2156  return __ret;
2157}
2158#endif
2159
2160#ifdef __LITTLE_ENDIAN__
2161__ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
2162  int32x2_t __ret;
2163  __ret = (int32x2_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 2);
2164  return __ret;
2165}
2166#else
2167__ai int32x2_t vbsl_s32(uint32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
2168  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2169  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2170  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
2171  int32x2_t __ret;
2172  __ret = (int32x2_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 2);
2173  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2174  return __ret;
2175}
2176#endif
2177
2178#ifdef __LITTLE_ENDIAN__
2179__ai int64x1_t vbsl_s64(uint64x1_t __p0, int64x1_t __p1, int64x1_t __p2) {
2180  int64x1_t __ret;
2181  __ret = (int64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 3);
2182  return __ret;
2183}
2184#else
2185__ai int64x1_t vbsl_s64(uint64x1_t __p0, int64x1_t __p1, int64x1_t __p2) {
2186  int64x1_t __ret;
2187  __ret = (int64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 3);
2188  return __ret;
2189}
2190#endif
2191
2192#ifdef __LITTLE_ENDIAN__
2193__ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
2194  int16x4_t __ret;
2195  __ret = (int16x4_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 1);
2196  return __ret;
2197}
2198#else
2199__ai int16x4_t vbsl_s16(uint16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
2200  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2201  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2202  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
2203  int16x4_t __ret;
2204  __ret = (int16x4_t) __builtin_neon_vbsl_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 1);
2205  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2206  return __ret;
2207}
2208#endif
2209
2210#ifdef __LITTLE_ENDIAN__
2211__ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1) {
2212  uint32x4_t __ret;
2213  __ret = (uint32x4_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2214  return __ret;
2215}
2216#else
2217__ai uint32x4_t vcageq_f32(float32x4_t __p0, float32x4_t __p1) {
2218  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2219  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2220  uint32x4_t __ret;
2221  __ret = (uint32x4_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2222  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2223  return __ret;
2224}
2225#endif
2226
2227#ifdef __LITTLE_ENDIAN__
2228__ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1) {
2229  uint32x2_t __ret;
2230  __ret = (uint32x2_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2231  return __ret;
2232}
2233#else
2234__ai uint32x2_t vcage_f32(float32x2_t __p0, float32x2_t __p1) {
2235  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2236  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2237  uint32x2_t __ret;
2238  __ret = (uint32x2_t) __builtin_neon_vcage_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2239  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2240  return __ret;
2241}
2242#endif
2243
2244#ifdef __LITTLE_ENDIAN__
2245__ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1) {
2246  uint32x4_t __ret;
2247  __ret = (uint32x4_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2248  return __ret;
2249}
2250#else
2251__ai uint32x4_t vcagtq_f32(float32x4_t __p0, float32x4_t __p1) {
2252  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2253  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2254  uint32x4_t __ret;
2255  __ret = (uint32x4_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2256  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2257  return __ret;
2258}
2259#endif
2260
2261#ifdef __LITTLE_ENDIAN__
2262__ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1) {
2263  uint32x2_t __ret;
2264  __ret = (uint32x2_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2265  return __ret;
2266}
2267#else
2268__ai uint32x2_t vcagt_f32(float32x2_t __p0, float32x2_t __p1) {
2269  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2270  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2271  uint32x2_t __ret;
2272  __ret = (uint32x2_t) __builtin_neon_vcagt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2273  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2274  return __ret;
2275}
2276#endif
2277
2278#ifdef __LITTLE_ENDIAN__
2279__ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1) {
2280  uint32x4_t __ret;
2281  __ret = (uint32x4_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2282  return __ret;
2283}
2284#else
2285__ai uint32x4_t vcaleq_f32(float32x4_t __p0, float32x4_t __p1) {
2286  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2287  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2288  uint32x4_t __ret;
2289  __ret = (uint32x4_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2290  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2291  return __ret;
2292}
2293#endif
2294
2295#ifdef __LITTLE_ENDIAN__
2296__ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1) {
2297  uint32x2_t __ret;
2298  __ret = (uint32x2_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2299  return __ret;
2300}
2301#else
2302__ai uint32x2_t vcale_f32(float32x2_t __p0, float32x2_t __p1) {
2303  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2304  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2305  uint32x2_t __ret;
2306  __ret = (uint32x2_t) __builtin_neon_vcale_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2307  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2308  return __ret;
2309}
2310#endif
2311
2312#ifdef __LITTLE_ENDIAN__
2313__ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1) {
2314  uint32x4_t __ret;
2315  __ret = (uint32x4_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
2316  return __ret;
2317}
2318#else
2319__ai uint32x4_t vcaltq_f32(float32x4_t __p0, float32x4_t __p1) {
2320  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2321  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2322  uint32x4_t __ret;
2323  __ret = (uint32x4_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
2324  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2325  return __ret;
2326}
2327#endif
2328
2329#ifdef __LITTLE_ENDIAN__
2330__ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1) {
2331  uint32x2_t __ret;
2332  __ret = (uint32x2_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
2333  return __ret;
2334}
2335#else
2336__ai uint32x2_t vcalt_f32(float32x2_t __p0, float32x2_t __p1) {
2337  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2338  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2339  uint32x2_t __ret;
2340  __ret = (uint32x2_t) __builtin_neon_vcalt_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
2341  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2342  return __ret;
2343}
2344#endif
2345
2346#ifdef __LITTLE_ENDIAN__
2347__ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1) {
2348  uint8x8_t __ret;
2349  __ret = (uint8x8_t)(__p0 == __p1);
2350  return __ret;
2351}
2352#else
2353__ai uint8x8_t vceq_p8(poly8x8_t __p0, poly8x8_t __p1) {
2354  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2355  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2356  uint8x8_t __ret;
2357  __ret = (uint8x8_t)(__rev0 == __rev1);
2358  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2359  return __ret;
2360}
2361#endif
2362
2363#ifdef __LITTLE_ENDIAN__
2364__ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1) {
2365  uint8x16_t __ret;
2366  __ret = (uint8x16_t)(__p0 == __p1);
2367  return __ret;
2368}
2369#else
2370__ai uint8x16_t vceqq_p8(poly8x16_t __p0, poly8x16_t __p1) {
2371  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2372  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2373  uint8x16_t __ret;
2374  __ret = (uint8x16_t)(__rev0 == __rev1);
2375  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2376  return __ret;
2377}
2378#endif
2379
2380#ifdef __LITTLE_ENDIAN__
2381__ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2382  uint8x16_t __ret;
2383  __ret = (uint8x16_t)(__p0 == __p1);
2384  return __ret;
2385}
2386#else
2387__ai uint8x16_t vceqq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2388  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2389  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2390  uint8x16_t __ret;
2391  __ret = (uint8x16_t)(__rev0 == __rev1);
2392  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2393  return __ret;
2394}
2395#endif
2396
2397#ifdef __LITTLE_ENDIAN__
2398__ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2399  uint32x4_t __ret;
2400  __ret = (uint32x4_t)(__p0 == __p1);
2401  return __ret;
2402}
2403#else
2404__ai uint32x4_t vceqq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2405  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2406  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2407  uint32x4_t __ret;
2408  __ret = (uint32x4_t)(__rev0 == __rev1);
2409  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2410  return __ret;
2411}
2412#endif
2413
2414#ifdef __LITTLE_ENDIAN__
2415__ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2416  uint16x8_t __ret;
2417  __ret = (uint16x8_t)(__p0 == __p1);
2418  return __ret;
2419}
2420#else
2421__ai uint16x8_t vceqq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2422  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2423  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2424  uint16x8_t __ret;
2425  __ret = (uint16x8_t)(__rev0 == __rev1);
2426  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2427  return __ret;
2428}
2429#endif
2430
2431#ifdef __LITTLE_ENDIAN__
2432__ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1) {
2433  uint8x16_t __ret;
2434  __ret = (uint8x16_t)(__p0 == __p1);
2435  return __ret;
2436}
2437#else
2438__ai uint8x16_t vceqq_s8(int8x16_t __p0, int8x16_t __p1) {
2439  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2440  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2441  uint8x16_t __ret;
2442  __ret = (uint8x16_t)(__rev0 == __rev1);
2443  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2444  return __ret;
2445}
2446#endif
2447
2448#ifdef __LITTLE_ENDIAN__
2449__ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1) {
2450  uint32x4_t __ret;
2451  __ret = (uint32x4_t)(__p0 == __p1);
2452  return __ret;
2453}
2454#else
2455__ai uint32x4_t vceqq_f32(float32x4_t __p0, float32x4_t __p1) {
2456  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2457  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2458  uint32x4_t __ret;
2459  __ret = (uint32x4_t)(__rev0 == __rev1);
2460  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2461  return __ret;
2462}
2463#endif
2464
2465#ifdef __LITTLE_ENDIAN__
2466__ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1) {
2467  uint32x4_t __ret;
2468  __ret = (uint32x4_t)(__p0 == __p1);
2469  return __ret;
2470}
2471#else
2472__ai uint32x4_t vceqq_s32(int32x4_t __p0, int32x4_t __p1) {
2473  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2474  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2475  uint32x4_t __ret;
2476  __ret = (uint32x4_t)(__rev0 == __rev1);
2477  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2478  return __ret;
2479}
2480#endif
2481
2482#ifdef __LITTLE_ENDIAN__
2483__ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1) {
2484  uint16x8_t __ret;
2485  __ret = (uint16x8_t)(__p0 == __p1);
2486  return __ret;
2487}
2488#else
2489__ai uint16x8_t vceqq_s16(int16x8_t __p0, int16x8_t __p1) {
2490  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2491  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2492  uint16x8_t __ret;
2493  __ret = (uint16x8_t)(__rev0 == __rev1);
2494  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2495  return __ret;
2496}
2497#endif
2498
2499#ifdef __LITTLE_ENDIAN__
2500__ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1) {
2501  uint8x8_t __ret;
2502  __ret = (uint8x8_t)(__p0 == __p1);
2503  return __ret;
2504}
2505#else
2506__ai uint8x8_t vceq_u8(uint8x8_t __p0, uint8x8_t __p1) {
2507  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2508  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2509  uint8x8_t __ret;
2510  __ret = (uint8x8_t)(__rev0 == __rev1);
2511  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2512  return __ret;
2513}
2514#endif
2515
2516#ifdef __LITTLE_ENDIAN__
2517__ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1) {
2518  uint32x2_t __ret;
2519  __ret = (uint32x2_t)(__p0 == __p1);
2520  return __ret;
2521}
2522#else
2523__ai uint32x2_t vceq_u32(uint32x2_t __p0, uint32x2_t __p1) {
2524  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2525  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2526  uint32x2_t __ret;
2527  __ret = (uint32x2_t)(__rev0 == __rev1);
2528  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2529  return __ret;
2530}
2531#endif
2532
2533#ifdef __LITTLE_ENDIAN__
2534__ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1) {
2535  uint16x4_t __ret;
2536  __ret = (uint16x4_t)(__p0 == __p1);
2537  return __ret;
2538}
2539#else
2540__ai uint16x4_t vceq_u16(uint16x4_t __p0, uint16x4_t __p1) {
2541  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2542  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2543  uint16x4_t __ret;
2544  __ret = (uint16x4_t)(__rev0 == __rev1);
2545  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2546  return __ret;
2547}
2548#endif
2549
2550#ifdef __LITTLE_ENDIAN__
2551__ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1) {
2552  uint8x8_t __ret;
2553  __ret = (uint8x8_t)(__p0 == __p1);
2554  return __ret;
2555}
2556#else
2557__ai uint8x8_t vceq_s8(int8x8_t __p0, int8x8_t __p1) {
2558  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2559  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2560  uint8x8_t __ret;
2561  __ret = (uint8x8_t)(__rev0 == __rev1);
2562  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2563  return __ret;
2564}
2565#endif
2566
2567#ifdef __LITTLE_ENDIAN__
2568__ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1) {
2569  uint32x2_t __ret;
2570  __ret = (uint32x2_t)(__p0 == __p1);
2571  return __ret;
2572}
2573#else
2574__ai uint32x2_t vceq_f32(float32x2_t __p0, float32x2_t __p1) {
2575  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2576  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2577  uint32x2_t __ret;
2578  __ret = (uint32x2_t)(__rev0 == __rev1);
2579  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2580  return __ret;
2581}
2582#endif
2583
2584#ifdef __LITTLE_ENDIAN__
2585__ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1) {
2586  uint32x2_t __ret;
2587  __ret = (uint32x2_t)(__p0 == __p1);
2588  return __ret;
2589}
2590#else
2591__ai uint32x2_t vceq_s32(int32x2_t __p0, int32x2_t __p1) {
2592  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2593  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2594  uint32x2_t __ret;
2595  __ret = (uint32x2_t)(__rev0 == __rev1);
2596  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2597  return __ret;
2598}
2599#endif
2600
2601#ifdef __LITTLE_ENDIAN__
2602__ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1) {
2603  uint16x4_t __ret;
2604  __ret = (uint16x4_t)(__p0 == __p1);
2605  return __ret;
2606}
2607#else
2608__ai uint16x4_t vceq_s16(int16x4_t __p0, int16x4_t __p1) {
2609  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2610  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2611  uint16x4_t __ret;
2612  __ret = (uint16x4_t)(__rev0 == __rev1);
2613  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2614  return __ret;
2615}
2616#endif
2617
2618#ifdef __LITTLE_ENDIAN__
2619__ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2620  uint8x16_t __ret;
2621  __ret = (uint8x16_t)(__p0 >= __p1);
2622  return __ret;
2623}
2624#else
2625__ai uint8x16_t vcgeq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2626  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2627  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2628  uint8x16_t __ret;
2629  __ret = (uint8x16_t)(__rev0 >= __rev1);
2630  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2631  return __ret;
2632}
2633#endif
2634
2635#ifdef __LITTLE_ENDIAN__
2636__ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2637  uint32x4_t __ret;
2638  __ret = (uint32x4_t)(__p0 >= __p1);
2639  return __ret;
2640}
2641#else
2642__ai uint32x4_t vcgeq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2643  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2644  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2645  uint32x4_t __ret;
2646  __ret = (uint32x4_t)(__rev0 >= __rev1);
2647  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2648  return __ret;
2649}
2650#endif
2651
2652#ifdef __LITTLE_ENDIAN__
2653__ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2654  uint16x8_t __ret;
2655  __ret = (uint16x8_t)(__p0 >= __p1);
2656  return __ret;
2657}
2658#else
2659__ai uint16x8_t vcgeq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2660  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2661  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2662  uint16x8_t __ret;
2663  __ret = (uint16x8_t)(__rev0 >= __rev1);
2664  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2665  return __ret;
2666}
2667#endif
2668
2669#ifdef __LITTLE_ENDIAN__
2670__ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1) {
2671  uint8x16_t __ret;
2672  __ret = (uint8x16_t)(__p0 >= __p1);
2673  return __ret;
2674}
2675#else
2676__ai uint8x16_t vcgeq_s8(int8x16_t __p0, int8x16_t __p1) {
2677  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2678  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2679  uint8x16_t __ret;
2680  __ret = (uint8x16_t)(__rev0 >= __rev1);
2681  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2682  return __ret;
2683}
2684#endif
2685
2686#ifdef __LITTLE_ENDIAN__
2687__ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1) {
2688  uint32x4_t __ret;
2689  __ret = (uint32x4_t)(__p0 >= __p1);
2690  return __ret;
2691}
2692#else
2693__ai uint32x4_t vcgeq_f32(float32x4_t __p0, float32x4_t __p1) {
2694  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2695  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2696  uint32x4_t __ret;
2697  __ret = (uint32x4_t)(__rev0 >= __rev1);
2698  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2699  return __ret;
2700}
2701#endif
2702
2703#ifdef __LITTLE_ENDIAN__
2704__ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1) {
2705  uint32x4_t __ret;
2706  __ret = (uint32x4_t)(__p0 >= __p1);
2707  return __ret;
2708}
2709#else
2710__ai uint32x4_t vcgeq_s32(int32x4_t __p0, int32x4_t __p1) {
2711  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2712  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2713  uint32x4_t __ret;
2714  __ret = (uint32x4_t)(__rev0 >= __rev1);
2715  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2716  return __ret;
2717}
2718#endif
2719
2720#ifdef __LITTLE_ENDIAN__
2721__ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1) {
2722  uint16x8_t __ret;
2723  __ret = (uint16x8_t)(__p0 >= __p1);
2724  return __ret;
2725}
2726#else
2727__ai uint16x8_t vcgeq_s16(int16x8_t __p0, int16x8_t __p1) {
2728  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2729  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2730  uint16x8_t __ret;
2731  __ret = (uint16x8_t)(__rev0 >= __rev1);
2732  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2733  return __ret;
2734}
2735#endif
2736
2737#ifdef __LITTLE_ENDIAN__
2738__ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1) {
2739  uint8x8_t __ret;
2740  __ret = (uint8x8_t)(__p0 >= __p1);
2741  return __ret;
2742}
2743#else
2744__ai uint8x8_t vcge_u8(uint8x8_t __p0, uint8x8_t __p1) {
2745  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2746  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2747  uint8x8_t __ret;
2748  __ret = (uint8x8_t)(__rev0 >= __rev1);
2749  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2750  return __ret;
2751}
2752#endif
2753
2754#ifdef __LITTLE_ENDIAN__
2755__ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1) {
2756  uint32x2_t __ret;
2757  __ret = (uint32x2_t)(__p0 >= __p1);
2758  return __ret;
2759}
2760#else
2761__ai uint32x2_t vcge_u32(uint32x2_t __p0, uint32x2_t __p1) {
2762  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2763  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2764  uint32x2_t __ret;
2765  __ret = (uint32x2_t)(__rev0 >= __rev1);
2766  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2767  return __ret;
2768}
2769#endif
2770
2771#ifdef __LITTLE_ENDIAN__
2772__ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1) {
2773  uint16x4_t __ret;
2774  __ret = (uint16x4_t)(__p0 >= __p1);
2775  return __ret;
2776}
2777#else
2778__ai uint16x4_t vcge_u16(uint16x4_t __p0, uint16x4_t __p1) {
2779  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2780  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2781  uint16x4_t __ret;
2782  __ret = (uint16x4_t)(__rev0 >= __rev1);
2783  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2784  return __ret;
2785}
2786#endif
2787
2788#ifdef __LITTLE_ENDIAN__
2789__ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1) {
2790  uint8x8_t __ret;
2791  __ret = (uint8x8_t)(__p0 >= __p1);
2792  return __ret;
2793}
2794#else
2795__ai uint8x8_t vcge_s8(int8x8_t __p0, int8x8_t __p1) {
2796  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2797  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2798  uint8x8_t __ret;
2799  __ret = (uint8x8_t)(__rev0 >= __rev1);
2800  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2801  return __ret;
2802}
2803#endif
2804
2805#ifdef __LITTLE_ENDIAN__
2806__ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1) {
2807  uint32x2_t __ret;
2808  __ret = (uint32x2_t)(__p0 >= __p1);
2809  return __ret;
2810}
2811#else
2812__ai uint32x2_t vcge_f32(float32x2_t __p0, float32x2_t __p1) {
2813  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2814  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2815  uint32x2_t __ret;
2816  __ret = (uint32x2_t)(__rev0 >= __rev1);
2817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2818  return __ret;
2819}
2820#endif
2821
2822#ifdef __LITTLE_ENDIAN__
2823__ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1) {
2824  uint32x2_t __ret;
2825  __ret = (uint32x2_t)(__p0 >= __p1);
2826  return __ret;
2827}
2828#else
2829__ai uint32x2_t vcge_s32(int32x2_t __p0, int32x2_t __p1) {
2830  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
2831  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
2832  uint32x2_t __ret;
2833  __ret = (uint32x2_t)(__rev0 >= __rev1);
2834  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
2835  return __ret;
2836}
2837#endif
2838
2839#ifdef __LITTLE_ENDIAN__
2840__ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1) {
2841  uint16x4_t __ret;
2842  __ret = (uint16x4_t)(__p0 >= __p1);
2843  return __ret;
2844}
2845#else
2846__ai uint16x4_t vcge_s16(int16x4_t __p0, int16x4_t __p1) {
2847  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2848  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2849  uint16x4_t __ret;
2850  __ret = (uint16x4_t)(__rev0 >= __rev1);
2851  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2852  return __ret;
2853}
2854#endif
2855
2856#ifdef __LITTLE_ENDIAN__
2857__ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2858  uint8x16_t __ret;
2859  __ret = (uint8x16_t)(__p0 > __p1);
2860  return __ret;
2861}
2862#else
2863__ai uint8x16_t vcgtq_u8(uint8x16_t __p0, uint8x16_t __p1) {
2864  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2865  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2866  uint8x16_t __ret;
2867  __ret = (uint8x16_t)(__rev0 > __rev1);
2868  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2869  return __ret;
2870}
2871#endif
2872
2873#ifdef __LITTLE_ENDIAN__
2874__ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2875  uint32x4_t __ret;
2876  __ret = (uint32x4_t)(__p0 > __p1);
2877  return __ret;
2878}
2879#else
2880__ai uint32x4_t vcgtq_u32(uint32x4_t __p0, uint32x4_t __p1) {
2881  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2882  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2883  uint32x4_t __ret;
2884  __ret = (uint32x4_t)(__rev0 > __rev1);
2885  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2886  return __ret;
2887}
2888#endif
2889
2890#ifdef __LITTLE_ENDIAN__
2891__ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2892  uint16x8_t __ret;
2893  __ret = (uint16x8_t)(__p0 > __p1);
2894  return __ret;
2895}
2896#else
2897__ai uint16x8_t vcgtq_u16(uint16x8_t __p0, uint16x8_t __p1) {
2898  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2899  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2900  uint16x8_t __ret;
2901  __ret = (uint16x8_t)(__rev0 > __rev1);
2902  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2903  return __ret;
2904}
2905#endif
2906
2907#ifdef __LITTLE_ENDIAN__
2908__ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1) {
2909  uint8x16_t __ret;
2910  __ret = (uint8x16_t)(__p0 > __p1);
2911  return __ret;
2912}
2913#else
2914__ai uint8x16_t vcgtq_s8(int8x16_t __p0, int8x16_t __p1) {
2915  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2916  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2917  uint8x16_t __ret;
2918  __ret = (uint8x16_t)(__rev0 > __rev1);
2919  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
2920  return __ret;
2921}
2922#endif
2923
2924#ifdef __LITTLE_ENDIAN__
2925__ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1) {
2926  uint32x4_t __ret;
2927  __ret = (uint32x4_t)(__p0 > __p1);
2928  return __ret;
2929}
2930#else
2931__ai uint32x4_t vcgtq_f32(float32x4_t __p0, float32x4_t __p1) {
2932  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2933  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2934  uint32x4_t __ret;
2935  __ret = (uint32x4_t)(__rev0 > __rev1);
2936  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2937  return __ret;
2938}
2939#endif
2940
2941#ifdef __LITTLE_ENDIAN__
2942__ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1) {
2943  uint32x4_t __ret;
2944  __ret = (uint32x4_t)(__p0 > __p1);
2945  return __ret;
2946}
2947#else
2948__ai uint32x4_t vcgtq_s32(int32x4_t __p0, int32x4_t __p1) {
2949  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
2950  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
2951  uint32x4_t __ret;
2952  __ret = (uint32x4_t)(__rev0 > __rev1);
2953  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
2954  return __ret;
2955}
2956#endif
2957
2958#ifdef __LITTLE_ENDIAN__
2959__ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1) {
2960  uint16x8_t __ret;
2961  __ret = (uint16x8_t)(__p0 > __p1);
2962  return __ret;
2963}
2964#else
2965__ai uint16x8_t vcgtq_s16(int16x8_t __p0, int16x8_t __p1) {
2966  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2967  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2968  uint16x8_t __ret;
2969  __ret = (uint16x8_t)(__rev0 > __rev1);
2970  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2971  return __ret;
2972}
2973#endif
2974
2975#ifdef __LITTLE_ENDIAN__
2976__ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1) {
2977  uint8x8_t __ret;
2978  __ret = (uint8x8_t)(__p0 > __p1);
2979  return __ret;
2980}
2981#else
2982__ai uint8x8_t vcgt_u8(uint8x8_t __p0, uint8x8_t __p1) {
2983  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
2984  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
2985  uint8x8_t __ret;
2986  __ret = (uint8x8_t)(__rev0 > __rev1);
2987  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
2988  return __ret;
2989}
2990#endif
2991
2992#ifdef __LITTLE_ENDIAN__
2993__ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1) {
2994  uint32x2_t __ret;
2995  __ret = (uint32x2_t)(__p0 > __p1);
2996  return __ret;
2997}
2998#else
2999__ai uint32x2_t vcgt_u32(uint32x2_t __p0, uint32x2_t __p1) {
3000  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3001  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3002  uint32x2_t __ret;
3003  __ret = (uint32x2_t)(__rev0 > __rev1);
3004  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3005  return __ret;
3006}
3007#endif
3008
3009#ifdef __LITTLE_ENDIAN__
3010__ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3011  uint16x4_t __ret;
3012  __ret = (uint16x4_t)(__p0 > __p1);
3013  return __ret;
3014}
3015#else
3016__ai uint16x4_t vcgt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3017  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3018  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3019  uint16x4_t __ret;
3020  __ret = (uint16x4_t)(__rev0 > __rev1);
3021  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3022  return __ret;
3023}
3024#endif
3025
3026#ifdef __LITTLE_ENDIAN__
3027__ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1) {
3028  uint8x8_t __ret;
3029  __ret = (uint8x8_t)(__p0 > __p1);
3030  return __ret;
3031}
3032#else
3033__ai uint8x8_t vcgt_s8(int8x8_t __p0, int8x8_t __p1) {
3034  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3035  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3036  uint8x8_t __ret;
3037  __ret = (uint8x8_t)(__rev0 > __rev1);
3038  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3039  return __ret;
3040}
3041#endif
3042
3043#ifdef __LITTLE_ENDIAN__
3044__ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1) {
3045  uint32x2_t __ret;
3046  __ret = (uint32x2_t)(__p0 > __p1);
3047  return __ret;
3048}
3049#else
3050__ai uint32x2_t vcgt_f32(float32x2_t __p0, float32x2_t __p1) {
3051  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3052  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3053  uint32x2_t __ret;
3054  __ret = (uint32x2_t)(__rev0 > __rev1);
3055  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3056  return __ret;
3057}
3058#endif
3059
3060#ifdef __LITTLE_ENDIAN__
3061__ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1) {
3062  uint32x2_t __ret;
3063  __ret = (uint32x2_t)(__p0 > __p1);
3064  return __ret;
3065}
3066#else
3067__ai uint32x2_t vcgt_s32(int32x2_t __p0, int32x2_t __p1) {
3068  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3069  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3070  uint32x2_t __ret;
3071  __ret = (uint32x2_t)(__rev0 > __rev1);
3072  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3073  return __ret;
3074}
3075#endif
3076
3077#ifdef __LITTLE_ENDIAN__
3078__ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1) {
3079  uint16x4_t __ret;
3080  __ret = (uint16x4_t)(__p0 > __p1);
3081  return __ret;
3082}
3083#else
3084__ai uint16x4_t vcgt_s16(int16x4_t __p0, int16x4_t __p1) {
3085  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3086  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3087  uint16x4_t __ret;
3088  __ret = (uint16x4_t)(__rev0 > __rev1);
3089  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3090  return __ret;
3091}
3092#endif
3093
3094#ifdef __LITTLE_ENDIAN__
3095__ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3096  uint8x16_t __ret;
3097  __ret = (uint8x16_t)(__p0 <= __p1);
3098  return __ret;
3099}
3100#else
3101__ai uint8x16_t vcleq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3102  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3103  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3104  uint8x16_t __ret;
3105  __ret = (uint8x16_t)(__rev0 <= __rev1);
3106  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3107  return __ret;
3108}
3109#endif
3110
3111#ifdef __LITTLE_ENDIAN__
3112__ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3113  uint32x4_t __ret;
3114  __ret = (uint32x4_t)(__p0 <= __p1);
3115  return __ret;
3116}
3117#else
3118__ai uint32x4_t vcleq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3119  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3120  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3121  uint32x4_t __ret;
3122  __ret = (uint32x4_t)(__rev0 <= __rev1);
3123  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3124  return __ret;
3125}
3126#endif
3127
3128#ifdef __LITTLE_ENDIAN__
3129__ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3130  uint16x8_t __ret;
3131  __ret = (uint16x8_t)(__p0 <= __p1);
3132  return __ret;
3133}
3134#else
3135__ai uint16x8_t vcleq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3136  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3137  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3138  uint16x8_t __ret;
3139  __ret = (uint16x8_t)(__rev0 <= __rev1);
3140  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3141  return __ret;
3142}
3143#endif
3144
3145#ifdef __LITTLE_ENDIAN__
3146__ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1) {
3147  uint8x16_t __ret;
3148  __ret = (uint8x16_t)(__p0 <= __p1);
3149  return __ret;
3150}
3151#else
3152__ai uint8x16_t vcleq_s8(int8x16_t __p0, int8x16_t __p1) {
3153  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3154  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3155  uint8x16_t __ret;
3156  __ret = (uint8x16_t)(__rev0 <= __rev1);
3157  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3158  return __ret;
3159}
3160#endif
3161
3162#ifdef __LITTLE_ENDIAN__
3163__ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1) {
3164  uint32x4_t __ret;
3165  __ret = (uint32x4_t)(__p0 <= __p1);
3166  return __ret;
3167}
3168#else
3169__ai uint32x4_t vcleq_f32(float32x4_t __p0, float32x4_t __p1) {
3170  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3171  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3172  uint32x4_t __ret;
3173  __ret = (uint32x4_t)(__rev0 <= __rev1);
3174  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3175  return __ret;
3176}
3177#endif
3178
3179#ifdef __LITTLE_ENDIAN__
3180__ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1) {
3181  uint32x4_t __ret;
3182  __ret = (uint32x4_t)(__p0 <= __p1);
3183  return __ret;
3184}
3185#else
3186__ai uint32x4_t vcleq_s32(int32x4_t __p0, int32x4_t __p1) {
3187  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3188  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3189  uint32x4_t __ret;
3190  __ret = (uint32x4_t)(__rev0 <= __rev1);
3191  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3192  return __ret;
3193}
3194#endif
3195
3196#ifdef __LITTLE_ENDIAN__
3197__ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1) {
3198  uint16x8_t __ret;
3199  __ret = (uint16x8_t)(__p0 <= __p1);
3200  return __ret;
3201}
3202#else
3203__ai uint16x8_t vcleq_s16(int16x8_t __p0, int16x8_t __p1) {
3204  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3205  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3206  uint16x8_t __ret;
3207  __ret = (uint16x8_t)(__rev0 <= __rev1);
3208  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3209  return __ret;
3210}
3211#endif
3212
3213#ifdef __LITTLE_ENDIAN__
3214__ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1) {
3215  uint8x8_t __ret;
3216  __ret = (uint8x8_t)(__p0 <= __p1);
3217  return __ret;
3218}
3219#else
3220__ai uint8x8_t vcle_u8(uint8x8_t __p0, uint8x8_t __p1) {
3221  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3222  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3223  uint8x8_t __ret;
3224  __ret = (uint8x8_t)(__rev0 <= __rev1);
3225  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3226  return __ret;
3227}
3228#endif
3229
3230#ifdef __LITTLE_ENDIAN__
3231__ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1) {
3232  uint32x2_t __ret;
3233  __ret = (uint32x2_t)(__p0 <= __p1);
3234  return __ret;
3235}
3236#else
3237__ai uint32x2_t vcle_u32(uint32x2_t __p0, uint32x2_t __p1) {
3238  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3239  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3240  uint32x2_t __ret;
3241  __ret = (uint32x2_t)(__rev0 <= __rev1);
3242  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3243  return __ret;
3244}
3245#endif
3246
3247#ifdef __LITTLE_ENDIAN__
3248__ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1) {
3249  uint16x4_t __ret;
3250  __ret = (uint16x4_t)(__p0 <= __p1);
3251  return __ret;
3252}
3253#else
3254__ai uint16x4_t vcle_u16(uint16x4_t __p0, uint16x4_t __p1) {
3255  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3256  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3257  uint16x4_t __ret;
3258  __ret = (uint16x4_t)(__rev0 <= __rev1);
3259  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3260  return __ret;
3261}
3262#endif
3263
3264#ifdef __LITTLE_ENDIAN__
3265__ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1) {
3266  uint8x8_t __ret;
3267  __ret = (uint8x8_t)(__p0 <= __p1);
3268  return __ret;
3269}
3270#else
3271__ai uint8x8_t vcle_s8(int8x8_t __p0, int8x8_t __p1) {
3272  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3273  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3274  uint8x8_t __ret;
3275  __ret = (uint8x8_t)(__rev0 <= __rev1);
3276  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3277  return __ret;
3278}
3279#endif
3280
3281#ifdef __LITTLE_ENDIAN__
3282__ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1) {
3283  uint32x2_t __ret;
3284  __ret = (uint32x2_t)(__p0 <= __p1);
3285  return __ret;
3286}
3287#else
3288__ai uint32x2_t vcle_f32(float32x2_t __p0, float32x2_t __p1) {
3289  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3290  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3291  uint32x2_t __ret;
3292  __ret = (uint32x2_t)(__rev0 <= __rev1);
3293  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3294  return __ret;
3295}
3296#endif
3297
3298#ifdef __LITTLE_ENDIAN__
3299__ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1) {
3300  uint32x2_t __ret;
3301  __ret = (uint32x2_t)(__p0 <= __p1);
3302  return __ret;
3303}
3304#else
3305__ai uint32x2_t vcle_s32(int32x2_t __p0, int32x2_t __p1) {
3306  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3307  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3308  uint32x2_t __ret;
3309  __ret = (uint32x2_t)(__rev0 <= __rev1);
3310  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3311  return __ret;
3312}
3313#endif
3314
3315#ifdef __LITTLE_ENDIAN__
3316__ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) {
3317  uint16x4_t __ret;
3318  __ret = (uint16x4_t)(__p0 <= __p1);
3319  return __ret;
3320}
3321#else
3322__ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) {
3323  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3324  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3325  uint16x4_t __ret;
3326  __ret = (uint16x4_t)(__rev0 <= __rev1);
3327  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3328  return __ret;
3329}
3330#endif
3331
3332#ifdef __LITTLE_ENDIAN__
3333__ai int8x16_t vclsq_s8(int8x16_t __p0) {
3334  int8x16_t __ret;
3335  __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 32);
3336  return __ret;
3337}
3338#else
3339__ai int8x16_t vclsq_s8(int8x16_t __p0) {
3340  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3341  int8x16_t __ret;
3342  __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 32);
3343  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3344  return __ret;
3345}
3346#endif
3347
3348#ifdef __LITTLE_ENDIAN__
3349__ai int32x4_t vclsq_s32(int32x4_t __p0) {
3350  int32x4_t __ret;
3351  __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 34);
3352  return __ret;
3353}
3354#else
3355__ai int32x4_t vclsq_s32(int32x4_t __p0) {
3356  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3357  int32x4_t __ret;
3358  __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 34);
3359  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3360  return __ret;
3361}
3362#endif
3363
3364#ifdef __LITTLE_ENDIAN__
3365__ai int16x8_t vclsq_s16(int16x8_t __p0) {
3366  int16x8_t __ret;
3367  __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 33);
3368  return __ret;
3369}
3370#else
3371__ai int16x8_t vclsq_s16(int16x8_t __p0) {
3372  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3373  int16x8_t __ret;
3374  __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 33);
3375  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3376  return __ret;
3377}
3378#endif
3379
3380#ifdef __LITTLE_ENDIAN__
3381__ai int8x8_t vcls_s8(int8x8_t __p0) {
3382  int8x8_t __ret;
3383  __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__p0, 0);
3384  return __ret;
3385}
3386#else
3387__ai int8x8_t vcls_s8(int8x8_t __p0) {
3388  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3389  int8x8_t __ret;
3390  __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 0);
3391  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3392  return __ret;
3393}
3394#endif
3395
3396#ifdef __LITTLE_ENDIAN__
3397__ai int32x2_t vcls_s32(int32x2_t __p0) {
3398  int32x2_t __ret;
3399  __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__p0, 2);
3400  return __ret;
3401}
3402#else
3403__ai int32x2_t vcls_s32(int32x2_t __p0) {
3404  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3405  int32x2_t __ret;
3406  __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 2);
3407  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3408  return __ret;
3409}
3410#endif
3411
3412#ifdef __LITTLE_ENDIAN__
3413__ai int16x4_t vcls_s16(int16x4_t __p0) {
3414  int16x4_t __ret;
3415  __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__p0, 1);
3416  return __ret;
3417}
3418#else
3419__ai int16x4_t vcls_s16(int16x4_t __p0) {
3420  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3421  int16x4_t __ret;
3422  __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 1);
3423  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3424  return __ret;
3425}
3426#endif
3427
3428#ifdef __LITTLE_ENDIAN__
3429__ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3430  uint8x16_t __ret;
3431  __ret = (uint8x16_t)(__p0 < __p1);
3432  return __ret;
3433}
3434#else
3435__ai uint8x16_t vcltq_u8(uint8x16_t __p0, uint8x16_t __p1) {
3436  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3437  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3438  uint8x16_t __ret;
3439  __ret = (uint8x16_t)(__rev0 < __rev1);
3440  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3441  return __ret;
3442}
3443#endif
3444
3445#ifdef __LITTLE_ENDIAN__
3446__ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3447  uint32x4_t __ret;
3448  __ret = (uint32x4_t)(__p0 < __p1);
3449  return __ret;
3450}
3451#else
3452__ai uint32x4_t vcltq_u32(uint32x4_t __p0, uint32x4_t __p1) {
3453  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3454  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3455  uint32x4_t __ret;
3456  __ret = (uint32x4_t)(__rev0 < __rev1);
3457  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3458  return __ret;
3459}
3460#endif
3461
3462#ifdef __LITTLE_ENDIAN__
3463__ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3464  uint16x8_t __ret;
3465  __ret = (uint16x8_t)(__p0 < __p1);
3466  return __ret;
3467}
3468#else
3469__ai uint16x8_t vcltq_u16(uint16x8_t __p0, uint16x8_t __p1) {
3470  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3471  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3472  uint16x8_t __ret;
3473  __ret = (uint16x8_t)(__rev0 < __rev1);
3474  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3475  return __ret;
3476}
3477#endif
3478
3479#ifdef __LITTLE_ENDIAN__
3480__ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1) {
3481  uint8x16_t __ret;
3482  __ret = (uint8x16_t)(__p0 < __p1);
3483  return __ret;
3484}
3485#else
3486__ai uint8x16_t vcltq_s8(int8x16_t __p0, int8x16_t __p1) {
3487  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3488  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3489  uint8x16_t __ret;
3490  __ret = (uint8x16_t)(__rev0 < __rev1);
3491  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3492  return __ret;
3493}
3494#endif
3495
3496#ifdef __LITTLE_ENDIAN__
3497__ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1) {
3498  uint32x4_t __ret;
3499  __ret = (uint32x4_t)(__p0 < __p1);
3500  return __ret;
3501}
3502#else
3503__ai uint32x4_t vcltq_f32(float32x4_t __p0, float32x4_t __p1) {
3504  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3505  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3506  uint32x4_t __ret;
3507  __ret = (uint32x4_t)(__rev0 < __rev1);
3508  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3509  return __ret;
3510}
3511#endif
3512
3513#ifdef __LITTLE_ENDIAN__
3514__ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1) {
3515  uint32x4_t __ret;
3516  __ret = (uint32x4_t)(__p0 < __p1);
3517  return __ret;
3518}
3519#else
3520__ai uint32x4_t vcltq_s32(int32x4_t __p0, int32x4_t __p1) {
3521  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3522  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3523  uint32x4_t __ret;
3524  __ret = (uint32x4_t)(__rev0 < __rev1);
3525  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3526  return __ret;
3527}
3528#endif
3529
3530#ifdef __LITTLE_ENDIAN__
3531__ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1) {
3532  uint16x8_t __ret;
3533  __ret = (uint16x8_t)(__p0 < __p1);
3534  return __ret;
3535}
3536#else
3537__ai uint16x8_t vcltq_s16(int16x8_t __p0, int16x8_t __p1) {
3538  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3539  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3540  uint16x8_t __ret;
3541  __ret = (uint16x8_t)(__rev0 < __rev1);
3542  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3543  return __ret;
3544}
3545#endif
3546
3547#ifdef __LITTLE_ENDIAN__
3548__ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1) {
3549  uint8x8_t __ret;
3550  __ret = (uint8x8_t)(__p0 < __p1);
3551  return __ret;
3552}
3553#else
3554__ai uint8x8_t vclt_u8(uint8x8_t __p0, uint8x8_t __p1) {
3555  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3556  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3557  uint8x8_t __ret;
3558  __ret = (uint8x8_t)(__rev0 < __rev1);
3559  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3560  return __ret;
3561}
3562#endif
3563
3564#ifdef __LITTLE_ENDIAN__
3565__ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1) {
3566  uint32x2_t __ret;
3567  __ret = (uint32x2_t)(__p0 < __p1);
3568  return __ret;
3569}
3570#else
3571__ai uint32x2_t vclt_u32(uint32x2_t __p0, uint32x2_t __p1) {
3572  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3573  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3574  uint32x2_t __ret;
3575  __ret = (uint32x2_t)(__rev0 < __rev1);
3576  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3577  return __ret;
3578}
3579#endif
3580
3581#ifdef __LITTLE_ENDIAN__
3582__ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3583  uint16x4_t __ret;
3584  __ret = (uint16x4_t)(__p0 < __p1);
3585  return __ret;
3586}
3587#else
3588__ai uint16x4_t vclt_u16(uint16x4_t __p0, uint16x4_t __p1) {
3589  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3590  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3591  uint16x4_t __ret;
3592  __ret = (uint16x4_t)(__rev0 < __rev1);
3593  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3594  return __ret;
3595}
3596#endif
3597
3598#ifdef __LITTLE_ENDIAN__
3599__ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1) {
3600  uint8x8_t __ret;
3601  __ret = (uint8x8_t)(__p0 < __p1);
3602  return __ret;
3603}
3604#else
3605__ai uint8x8_t vclt_s8(int8x8_t __p0, int8x8_t __p1) {
3606  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3607  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3608  uint8x8_t __ret;
3609  __ret = (uint8x8_t)(__rev0 < __rev1);
3610  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3611  return __ret;
3612}
3613#endif
3614
3615#ifdef __LITTLE_ENDIAN__
3616__ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1) {
3617  uint32x2_t __ret;
3618  __ret = (uint32x2_t)(__p0 < __p1);
3619  return __ret;
3620}
3621#else
3622__ai uint32x2_t vclt_f32(float32x2_t __p0, float32x2_t __p1) {
3623  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3624  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3625  uint32x2_t __ret;
3626  __ret = (uint32x2_t)(__rev0 < __rev1);
3627  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3628  return __ret;
3629}
3630#endif
3631
3632#ifdef __LITTLE_ENDIAN__
3633__ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1) {
3634  uint32x2_t __ret;
3635  __ret = (uint32x2_t)(__p0 < __p1);
3636  return __ret;
3637}
3638#else
3639__ai uint32x2_t vclt_s32(int32x2_t __p0, int32x2_t __p1) {
3640  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3641  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
3642  uint32x2_t __ret;
3643  __ret = (uint32x2_t)(__rev0 < __rev1);
3644  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3645  return __ret;
3646}
3647#endif
3648
3649#ifdef __LITTLE_ENDIAN__
3650__ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1) {
3651  uint16x4_t __ret;
3652  __ret = (uint16x4_t)(__p0 < __p1);
3653  return __ret;
3654}
3655#else
3656__ai uint16x4_t vclt_s16(int16x4_t __p0, int16x4_t __p1) {
3657  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3658  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3659  uint16x4_t __ret;
3660  __ret = (uint16x4_t)(__rev0 < __rev1);
3661  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3662  return __ret;
3663}
3664#endif
3665
3666#ifdef __LITTLE_ENDIAN__
3667__ai uint8x16_t vclzq_u8(uint8x16_t __p0) {
3668  uint8x16_t __ret;
3669  __ret = (uint8x16_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 48);
3670  return __ret;
3671}
3672#else
3673__ai uint8x16_t vclzq_u8(uint8x16_t __p0) {
3674  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3675  uint8x16_t __ret;
3676  __ret = (uint8x16_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 48);
3677  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3678  return __ret;
3679}
3680#endif
3681
3682#ifdef __LITTLE_ENDIAN__
3683__ai uint32x4_t vclzq_u32(uint32x4_t __p0) {
3684  uint32x4_t __ret;
3685  __ret = (uint32x4_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 50);
3686  return __ret;
3687}
3688#else
3689__ai uint32x4_t vclzq_u32(uint32x4_t __p0) {
3690  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3691  uint32x4_t __ret;
3692  __ret = (uint32x4_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 50);
3693  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3694  return __ret;
3695}
3696#endif
3697
3698#ifdef __LITTLE_ENDIAN__
3699__ai uint16x8_t vclzq_u16(uint16x8_t __p0) {
3700  uint16x8_t __ret;
3701  __ret = (uint16x8_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 49);
3702  return __ret;
3703}
3704#else
3705__ai uint16x8_t vclzq_u16(uint16x8_t __p0) {
3706  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3707  uint16x8_t __ret;
3708  __ret = (uint16x8_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 49);
3709  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3710  return __ret;
3711}
3712#endif
3713
3714#ifdef __LITTLE_ENDIAN__
3715__ai int8x16_t vclzq_s8(int8x16_t __p0) {
3716  int8x16_t __ret;
3717  __ret = (int8x16_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 32);
3718  return __ret;
3719}
3720#else
3721__ai int8x16_t vclzq_s8(int8x16_t __p0) {
3722  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3723  int8x16_t __ret;
3724  __ret = (int8x16_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 32);
3725  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3726  return __ret;
3727}
3728#endif
3729
3730#ifdef __LITTLE_ENDIAN__
3731__ai int32x4_t vclzq_s32(int32x4_t __p0) {
3732  int32x4_t __ret;
3733  __ret = (int32x4_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 34);
3734  return __ret;
3735}
3736#else
3737__ai int32x4_t vclzq_s32(int32x4_t __p0) {
3738  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3739  int32x4_t __ret;
3740  __ret = (int32x4_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 34);
3741  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3742  return __ret;
3743}
3744#endif
3745
3746#ifdef __LITTLE_ENDIAN__
3747__ai int16x8_t vclzq_s16(int16x8_t __p0) {
3748  int16x8_t __ret;
3749  __ret = (int16x8_t) __builtin_neon_vclzq_v((int8x16_t)__p0, 33);
3750  return __ret;
3751}
3752#else
3753__ai int16x8_t vclzq_s16(int16x8_t __p0) {
3754  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3755  int16x8_t __ret;
3756  __ret = (int16x8_t) __builtin_neon_vclzq_v((int8x16_t)__rev0, 33);
3757  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3758  return __ret;
3759}
3760#endif
3761
3762#ifdef __LITTLE_ENDIAN__
3763__ai uint8x8_t vclz_u8(uint8x8_t __p0) {
3764  uint8x8_t __ret;
3765  __ret = (uint8x8_t) __builtin_neon_vclz_v((int8x8_t)__p0, 16);
3766  return __ret;
3767}
3768#else
3769__ai uint8x8_t vclz_u8(uint8x8_t __p0) {
3770  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3771  uint8x8_t __ret;
3772  __ret = (uint8x8_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 16);
3773  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3774  return __ret;
3775}
3776#endif
3777
3778#ifdef __LITTLE_ENDIAN__
3779__ai uint32x2_t vclz_u32(uint32x2_t __p0) {
3780  uint32x2_t __ret;
3781  __ret = (uint32x2_t) __builtin_neon_vclz_v((int8x8_t)__p0, 18);
3782  return __ret;
3783}
3784#else
3785__ai uint32x2_t vclz_u32(uint32x2_t __p0) {
3786  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3787  uint32x2_t __ret;
3788  __ret = (uint32x2_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 18);
3789  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3790  return __ret;
3791}
3792#endif
3793
3794#ifdef __LITTLE_ENDIAN__
3795__ai uint16x4_t vclz_u16(uint16x4_t __p0) {
3796  uint16x4_t __ret;
3797  __ret = (uint16x4_t) __builtin_neon_vclz_v((int8x8_t)__p0, 17);
3798  return __ret;
3799}
3800#else
3801__ai uint16x4_t vclz_u16(uint16x4_t __p0) {
3802  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3803  uint16x4_t __ret;
3804  __ret = (uint16x4_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 17);
3805  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3806  return __ret;
3807}
3808#endif
3809
3810#ifdef __LITTLE_ENDIAN__
3811__ai int8x8_t vclz_s8(int8x8_t __p0) {
3812  int8x8_t __ret;
3813  __ret = (int8x8_t) __builtin_neon_vclz_v((int8x8_t)__p0, 0);
3814  return __ret;
3815}
3816#else
3817__ai int8x8_t vclz_s8(int8x8_t __p0) {
3818  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3819  int8x8_t __ret;
3820  __ret = (int8x8_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 0);
3821  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3822  return __ret;
3823}
3824#endif
3825
3826#ifdef __LITTLE_ENDIAN__
3827__ai int32x2_t vclz_s32(int32x2_t __p0) {
3828  int32x2_t __ret;
3829  __ret = (int32x2_t) __builtin_neon_vclz_v((int8x8_t)__p0, 2);
3830  return __ret;
3831}
3832#else
3833__ai int32x2_t vclz_s32(int32x2_t __p0) {
3834  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
3835  int32x2_t __ret;
3836  __ret = (int32x2_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 2);
3837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
3838  return __ret;
3839}
3840#endif
3841
3842#ifdef __LITTLE_ENDIAN__
3843__ai int16x4_t vclz_s16(int16x4_t __p0) {
3844  int16x4_t __ret;
3845  __ret = (int16x4_t) __builtin_neon_vclz_v((int8x8_t)__p0, 1);
3846  return __ret;
3847}
3848#else
3849__ai int16x4_t vclz_s16(int16x4_t __p0) {
3850  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3851  int16x4_t __ret;
3852  __ret = (int16x4_t) __builtin_neon_vclz_v((int8x8_t)__rev0, 1);
3853  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
3854  return __ret;
3855}
3856#endif
3857
3858#ifdef __LITTLE_ENDIAN__
3859__ai poly8x8_t vcnt_p8(poly8x8_t __p0) {
3860  poly8x8_t __ret;
3861  __ret = (poly8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 4);
3862  return __ret;
3863}
3864#else
3865__ai poly8x8_t vcnt_p8(poly8x8_t __p0) {
3866  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3867  poly8x8_t __ret;
3868  __ret = (poly8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 4);
3869  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3870  return __ret;
3871}
3872#endif
3873
3874#ifdef __LITTLE_ENDIAN__
3875__ai poly8x16_t vcntq_p8(poly8x16_t __p0) {
3876  poly8x16_t __ret;
3877  __ret = (poly8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 36);
3878  return __ret;
3879}
3880#else
3881__ai poly8x16_t vcntq_p8(poly8x16_t __p0) {
3882  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3883  poly8x16_t __ret;
3884  __ret = (poly8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 36);
3885  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3886  return __ret;
3887}
3888#endif
3889
3890#ifdef __LITTLE_ENDIAN__
3891__ai uint8x16_t vcntq_u8(uint8x16_t __p0) {
3892  uint8x16_t __ret;
3893  __ret = (uint8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 48);
3894  return __ret;
3895}
3896#else
3897__ai uint8x16_t vcntq_u8(uint8x16_t __p0) {
3898  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3899  uint8x16_t __ret;
3900  __ret = (uint8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 48);
3901  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3902  return __ret;
3903}
3904#endif
3905
3906#ifdef __LITTLE_ENDIAN__
3907__ai int8x16_t vcntq_s8(int8x16_t __p0) {
3908  int8x16_t __ret;
3909  __ret = (int8x16_t) __builtin_neon_vcntq_v((int8x16_t)__p0, 32);
3910  return __ret;
3911}
3912#else
3913__ai int8x16_t vcntq_s8(int8x16_t __p0) {
3914  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3915  int8x16_t __ret;
3916  __ret = (int8x16_t) __builtin_neon_vcntq_v((int8x16_t)__rev0, 32);
3917  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3918  return __ret;
3919}
3920#endif
3921
3922#ifdef __LITTLE_ENDIAN__
3923__ai uint8x8_t vcnt_u8(uint8x8_t __p0) {
3924  uint8x8_t __ret;
3925  __ret = (uint8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 16);
3926  return __ret;
3927}
3928#else
3929__ai uint8x8_t vcnt_u8(uint8x8_t __p0) {
3930  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3931  uint8x8_t __ret;
3932  __ret = (uint8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 16);
3933  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3934  return __ret;
3935}
3936#endif
3937
3938#ifdef __LITTLE_ENDIAN__
3939__ai int8x8_t vcnt_s8(int8x8_t __p0) {
3940  int8x8_t __ret;
3941  __ret = (int8x8_t) __builtin_neon_vcnt_v((int8x8_t)__p0, 0);
3942  return __ret;
3943}
3944#else
3945__ai int8x8_t vcnt_s8(int8x8_t __p0) {
3946  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3947  int8x8_t __ret;
3948  __ret = (int8x8_t) __builtin_neon_vcnt_v((int8x8_t)__rev0, 0);
3949  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3950  return __ret;
3951}
3952#endif
3953
3954#ifdef __LITTLE_ENDIAN__
3955__ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1) {
3956  poly8x16_t __ret;
3957  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
3958  return __ret;
3959}
3960#else
3961__ai poly8x16_t vcombine_p8(poly8x8_t __p0, poly8x8_t __p1) {
3962  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3963  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3964  poly8x16_t __ret;
3965  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
3966  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
3967  return __ret;
3968}
3969#endif
3970
3971#ifdef __LITTLE_ENDIAN__
3972__ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1) {
3973  poly16x8_t __ret;
3974  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
3975  return __ret;
3976}
3977#else
3978__ai poly16x8_t vcombine_p16(poly16x4_t __p0, poly16x4_t __p1) {
3979  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
3980  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
3981  poly16x8_t __ret;
3982  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
3983  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
3984  return __ret;
3985}
3986#endif
3987
3988#ifdef __LITTLE_ENDIAN__
3989__ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
3990  uint8x16_t __ret;
3991  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
3992  return __ret;
3993}
3994#else
3995__ai uint8x16_t vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
3996  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
3997  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
3998  uint8x16_t __ret;
3999  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4000  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4001  return __ret;
4002}
4003__ai uint8x16_t __noswap_vcombine_u8(uint8x8_t __p0, uint8x8_t __p1) {
4004  uint8x16_t __ret;
4005  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4006  return __ret;
4007}
4008#endif
4009
4010#ifdef __LITTLE_ENDIAN__
4011__ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
4012  uint32x4_t __ret;
4013  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4014  return __ret;
4015}
4016#else
4017__ai uint32x4_t vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
4018  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4019  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4020  uint32x4_t __ret;
4021  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
4022  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4023  return __ret;
4024}
4025__ai uint32x4_t __noswap_vcombine_u32(uint32x2_t __p0, uint32x2_t __p1) {
4026  uint32x4_t __ret;
4027  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4028  return __ret;
4029}
4030#endif
4031
4032#ifdef __LITTLE_ENDIAN__
4033__ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1) {
4034  uint64x2_t __ret;
4035  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4036  return __ret;
4037}
4038#else
4039__ai uint64x2_t vcombine_u64(uint64x1_t __p0, uint64x1_t __p1) {
4040  uint64x2_t __ret;
4041  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4042  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4043  return __ret;
4044}
4045#endif
4046
4047#ifdef __LITTLE_ENDIAN__
4048__ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
4049  uint16x8_t __ret;
4050  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4051  return __ret;
4052}
4053#else
4054__ai uint16x8_t vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
4055  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4056  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4057  uint16x8_t __ret;
4058  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
4059  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4060  return __ret;
4061}
4062__ai uint16x8_t __noswap_vcombine_u16(uint16x4_t __p0, uint16x4_t __p1) {
4063  uint16x8_t __ret;
4064  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4065  return __ret;
4066}
4067#endif
4068
4069#ifdef __LITTLE_ENDIAN__
4070__ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
4071  int8x16_t __ret;
4072  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4073  return __ret;
4074}
4075#else
4076__ai int8x16_t vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
4077  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
4078  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
4079  int8x16_t __ret;
4080  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4081  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
4082  return __ret;
4083}
4084__ai int8x16_t __noswap_vcombine_s8(int8x8_t __p0, int8x8_t __p1) {
4085  int8x16_t __ret;
4086  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15);
4087  return __ret;
4088}
4089#endif
4090
4091#ifdef __LITTLE_ENDIAN__
4092__ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
4093  float32x4_t __ret;
4094  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4095  return __ret;
4096}
4097#else
4098__ai float32x4_t vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
4099  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4100  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4101  float32x4_t __ret;
4102  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
4103  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4104  return __ret;
4105}
4106__ai float32x4_t __noswap_vcombine_f32(float32x2_t __p0, float32x2_t __p1) {
4107  float32x4_t __ret;
4108  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4109  return __ret;
4110}
4111#endif
4112
4113#ifdef __LITTLE_ENDIAN__
4114__ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
4115  float16x8_t __ret;
4116  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4117  return __ret;
4118}
4119#else
4120__ai float16x8_t vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
4121  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4122  float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4123  float16x8_t __ret;
4124  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
4125  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4126  return __ret;
4127}
4128__ai float16x8_t __noswap_vcombine_f16(float16x4_t __p0, float16x4_t __p1) {
4129  float16x8_t __ret;
4130  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4131  return __ret;
4132}
4133#endif
4134
4135#ifdef __LITTLE_ENDIAN__
4136__ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
4137  int32x4_t __ret;
4138  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4139  return __ret;
4140}
4141#else
4142__ai int32x4_t vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
4143  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4144  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
4145  int32x4_t __ret;
4146  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3);
4147  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4148  return __ret;
4149}
4150__ai int32x4_t __noswap_vcombine_s32(int32x2_t __p0, int32x2_t __p1) {
4151  int32x4_t __ret;
4152  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3);
4153  return __ret;
4154}
4155#endif
4156
4157#ifdef __LITTLE_ENDIAN__
4158__ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1) {
4159  int64x2_t __ret;
4160  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4161  return __ret;
4162}
4163#else
4164__ai int64x2_t vcombine_s64(int64x1_t __p0, int64x1_t __p1) {
4165  int64x2_t __ret;
4166  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
4167  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4168  return __ret;
4169}
4170#endif
4171
4172#ifdef __LITTLE_ENDIAN__
4173__ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
4174  int16x8_t __ret;
4175  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4176  return __ret;
4177}
4178#else
4179__ai int16x8_t vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
4180  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4181  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
4182  int16x8_t __ret;
4183  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 1, 2, 3, 4, 5, 6, 7);
4184  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
4185  return __ret;
4186}
4187__ai int16x8_t __noswap_vcombine_s16(int16x4_t __p0, int16x4_t __p1) {
4188  int16x8_t __ret;
4189  __ret = __builtin_shufflevector(__p0, __p1, 0, 1, 2, 3, 4, 5, 6, 7);
4190  return __ret;
4191}
4192#endif
4193
4194#ifdef __LITTLE_ENDIAN__
4195__ai poly8x8_t vcreate_p8(uint64_t __p0) {
4196  poly8x8_t __ret;
4197  __ret = (poly8x8_t)(__p0);
4198  return __ret;
4199}
4200#else
4201__ai poly8x8_t vcreate_p8(uint64_t __p0) {
4202  poly8x8_t __ret;
4203  __ret = (poly8x8_t)(__p0);
4204  return __ret;
4205}
4206#endif
4207
4208#ifdef __LITTLE_ENDIAN__
4209__ai poly16x4_t vcreate_p16(uint64_t __p0) {
4210  poly16x4_t __ret;
4211  __ret = (poly16x4_t)(__p0);
4212  return __ret;
4213}
4214#else
4215__ai poly16x4_t vcreate_p16(uint64_t __p0) {
4216  poly16x4_t __ret;
4217  __ret = (poly16x4_t)(__p0);
4218  return __ret;
4219}
4220#endif
4221
4222#ifdef __LITTLE_ENDIAN__
4223__ai uint8x8_t vcreate_u8(uint64_t __p0) {
4224  uint8x8_t __ret;
4225  __ret = (uint8x8_t)(__p0);
4226  return __ret;
4227}
4228#else
4229__ai uint8x8_t vcreate_u8(uint64_t __p0) {
4230  uint8x8_t __ret;
4231  __ret = (uint8x8_t)(__p0);
4232  return __ret;
4233}
4234#endif
4235
4236#ifdef __LITTLE_ENDIAN__
4237__ai uint32x2_t vcreate_u32(uint64_t __p0) {
4238  uint32x2_t __ret;
4239  __ret = (uint32x2_t)(__p0);
4240  return __ret;
4241}
4242#else
4243__ai uint32x2_t vcreate_u32(uint64_t __p0) {
4244  uint32x2_t __ret;
4245  __ret = (uint32x2_t)(__p0);
4246  return __ret;
4247}
4248#endif
4249
4250#ifdef __LITTLE_ENDIAN__
4251__ai uint64x1_t vcreate_u64(uint64_t __p0) {
4252  uint64x1_t __ret;
4253  __ret = (uint64x1_t)(__p0);
4254  return __ret;
4255}
4256#else
4257__ai uint64x1_t vcreate_u64(uint64_t __p0) {
4258  uint64x1_t __ret;
4259  __ret = (uint64x1_t)(__p0);
4260  return __ret;
4261}
4262#endif
4263
4264#ifdef __LITTLE_ENDIAN__
4265__ai uint16x4_t vcreate_u16(uint64_t __p0) {
4266  uint16x4_t __ret;
4267  __ret = (uint16x4_t)(__p0);
4268  return __ret;
4269}
4270#else
4271__ai uint16x4_t vcreate_u16(uint64_t __p0) {
4272  uint16x4_t __ret;
4273  __ret = (uint16x4_t)(__p0);
4274  return __ret;
4275}
4276#endif
4277
4278#ifdef __LITTLE_ENDIAN__
4279__ai int8x8_t vcreate_s8(uint64_t __p0) {
4280  int8x8_t __ret;
4281  __ret = (int8x8_t)(__p0);
4282  return __ret;
4283}
4284#else
4285__ai int8x8_t vcreate_s8(uint64_t __p0) {
4286  int8x8_t __ret;
4287  __ret = (int8x8_t)(__p0);
4288  return __ret;
4289}
4290#endif
4291
4292#ifdef __LITTLE_ENDIAN__
4293__ai float32x2_t vcreate_f32(uint64_t __p0) {
4294  float32x2_t __ret;
4295  __ret = (float32x2_t)(__p0);
4296  return __ret;
4297}
4298#else
4299__ai float32x2_t vcreate_f32(uint64_t __p0) {
4300  float32x2_t __ret;
4301  __ret = (float32x2_t)(__p0);
4302  return __ret;
4303}
4304#endif
4305
4306#ifdef __LITTLE_ENDIAN__
4307__ai float16x4_t vcreate_f16(uint64_t __p0) {
4308  float16x4_t __ret;
4309  __ret = (float16x4_t)(__p0);
4310  return __ret;
4311}
4312#else
4313__ai float16x4_t vcreate_f16(uint64_t __p0) {
4314  float16x4_t __ret;
4315  __ret = (float16x4_t)(__p0);
4316  return __ret;
4317}
4318#endif
4319
4320#ifdef __LITTLE_ENDIAN__
4321__ai int32x2_t vcreate_s32(uint64_t __p0) {
4322  int32x2_t __ret;
4323  __ret = (int32x2_t)(__p0);
4324  return __ret;
4325}
4326#else
4327__ai int32x2_t vcreate_s32(uint64_t __p0) {
4328  int32x2_t __ret;
4329  __ret = (int32x2_t)(__p0);
4330  return __ret;
4331}
4332#endif
4333
4334#ifdef __LITTLE_ENDIAN__
4335__ai int64x1_t vcreate_s64(uint64_t __p0) {
4336  int64x1_t __ret;
4337  __ret = (int64x1_t)(__p0);
4338  return __ret;
4339}
4340#else
4341__ai int64x1_t vcreate_s64(uint64_t __p0) {
4342  int64x1_t __ret;
4343  __ret = (int64x1_t)(__p0);
4344  return __ret;
4345}
4346#endif
4347
4348#ifdef __LITTLE_ENDIAN__
4349__ai int16x4_t vcreate_s16(uint64_t __p0) {
4350  int16x4_t __ret;
4351  __ret = (int16x4_t)(__p0);
4352  return __ret;
4353}
4354#else
4355__ai int16x4_t vcreate_s16(uint64_t __p0) {
4356  int16x4_t __ret;
4357  __ret = (int16x4_t)(__p0);
4358  return __ret;
4359}
4360#endif
4361
4362#ifdef __LITTLE_ENDIAN__
4363__ai float16x4_t vcvt_f16_f32(float32x4_t __p0) {
4364  float16x4_t __ret;
4365  __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__p0, 8);
4366  return __ret;
4367}
4368#else
4369__ai float16x4_t vcvt_f16_f32(float32x4_t __p0) {
4370  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4371  float16x4_t __ret;
4372  __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__rev0, 8);
4373  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4374  return __ret;
4375}
4376__ai float16x4_t __noswap_vcvt_f16_f32(float32x4_t __p0) {
4377  float16x4_t __ret;
4378  __ret = (float16x4_t) __builtin_neon_vcvt_f16_f32((int8x16_t)__p0, 8);
4379  return __ret;
4380}
4381#endif
4382
4383#ifdef __LITTLE_ENDIAN__
4384__ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0) {
4385  float32x4_t __ret;
4386  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__p0, 50);
4387  return __ret;
4388}
4389#else
4390__ai float32x4_t vcvtq_f32_u32(uint32x4_t __p0) {
4391  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4392  float32x4_t __ret;
4393  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__rev0, 50);
4394  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4395  return __ret;
4396}
4397#endif
4398
4399#ifdef __LITTLE_ENDIAN__
4400__ai float32x4_t vcvtq_f32_s32(int32x4_t __p0) {
4401  float32x4_t __ret;
4402  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__p0, 34);
4403  return __ret;
4404}
4405#else
4406__ai float32x4_t vcvtq_f32_s32(int32x4_t __p0) {
4407  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4408  float32x4_t __ret;
4409  __ret = (float32x4_t) __builtin_neon_vcvtq_f32_v((int8x16_t)__rev0, 34);
4410  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4411  return __ret;
4412}
4413#endif
4414
4415#ifdef __LITTLE_ENDIAN__
4416__ai float32x2_t vcvt_f32_u32(uint32x2_t __p0) {
4417  float32x2_t __ret;
4418  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__p0, 18);
4419  return __ret;
4420}
4421#else
4422__ai float32x2_t vcvt_f32_u32(uint32x2_t __p0) {
4423  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4424  float32x2_t __ret;
4425  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__rev0, 18);
4426  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4427  return __ret;
4428}
4429#endif
4430
4431#ifdef __LITTLE_ENDIAN__
4432__ai float32x2_t vcvt_f32_s32(int32x2_t __p0) {
4433  float32x2_t __ret;
4434  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__p0, 2);
4435  return __ret;
4436}
4437#else
4438__ai float32x2_t vcvt_f32_s32(int32x2_t __p0) {
4439  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4440  float32x2_t __ret;
4441  __ret = (float32x2_t) __builtin_neon_vcvt_f32_v((int8x8_t)__rev0, 2);
4442  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4443  return __ret;
4444}
4445#endif
4446
4447#ifdef __LITTLE_ENDIAN__
4448__ai float32x4_t vcvt_f32_f16(float16x4_t __p0) {
4449  float32x4_t __ret;
4450  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__p0, 41);
4451  return __ret;
4452}
4453#else
4454__ai float32x4_t vcvt_f32_f16(float16x4_t __p0) {
4455  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4456  float32x4_t __ret;
4457  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__rev0, 41);
4458  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4459  return __ret;
4460}
4461__ai float32x4_t __noswap_vcvt_f32_f16(float16x4_t __p0) {
4462  float32x4_t __ret;
4463  __ret = (float32x4_t) __builtin_neon_vcvt_f32_f16((int8x8_t)__p0, 41);
4464  return __ret;
4465}
4466#endif
4467
4468#ifdef __LITTLE_ENDIAN__
4469#define vcvtq_n_f32_u32(__p0, __p1) __extension__ ({ \
4470  uint32x4_t __s0 = __p0; \
4471  float32x4_t __ret; \
4472  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__s0, __p1, 50); \
4473  __ret; \
4474})
4475#else
4476#define vcvtq_n_f32_u32(__p0, __p1) __extension__ ({ \
4477  uint32x4_t __s0 = __p0; \
4478  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4479  float32x4_t __ret; \
4480  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__rev0, __p1, 50); \
4481  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4482  __ret; \
4483})
4484#endif
4485
4486#ifdef __LITTLE_ENDIAN__
4487#define vcvtq_n_f32_s32(__p0, __p1) __extension__ ({ \
4488  int32x4_t __s0 = __p0; \
4489  float32x4_t __ret; \
4490  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__s0, __p1, 34); \
4491  __ret; \
4492})
4493#else
4494#define vcvtq_n_f32_s32(__p0, __p1) __extension__ ({ \
4495  int32x4_t __s0 = __p0; \
4496  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4497  float32x4_t __ret; \
4498  __ret = (float32x4_t) __builtin_neon_vcvtq_n_f32_v((int8x16_t)__rev0, __p1, 34); \
4499  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4500  __ret; \
4501})
4502#endif
4503
4504#ifdef __LITTLE_ENDIAN__
4505#define vcvt_n_f32_u32(__p0, __p1) __extension__ ({ \
4506  uint32x2_t __s0 = __p0; \
4507  float32x2_t __ret; \
4508  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__s0, __p1, 18); \
4509  __ret; \
4510})
4511#else
4512#define vcvt_n_f32_u32(__p0, __p1) __extension__ ({ \
4513  uint32x2_t __s0 = __p0; \
4514  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4515  float32x2_t __ret; \
4516  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__rev0, __p1, 18); \
4517  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4518  __ret; \
4519})
4520#endif
4521
4522#ifdef __LITTLE_ENDIAN__
4523#define vcvt_n_f32_s32(__p0, __p1) __extension__ ({ \
4524  int32x2_t __s0 = __p0; \
4525  float32x2_t __ret; \
4526  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__s0, __p1, 2); \
4527  __ret; \
4528})
4529#else
4530#define vcvt_n_f32_s32(__p0, __p1) __extension__ ({ \
4531  int32x2_t __s0 = __p0; \
4532  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4533  float32x2_t __ret; \
4534  __ret = (float32x2_t) __builtin_neon_vcvt_n_f32_v((int8x8_t)__rev0, __p1, 2); \
4535  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4536  __ret; \
4537})
4538#endif
4539
4540#ifdef __LITTLE_ENDIAN__
4541#define vcvtq_n_s32_f32(__p0, __p1) __extension__ ({ \
4542  float32x4_t __s0 = __p0; \
4543  int32x4_t __ret; \
4544  __ret = (int32x4_t) __builtin_neon_vcvtq_n_s32_v((int8x16_t)__s0, __p1, 34); \
4545  __ret; \
4546})
4547#else
4548#define vcvtq_n_s32_f32(__p0, __p1) __extension__ ({ \
4549  float32x4_t __s0 = __p0; \
4550  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4551  int32x4_t __ret; \
4552  __ret = (int32x4_t) __builtin_neon_vcvtq_n_s32_v((int8x16_t)__rev0, __p1, 34); \
4553  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4554  __ret; \
4555})
4556#endif
4557
4558#ifdef __LITTLE_ENDIAN__
4559#define vcvt_n_s32_f32(__p0, __p1) __extension__ ({ \
4560  float32x2_t __s0 = __p0; \
4561  int32x2_t __ret; \
4562  __ret = (int32x2_t) __builtin_neon_vcvt_n_s32_v((int8x8_t)__s0, __p1, 2); \
4563  __ret; \
4564})
4565#else
4566#define vcvt_n_s32_f32(__p0, __p1) __extension__ ({ \
4567  float32x2_t __s0 = __p0; \
4568  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4569  int32x2_t __ret; \
4570  __ret = (int32x2_t) __builtin_neon_vcvt_n_s32_v((int8x8_t)__rev0, __p1, 2); \
4571  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4572  __ret; \
4573})
4574#endif
4575
4576#ifdef __LITTLE_ENDIAN__
4577#define vcvtq_n_u32_f32(__p0, __p1) __extension__ ({ \
4578  float32x4_t __s0 = __p0; \
4579  uint32x4_t __ret; \
4580  __ret = (uint32x4_t) __builtin_neon_vcvtq_n_u32_v((int8x16_t)__s0, __p1, 50); \
4581  __ret; \
4582})
4583#else
4584#define vcvtq_n_u32_f32(__p0, __p1) __extension__ ({ \
4585  float32x4_t __s0 = __p0; \
4586  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4587  uint32x4_t __ret; \
4588  __ret = (uint32x4_t) __builtin_neon_vcvtq_n_u32_v((int8x16_t)__rev0, __p1, 50); \
4589  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4590  __ret; \
4591})
4592#endif
4593
4594#ifdef __LITTLE_ENDIAN__
4595#define vcvt_n_u32_f32(__p0, __p1) __extension__ ({ \
4596  float32x2_t __s0 = __p0; \
4597  uint32x2_t __ret; \
4598  __ret = (uint32x2_t) __builtin_neon_vcvt_n_u32_v((int8x8_t)__s0, __p1, 18); \
4599  __ret; \
4600})
4601#else
4602#define vcvt_n_u32_f32(__p0, __p1) __extension__ ({ \
4603  float32x2_t __s0 = __p0; \
4604  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4605  uint32x2_t __ret; \
4606  __ret = (uint32x2_t) __builtin_neon_vcvt_n_u32_v((int8x8_t)__rev0, __p1, 18); \
4607  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4608  __ret; \
4609})
4610#endif
4611
4612#ifdef __LITTLE_ENDIAN__
4613__ai int32x4_t vcvtq_s32_f32(float32x4_t __p0) {
4614  int32x4_t __ret;
4615  __ret = (int32x4_t) __builtin_neon_vcvtq_s32_v((int8x16_t)__p0, 34);
4616  return __ret;
4617}
4618#else
4619__ai int32x4_t vcvtq_s32_f32(float32x4_t __p0) {
4620  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4621  int32x4_t __ret;
4622  __ret = (int32x4_t) __builtin_neon_vcvtq_s32_v((int8x16_t)__rev0, 34);
4623  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4624  return __ret;
4625}
4626#endif
4627
4628#ifdef __LITTLE_ENDIAN__
4629__ai int32x2_t vcvt_s32_f32(float32x2_t __p0) {
4630  int32x2_t __ret;
4631  __ret = (int32x2_t) __builtin_neon_vcvt_s32_v((int8x8_t)__p0, 2);
4632  return __ret;
4633}
4634#else
4635__ai int32x2_t vcvt_s32_f32(float32x2_t __p0) {
4636  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4637  int32x2_t __ret;
4638  __ret = (int32x2_t) __builtin_neon_vcvt_s32_v((int8x8_t)__rev0, 2);
4639  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4640  return __ret;
4641}
4642#endif
4643
4644#ifdef __LITTLE_ENDIAN__
4645__ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0) {
4646  uint32x4_t __ret;
4647  __ret = (uint32x4_t) __builtin_neon_vcvtq_u32_v((int8x16_t)__p0, 50);
4648  return __ret;
4649}
4650#else
4651__ai uint32x4_t vcvtq_u32_f32(float32x4_t __p0) {
4652  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
4653  uint32x4_t __ret;
4654  __ret = (uint32x4_t) __builtin_neon_vcvtq_u32_v((int8x16_t)__rev0, 50);
4655  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
4656  return __ret;
4657}
4658#endif
4659
4660#ifdef __LITTLE_ENDIAN__
4661__ai uint32x2_t vcvt_u32_f32(float32x2_t __p0) {
4662  uint32x2_t __ret;
4663  __ret = (uint32x2_t) __builtin_neon_vcvt_u32_v((int8x8_t)__p0, 18);
4664  return __ret;
4665}
4666#else
4667__ai uint32x2_t vcvt_u32_f32(float32x2_t __p0) {
4668  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
4669  uint32x2_t __ret;
4670  __ret = (uint32x2_t) __builtin_neon_vcvt_u32_v((int8x8_t)__rev0, 18);
4671  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
4672  return __ret;
4673}
4674#endif
4675
4676#ifdef __LITTLE_ENDIAN__
4677#define vdup_lane_p8(__p0, __p1) __extension__ ({ \
4678  poly8x8_t __s0 = __p0; \
4679  poly8x8_t __ret; \
4680  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4681  __ret; \
4682})
4683#else
4684#define vdup_lane_p8(__p0, __p1) __extension__ ({ \
4685  poly8x8_t __s0 = __p0; \
4686  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4687  poly8x8_t __ret; \
4688  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4689  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4690  __ret; \
4691})
4692#endif
4693
4694#ifdef __LITTLE_ENDIAN__
4695#define vdup_lane_p16(__p0, __p1) __extension__ ({ \
4696  poly16x4_t __s0 = __p0; \
4697  poly16x4_t __ret; \
4698  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4699  __ret; \
4700})
4701#else
4702#define vdup_lane_p16(__p0, __p1) __extension__ ({ \
4703  poly16x4_t __s0 = __p0; \
4704  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4705  poly16x4_t __ret; \
4706  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4707  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4708  __ret; \
4709})
4710#endif
4711
4712#ifdef __LITTLE_ENDIAN__
4713#define vdupq_lane_p8(__p0, __p1) __extension__ ({ \
4714  poly8x8_t __s0 = __p0; \
4715  poly8x16_t __ret; \
4716  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4717  __ret; \
4718})
4719#else
4720#define vdupq_lane_p8(__p0, __p1) __extension__ ({ \
4721  poly8x8_t __s0 = __p0; \
4722  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4723  poly8x16_t __ret; \
4724  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4725  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
4726  __ret; \
4727})
4728#endif
4729
4730#ifdef __LITTLE_ENDIAN__
4731#define vdupq_lane_p16(__p0, __p1) __extension__ ({ \
4732  poly16x4_t __s0 = __p0; \
4733  poly16x8_t __ret; \
4734  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4735  __ret; \
4736})
4737#else
4738#define vdupq_lane_p16(__p0, __p1) __extension__ ({ \
4739  poly16x4_t __s0 = __p0; \
4740  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4741  poly16x8_t __ret; \
4742  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4743  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4744  __ret; \
4745})
4746#endif
4747
4748#ifdef __LITTLE_ENDIAN__
4749#define vdupq_lane_u8(__p0, __p1) __extension__ ({ \
4750  uint8x8_t __s0 = __p0; \
4751  uint8x16_t __ret; \
4752  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4753  __ret; \
4754})
4755#else
4756#define vdupq_lane_u8(__p0, __p1) __extension__ ({ \
4757  uint8x8_t __s0 = __p0; \
4758  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4759  uint8x16_t __ret; \
4760  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4761  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
4762  __ret; \
4763})
4764#endif
4765
4766#ifdef __LITTLE_ENDIAN__
4767#define vdupq_lane_u32(__p0, __p1) __extension__ ({ \
4768  uint32x2_t __s0 = __p0; \
4769  uint32x4_t __ret; \
4770  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4771  __ret; \
4772})
4773#else
4774#define vdupq_lane_u32(__p0, __p1) __extension__ ({ \
4775  uint32x2_t __s0 = __p0; \
4776  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4777  uint32x4_t __ret; \
4778  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4779  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4780  __ret; \
4781})
4782#endif
4783
4784#ifdef __LITTLE_ENDIAN__
4785#define vdupq_lane_u64(__p0, __p1) __extension__ ({ \
4786  uint64x1_t __s0 = __p0; \
4787  uint64x2_t __ret; \
4788  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4789  __ret; \
4790})
4791#else
4792#define vdupq_lane_u64(__p0, __p1) __extension__ ({ \
4793  uint64x1_t __s0 = __p0; \
4794  uint64x2_t __ret; \
4795  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4796  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4797  __ret; \
4798})
4799#endif
4800
4801#ifdef __LITTLE_ENDIAN__
4802#define vdupq_lane_u16(__p0, __p1) __extension__ ({ \
4803  uint16x4_t __s0 = __p0; \
4804  uint16x8_t __ret; \
4805  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4806  __ret; \
4807})
4808#else
4809#define vdupq_lane_u16(__p0, __p1) __extension__ ({ \
4810  uint16x4_t __s0 = __p0; \
4811  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4812  uint16x8_t __ret; \
4813  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4814  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4815  __ret; \
4816})
4817#endif
4818
4819#ifdef __LITTLE_ENDIAN__
4820#define vdupq_lane_s8(__p0, __p1) __extension__ ({ \
4821  int8x8_t __s0 = __p0; \
4822  int8x16_t __ret; \
4823  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4824  __ret; \
4825})
4826#else
4827#define vdupq_lane_s8(__p0, __p1) __extension__ ({ \
4828  int8x8_t __s0 = __p0; \
4829  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4830  int8x16_t __ret; \
4831  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4832  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
4833  __ret; \
4834})
4835#endif
4836
4837#ifdef __LITTLE_ENDIAN__
4838#define vdupq_lane_f32(__p0, __p1) __extension__ ({ \
4839  float32x2_t __s0 = __p0; \
4840  float32x4_t __ret; \
4841  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4842  __ret; \
4843})
4844#else
4845#define vdupq_lane_f32(__p0, __p1) __extension__ ({ \
4846  float32x2_t __s0 = __p0; \
4847  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4848  float32x4_t __ret; \
4849  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4850  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4851  __ret; \
4852})
4853#endif
4854
4855#ifdef __LITTLE_ENDIAN__
4856#define vdupq_lane_s32(__p0, __p1) __extension__ ({ \
4857  int32x2_t __s0 = __p0; \
4858  int32x4_t __ret; \
4859  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4860  __ret; \
4861})
4862#else
4863#define vdupq_lane_s32(__p0, __p1) __extension__ ({ \
4864  int32x2_t __s0 = __p0; \
4865  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4866  int32x4_t __ret; \
4867  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4868  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4869  __ret; \
4870})
4871#endif
4872
4873#ifdef __LITTLE_ENDIAN__
4874#define vdupq_lane_s64(__p0, __p1) __extension__ ({ \
4875  int64x1_t __s0 = __p0; \
4876  int64x2_t __ret; \
4877  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4878  __ret; \
4879})
4880#else
4881#define vdupq_lane_s64(__p0, __p1) __extension__ ({ \
4882  int64x1_t __s0 = __p0; \
4883  int64x2_t __ret; \
4884  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4885  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4886  __ret; \
4887})
4888#endif
4889
4890#ifdef __LITTLE_ENDIAN__
4891#define vdupq_lane_s16(__p0, __p1) __extension__ ({ \
4892  int16x4_t __s0 = __p0; \
4893  int16x8_t __ret; \
4894  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4895  __ret; \
4896})
4897#else
4898#define vdupq_lane_s16(__p0, __p1) __extension__ ({ \
4899  int16x4_t __s0 = __p0; \
4900  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4901  int16x8_t __ret; \
4902  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4903  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4904  __ret; \
4905})
4906#endif
4907
4908#ifdef __LITTLE_ENDIAN__
4909#define vdup_lane_u8(__p0, __p1) __extension__ ({ \
4910  uint8x8_t __s0 = __p0; \
4911  uint8x8_t __ret; \
4912  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4913  __ret; \
4914})
4915#else
4916#define vdup_lane_u8(__p0, __p1) __extension__ ({ \
4917  uint8x8_t __s0 = __p0; \
4918  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4919  uint8x8_t __ret; \
4920  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4921  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4922  __ret; \
4923})
4924#endif
4925
4926#ifdef __LITTLE_ENDIAN__
4927#define vdup_lane_u32(__p0, __p1) __extension__ ({ \
4928  uint32x2_t __s0 = __p0; \
4929  uint32x2_t __ret; \
4930  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
4931  __ret; \
4932})
4933#else
4934#define vdup_lane_u32(__p0, __p1) __extension__ ({ \
4935  uint32x2_t __s0 = __p0; \
4936  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
4937  uint32x2_t __ret; \
4938  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
4939  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
4940  __ret; \
4941})
4942#endif
4943
4944#ifdef __LITTLE_ENDIAN__
4945#define vdup_lane_u64(__p0, __p1) __extension__ ({ \
4946  uint64x1_t __s0 = __p0; \
4947  uint64x1_t __ret; \
4948  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
4949  __ret; \
4950})
4951#else
4952#define vdup_lane_u64(__p0, __p1) __extension__ ({ \
4953  uint64x1_t __s0 = __p0; \
4954  uint64x1_t __ret; \
4955  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
4956  __ret; \
4957})
4958#endif
4959
4960#ifdef __LITTLE_ENDIAN__
4961#define vdup_lane_u16(__p0, __p1) __extension__ ({ \
4962  uint16x4_t __s0 = __p0; \
4963  uint16x4_t __ret; \
4964  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
4965  __ret; \
4966})
4967#else
4968#define vdup_lane_u16(__p0, __p1) __extension__ ({ \
4969  uint16x4_t __s0 = __p0; \
4970  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
4971  uint16x4_t __ret; \
4972  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
4973  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
4974  __ret; \
4975})
4976#endif
4977
4978#ifdef __LITTLE_ENDIAN__
4979#define vdup_lane_s8(__p0, __p1) __extension__ ({ \
4980  int8x8_t __s0 = __p0; \
4981  int8x8_t __ret; \
4982  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4983  __ret; \
4984})
4985#else
4986#define vdup_lane_s8(__p0, __p1) __extension__ ({ \
4987  int8x8_t __s0 = __p0; \
4988  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
4989  int8x8_t __ret; \
4990  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
4991  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
4992  __ret; \
4993})
4994#endif
4995
4996#ifdef __LITTLE_ENDIAN__
4997#define vdup_lane_f32(__p0, __p1) __extension__ ({ \
4998  float32x2_t __s0 = __p0; \
4999  float32x2_t __ret; \
5000  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
5001  __ret; \
5002})
5003#else
5004#define vdup_lane_f32(__p0, __p1) __extension__ ({ \
5005  float32x2_t __s0 = __p0; \
5006  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5007  float32x2_t __ret; \
5008  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
5009  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5010  __ret; \
5011})
5012#endif
5013
5014#ifdef __LITTLE_ENDIAN__
5015#define vdup_lane_s32(__p0, __p1) __extension__ ({ \
5016  int32x2_t __s0 = __p0; \
5017  int32x2_t __ret; \
5018  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
5019  __ret; \
5020})
5021#else
5022#define vdup_lane_s32(__p0, __p1) __extension__ ({ \
5023  int32x2_t __s0 = __p0; \
5024  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5025  int32x2_t __ret; \
5026  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
5027  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5028  __ret; \
5029})
5030#endif
5031
5032#ifdef __LITTLE_ENDIAN__
5033#define vdup_lane_s64(__p0, __p1) __extension__ ({ \
5034  int64x1_t __s0 = __p0; \
5035  int64x1_t __ret; \
5036  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
5037  __ret; \
5038})
5039#else
5040#define vdup_lane_s64(__p0, __p1) __extension__ ({ \
5041  int64x1_t __s0 = __p0; \
5042  int64x1_t __ret; \
5043  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
5044  __ret; \
5045})
5046#endif
5047
5048#ifdef __LITTLE_ENDIAN__
5049#define vdup_lane_s16(__p0, __p1) __extension__ ({ \
5050  int16x4_t __s0 = __p0; \
5051  int16x4_t __ret; \
5052  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
5053  __ret; \
5054})
5055#else
5056#define vdup_lane_s16(__p0, __p1) __extension__ ({ \
5057  int16x4_t __s0 = __p0; \
5058  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5059  int16x4_t __ret; \
5060  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
5061  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5062  __ret; \
5063})
5064#endif
5065
5066#ifdef __LITTLE_ENDIAN__
5067__ai poly8x8_t vdup_n_p8(poly8_t __p0) {
5068  poly8x8_t __ret;
5069  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5070  return __ret;
5071}
5072#else
5073__ai poly8x8_t vdup_n_p8(poly8_t __p0) {
5074  poly8x8_t __ret;
5075  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5076  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5077  return __ret;
5078}
5079#endif
5080
5081#ifdef __LITTLE_ENDIAN__
5082__ai poly16x4_t vdup_n_p16(poly16_t __p0) {
5083  poly16x4_t __ret;
5084  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
5085  return __ret;
5086}
5087#else
5088__ai poly16x4_t vdup_n_p16(poly16_t __p0) {
5089  poly16x4_t __ret;
5090  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
5091  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5092  return __ret;
5093}
5094#endif
5095
5096#ifdef __LITTLE_ENDIAN__
5097__ai poly8x16_t vdupq_n_p8(poly8_t __p0) {
5098  poly8x16_t __ret;
5099  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5100  return __ret;
5101}
5102#else
5103__ai poly8x16_t vdupq_n_p8(poly8_t __p0) {
5104  poly8x16_t __ret;
5105  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5106  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5107  return __ret;
5108}
5109#endif
5110
5111#ifdef __LITTLE_ENDIAN__
5112__ai poly16x8_t vdupq_n_p16(poly16_t __p0) {
5113  poly16x8_t __ret;
5114  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5115  return __ret;
5116}
5117#else
5118__ai poly16x8_t vdupq_n_p16(poly16_t __p0) {
5119  poly16x8_t __ret;
5120  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5121  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5122  return __ret;
5123}
5124#endif
5125
5126#ifdef __LITTLE_ENDIAN__
5127__ai uint8x16_t vdupq_n_u8(uint8_t __p0) {
5128  uint8x16_t __ret;
5129  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5130  return __ret;
5131}
5132#else
5133__ai uint8x16_t vdupq_n_u8(uint8_t __p0) {
5134  uint8x16_t __ret;
5135  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5136  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5137  return __ret;
5138}
5139#endif
5140
5141#ifdef __LITTLE_ENDIAN__
5142__ai uint32x4_t vdupq_n_u32(uint32_t __p0) {
5143  uint32x4_t __ret;
5144  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
5145  return __ret;
5146}
5147#else
5148__ai uint32x4_t vdupq_n_u32(uint32_t __p0) {
5149  uint32x4_t __ret;
5150  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
5151  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5152  return __ret;
5153}
5154#endif
5155
5156#ifdef __LITTLE_ENDIAN__
5157__ai uint64x2_t vdupq_n_u64(uint64_t __p0) {
5158  uint64x2_t __ret;
5159  __ret = (uint64x2_t) {__p0, __p0};
5160  return __ret;
5161}
5162#else
5163__ai uint64x2_t vdupq_n_u64(uint64_t __p0) {
5164  uint64x2_t __ret;
5165  __ret = (uint64x2_t) {__p0, __p0};
5166  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5167  return __ret;
5168}
5169#endif
5170
5171#ifdef __LITTLE_ENDIAN__
5172__ai uint16x8_t vdupq_n_u16(uint16_t __p0) {
5173  uint16x8_t __ret;
5174  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5175  return __ret;
5176}
5177#else
5178__ai uint16x8_t vdupq_n_u16(uint16_t __p0) {
5179  uint16x8_t __ret;
5180  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5181  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5182  return __ret;
5183}
5184#endif
5185
5186#ifdef __LITTLE_ENDIAN__
5187__ai int8x16_t vdupq_n_s8(int8_t __p0) {
5188  int8x16_t __ret;
5189  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5190  return __ret;
5191}
5192#else
5193__ai int8x16_t vdupq_n_s8(int8_t __p0) {
5194  int8x16_t __ret;
5195  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5196  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5197  return __ret;
5198}
5199#endif
5200
5201#ifdef __LITTLE_ENDIAN__
5202__ai float32x4_t vdupq_n_f32(float32_t __p0) {
5203  float32x4_t __ret;
5204  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
5205  return __ret;
5206}
5207#else
5208__ai float32x4_t vdupq_n_f32(float32_t __p0) {
5209  float32x4_t __ret;
5210  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
5211  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5212  return __ret;
5213}
5214#endif
5215
5216#ifdef __LITTLE_ENDIAN__
5217#define vdupq_n_f16(__p0) __extension__ ({ \
5218  float16_t __s0 = __p0; \
5219  float16x8_t __ret; \
5220  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
5221  __ret; \
5222})
5223#else
5224#define vdupq_n_f16(__p0) __extension__ ({ \
5225  float16_t __s0 = __p0; \
5226  float16x8_t __ret; \
5227  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
5228  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5229  __ret; \
5230})
5231#endif
5232
5233#ifdef __LITTLE_ENDIAN__
5234__ai int32x4_t vdupq_n_s32(int32_t __p0) {
5235  int32x4_t __ret;
5236  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
5237  return __ret;
5238}
5239#else
5240__ai int32x4_t vdupq_n_s32(int32_t __p0) {
5241  int32x4_t __ret;
5242  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
5243  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5244  return __ret;
5245}
5246#endif
5247
5248#ifdef __LITTLE_ENDIAN__
5249__ai int64x2_t vdupq_n_s64(int64_t __p0) {
5250  int64x2_t __ret;
5251  __ret = (int64x2_t) {__p0, __p0};
5252  return __ret;
5253}
5254#else
5255__ai int64x2_t vdupq_n_s64(int64_t __p0) {
5256  int64x2_t __ret;
5257  __ret = (int64x2_t) {__p0, __p0};
5258  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5259  return __ret;
5260}
5261#endif
5262
5263#ifdef __LITTLE_ENDIAN__
5264__ai int16x8_t vdupq_n_s16(int16_t __p0) {
5265  int16x8_t __ret;
5266  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5267  return __ret;
5268}
5269#else
5270__ai int16x8_t vdupq_n_s16(int16_t __p0) {
5271  int16x8_t __ret;
5272  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5273  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5274  return __ret;
5275}
5276#endif
5277
5278#ifdef __LITTLE_ENDIAN__
5279__ai uint8x8_t vdup_n_u8(uint8_t __p0) {
5280  uint8x8_t __ret;
5281  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5282  return __ret;
5283}
5284#else
5285__ai uint8x8_t vdup_n_u8(uint8_t __p0) {
5286  uint8x8_t __ret;
5287  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5288  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5289  return __ret;
5290}
5291#endif
5292
5293#ifdef __LITTLE_ENDIAN__
5294__ai uint32x2_t vdup_n_u32(uint32_t __p0) {
5295  uint32x2_t __ret;
5296  __ret = (uint32x2_t) {__p0, __p0};
5297  return __ret;
5298}
5299#else
5300__ai uint32x2_t vdup_n_u32(uint32_t __p0) {
5301  uint32x2_t __ret;
5302  __ret = (uint32x2_t) {__p0, __p0};
5303  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5304  return __ret;
5305}
5306#endif
5307
5308#ifdef __LITTLE_ENDIAN__
5309__ai uint64x1_t vdup_n_u64(uint64_t __p0) {
5310  uint64x1_t __ret;
5311  __ret = (uint64x1_t) {__p0};
5312  return __ret;
5313}
5314#else
5315__ai uint64x1_t vdup_n_u64(uint64_t __p0) {
5316  uint64x1_t __ret;
5317  __ret = (uint64x1_t) {__p0};
5318  return __ret;
5319}
5320#endif
5321
5322#ifdef __LITTLE_ENDIAN__
5323__ai uint16x4_t vdup_n_u16(uint16_t __p0) {
5324  uint16x4_t __ret;
5325  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
5326  return __ret;
5327}
5328#else
5329__ai uint16x4_t vdup_n_u16(uint16_t __p0) {
5330  uint16x4_t __ret;
5331  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
5332  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5333  return __ret;
5334}
5335#endif
5336
5337#ifdef __LITTLE_ENDIAN__
5338__ai int8x8_t vdup_n_s8(int8_t __p0) {
5339  int8x8_t __ret;
5340  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5341  return __ret;
5342}
5343#else
5344__ai int8x8_t vdup_n_s8(int8_t __p0) {
5345  int8x8_t __ret;
5346  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
5347  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5348  return __ret;
5349}
5350#endif
5351
5352#ifdef __LITTLE_ENDIAN__
5353__ai float32x2_t vdup_n_f32(float32_t __p0) {
5354  float32x2_t __ret;
5355  __ret = (float32x2_t) {__p0, __p0};
5356  return __ret;
5357}
5358#else
5359__ai float32x2_t vdup_n_f32(float32_t __p0) {
5360  float32x2_t __ret;
5361  __ret = (float32x2_t) {__p0, __p0};
5362  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5363  return __ret;
5364}
5365#endif
5366
5367#ifdef __LITTLE_ENDIAN__
5368#define vdup_n_f16(__p0) __extension__ ({ \
5369  float16_t __s0 = __p0; \
5370  float16x4_t __ret; \
5371  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
5372  __ret; \
5373})
5374#else
5375#define vdup_n_f16(__p0) __extension__ ({ \
5376  float16_t __s0 = __p0; \
5377  float16x4_t __ret; \
5378  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
5379  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5380  __ret; \
5381})
5382#endif
5383
5384#ifdef __LITTLE_ENDIAN__
5385__ai int32x2_t vdup_n_s32(int32_t __p0) {
5386  int32x2_t __ret;
5387  __ret = (int32x2_t) {__p0, __p0};
5388  return __ret;
5389}
5390#else
5391__ai int32x2_t vdup_n_s32(int32_t __p0) {
5392  int32x2_t __ret;
5393  __ret = (int32x2_t) {__p0, __p0};
5394  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5395  return __ret;
5396}
5397#endif
5398
5399#ifdef __LITTLE_ENDIAN__
5400__ai int64x1_t vdup_n_s64(int64_t __p0) {
5401  int64x1_t __ret;
5402  __ret = (int64x1_t) {__p0};
5403  return __ret;
5404}
5405#else
5406__ai int64x1_t vdup_n_s64(int64_t __p0) {
5407  int64x1_t __ret;
5408  __ret = (int64x1_t) {__p0};
5409  return __ret;
5410}
5411#endif
5412
5413#ifdef __LITTLE_ENDIAN__
5414__ai int16x4_t vdup_n_s16(int16_t __p0) {
5415  int16x4_t __ret;
5416  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
5417  return __ret;
5418}
5419#else
5420__ai int16x4_t vdup_n_s16(int16_t __p0) {
5421  int16x4_t __ret;
5422  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
5423  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5424  return __ret;
5425}
5426#endif
5427
5428#ifdef __LITTLE_ENDIAN__
5429__ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1) {
5430  uint8x16_t __ret;
5431  __ret = __p0 ^ __p1;
5432  return __ret;
5433}
5434#else
5435__ai uint8x16_t veorq_u8(uint8x16_t __p0, uint8x16_t __p1) {
5436  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5437  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5438  uint8x16_t __ret;
5439  __ret = __rev0 ^ __rev1;
5440  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5441  return __ret;
5442}
5443#endif
5444
5445#ifdef __LITTLE_ENDIAN__
5446__ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1) {
5447  uint32x4_t __ret;
5448  __ret = __p0 ^ __p1;
5449  return __ret;
5450}
5451#else
5452__ai uint32x4_t veorq_u32(uint32x4_t __p0, uint32x4_t __p1) {
5453  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5454  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5455  uint32x4_t __ret;
5456  __ret = __rev0 ^ __rev1;
5457  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5458  return __ret;
5459}
5460#endif
5461
5462#ifdef __LITTLE_ENDIAN__
5463__ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1) {
5464  uint64x2_t __ret;
5465  __ret = __p0 ^ __p1;
5466  return __ret;
5467}
5468#else
5469__ai uint64x2_t veorq_u64(uint64x2_t __p0, uint64x2_t __p1) {
5470  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5471  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5472  uint64x2_t __ret;
5473  __ret = __rev0 ^ __rev1;
5474  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5475  return __ret;
5476}
5477#endif
5478
5479#ifdef __LITTLE_ENDIAN__
5480__ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1) {
5481  uint16x8_t __ret;
5482  __ret = __p0 ^ __p1;
5483  return __ret;
5484}
5485#else
5486__ai uint16x8_t veorq_u16(uint16x8_t __p0, uint16x8_t __p1) {
5487  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5488  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5489  uint16x8_t __ret;
5490  __ret = __rev0 ^ __rev1;
5491  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5492  return __ret;
5493}
5494#endif
5495
5496#ifdef __LITTLE_ENDIAN__
5497__ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1) {
5498  int8x16_t __ret;
5499  __ret = __p0 ^ __p1;
5500  return __ret;
5501}
5502#else
5503__ai int8x16_t veorq_s8(int8x16_t __p0, int8x16_t __p1) {
5504  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5505  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5506  int8x16_t __ret;
5507  __ret = __rev0 ^ __rev1;
5508  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
5509  return __ret;
5510}
5511#endif
5512
5513#ifdef __LITTLE_ENDIAN__
5514__ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1) {
5515  int32x4_t __ret;
5516  __ret = __p0 ^ __p1;
5517  return __ret;
5518}
5519#else
5520__ai int32x4_t veorq_s32(int32x4_t __p0, int32x4_t __p1) {
5521  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5522  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5523  int32x4_t __ret;
5524  __ret = __rev0 ^ __rev1;
5525  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5526  return __ret;
5527}
5528#endif
5529
5530#ifdef __LITTLE_ENDIAN__
5531__ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1) {
5532  int64x2_t __ret;
5533  __ret = __p0 ^ __p1;
5534  return __ret;
5535}
5536#else
5537__ai int64x2_t veorq_s64(int64x2_t __p0, int64x2_t __p1) {
5538  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5539  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5540  int64x2_t __ret;
5541  __ret = __rev0 ^ __rev1;
5542  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5543  return __ret;
5544}
5545#endif
5546
5547#ifdef __LITTLE_ENDIAN__
5548__ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1) {
5549  int16x8_t __ret;
5550  __ret = __p0 ^ __p1;
5551  return __ret;
5552}
5553#else
5554__ai int16x8_t veorq_s16(int16x8_t __p0, int16x8_t __p1) {
5555  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5556  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5557  int16x8_t __ret;
5558  __ret = __rev0 ^ __rev1;
5559  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5560  return __ret;
5561}
5562#endif
5563
5564#ifdef __LITTLE_ENDIAN__
5565__ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1) {
5566  uint8x8_t __ret;
5567  __ret = __p0 ^ __p1;
5568  return __ret;
5569}
5570#else
5571__ai uint8x8_t veor_u8(uint8x8_t __p0, uint8x8_t __p1) {
5572  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5573  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5574  uint8x8_t __ret;
5575  __ret = __rev0 ^ __rev1;
5576  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5577  return __ret;
5578}
5579#endif
5580
5581#ifdef __LITTLE_ENDIAN__
5582__ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1) {
5583  uint32x2_t __ret;
5584  __ret = __p0 ^ __p1;
5585  return __ret;
5586}
5587#else
5588__ai uint32x2_t veor_u32(uint32x2_t __p0, uint32x2_t __p1) {
5589  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5590  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5591  uint32x2_t __ret;
5592  __ret = __rev0 ^ __rev1;
5593  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5594  return __ret;
5595}
5596#endif
5597
5598#ifdef __LITTLE_ENDIAN__
5599__ai uint64x1_t veor_u64(uint64x1_t __p0, uint64x1_t __p1) {
5600  uint64x1_t __ret;
5601  __ret = __p0 ^ __p1;
5602  return __ret;
5603}
5604#else
5605__ai uint64x1_t veor_u64(uint64x1_t __p0, uint64x1_t __p1) {
5606  uint64x1_t __ret;
5607  __ret = __p0 ^ __p1;
5608  return __ret;
5609}
5610#endif
5611
5612#ifdef __LITTLE_ENDIAN__
5613__ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1) {
5614  uint16x4_t __ret;
5615  __ret = __p0 ^ __p1;
5616  return __ret;
5617}
5618#else
5619__ai uint16x4_t veor_u16(uint16x4_t __p0, uint16x4_t __p1) {
5620  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5621  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5622  uint16x4_t __ret;
5623  __ret = __rev0 ^ __rev1;
5624  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5625  return __ret;
5626}
5627#endif
5628
5629#ifdef __LITTLE_ENDIAN__
5630__ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1) {
5631  int8x8_t __ret;
5632  __ret = __p0 ^ __p1;
5633  return __ret;
5634}
5635#else
5636__ai int8x8_t veor_s8(int8x8_t __p0, int8x8_t __p1) {
5637  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
5638  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
5639  int8x8_t __ret;
5640  __ret = __rev0 ^ __rev1;
5641  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
5642  return __ret;
5643}
5644#endif
5645
5646#ifdef __LITTLE_ENDIAN__
5647__ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1) {
5648  int32x2_t __ret;
5649  __ret = __p0 ^ __p1;
5650  return __ret;
5651}
5652#else
5653__ai int32x2_t veor_s32(int32x2_t __p0, int32x2_t __p1) {
5654  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
5655  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
5656  int32x2_t __ret;
5657  __ret = __rev0 ^ __rev1;
5658  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
5659  return __ret;
5660}
5661#endif
5662
5663#ifdef __LITTLE_ENDIAN__
5664__ai int64x1_t veor_s64(int64x1_t __p0, int64x1_t __p1) {
5665  int64x1_t __ret;
5666  __ret = __p0 ^ __p1;
5667  return __ret;
5668}
5669#else
5670__ai int64x1_t veor_s64(int64x1_t __p0, int64x1_t __p1) {
5671  int64x1_t __ret;
5672  __ret = __p0 ^ __p1;
5673  return __ret;
5674}
5675#endif
5676
5677#ifdef __LITTLE_ENDIAN__
5678__ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1) {
5679  int16x4_t __ret;
5680  __ret = __p0 ^ __p1;
5681  return __ret;
5682}
5683#else
5684__ai int16x4_t veor_s16(int16x4_t __p0, int16x4_t __p1) {
5685  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
5686  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
5687  int16x4_t __ret;
5688  __ret = __rev0 ^ __rev1;
5689  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
5690  return __ret;
5691}
5692#endif
5693
5694#ifdef __LITTLE_ENDIAN__
5695#define vext_p8(__p0, __p1, __p2) __extension__ ({ \
5696  poly8x8_t __s0 = __p0; \
5697  poly8x8_t __s1 = __p1; \
5698  poly8x8_t __ret; \
5699  __ret = (poly8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
5700  __ret; \
5701})
5702#else
5703#define vext_p8(__p0, __p1, __p2) __extension__ ({ \
5704  poly8x8_t __s0 = __p0; \
5705  poly8x8_t __s1 = __p1; \
5706  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5707  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5708  poly8x8_t __ret; \
5709  __ret = (poly8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
5710  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5711  __ret; \
5712})
5713#endif
5714
5715#ifdef __LITTLE_ENDIAN__
5716#define vext_p16(__p0, __p1, __p2) __extension__ ({ \
5717  poly16x4_t __s0 = __p0; \
5718  poly16x4_t __s1 = __p1; \
5719  poly16x4_t __ret; \
5720  __ret = (poly16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
5721  __ret; \
5722})
5723#else
5724#define vext_p16(__p0, __p1, __p2) __extension__ ({ \
5725  poly16x4_t __s0 = __p0; \
5726  poly16x4_t __s1 = __p1; \
5727  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5728  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5729  poly16x4_t __ret; \
5730  __ret = (poly16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
5731  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5732  __ret; \
5733})
5734#endif
5735
5736#ifdef __LITTLE_ENDIAN__
5737#define vextq_p8(__p0, __p1, __p2) __extension__ ({ \
5738  poly8x16_t __s0 = __p0; \
5739  poly8x16_t __s1 = __p1; \
5740  poly8x16_t __ret; \
5741  __ret = (poly8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
5742  __ret; \
5743})
5744#else
5745#define vextq_p8(__p0, __p1, __p2) __extension__ ({ \
5746  poly8x16_t __s0 = __p0; \
5747  poly8x16_t __s1 = __p1; \
5748  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5749  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5750  poly8x16_t __ret; \
5751  __ret = (poly8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
5752  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5753  __ret; \
5754})
5755#endif
5756
5757#ifdef __LITTLE_ENDIAN__
5758#define vextq_p16(__p0, __p1, __p2) __extension__ ({ \
5759  poly16x8_t __s0 = __p0; \
5760  poly16x8_t __s1 = __p1; \
5761  poly16x8_t __ret; \
5762  __ret = (poly16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
5763  __ret; \
5764})
5765#else
5766#define vextq_p16(__p0, __p1, __p2) __extension__ ({ \
5767  poly16x8_t __s0 = __p0; \
5768  poly16x8_t __s1 = __p1; \
5769  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5770  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5771  poly16x8_t __ret; \
5772  __ret = (poly16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
5773  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5774  __ret; \
5775})
5776#endif
5777
5778#ifdef __LITTLE_ENDIAN__
5779#define vextq_u8(__p0, __p1, __p2) __extension__ ({ \
5780  uint8x16_t __s0 = __p0; \
5781  uint8x16_t __s1 = __p1; \
5782  uint8x16_t __ret; \
5783  __ret = (uint8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
5784  __ret; \
5785})
5786#else
5787#define vextq_u8(__p0, __p1, __p2) __extension__ ({ \
5788  uint8x16_t __s0 = __p0; \
5789  uint8x16_t __s1 = __p1; \
5790  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5791  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5792  uint8x16_t __ret; \
5793  __ret = (uint8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
5794  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5795  __ret; \
5796})
5797#endif
5798
5799#ifdef __LITTLE_ENDIAN__
5800#define vextq_u32(__p0, __p1, __p2) __extension__ ({ \
5801  uint32x4_t __s0 = __p0; \
5802  uint32x4_t __s1 = __p1; \
5803  uint32x4_t __ret; \
5804  __ret = (uint32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
5805  __ret; \
5806})
5807#else
5808#define vextq_u32(__p0, __p1, __p2) __extension__ ({ \
5809  uint32x4_t __s0 = __p0; \
5810  uint32x4_t __s1 = __p1; \
5811  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5812  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5813  uint32x4_t __ret; \
5814  __ret = (uint32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
5815  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5816  __ret; \
5817})
5818#endif
5819
5820#ifdef __LITTLE_ENDIAN__
5821#define vextq_u64(__p0, __p1, __p2) __extension__ ({ \
5822  uint64x2_t __s0 = __p0; \
5823  uint64x2_t __s1 = __p1; \
5824  uint64x2_t __ret; \
5825  __ret = (uint64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
5826  __ret; \
5827})
5828#else
5829#define vextq_u64(__p0, __p1, __p2) __extension__ ({ \
5830  uint64x2_t __s0 = __p0; \
5831  uint64x2_t __s1 = __p1; \
5832  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5833  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
5834  uint64x2_t __ret; \
5835  __ret = (uint64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
5836  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5837  __ret; \
5838})
5839#endif
5840
5841#ifdef __LITTLE_ENDIAN__
5842#define vextq_u16(__p0, __p1, __p2) __extension__ ({ \
5843  uint16x8_t __s0 = __p0; \
5844  uint16x8_t __s1 = __p1; \
5845  uint16x8_t __ret; \
5846  __ret = (uint16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
5847  __ret; \
5848})
5849#else
5850#define vextq_u16(__p0, __p1, __p2) __extension__ ({ \
5851  uint16x8_t __s0 = __p0; \
5852  uint16x8_t __s1 = __p1; \
5853  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5854  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5855  uint16x8_t __ret; \
5856  __ret = (uint16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
5857  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5858  __ret; \
5859})
5860#endif
5861
5862#ifdef __LITTLE_ENDIAN__
5863#define vextq_s8(__p0, __p1, __p2) __extension__ ({ \
5864  int8x16_t __s0 = __p0; \
5865  int8x16_t __s1 = __p1; \
5866  int8x16_t __ret; \
5867  __ret = (int8x16_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
5868  __ret; \
5869})
5870#else
5871#define vextq_s8(__p0, __p1, __p2) __extension__ ({ \
5872  int8x16_t __s0 = __p0; \
5873  int8x16_t __s1 = __p1; \
5874  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5875  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5876  int8x16_t __ret; \
5877  __ret = (int8x16_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
5878  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
5879  __ret; \
5880})
5881#endif
5882
5883#ifdef __LITTLE_ENDIAN__
5884#define vextq_f32(__p0, __p1, __p2) __extension__ ({ \
5885  float32x4_t __s0 = __p0; \
5886  float32x4_t __s1 = __p1; \
5887  float32x4_t __ret; \
5888  __ret = (float32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 41); \
5889  __ret; \
5890})
5891#else
5892#define vextq_f32(__p0, __p1, __p2) __extension__ ({ \
5893  float32x4_t __s0 = __p0; \
5894  float32x4_t __s1 = __p1; \
5895  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5896  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5897  float32x4_t __ret; \
5898  __ret = (float32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 41); \
5899  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5900  __ret; \
5901})
5902#endif
5903
5904#ifdef __LITTLE_ENDIAN__
5905#define vextq_s32(__p0, __p1, __p2) __extension__ ({ \
5906  int32x4_t __s0 = __p0; \
5907  int32x4_t __s1 = __p1; \
5908  int32x4_t __ret; \
5909  __ret = (int32x4_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
5910  __ret; \
5911})
5912#else
5913#define vextq_s32(__p0, __p1, __p2) __extension__ ({ \
5914  int32x4_t __s0 = __p0; \
5915  int32x4_t __s1 = __p1; \
5916  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
5917  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
5918  int32x4_t __ret; \
5919  __ret = (int32x4_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
5920  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
5921  __ret; \
5922})
5923#endif
5924
5925#ifdef __LITTLE_ENDIAN__
5926#define vextq_s64(__p0, __p1, __p2) __extension__ ({ \
5927  int64x2_t __s0 = __p0; \
5928  int64x2_t __s1 = __p1; \
5929  int64x2_t __ret; \
5930  __ret = (int64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
5931  __ret; \
5932})
5933#else
5934#define vextq_s64(__p0, __p1, __p2) __extension__ ({ \
5935  int64x2_t __s0 = __p0; \
5936  int64x2_t __s1 = __p1; \
5937  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
5938  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
5939  int64x2_t __ret; \
5940  __ret = (int64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
5941  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
5942  __ret; \
5943})
5944#endif
5945
5946#ifdef __LITTLE_ENDIAN__
5947#define vextq_s16(__p0, __p1, __p2) __extension__ ({ \
5948  int16x8_t __s0 = __p0; \
5949  int16x8_t __s1 = __p1; \
5950  int16x8_t __ret; \
5951  __ret = (int16x8_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
5952  __ret; \
5953})
5954#else
5955#define vextq_s16(__p0, __p1, __p2) __extension__ ({ \
5956  int16x8_t __s0 = __p0; \
5957  int16x8_t __s1 = __p1; \
5958  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5959  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5960  int16x8_t __ret; \
5961  __ret = (int16x8_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
5962  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5963  __ret; \
5964})
5965#endif
5966
5967#ifdef __LITTLE_ENDIAN__
5968#define vext_u8(__p0, __p1, __p2) __extension__ ({ \
5969  uint8x8_t __s0 = __p0; \
5970  uint8x8_t __s1 = __p1; \
5971  uint8x8_t __ret; \
5972  __ret = (uint8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
5973  __ret; \
5974})
5975#else
5976#define vext_u8(__p0, __p1, __p2) __extension__ ({ \
5977  uint8x8_t __s0 = __p0; \
5978  uint8x8_t __s1 = __p1; \
5979  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
5980  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
5981  uint8x8_t __ret; \
5982  __ret = (uint8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
5983  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
5984  __ret; \
5985})
5986#endif
5987
5988#ifdef __LITTLE_ENDIAN__
5989#define vext_u32(__p0, __p1, __p2) __extension__ ({ \
5990  uint32x2_t __s0 = __p0; \
5991  uint32x2_t __s1 = __p1; \
5992  uint32x2_t __ret; \
5993  __ret = (uint32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
5994  __ret; \
5995})
5996#else
5997#define vext_u32(__p0, __p1, __p2) __extension__ ({ \
5998  uint32x2_t __s0 = __p0; \
5999  uint32x2_t __s1 = __p1; \
6000  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6001  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6002  uint32x2_t __ret; \
6003  __ret = (uint32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
6004  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6005  __ret; \
6006})
6007#endif
6008
6009#ifdef __LITTLE_ENDIAN__
6010#define vext_u64(__p0, __p1, __p2) __extension__ ({ \
6011  uint64x1_t __s0 = __p0; \
6012  uint64x1_t __s1 = __p1; \
6013  uint64x1_t __ret; \
6014  __ret = (uint64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
6015  __ret; \
6016})
6017#else
6018#define vext_u64(__p0, __p1, __p2) __extension__ ({ \
6019  uint64x1_t __s0 = __p0; \
6020  uint64x1_t __s1 = __p1; \
6021  uint64x1_t __ret; \
6022  __ret = (uint64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
6023  __ret; \
6024})
6025#endif
6026
6027#ifdef __LITTLE_ENDIAN__
6028#define vext_u16(__p0, __p1, __p2) __extension__ ({ \
6029  uint16x4_t __s0 = __p0; \
6030  uint16x4_t __s1 = __p1; \
6031  uint16x4_t __ret; \
6032  __ret = (uint16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
6033  __ret; \
6034})
6035#else
6036#define vext_u16(__p0, __p1, __p2) __extension__ ({ \
6037  uint16x4_t __s0 = __p0; \
6038  uint16x4_t __s1 = __p1; \
6039  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6040  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6041  uint16x4_t __ret; \
6042  __ret = (uint16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
6043  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6044  __ret; \
6045})
6046#endif
6047
6048#ifdef __LITTLE_ENDIAN__
6049#define vext_s8(__p0, __p1, __p2) __extension__ ({ \
6050  int8x8_t __s0 = __p0; \
6051  int8x8_t __s1 = __p1; \
6052  int8x8_t __ret; \
6053  __ret = (int8x8_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
6054  __ret; \
6055})
6056#else
6057#define vext_s8(__p0, __p1, __p2) __extension__ ({ \
6058  int8x8_t __s0 = __p0; \
6059  int8x8_t __s1 = __p1; \
6060  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6061  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
6062  int8x8_t __ret; \
6063  __ret = (int8x8_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
6064  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
6065  __ret; \
6066})
6067#endif
6068
6069#ifdef __LITTLE_ENDIAN__
6070#define vext_f32(__p0, __p1, __p2) __extension__ ({ \
6071  float32x2_t __s0 = __p0; \
6072  float32x2_t __s1 = __p1; \
6073  float32x2_t __ret; \
6074  __ret = (float32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 9); \
6075  __ret; \
6076})
6077#else
6078#define vext_f32(__p0, __p1, __p2) __extension__ ({ \
6079  float32x2_t __s0 = __p0; \
6080  float32x2_t __s1 = __p1; \
6081  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6082  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6083  float32x2_t __ret; \
6084  __ret = (float32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 9); \
6085  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6086  __ret; \
6087})
6088#endif
6089
6090#ifdef __LITTLE_ENDIAN__
6091#define vext_s32(__p0, __p1, __p2) __extension__ ({ \
6092  int32x2_t __s0 = __p0; \
6093  int32x2_t __s1 = __p1; \
6094  int32x2_t __ret; \
6095  __ret = (int32x2_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
6096  __ret; \
6097})
6098#else
6099#define vext_s32(__p0, __p1, __p2) __extension__ ({ \
6100  int32x2_t __s0 = __p0; \
6101  int32x2_t __s1 = __p1; \
6102  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6103  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
6104  int32x2_t __ret; \
6105  __ret = (int32x2_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
6106  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
6107  __ret; \
6108})
6109#endif
6110
6111#ifdef __LITTLE_ENDIAN__
6112#define vext_s64(__p0, __p1, __p2) __extension__ ({ \
6113  int64x1_t __s0 = __p0; \
6114  int64x1_t __s1 = __p1; \
6115  int64x1_t __ret; \
6116  __ret = (int64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
6117  __ret; \
6118})
6119#else
6120#define vext_s64(__p0, __p1, __p2) __extension__ ({ \
6121  int64x1_t __s0 = __p0; \
6122  int64x1_t __s1 = __p1; \
6123  int64x1_t __ret; \
6124  __ret = (int64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
6125  __ret; \
6126})
6127#endif
6128
6129#ifdef __LITTLE_ENDIAN__
6130#define vext_s16(__p0, __p1, __p2) __extension__ ({ \
6131  int16x4_t __s0 = __p0; \
6132  int16x4_t __s1 = __p1; \
6133  int16x4_t __ret; \
6134  __ret = (int16x4_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
6135  __ret; \
6136})
6137#else
6138#define vext_s16(__p0, __p1, __p2) __extension__ ({ \
6139  int16x4_t __s0 = __p0; \
6140  int16x4_t __s1 = __p1; \
6141  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6142  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
6143  int16x4_t __ret; \
6144  __ret = (int16x4_t) __builtin_neon_vext_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
6145  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
6146  __ret; \
6147})
6148#endif
6149
6150#ifdef __LITTLE_ENDIAN__
6151__ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
6152  float32x4_t __ret;
6153  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
6154  return __ret;
6155}
6156#else
6157__ai float32x4_t vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
6158  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6159  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
6160  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
6161  float32x4_t __ret;
6162  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
6163  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6164  return __ret;
6165}
6166__ai float32x4_t __noswap_vfmaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
6167  float32x4_t __ret;
6168  __ret = (float32x4_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
6169  return __ret;
6170}
6171#endif
6172
6173#ifdef __LITTLE_ENDIAN__
6174__ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
6175  float32x2_t __ret;
6176  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
6177  return __ret;
6178}
6179#else
6180__ai float32x2_t vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
6181  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6182  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
6183  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
6184  float32x2_t __ret;
6185  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
6186  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6187  return __ret;
6188}
6189__ai float32x2_t __noswap_vfma_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
6190  float32x2_t __ret;
6191  __ret = (float32x2_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
6192  return __ret;
6193}
6194#endif
6195
6196#ifdef __LITTLE_ENDIAN__
6197__ai poly8x8_t vget_high_p8(poly8x16_t __p0) {
6198  poly8x8_t __ret;
6199  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6200  return __ret;
6201}
6202#else
6203__ai poly8x8_t vget_high_p8(poly8x16_t __p0) {
6204  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6205  poly8x8_t __ret;
6206  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
6207  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6208  return __ret;
6209}
6210__ai poly8x8_t __noswap_vget_high_p8(poly8x16_t __p0) {
6211  poly8x8_t __ret;
6212  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6213  return __ret;
6214}
6215#endif
6216
6217#ifdef __LITTLE_ENDIAN__
6218__ai poly16x4_t vget_high_p16(poly16x8_t __p0) {
6219  poly16x4_t __ret;
6220  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6221  return __ret;
6222}
6223#else
6224__ai poly16x4_t vget_high_p16(poly16x8_t __p0) {
6225  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6226  poly16x4_t __ret;
6227  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6228  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6229  return __ret;
6230}
6231#endif
6232
6233#ifdef __LITTLE_ENDIAN__
6234__ai uint8x8_t vget_high_u8(uint8x16_t __p0) {
6235  uint8x8_t __ret;
6236  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6237  return __ret;
6238}
6239#else
6240__ai uint8x8_t vget_high_u8(uint8x16_t __p0) {
6241  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6242  uint8x8_t __ret;
6243  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
6244  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6245  return __ret;
6246}
6247__ai uint8x8_t __noswap_vget_high_u8(uint8x16_t __p0) {
6248  uint8x8_t __ret;
6249  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6250  return __ret;
6251}
6252#endif
6253
6254#ifdef __LITTLE_ENDIAN__
6255__ai uint32x2_t vget_high_u32(uint32x4_t __p0) {
6256  uint32x2_t __ret;
6257  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6258  return __ret;
6259}
6260#else
6261__ai uint32x2_t vget_high_u32(uint32x4_t __p0) {
6262  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6263  uint32x2_t __ret;
6264  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
6265  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6266  return __ret;
6267}
6268__ai uint32x2_t __noswap_vget_high_u32(uint32x4_t __p0) {
6269  uint32x2_t __ret;
6270  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6271  return __ret;
6272}
6273#endif
6274
6275#ifdef __LITTLE_ENDIAN__
6276__ai uint64x1_t vget_high_u64(uint64x2_t __p0) {
6277  uint64x1_t __ret;
6278  __ret = __builtin_shufflevector(__p0, __p0, 1);
6279  return __ret;
6280}
6281#else
6282__ai uint64x1_t vget_high_u64(uint64x2_t __p0) {
6283  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6284  uint64x1_t __ret;
6285  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
6286  return __ret;
6287}
6288#endif
6289
6290#ifdef __LITTLE_ENDIAN__
6291__ai uint16x4_t vget_high_u16(uint16x8_t __p0) {
6292  uint16x4_t __ret;
6293  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6294  return __ret;
6295}
6296#else
6297__ai uint16x4_t vget_high_u16(uint16x8_t __p0) {
6298  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6299  uint16x4_t __ret;
6300  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6301  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6302  return __ret;
6303}
6304__ai uint16x4_t __noswap_vget_high_u16(uint16x8_t __p0) {
6305  uint16x4_t __ret;
6306  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6307  return __ret;
6308}
6309#endif
6310
6311#ifdef __LITTLE_ENDIAN__
6312__ai int8x8_t vget_high_s8(int8x16_t __p0) {
6313  int8x8_t __ret;
6314  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6315  return __ret;
6316}
6317#else
6318__ai int8x8_t vget_high_s8(int8x16_t __p0) {
6319  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6320  int8x8_t __ret;
6321  __ret = __builtin_shufflevector(__rev0, __rev0, 8, 9, 10, 11, 12, 13, 14, 15);
6322  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6323  return __ret;
6324}
6325__ai int8x8_t __noswap_vget_high_s8(int8x16_t __p0) {
6326  int8x8_t __ret;
6327  __ret = __builtin_shufflevector(__p0, __p0, 8, 9, 10, 11, 12, 13, 14, 15);
6328  return __ret;
6329}
6330#endif
6331
6332#ifdef __LITTLE_ENDIAN__
6333__ai float32x2_t vget_high_f32(float32x4_t __p0) {
6334  float32x2_t __ret;
6335  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6336  return __ret;
6337}
6338#else
6339__ai float32x2_t vget_high_f32(float32x4_t __p0) {
6340  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6341  float32x2_t __ret;
6342  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
6343  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6344  return __ret;
6345}
6346__ai float32x2_t __noswap_vget_high_f32(float32x4_t __p0) {
6347  float32x2_t __ret;
6348  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6349  return __ret;
6350}
6351#endif
6352
6353#ifdef __LITTLE_ENDIAN__
6354__ai float16x4_t vget_high_f16(float16x8_t __p0) {
6355  float16x4_t __ret;
6356  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6357  return __ret;
6358}
6359#else
6360__ai float16x4_t vget_high_f16(float16x8_t __p0) {
6361  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6362  float16x4_t __ret;
6363  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6364  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6365  return __ret;
6366}
6367__ai float16x4_t __noswap_vget_high_f16(float16x8_t __p0) {
6368  float16x4_t __ret;
6369  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6370  return __ret;
6371}
6372#endif
6373
6374#ifdef __LITTLE_ENDIAN__
6375__ai int32x2_t vget_high_s32(int32x4_t __p0) {
6376  int32x2_t __ret;
6377  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6378  return __ret;
6379}
6380#else
6381__ai int32x2_t vget_high_s32(int32x4_t __p0) {
6382  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6383  int32x2_t __ret;
6384  __ret = __builtin_shufflevector(__rev0, __rev0, 2, 3);
6385  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6386  return __ret;
6387}
6388__ai int32x2_t __noswap_vget_high_s32(int32x4_t __p0) {
6389  int32x2_t __ret;
6390  __ret = __builtin_shufflevector(__p0, __p0, 2, 3);
6391  return __ret;
6392}
6393#endif
6394
6395#ifdef __LITTLE_ENDIAN__
6396__ai int64x1_t vget_high_s64(int64x2_t __p0) {
6397  int64x1_t __ret;
6398  __ret = __builtin_shufflevector(__p0, __p0, 1);
6399  return __ret;
6400}
6401#else
6402__ai int64x1_t vget_high_s64(int64x2_t __p0) {
6403  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
6404  int64x1_t __ret;
6405  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
6406  return __ret;
6407}
6408#endif
6409
6410#ifdef __LITTLE_ENDIAN__
6411__ai int16x4_t vget_high_s16(int16x8_t __p0) {
6412  int16x4_t __ret;
6413  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6414  return __ret;
6415}
6416#else
6417__ai int16x4_t vget_high_s16(int16x8_t __p0) {
6418  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6419  int16x4_t __ret;
6420  __ret = __builtin_shufflevector(__rev0, __rev0, 4, 5, 6, 7);
6421  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6422  return __ret;
6423}
6424__ai int16x4_t __noswap_vget_high_s16(int16x8_t __p0) {
6425  int16x4_t __ret;
6426  __ret = __builtin_shufflevector(__p0, __p0, 4, 5, 6, 7);
6427  return __ret;
6428}
6429#endif
6430
6431#ifdef __LITTLE_ENDIAN__
6432#define vget_lane_p8(__p0, __p1) __extension__ ({ \
6433  poly8x8_t __s0 = __p0; \
6434  poly8_t __ret; \
6435  __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6436  __ret; \
6437})
6438#else
6439#define vget_lane_p8(__p0, __p1) __extension__ ({ \
6440  poly8x8_t __s0 = __p0; \
6441  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6442  poly8_t __ret; \
6443  __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
6444  __ret; \
6445})
6446#define __noswap_vget_lane_p8(__p0, __p1) __extension__ ({ \
6447  poly8x8_t __s0 = __p0; \
6448  poly8_t __ret; \
6449  __ret = (poly8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6450  __ret; \
6451})
6452#endif
6453
6454#ifdef __LITTLE_ENDIAN__
6455#define vget_lane_p16(__p0, __p1) __extension__ ({ \
6456  poly16x4_t __s0 = __p0; \
6457  poly16_t __ret; \
6458  __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6459  __ret; \
6460})
6461#else
6462#define vget_lane_p16(__p0, __p1) __extension__ ({ \
6463  poly16x4_t __s0 = __p0; \
6464  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6465  poly16_t __ret; \
6466  __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
6467  __ret; \
6468})
6469#define __noswap_vget_lane_p16(__p0, __p1) __extension__ ({ \
6470  poly16x4_t __s0 = __p0; \
6471  poly16_t __ret; \
6472  __ret = (poly16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6473  __ret; \
6474})
6475#endif
6476
6477#ifdef __LITTLE_ENDIAN__
6478#define vgetq_lane_p8(__p0, __p1) __extension__ ({ \
6479  poly8x16_t __s0 = __p0; \
6480  poly8_t __ret; \
6481  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6482  __ret; \
6483})
6484#else
6485#define vgetq_lane_p8(__p0, __p1) __extension__ ({ \
6486  poly8x16_t __s0 = __p0; \
6487  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6488  poly8_t __ret; \
6489  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
6490  __ret; \
6491})
6492#define __noswap_vgetq_lane_p8(__p0, __p1) __extension__ ({ \
6493  poly8x16_t __s0 = __p0; \
6494  poly8_t __ret; \
6495  __ret = (poly8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6496  __ret; \
6497})
6498#endif
6499
6500#ifdef __LITTLE_ENDIAN__
6501#define vgetq_lane_p16(__p0, __p1) __extension__ ({ \
6502  poly16x8_t __s0 = __p0; \
6503  poly16_t __ret; \
6504  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6505  __ret; \
6506})
6507#else
6508#define vgetq_lane_p16(__p0, __p1) __extension__ ({ \
6509  poly16x8_t __s0 = __p0; \
6510  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6511  poly16_t __ret; \
6512  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
6513  __ret; \
6514})
6515#define __noswap_vgetq_lane_p16(__p0, __p1) __extension__ ({ \
6516  poly16x8_t __s0 = __p0; \
6517  poly16_t __ret; \
6518  __ret = (poly16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6519  __ret; \
6520})
6521#endif
6522
6523#ifdef __LITTLE_ENDIAN__
6524#define vgetq_lane_u8(__p0, __p1) __extension__ ({ \
6525  uint8x16_t __s0 = __p0; \
6526  uint8_t __ret; \
6527  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6528  __ret; \
6529})
6530#else
6531#define vgetq_lane_u8(__p0, __p1) __extension__ ({ \
6532  uint8x16_t __s0 = __p0; \
6533  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6534  uint8_t __ret; \
6535  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
6536  __ret; \
6537})
6538#define __noswap_vgetq_lane_u8(__p0, __p1) __extension__ ({ \
6539  uint8x16_t __s0 = __p0; \
6540  uint8_t __ret; \
6541  __ret = (uint8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6542  __ret; \
6543})
6544#endif
6545
6546#ifdef __LITTLE_ENDIAN__
6547#define vgetq_lane_u32(__p0, __p1) __extension__ ({ \
6548  uint32x4_t __s0 = __p0; \
6549  uint32_t __ret; \
6550  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6551  __ret; \
6552})
6553#else
6554#define vgetq_lane_u32(__p0, __p1) __extension__ ({ \
6555  uint32x4_t __s0 = __p0; \
6556  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6557  uint32_t __ret; \
6558  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__rev0, __p1); \
6559  __ret; \
6560})
6561#define __noswap_vgetq_lane_u32(__p0, __p1) __extension__ ({ \
6562  uint32x4_t __s0 = __p0; \
6563  uint32_t __ret; \
6564  __ret = (uint32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6565  __ret; \
6566})
6567#endif
6568
6569#ifdef __LITTLE_ENDIAN__
6570#define vgetq_lane_u64(__p0, __p1) __extension__ ({ \
6571  uint64x2_t __s0 = __p0; \
6572  uint64_t __ret; \
6573  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6574  __ret; \
6575})
6576#else
6577#define vgetq_lane_u64(__p0, __p1) __extension__ ({ \
6578  uint64x2_t __s0 = __p0; \
6579  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6580  uint64_t __ret; \
6581  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
6582  __ret; \
6583})
6584#define __noswap_vgetq_lane_u64(__p0, __p1) __extension__ ({ \
6585  uint64x2_t __s0 = __p0; \
6586  uint64_t __ret; \
6587  __ret = (uint64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6588  __ret; \
6589})
6590#endif
6591
6592#ifdef __LITTLE_ENDIAN__
6593#define vgetq_lane_u16(__p0, __p1) __extension__ ({ \
6594  uint16x8_t __s0 = __p0; \
6595  uint16_t __ret; \
6596  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6597  __ret; \
6598})
6599#else
6600#define vgetq_lane_u16(__p0, __p1) __extension__ ({ \
6601  uint16x8_t __s0 = __p0; \
6602  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6603  uint16_t __ret; \
6604  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
6605  __ret; \
6606})
6607#define __noswap_vgetq_lane_u16(__p0, __p1) __extension__ ({ \
6608  uint16x8_t __s0 = __p0; \
6609  uint16_t __ret; \
6610  __ret = (uint16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6611  __ret; \
6612})
6613#endif
6614
6615#ifdef __LITTLE_ENDIAN__
6616#define vgetq_lane_s8(__p0, __p1) __extension__ ({ \
6617  int8x16_t __s0 = __p0; \
6618  int8_t __ret; \
6619  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6620  __ret; \
6621})
6622#else
6623#define vgetq_lane_s8(__p0, __p1) __extension__ ({ \
6624  int8x16_t __s0 = __p0; \
6625  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
6626  int8_t __ret; \
6627  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__rev0, __p1); \
6628  __ret; \
6629})
6630#define __noswap_vgetq_lane_s8(__p0, __p1) __extension__ ({ \
6631  int8x16_t __s0 = __p0; \
6632  int8_t __ret; \
6633  __ret = (int8_t) __builtin_neon_vgetq_lane_i8((int8x16_t)__s0, __p1); \
6634  __ret; \
6635})
6636#endif
6637
6638#ifdef __LITTLE_ENDIAN__
6639#define vgetq_lane_f32(__p0, __p1) __extension__ ({ \
6640  float32x4_t __s0 = __p0; \
6641  float32_t __ret; \
6642  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__s0, __p1); \
6643  __ret; \
6644})
6645#else
6646#define vgetq_lane_f32(__p0, __p1) __extension__ ({ \
6647  float32x4_t __s0 = __p0; \
6648  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6649  float32_t __ret; \
6650  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__rev0, __p1); \
6651  __ret; \
6652})
6653#define __noswap_vgetq_lane_f32(__p0, __p1) __extension__ ({ \
6654  float32x4_t __s0 = __p0; \
6655  float32_t __ret; \
6656  __ret = (float32_t) __builtin_neon_vgetq_lane_f32((int8x16_t)__s0, __p1); \
6657  __ret; \
6658})
6659#endif
6660
6661#ifdef __LITTLE_ENDIAN__
6662#define vgetq_lane_s32(__p0, __p1) __extension__ ({ \
6663  int32x4_t __s0 = __p0; \
6664  int32_t __ret; \
6665  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6666  __ret; \
6667})
6668#else
6669#define vgetq_lane_s32(__p0, __p1) __extension__ ({ \
6670  int32x4_t __s0 = __p0; \
6671  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6672  int32_t __ret; \
6673  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__rev0, __p1); \
6674  __ret; \
6675})
6676#define __noswap_vgetq_lane_s32(__p0, __p1) __extension__ ({ \
6677  int32x4_t __s0 = __p0; \
6678  int32_t __ret; \
6679  __ret = (int32_t) __builtin_neon_vgetq_lane_i32((int8x16_t)__s0, __p1); \
6680  __ret; \
6681})
6682#endif
6683
6684#ifdef __LITTLE_ENDIAN__
6685#define vgetq_lane_s64(__p0, __p1) __extension__ ({ \
6686  int64x2_t __s0 = __p0; \
6687  int64_t __ret; \
6688  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6689  __ret; \
6690})
6691#else
6692#define vgetq_lane_s64(__p0, __p1) __extension__ ({ \
6693  int64x2_t __s0 = __p0; \
6694  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6695  int64_t __ret; \
6696  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
6697  __ret; \
6698})
6699#define __noswap_vgetq_lane_s64(__p0, __p1) __extension__ ({ \
6700  int64x2_t __s0 = __p0; \
6701  int64_t __ret; \
6702  __ret = (int64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
6703  __ret; \
6704})
6705#endif
6706
6707#ifdef __LITTLE_ENDIAN__
6708#define vgetq_lane_s16(__p0, __p1) __extension__ ({ \
6709  int16x8_t __s0 = __p0; \
6710  int16_t __ret; \
6711  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6712  __ret; \
6713})
6714#else
6715#define vgetq_lane_s16(__p0, __p1) __extension__ ({ \
6716  int16x8_t __s0 = __p0; \
6717  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6718  int16_t __ret; \
6719  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__rev0, __p1); \
6720  __ret; \
6721})
6722#define __noswap_vgetq_lane_s16(__p0, __p1) __extension__ ({ \
6723  int16x8_t __s0 = __p0; \
6724  int16_t __ret; \
6725  __ret = (int16_t) __builtin_neon_vgetq_lane_i16((int8x16_t)__s0, __p1); \
6726  __ret; \
6727})
6728#endif
6729
6730#ifdef __LITTLE_ENDIAN__
6731#define vget_lane_u8(__p0, __p1) __extension__ ({ \
6732  uint8x8_t __s0 = __p0; \
6733  uint8_t __ret; \
6734  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6735  __ret; \
6736})
6737#else
6738#define vget_lane_u8(__p0, __p1) __extension__ ({ \
6739  uint8x8_t __s0 = __p0; \
6740  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6741  uint8_t __ret; \
6742  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
6743  __ret; \
6744})
6745#define __noswap_vget_lane_u8(__p0, __p1) __extension__ ({ \
6746  uint8x8_t __s0 = __p0; \
6747  uint8_t __ret; \
6748  __ret = (uint8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6749  __ret; \
6750})
6751#endif
6752
6753#ifdef __LITTLE_ENDIAN__
6754#define vget_lane_u32(__p0, __p1) __extension__ ({ \
6755  uint32x2_t __s0 = __p0; \
6756  uint32_t __ret; \
6757  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6758  __ret; \
6759})
6760#else
6761#define vget_lane_u32(__p0, __p1) __extension__ ({ \
6762  uint32x2_t __s0 = __p0; \
6763  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6764  uint32_t __ret; \
6765  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__rev0, __p1); \
6766  __ret; \
6767})
6768#define __noswap_vget_lane_u32(__p0, __p1) __extension__ ({ \
6769  uint32x2_t __s0 = __p0; \
6770  uint32_t __ret; \
6771  __ret = (uint32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6772  __ret; \
6773})
6774#endif
6775
6776#ifdef __LITTLE_ENDIAN__
6777#define vget_lane_u64(__p0, __p1) __extension__ ({ \
6778  uint64x1_t __s0 = __p0; \
6779  uint64_t __ret; \
6780  __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6781  __ret; \
6782})
6783#else
6784#define vget_lane_u64(__p0, __p1) __extension__ ({ \
6785  uint64x1_t __s0 = __p0; \
6786  uint64_t __ret; \
6787  __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6788  __ret; \
6789})
6790#define __noswap_vget_lane_u64(__p0, __p1) __extension__ ({ \
6791  uint64x1_t __s0 = __p0; \
6792  uint64_t __ret; \
6793  __ret = (uint64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6794  __ret; \
6795})
6796#endif
6797
6798#ifdef __LITTLE_ENDIAN__
6799#define vget_lane_u16(__p0, __p1) __extension__ ({ \
6800  uint16x4_t __s0 = __p0; \
6801  uint16_t __ret; \
6802  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6803  __ret; \
6804})
6805#else
6806#define vget_lane_u16(__p0, __p1) __extension__ ({ \
6807  uint16x4_t __s0 = __p0; \
6808  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6809  uint16_t __ret; \
6810  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
6811  __ret; \
6812})
6813#define __noswap_vget_lane_u16(__p0, __p1) __extension__ ({ \
6814  uint16x4_t __s0 = __p0; \
6815  uint16_t __ret; \
6816  __ret = (uint16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6817  __ret; \
6818})
6819#endif
6820
6821#ifdef __LITTLE_ENDIAN__
6822#define vget_lane_s8(__p0, __p1) __extension__ ({ \
6823  int8x8_t __s0 = __p0; \
6824  int8_t __ret; \
6825  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6826  __ret; \
6827})
6828#else
6829#define vget_lane_s8(__p0, __p1) __extension__ ({ \
6830  int8x8_t __s0 = __p0; \
6831  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
6832  int8_t __ret; \
6833  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__rev0, __p1); \
6834  __ret; \
6835})
6836#define __noswap_vget_lane_s8(__p0, __p1) __extension__ ({ \
6837  int8x8_t __s0 = __p0; \
6838  int8_t __ret; \
6839  __ret = (int8_t) __builtin_neon_vget_lane_i8((int8x8_t)__s0, __p1); \
6840  __ret; \
6841})
6842#endif
6843
6844#ifdef __LITTLE_ENDIAN__
6845#define vget_lane_f32(__p0, __p1) __extension__ ({ \
6846  float32x2_t __s0 = __p0; \
6847  float32_t __ret; \
6848  __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__s0, __p1); \
6849  __ret; \
6850})
6851#else
6852#define vget_lane_f32(__p0, __p1) __extension__ ({ \
6853  float32x2_t __s0 = __p0; \
6854  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6855  float32_t __ret; \
6856  __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__rev0, __p1); \
6857  __ret; \
6858})
6859#define __noswap_vget_lane_f32(__p0, __p1) __extension__ ({ \
6860  float32x2_t __s0 = __p0; \
6861  float32_t __ret; \
6862  __ret = (float32_t) __builtin_neon_vget_lane_f32((int8x8_t)__s0, __p1); \
6863  __ret; \
6864})
6865#endif
6866
6867#ifdef __LITTLE_ENDIAN__
6868#define vget_lane_s32(__p0, __p1) __extension__ ({ \
6869  int32x2_t __s0 = __p0; \
6870  int32_t __ret; \
6871  __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6872  __ret; \
6873})
6874#else
6875#define vget_lane_s32(__p0, __p1) __extension__ ({ \
6876  int32x2_t __s0 = __p0; \
6877  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
6878  int32_t __ret; \
6879  __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__rev0, __p1); \
6880  __ret; \
6881})
6882#define __noswap_vget_lane_s32(__p0, __p1) __extension__ ({ \
6883  int32x2_t __s0 = __p0; \
6884  int32_t __ret; \
6885  __ret = (int32_t) __builtin_neon_vget_lane_i32((int8x8_t)__s0, __p1); \
6886  __ret; \
6887})
6888#endif
6889
6890#ifdef __LITTLE_ENDIAN__
6891#define vget_lane_s64(__p0, __p1) __extension__ ({ \
6892  int64x1_t __s0 = __p0; \
6893  int64_t __ret; \
6894  __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6895  __ret; \
6896})
6897#else
6898#define vget_lane_s64(__p0, __p1) __extension__ ({ \
6899  int64x1_t __s0 = __p0; \
6900  int64_t __ret; \
6901  __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6902  __ret; \
6903})
6904#define __noswap_vget_lane_s64(__p0, __p1) __extension__ ({ \
6905  int64x1_t __s0 = __p0; \
6906  int64_t __ret; \
6907  __ret = (int64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
6908  __ret; \
6909})
6910#endif
6911
6912#ifdef __LITTLE_ENDIAN__
6913#define vget_lane_s16(__p0, __p1) __extension__ ({ \
6914  int16x4_t __s0 = __p0; \
6915  int16_t __ret; \
6916  __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6917  __ret; \
6918})
6919#else
6920#define vget_lane_s16(__p0, __p1) __extension__ ({ \
6921  int16x4_t __s0 = __p0; \
6922  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
6923  int16_t __ret; \
6924  __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__rev0, __p1); \
6925  __ret; \
6926})
6927#define __noswap_vget_lane_s16(__p0, __p1) __extension__ ({ \
6928  int16x4_t __s0 = __p0; \
6929  int16_t __ret; \
6930  __ret = (int16_t) __builtin_neon_vget_lane_i16((int8x8_t)__s0, __p1); \
6931  __ret; \
6932})
6933#endif
6934
6935#ifdef __LITTLE_ENDIAN__
6936__ai poly8x8_t vget_low_p8(poly8x16_t __p0) {
6937  poly8x8_t __ret;
6938  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
6939  return __ret;
6940}
6941#else
6942__ai poly8x8_t vget_low_p8(poly8x16_t __p0) {
6943  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6944  poly8x8_t __ret;
6945  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
6946  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6947  return __ret;
6948}
6949#endif
6950
6951#ifdef __LITTLE_ENDIAN__
6952__ai poly16x4_t vget_low_p16(poly16x8_t __p0) {
6953  poly16x4_t __ret;
6954  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
6955  return __ret;
6956}
6957#else
6958__ai poly16x4_t vget_low_p16(poly16x8_t __p0) {
6959  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
6960  poly16x4_t __ret;
6961  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
6962  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
6963  return __ret;
6964}
6965#endif
6966
6967#ifdef __LITTLE_ENDIAN__
6968__ai uint8x8_t vget_low_u8(uint8x16_t __p0) {
6969  uint8x8_t __ret;
6970  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
6971  return __ret;
6972}
6973#else
6974__ai uint8x8_t vget_low_u8(uint8x16_t __p0) {
6975  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
6976  uint8x8_t __ret;
6977  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
6978  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
6979  return __ret;
6980}
6981#endif
6982
6983#ifdef __LITTLE_ENDIAN__
6984__ai uint32x2_t vget_low_u32(uint32x4_t __p0) {
6985  uint32x2_t __ret;
6986  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
6987  return __ret;
6988}
6989#else
6990__ai uint32x2_t vget_low_u32(uint32x4_t __p0) {
6991  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
6992  uint32x2_t __ret;
6993  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
6994  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
6995  return __ret;
6996}
6997#endif
6998
6999#ifdef __LITTLE_ENDIAN__
7000__ai uint64x1_t vget_low_u64(uint64x2_t __p0) {
7001  uint64x1_t __ret;
7002  __ret = __builtin_shufflevector(__p0, __p0, 0);
7003  return __ret;
7004}
7005#else
7006__ai uint64x1_t vget_low_u64(uint64x2_t __p0) {
7007  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7008  uint64x1_t __ret;
7009  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
7010  return __ret;
7011}
7012#endif
7013
7014#ifdef __LITTLE_ENDIAN__
7015__ai uint16x4_t vget_low_u16(uint16x8_t __p0) {
7016  uint16x4_t __ret;
7017  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7018  return __ret;
7019}
7020#else
7021__ai uint16x4_t vget_low_u16(uint16x8_t __p0) {
7022  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7023  uint16x4_t __ret;
7024  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7025  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7026  return __ret;
7027}
7028#endif
7029
7030#ifdef __LITTLE_ENDIAN__
7031__ai int8x8_t vget_low_s8(int8x16_t __p0) {
7032  int8x8_t __ret;
7033  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3, 4, 5, 6, 7);
7034  return __ret;
7035}
7036#else
7037__ai int8x8_t vget_low_s8(int8x16_t __p0) {
7038  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7039  int8x8_t __ret;
7040  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3, 4, 5, 6, 7);
7041  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7042  return __ret;
7043}
7044#endif
7045
7046#ifdef __LITTLE_ENDIAN__
7047__ai float32x2_t vget_low_f32(float32x4_t __p0) {
7048  float32x2_t __ret;
7049  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
7050  return __ret;
7051}
7052#else
7053__ai float32x2_t vget_low_f32(float32x4_t __p0) {
7054  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7055  float32x2_t __ret;
7056  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
7057  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7058  return __ret;
7059}
7060#endif
7061
7062#ifdef __LITTLE_ENDIAN__
7063__ai float16x4_t vget_low_f16(float16x8_t __p0) {
7064  float16x4_t __ret;
7065  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7066  return __ret;
7067}
7068#else
7069__ai float16x4_t vget_low_f16(float16x8_t __p0) {
7070  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7071  float16x4_t __ret;
7072  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7073  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7074  return __ret;
7075}
7076#endif
7077
7078#ifdef __LITTLE_ENDIAN__
7079__ai int32x2_t vget_low_s32(int32x4_t __p0) {
7080  int32x2_t __ret;
7081  __ret = __builtin_shufflevector(__p0, __p0, 0, 1);
7082  return __ret;
7083}
7084#else
7085__ai int32x2_t vget_low_s32(int32x4_t __p0) {
7086  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7087  int32x2_t __ret;
7088  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1);
7089  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7090  return __ret;
7091}
7092#endif
7093
7094#ifdef __LITTLE_ENDIAN__
7095__ai int64x1_t vget_low_s64(int64x2_t __p0) {
7096  int64x1_t __ret;
7097  __ret = __builtin_shufflevector(__p0, __p0, 0);
7098  return __ret;
7099}
7100#else
7101__ai int64x1_t vget_low_s64(int64x2_t __p0) {
7102  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7103  int64x1_t __ret;
7104  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
7105  return __ret;
7106}
7107#endif
7108
7109#ifdef __LITTLE_ENDIAN__
7110__ai int16x4_t vget_low_s16(int16x8_t __p0) {
7111  int16x4_t __ret;
7112  __ret = __builtin_shufflevector(__p0, __p0, 0, 1, 2, 3);
7113  return __ret;
7114}
7115#else
7116__ai int16x4_t vget_low_s16(int16x8_t __p0) {
7117  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7118  int16x4_t __ret;
7119  __ret = __builtin_shufflevector(__rev0, __rev0, 0, 1, 2, 3);
7120  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7121  return __ret;
7122}
7123#endif
7124
7125#ifdef __LITTLE_ENDIAN__
7126__ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7127  uint8x16_t __ret;
7128  __ret = (uint8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
7129  return __ret;
7130}
7131#else
7132__ai uint8x16_t vhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7133  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7134  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7135  uint8x16_t __ret;
7136  __ret = (uint8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
7137  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7138  return __ret;
7139}
7140#endif
7141
7142#ifdef __LITTLE_ENDIAN__
7143__ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7144  uint32x4_t __ret;
7145  __ret = (uint32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
7146  return __ret;
7147}
7148#else
7149__ai uint32x4_t vhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7150  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7151  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7152  uint32x4_t __ret;
7153  __ret = (uint32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
7154  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7155  return __ret;
7156}
7157#endif
7158
7159#ifdef __LITTLE_ENDIAN__
7160__ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7161  uint16x8_t __ret;
7162  __ret = (uint16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
7163  return __ret;
7164}
7165#else
7166__ai uint16x8_t vhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7167  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7168  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7169  uint16x8_t __ret;
7170  __ret = (uint16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
7171  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7172  return __ret;
7173}
7174#endif
7175
7176#ifdef __LITTLE_ENDIAN__
7177__ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
7178  int8x16_t __ret;
7179  __ret = (int8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
7180  return __ret;
7181}
7182#else
7183__ai int8x16_t vhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
7184  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7185  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7186  int8x16_t __ret;
7187  __ret = (int8x16_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
7188  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7189  return __ret;
7190}
7191#endif
7192
7193#ifdef __LITTLE_ENDIAN__
7194__ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
7195  int32x4_t __ret;
7196  __ret = (int32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
7197  return __ret;
7198}
7199#else
7200__ai int32x4_t vhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
7201  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7202  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7203  int32x4_t __ret;
7204  __ret = (int32x4_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
7205  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7206  return __ret;
7207}
7208#endif
7209
7210#ifdef __LITTLE_ENDIAN__
7211__ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
7212  int16x8_t __ret;
7213  __ret = (int16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
7214  return __ret;
7215}
7216#else
7217__ai int16x8_t vhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
7218  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7219  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7220  int16x8_t __ret;
7221  __ret = (int16x8_t) __builtin_neon_vhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
7222  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7223  return __ret;
7224}
7225#endif
7226
7227#ifdef __LITTLE_ENDIAN__
7228__ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
7229  uint8x8_t __ret;
7230  __ret = (uint8x8_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
7231  return __ret;
7232}
7233#else
7234__ai uint8x8_t vhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
7235  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7236  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7237  uint8x8_t __ret;
7238  __ret = (uint8x8_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
7239  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7240  return __ret;
7241}
7242#endif
7243
7244#ifdef __LITTLE_ENDIAN__
7245__ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
7246  uint32x2_t __ret;
7247  __ret = (uint32x2_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
7248  return __ret;
7249}
7250#else
7251__ai uint32x2_t vhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
7252  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7253  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7254  uint32x2_t __ret;
7255  __ret = (uint32x2_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
7256  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7257  return __ret;
7258}
7259#endif
7260
7261#ifdef __LITTLE_ENDIAN__
7262__ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
7263  uint16x4_t __ret;
7264  __ret = (uint16x4_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
7265  return __ret;
7266}
7267#else
7268__ai uint16x4_t vhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
7269  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7270  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7271  uint16x4_t __ret;
7272  __ret = (uint16x4_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
7273  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7274  return __ret;
7275}
7276#endif
7277
7278#ifdef __LITTLE_ENDIAN__
7279__ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1) {
7280  int8x8_t __ret;
7281  __ret = (int8x8_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
7282  return __ret;
7283}
7284#else
7285__ai int8x8_t vhadd_s8(int8x8_t __p0, int8x8_t __p1) {
7286  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7287  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7288  int8x8_t __ret;
7289  __ret = (int8x8_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
7290  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7291  return __ret;
7292}
7293#endif
7294
7295#ifdef __LITTLE_ENDIAN__
7296__ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1) {
7297  int32x2_t __ret;
7298  __ret = (int32x2_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
7299  return __ret;
7300}
7301#else
7302__ai int32x2_t vhadd_s32(int32x2_t __p0, int32x2_t __p1) {
7303  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7304  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7305  int32x2_t __ret;
7306  __ret = (int32x2_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
7307  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7308  return __ret;
7309}
7310#endif
7311
7312#ifdef __LITTLE_ENDIAN__
7313__ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1) {
7314  int16x4_t __ret;
7315  __ret = (int16x4_t) __builtin_neon_vhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
7316  return __ret;
7317}
7318#else
7319__ai int16x4_t vhadd_s16(int16x4_t __p0, int16x4_t __p1) {
7320  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7321  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7322  int16x4_t __ret;
7323  __ret = (int16x4_t) __builtin_neon_vhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
7324  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7325  return __ret;
7326}
7327#endif
7328
7329#ifdef __LITTLE_ENDIAN__
7330__ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7331  uint8x16_t __ret;
7332  __ret = (uint8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
7333  return __ret;
7334}
7335#else
7336__ai uint8x16_t vhsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
7337  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7338  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7339  uint8x16_t __ret;
7340  __ret = (uint8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
7341  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7342  return __ret;
7343}
7344#endif
7345
7346#ifdef __LITTLE_ENDIAN__
7347__ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7348  uint32x4_t __ret;
7349  __ret = (uint32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
7350  return __ret;
7351}
7352#else
7353__ai uint32x4_t vhsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
7354  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7355  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7356  uint32x4_t __ret;
7357  __ret = (uint32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
7358  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7359  return __ret;
7360}
7361#endif
7362
7363#ifdef __LITTLE_ENDIAN__
7364__ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7365  uint16x8_t __ret;
7366  __ret = (uint16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
7367  return __ret;
7368}
7369#else
7370__ai uint16x8_t vhsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
7371  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7372  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7373  uint16x8_t __ret;
7374  __ret = (uint16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
7375  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7376  return __ret;
7377}
7378#endif
7379
7380#ifdef __LITTLE_ENDIAN__
7381__ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1) {
7382  int8x16_t __ret;
7383  __ret = (int8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
7384  return __ret;
7385}
7386#else
7387__ai int8x16_t vhsubq_s8(int8x16_t __p0, int8x16_t __p1) {
7388  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7389  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7390  int8x16_t __ret;
7391  __ret = (int8x16_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
7392  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
7393  return __ret;
7394}
7395#endif
7396
7397#ifdef __LITTLE_ENDIAN__
7398__ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1) {
7399  int32x4_t __ret;
7400  __ret = (int32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
7401  return __ret;
7402}
7403#else
7404__ai int32x4_t vhsubq_s32(int32x4_t __p0, int32x4_t __p1) {
7405  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7406  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7407  int32x4_t __ret;
7408  __ret = (int32x4_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
7409  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7410  return __ret;
7411}
7412#endif
7413
7414#ifdef __LITTLE_ENDIAN__
7415__ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1) {
7416  int16x8_t __ret;
7417  __ret = (int16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
7418  return __ret;
7419}
7420#else
7421__ai int16x8_t vhsubq_s16(int16x8_t __p0, int16x8_t __p1) {
7422  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7423  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7424  int16x8_t __ret;
7425  __ret = (int16x8_t) __builtin_neon_vhsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
7426  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7427  return __ret;
7428}
7429#endif
7430
7431#ifdef __LITTLE_ENDIAN__
7432__ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
7433  uint8x8_t __ret;
7434  __ret = (uint8x8_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
7435  return __ret;
7436}
7437#else
7438__ai uint8x8_t vhsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
7439  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7440  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7441  uint8x8_t __ret;
7442  __ret = (uint8x8_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
7443  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7444  return __ret;
7445}
7446#endif
7447
7448#ifdef __LITTLE_ENDIAN__
7449__ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
7450  uint32x2_t __ret;
7451  __ret = (uint32x2_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
7452  return __ret;
7453}
7454#else
7455__ai uint32x2_t vhsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
7456  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7457  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7458  uint32x2_t __ret;
7459  __ret = (uint32x2_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
7460  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7461  return __ret;
7462}
7463#endif
7464
7465#ifdef __LITTLE_ENDIAN__
7466__ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
7467  uint16x4_t __ret;
7468  __ret = (uint16x4_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
7469  return __ret;
7470}
7471#else
7472__ai uint16x4_t vhsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
7473  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7474  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7475  uint16x4_t __ret;
7476  __ret = (uint16x4_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
7477  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7478  return __ret;
7479}
7480#endif
7481
7482#ifdef __LITTLE_ENDIAN__
7483__ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1) {
7484  int8x8_t __ret;
7485  __ret = (int8x8_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
7486  return __ret;
7487}
7488#else
7489__ai int8x8_t vhsub_s8(int8x8_t __p0, int8x8_t __p1) {
7490  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
7491  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
7492  int8x8_t __ret;
7493  __ret = (int8x8_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
7494  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
7495  return __ret;
7496}
7497#endif
7498
7499#ifdef __LITTLE_ENDIAN__
7500__ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1) {
7501  int32x2_t __ret;
7502  __ret = (int32x2_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
7503  return __ret;
7504}
7505#else
7506__ai int32x2_t vhsub_s32(int32x2_t __p0, int32x2_t __p1) {
7507  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
7508  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
7509  int32x2_t __ret;
7510  __ret = (int32x2_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
7511  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
7512  return __ret;
7513}
7514#endif
7515
7516#ifdef __LITTLE_ENDIAN__
7517__ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1) {
7518  int16x4_t __ret;
7519  __ret = (int16x4_t) __builtin_neon_vhsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
7520  return __ret;
7521}
7522#else
7523__ai int16x4_t vhsub_s16(int16x4_t __p0, int16x4_t __p1) {
7524  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
7525  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
7526  int16x4_t __ret;
7527  __ret = (int16x4_t) __builtin_neon_vhsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
7528  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
7529  return __ret;
7530}
7531#endif
7532
7533#ifdef __LITTLE_ENDIAN__
7534#define vld1_p8(__p0) __extension__ ({ \
7535  poly8x8_t __ret; \
7536  __ret = (poly8x8_t) __builtin_neon_vld1_v(__p0, 4); \
7537  __ret; \
7538})
7539#else
7540#define vld1_p8(__p0) __extension__ ({ \
7541  poly8x8_t __ret; \
7542  __ret = (poly8x8_t) __builtin_neon_vld1_v(__p0, 4); \
7543  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7544  __ret; \
7545})
7546#endif
7547
7548#ifdef __LITTLE_ENDIAN__
7549#define vld1_p16(__p0) __extension__ ({ \
7550  poly16x4_t __ret; \
7551  __ret = (poly16x4_t) __builtin_neon_vld1_v(__p0, 5); \
7552  __ret; \
7553})
7554#else
7555#define vld1_p16(__p0) __extension__ ({ \
7556  poly16x4_t __ret; \
7557  __ret = (poly16x4_t) __builtin_neon_vld1_v(__p0, 5); \
7558  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7559  __ret; \
7560})
7561#endif
7562
7563#ifdef __LITTLE_ENDIAN__
7564#define vld1q_p8(__p0) __extension__ ({ \
7565  poly8x16_t __ret; \
7566  __ret = (poly8x16_t) __builtin_neon_vld1q_v(__p0, 36); \
7567  __ret; \
7568})
7569#else
7570#define vld1q_p8(__p0) __extension__ ({ \
7571  poly8x16_t __ret; \
7572  __ret = (poly8x16_t) __builtin_neon_vld1q_v(__p0, 36); \
7573  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7574  __ret; \
7575})
7576#endif
7577
7578#ifdef __LITTLE_ENDIAN__
7579#define vld1q_p16(__p0) __extension__ ({ \
7580  poly16x8_t __ret; \
7581  __ret = (poly16x8_t) __builtin_neon_vld1q_v(__p0, 37); \
7582  __ret; \
7583})
7584#else
7585#define vld1q_p16(__p0) __extension__ ({ \
7586  poly16x8_t __ret; \
7587  __ret = (poly16x8_t) __builtin_neon_vld1q_v(__p0, 37); \
7588  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7589  __ret; \
7590})
7591#endif
7592
7593#ifdef __LITTLE_ENDIAN__
7594#define vld1q_u8(__p0) __extension__ ({ \
7595  uint8x16_t __ret; \
7596  __ret = (uint8x16_t) __builtin_neon_vld1q_v(__p0, 48); \
7597  __ret; \
7598})
7599#else
7600#define vld1q_u8(__p0) __extension__ ({ \
7601  uint8x16_t __ret; \
7602  __ret = (uint8x16_t) __builtin_neon_vld1q_v(__p0, 48); \
7603  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7604  __ret; \
7605})
7606#endif
7607
7608#ifdef __LITTLE_ENDIAN__
7609#define vld1q_u32(__p0) __extension__ ({ \
7610  uint32x4_t __ret; \
7611  __ret = (uint32x4_t) __builtin_neon_vld1q_v(__p0, 50); \
7612  __ret; \
7613})
7614#else
7615#define vld1q_u32(__p0) __extension__ ({ \
7616  uint32x4_t __ret; \
7617  __ret = (uint32x4_t) __builtin_neon_vld1q_v(__p0, 50); \
7618  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7619  __ret; \
7620})
7621#endif
7622
7623#ifdef __LITTLE_ENDIAN__
7624#define vld1q_u64(__p0) __extension__ ({ \
7625  uint64x2_t __ret; \
7626  __ret = (uint64x2_t) __builtin_neon_vld1q_v(__p0, 51); \
7627  __ret; \
7628})
7629#else
7630#define vld1q_u64(__p0) __extension__ ({ \
7631  uint64x2_t __ret; \
7632  __ret = (uint64x2_t) __builtin_neon_vld1q_v(__p0, 51); \
7633  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7634  __ret; \
7635})
7636#endif
7637
7638#ifdef __LITTLE_ENDIAN__
7639#define vld1q_u16(__p0) __extension__ ({ \
7640  uint16x8_t __ret; \
7641  __ret = (uint16x8_t) __builtin_neon_vld1q_v(__p0, 49); \
7642  __ret; \
7643})
7644#else
7645#define vld1q_u16(__p0) __extension__ ({ \
7646  uint16x8_t __ret; \
7647  __ret = (uint16x8_t) __builtin_neon_vld1q_v(__p0, 49); \
7648  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7649  __ret; \
7650})
7651#endif
7652
7653#ifdef __LITTLE_ENDIAN__
7654#define vld1q_s8(__p0) __extension__ ({ \
7655  int8x16_t __ret; \
7656  __ret = (int8x16_t) __builtin_neon_vld1q_v(__p0, 32); \
7657  __ret; \
7658})
7659#else
7660#define vld1q_s8(__p0) __extension__ ({ \
7661  int8x16_t __ret; \
7662  __ret = (int8x16_t) __builtin_neon_vld1q_v(__p0, 32); \
7663  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7664  __ret; \
7665})
7666#endif
7667
7668#ifdef __LITTLE_ENDIAN__
7669#define vld1q_f32(__p0) __extension__ ({ \
7670  float32x4_t __ret; \
7671  __ret = (float32x4_t) __builtin_neon_vld1q_v(__p0, 41); \
7672  __ret; \
7673})
7674#else
7675#define vld1q_f32(__p0) __extension__ ({ \
7676  float32x4_t __ret; \
7677  __ret = (float32x4_t) __builtin_neon_vld1q_v(__p0, 41); \
7678  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7679  __ret; \
7680})
7681#endif
7682
7683#ifdef __LITTLE_ENDIAN__
7684#define vld1q_f16(__p0) __extension__ ({ \
7685  float16x8_t __ret; \
7686  __ret = (float16x8_t) __builtin_neon_vld1q_v(__p0, 40); \
7687  __ret; \
7688})
7689#else
7690#define vld1q_f16(__p0) __extension__ ({ \
7691  float16x8_t __ret; \
7692  __ret = (float16x8_t) __builtin_neon_vld1q_v(__p0, 40); \
7693  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7694  __ret; \
7695})
7696#endif
7697
7698#ifdef __LITTLE_ENDIAN__
7699#define vld1q_s32(__p0) __extension__ ({ \
7700  int32x4_t __ret; \
7701  __ret = (int32x4_t) __builtin_neon_vld1q_v(__p0, 34); \
7702  __ret; \
7703})
7704#else
7705#define vld1q_s32(__p0) __extension__ ({ \
7706  int32x4_t __ret; \
7707  __ret = (int32x4_t) __builtin_neon_vld1q_v(__p0, 34); \
7708  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7709  __ret; \
7710})
7711#endif
7712
7713#ifdef __LITTLE_ENDIAN__
7714#define vld1q_s64(__p0) __extension__ ({ \
7715  int64x2_t __ret; \
7716  __ret = (int64x2_t) __builtin_neon_vld1q_v(__p0, 35); \
7717  __ret; \
7718})
7719#else
7720#define vld1q_s64(__p0) __extension__ ({ \
7721  int64x2_t __ret; \
7722  __ret = (int64x2_t) __builtin_neon_vld1q_v(__p0, 35); \
7723  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7724  __ret; \
7725})
7726#endif
7727
7728#ifdef __LITTLE_ENDIAN__
7729#define vld1q_s16(__p0) __extension__ ({ \
7730  int16x8_t __ret; \
7731  __ret = (int16x8_t) __builtin_neon_vld1q_v(__p0, 33); \
7732  __ret; \
7733})
7734#else
7735#define vld1q_s16(__p0) __extension__ ({ \
7736  int16x8_t __ret; \
7737  __ret = (int16x8_t) __builtin_neon_vld1q_v(__p0, 33); \
7738  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7739  __ret; \
7740})
7741#endif
7742
7743#ifdef __LITTLE_ENDIAN__
7744#define vld1_u8(__p0) __extension__ ({ \
7745  uint8x8_t __ret; \
7746  __ret = (uint8x8_t) __builtin_neon_vld1_v(__p0, 16); \
7747  __ret; \
7748})
7749#else
7750#define vld1_u8(__p0) __extension__ ({ \
7751  uint8x8_t __ret; \
7752  __ret = (uint8x8_t) __builtin_neon_vld1_v(__p0, 16); \
7753  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7754  __ret; \
7755})
7756#endif
7757
7758#ifdef __LITTLE_ENDIAN__
7759#define vld1_u32(__p0) __extension__ ({ \
7760  uint32x2_t __ret; \
7761  __ret = (uint32x2_t) __builtin_neon_vld1_v(__p0, 18); \
7762  __ret; \
7763})
7764#else
7765#define vld1_u32(__p0) __extension__ ({ \
7766  uint32x2_t __ret; \
7767  __ret = (uint32x2_t) __builtin_neon_vld1_v(__p0, 18); \
7768  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7769  __ret; \
7770})
7771#endif
7772
7773#ifdef __LITTLE_ENDIAN__
7774#define vld1_u64(__p0) __extension__ ({ \
7775  uint64x1_t __ret; \
7776  __ret = (uint64x1_t) __builtin_neon_vld1_v(__p0, 19); \
7777  __ret; \
7778})
7779#else
7780#define vld1_u64(__p0) __extension__ ({ \
7781  uint64x1_t __ret; \
7782  __ret = (uint64x1_t) __builtin_neon_vld1_v(__p0, 19); \
7783  __ret; \
7784})
7785#endif
7786
7787#ifdef __LITTLE_ENDIAN__
7788#define vld1_u16(__p0) __extension__ ({ \
7789  uint16x4_t __ret; \
7790  __ret = (uint16x4_t) __builtin_neon_vld1_v(__p0, 17); \
7791  __ret; \
7792})
7793#else
7794#define vld1_u16(__p0) __extension__ ({ \
7795  uint16x4_t __ret; \
7796  __ret = (uint16x4_t) __builtin_neon_vld1_v(__p0, 17); \
7797  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7798  __ret; \
7799})
7800#endif
7801
7802#ifdef __LITTLE_ENDIAN__
7803#define vld1_s8(__p0) __extension__ ({ \
7804  int8x8_t __ret; \
7805  __ret = (int8x8_t) __builtin_neon_vld1_v(__p0, 0); \
7806  __ret; \
7807})
7808#else
7809#define vld1_s8(__p0) __extension__ ({ \
7810  int8x8_t __ret; \
7811  __ret = (int8x8_t) __builtin_neon_vld1_v(__p0, 0); \
7812  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7813  __ret; \
7814})
7815#endif
7816
7817#ifdef __LITTLE_ENDIAN__
7818#define vld1_f32(__p0) __extension__ ({ \
7819  float32x2_t __ret; \
7820  __ret = (float32x2_t) __builtin_neon_vld1_v(__p0, 9); \
7821  __ret; \
7822})
7823#else
7824#define vld1_f32(__p0) __extension__ ({ \
7825  float32x2_t __ret; \
7826  __ret = (float32x2_t) __builtin_neon_vld1_v(__p0, 9); \
7827  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7828  __ret; \
7829})
7830#endif
7831
7832#ifdef __LITTLE_ENDIAN__
7833#define vld1_f16(__p0) __extension__ ({ \
7834  float16x4_t __ret; \
7835  __ret = (float16x4_t) __builtin_neon_vld1_v(__p0, 8); \
7836  __ret; \
7837})
7838#else
7839#define vld1_f16(__p0) __extension__ ({ \
7840  float16x4_t __ret; \
7841  __ret = (float16x4_t) __builtin_neon_vld1_v(__p0, 8); \
7842  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7843  __ret; \
7844})
7845#endif
7846
7847#ifdef __LITTLE_ENDIAN__
7848#define vld1_s32(__p0) __extension__ ({ \
7849  int32x2_t __ret; \
7850  __ret = (int32x2_t) __builtin_neon_vld1_v(__p0, 2); \
7851  __ret; \
7852})
7853#else
7854#define vld1_s32(__p0) __extension__ ({ \
7855  int32x2_t __ret; \
7856  __ret = (int32x2_t) __builtin_neon_vld1_v(__p0, 2); \
7857  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7858  __ret; \
7859})
7860#endif
7861
7862#ifdef __LITTLE_ENDIAN__
7863#define vld1_s64(__p0) __extension__ ({ \
7864  int64x1_t __ret; \
7865  __ret = (int64x1_t) __builtin_neon_vld1_v(__p0, 3); \
7866  __ret; \
7867})
7868#else
7869#define vld1_s64(__p0) __extension__ ({ \
7870  int64x1_t __ret; \
7871  __ret = (int64x1_t) __builtin_neon_vld1_v(__p0, 3); \
7872  __ret; \
7873})
7874#endif
7875
7876#ifdef __LITTLE_ENDIAN__
7877#define vld1_s16(__p0) __extension__ ({ \
7878  int16x4_t __ret; \
7879  __ret = (int16x4_t) __builtin_neon_vld1_v(__p0, 1); \
7880  __ret; \
7881})
7882#else
7883#define vld1_s16(__p0) __extension__ ({ \
7884  int16x4_t __ret; \
7885  __ret = (int16x4_t) __builtin_neon_vld1_v(__p0, 1); \
7886  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7887  __ret; \
7888})
7889#endif
7890
7891#ifdef __LITTLE_ENDIAN__
7892#define vld1_dup_p8(__p0) __extension__ ({ \
7893  poly8x8_t __ret; \
7894  __ret = (poly8x8_t) __builtin_neon_vld1_dup_v(__p0, 4); \
7895  __ret; \
7896})
7897#else
7898#define vld1_dup_p8(__p0) __extension__ ({ \
7899  poly8x8_t __ret; \
7900  __ret = (poly8x8_t) __builtin_neon_vld1_dup_v(__p0, 4); \
7901  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7902  __ret; \
7903})
7904#endif
7905
7906#ifdef __LITTLE_ENDIAN__
7907#define vld1_dup_p16(__p0) __extension__ ({ \
7908  poly16x4_t __ret; \
7909  __ret = (poly16x4_t) __builtin_neon_vld1_dup_v(__p0, 5); \
7910  __ret; \
7911})
7912#else
7913#define vld1_dup_p16(__p0) __extension__ ({ \
7914  poly16x4_t __ret; \
7915  __ret = (poly16x4_t) __builtin_neon_vld1_dup_v(__p0, 5); \
7916  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7917  __ret; \
7918})
7919#endif
7920
7921#ifdef __LITTLE_ENDIAN__
7922#define vld1q_dup_p8(__p0) __extension__ ({ \
7923  poly8x16_t __ret; \
7924  __ret = (poly8x16_t) __builtin_neon_vld1q_dup_v(__p0, 36); \
7925  __ret; \
7926})
7927#else
7928#define vld1q_dup_p8(__p0) __extension__ ({ \
7929  poly8x16_t __ret; \
7930  __ret = (poly8x16_t) __builtin_neon_vld1q_dup_v(__p0, 36); \
7931  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7932  __ret; \
7933})
7934#endif
7935
7936#ifdef __LITTLE_ENDIAN__
7937#define vld1q_dup_p16(__p0) __extension__ ({ \
7938  poly16x8_t __ret; \
7939  __ret = (poly16x8_t) __builtin_neon_vld1q_dup_v(__p0, 37); \
7940  __ret; \
7941})
7942#else
7943#define vld1q_dup_p16(__p0) __extension__ ({ \
7944  poly16x8_t __ret; \
7945  __ret = (poly16x8_t) __builtin_neon_vld1q_dup_v(__p0, 37); \
7946  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
7947  __ret; \
7948})
7949#endif
7950
7951#ifdef __LITTLE_ENDIAN__
7952#define vld1q_dup_u8(__p0) __extension__ ({ \
7953  uint8x16_t __ret; \
7954  __ret = (uint8x16_t) __builtin_neon_vld1q_dup_v(__p0, 48); \
7955  __ret; \
7956})
7957#else
7958#define vld1q_dup_u8(__p0) __extension__ ({ \
7959  uint8x16_t __ret; \
7960  __ret = (uint8x16_t) __builtin_neon_vld1q_dup_v(__p0, 48); \
7961  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
7962  __ret; \
7963})
7964#endif
7965
7966#ifdef __LITTLE_ENDIAN__
7967#define vld1q_dup_u32(__p0) __extension__ ({ \
7968  uint32x4_t __ret; \
7969  __ret = (uint32x4_t) __builtin_neon_vld1q_dup_v(__p0, 50); \
7970  __ret; \
7971})
7972#else
7973#define vld1q_dup_u32(__p0) __extension__ ({ \
7974  uint32x4_t __ret; \
7975  __ret = (uint32x4_t) __builtin_neon_vld1q_dup_v(__p0, 50); \
7976  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
7977  __ret; \
7978})
7979#endif
7980
7981#ifdef __LITTLE_ENDIAN__
7982#define vld1q_dup_u64(__p0) __extension__ ({ \
7983  uint64x2_t __ret; \
7984  __ret = (uint64x2_t) __builtin_neon_vld1q_dup_v(__p0, 51); \
7985  __ret; \
7986})
7987#else
7988#define vld1q_dup_u64(__p0) __extension__ ({ \
7989  uint64x2_t __ret; \
7990  __ret = (uint64x2_t) __builtin_neon_vld1q_dup_v(__p0, 51); \
7991  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
7992  __ret; \
7993})
7994#endif
7995
7996#ifdef __LITTLE_ENDIAN__
7997#define vld1q_dup_u16(__p0) __extension__ ({ \
7998  uint16x8_t __ret; \
7999  __ret = (uint16x8_t) __builtin_neon_vld1q_dup_v(__p0, 49); \
8000  __ret; \
8001})
8002#else
8003#define vld1q_dup_u16(__p0) __extension__ ({ \
8004  uint16x8_t __ret; \
8005  __ret = (uint16x8_t) __builtin_neon_vld1q_dup_v(__p0, 49); \
8006  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8007  __ret; \
8008})
8009#endif
8010
8011#ifdef __LITTLE_ENDIAN__
8012#define vld1q_dup_s8(__p0) __extension__ ({ \
8013  int8x16_t __ret; \
8014  __ret = (int8x16_t) __builtin_neon_vld1q_dup_v(__p0, 32); \
8015  __ret; \
8016})
8017#else
8018#define vld1q_dup_s8(__p0) __extension__ ({ \
8019  int8x16_t __ret; \
8020  __ret = (int8x16_t) __builtin_neon_vld1q_dup_v(__p0, 32); \
8021  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8022  __ret; \
8023})
8024#endif
8025
8026#ifdef __LITTLE_ENDIAN__
8027#define vld1q_dup_f32(__p0) __extension__ ({ \
8028  float32x4_t __ret; \
8029  __ret = (float32x4_t) __builtin_neon_vld1q_dup_v(__p0, 41); \
8030  __ret; \
8031})
8032#else
8033#define vld1q_dup_f32(__p0) __extension__ ({ \
8034  float32x4_t __ret; \
8035  __ret = (float32x4_t) __builtin_neon_vld1q_dup_v(__p0, 41); \
8036  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8037  __ret; \
8038})
8039#endif
8040
8041#ifdef __LITTLE_ENDIAN__
8042#define vld1q_dup_f16(__p0) __extension__ ({ \
8043  float16x8_t __ret; \
8044  __ret = (float16x8_t) __builtin_neon_vld1q_dup_v(__p0, 40); \
8045  __ret; \
8046})
8047#else
8048#define vld1q_dup_f16(__p0) __extension__ ({ \
8049  float16x8_t __ret; \
8050  __ret = (float16x8_t) __builtin_neon_vld1q_dup_v(__p0, 40); \
8051  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8052  __ret; \
8053})
8054#endif
8055
8056#ifdef __LITTLE_ENDIAN__
8057#define vld1q_dup_s32(__p0) __extension__ ({ \
8058  int32x4_t __ret; \
8059  __ret = (int32x4_t) __builtin_neon_vld1q_dup_v(__p0, 34); \
8060  __ret; \
8061})
8062#else
8063#define vld1q_dup_s32(__p0) __extension__ ({ \
8064  int32x4_t __ret; \
8065  __ret = (int32x4_t) __builtin_neon_vld1q_dup_v(__p0, 34); \
8066  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8067  __ret; \
8068})
8069#endif
8070
8071#ifdef __LITTLE_ENDIAN__
8072#define vld1q_dup_s64(__p0) __extension__ ({ \
8073  int64x2_t __ret; \
8074  __ret = (int64x2_t) __builtin_neon_vld1q_dup_v(__p0, 35); \
8075  __ret; \
8076})
8077#else
8078#define vld1q_dup_s64(__p0) __extension__ ({ \
8079  int64x2_t __ret; \
8080  __ret = (int64x2_t) __builtin_neon_vld1q_dup_v(__p0, 35); \
8081  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8082  __ret; \
8083})
8084#endif
8085
8086#ifdef __LITTLE_ENDIAN__
8087#define vld1q_dup_s16(__p0) __extension__ ({ \
8088  int16x8_t __ret; \
8089  __ret = (int16x8_t) __builtin_neon_vld1q_dup_v(__p0, 33); \
8090  __ret; \
8091})
8092#else
8093#define vld1q_dup_s16(__p0) __extension__ ({ \
8094  int16x8_t __ret; \
8095  __ret = (int16x8_t) __builtin_neon_vld1q_dup_v(__p0, 33); \
8096  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8097  __ret; \
8098})
8099#endif
8100
8101#ifdef __LITTLE_ENDIAN__
8102#define vld1_dup_u8(__p0) __extension__ ({ \
8103  uint8x8_t __ret; \
8104  __ret = (uint8x8_t) __builtin_neon_vld1_dup_v(__p0, 16); \
8105  __ret; \
8106})
8107#else
8108#define vld1_dup_u8(__p0) __extension__ ({ \
8109  uint8x8_t __ret; \
8110  __ret = (uint8x8_t) __builtin_neon_vld1_dup_v(__p0, 16); \
8111  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8112  __ret; \
8113})
8114#endif
8115
8116#ifdef __LITTLE_ENDIAN__
8117#define vld1_dup_u32(__p0) __extension__ ({ \
8118  uint32x2_t __ret; \
8119  __ret = (uint32x2_t) __builtin_neon_vld1_dup_v(__p0, 18); \
8120  __ret; \
8121})
8122#else
8123#define vld1_dup_u32(__p0) __extension__ ({ \
8124  uint32x2_t __ret; \
8125  __ret = (uint32x2_t) __builtin_neon_vld1_dup_v(__p0, 18); \
8126  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8127  __ret; \
8128})
8129#endif
8130
8131#ifdef __LITTLE_ENDIAN__
8132#define vld1_dup_u64(__p0) __extension__ ({ \
8133  uint64x1_t __ret; \
8134  __ret = (uint64x1_t) __builtin_neon_vld1_dup_v(__p0, 19); \
8135  __ret; \
8136})
8137#else
8138#define vld1_dup_u64(__p0) __extension__ ({ \
8139  uint64x1_t __ret; \
8140  __ret = (uint64x1_t) __builtin_neon_vld1_dup_v(__p0, 19); \
8141  __ret; \
8142})
8143#endif
8144
8145#ifdef __LITTLE_ENDIAN__
8146#define vld1_dup_u16(__p0) __extension__ ({ \
8147  uint16x4_t __ret; \
8148  __ret = (uint16x4_t) __builtin_neon_vld1_dup_v(__p0, 17); \
8149  __ret; \
8150})
8151#else
8152#define vld1_dup_u16(__p0) __extension__ ({ \
8153  uint16x4_t __ret; \
8154  __ret = (uint16x4_t) __builtin_neon_vld1_dup_v(__p0, 17); \
8155  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8156  __ret; \
8157})
8158#endif
8159
8160#ifdef __LITTLE_ENDIAN__
8161#define vld1_dup_s8(__p0) __extension__ ({ \
8162  int8x8_t __ret; \
8163  __ret = (int8x8_t) __builtin_neon_vld1_dup_v(__p0, 0); \
8164  __ret; \
8165})
8166#else
8167#define vld1_dup_s8(__p0) __extension__ ({ \
8168  int8x8_t __ret; \
8169  __ret = (int8x8_t) __builtin_neon_vld1_dup_v(__p0, 0); \
8170  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8171  __ret; \
8172})
8173#endif
8174
8175#ifdef __LITTLE_ENDIAN__
8176#define vld1_dup_f32(__p0) __extension__ ({ \
8177  float32x2_t __ret; \
8178  __ret = (float32x2_t) __builtin_neon_vld1_dup_v(__p0, 9); \
8179  __ret; \
8180})
8181#else
8182#define vld1_dup_f32(__p0) __extension__ ({ \
8183  float32x2_t __ret; \
8184  __ret = (float32x2_t) __builtin_neon_vld1_dup_v(__p0, 9); \
8185  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8186  __ret; \
8187})
8188#endif
8189
8190#ifdef __LITTLE_ENDIAN__
8191#define vld1_dup_f16(__p0) __extension__ ({ \
8192  float16x4_t __ret; \
8193  __ret = (float16x4_t) __builtin_neon_vld1_dup_v(__p0, 8); \
8194  __ret; \
8195})
8196#else
8197#define vld1_dup_f16(__p0) __extension__ ({ \
8198  float16x4_t __ret; \
8199  __ret = (float16x4_t) __builtin_neon_vld1_dup_v(__p0, 8); \
8200  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8201  __ret; \
8202})
8203#endif
8204
8205#ifdef __LITTLE_ENDIAN__
8206#define vld1_dup_s32(__p0) __extension__ ({ \
8207  int32x2_t __ret; \
8208  __ret = (int32x2_t) __builtin_neon_vld1_dup_v(__p0, 2); \
8209  __ret; \
8210})
8211#else
8212#define vld1_dup_s32(__p0) __extension__ ({ \
8213  int32x2_t __ret; \
8214  __ret = (int32x2_t) __builtin_neon_vld1_dup_v(__p0, 2); \
8215  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8216  __ret; \
8217})
8218#endif
8219
8220#ifdef __LITTLE_ENDIAN__
8221#define vld1_dup_s64(__p0) __extension__ ({ \
8222  int64x1_t __ret; \
8223  __ret = (int64x1_t) __builtin_neon_vld1_dup_v(__p0, 3); \
8224  __ret; \
8225})
8226#else
8227#define vld1_dup_s64(__p0) __extension__ ({ \
8228  int64x1_t __ret; \
8229  __ret = (int64x1_t) __builtin_neon_vld1_dup_v(__p0, 3); \
8230  __ret; \
8231})
8232#endif
8233
8234#ifdef __LITTLE_ENDIAN__
8235#define vld1_dup_s16(__p0) __extension__ ({ \
8236  int16x4_t __ret; \
8237  __ret = (int16x4_t) __builtin_neon_vld1_dup_v(__p0, 1); \
8238  __ret; \
8239})
8240#else
8241#define vld1_dup_s16(__p0) __extension__ ({ \
8242  int16x4_t __ret; \
8243  __ret = (int16x4_t) __builtin_neon_vld1_dup_v(__p0, 1); \
8244  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8245  __ret; \
8246})
8247#endif
8248
8249#ifdef __LITTLE_ENDIAN__
8250#define vld1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8251  poly8x8_t __s1 = __p1; \
8252  poly8x8_t __ret; \
8253  __ret = (poly8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 4); \
8254  __ret; \
8255})
8256#else
8257#define vld1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8258  poly8x8_t __s1 = __p1; \
8259  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8260  poly8x8_t __ret; \
8261  __ret = (poly8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 4); \
8262  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8263  __ret; \
8264})
8265#endif
8266
8267#ifdef __LITTLE_ENDIAN__
8268#define vld1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8269  poly16x4_t __s1 = __p1; \
8270  poly16x4_t __ret; \
8271  __ret = (poly16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 5); \
8272  __ret; \
8273})
8274#else
8275#define vld1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8276  poly16x4_t __s1 = __p1; \
8277  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8278  poly16x4_t __ret; \
8279  __ret = (poly16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 5); \
8280  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8281  __ret; \
8282})
8283#endif
8284
8285#ifdef __LITTLE_ENDIAN__
8286#define vld1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8287  poly8x16_t __s1 = __p1; \
8288  poly8x16_t __ret; \
8289  __ret = (poly8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 36); \
8290  __ret; \
8291})
8292#else
8293#define vld1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
8294  poly8x16_t __s1 = __p1; \
8295  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8296  poly8x16_t __ret; \
8297  __ret = (poly8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 36); \
8298  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8299  __ret; \
8300})
8301#endif
8302
8303#ifdef __LITTLE_ENDIAN__
8304#define vld1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8305  poly16x8_t __s1 = __p1; \
8306  poly16x8_t __ret; \
8307  __ret = (poly16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 37); \
8308  __ret; \
8309})
8310#else
8311#define vld1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
8312  poly16x8_t __s1 = __p1; \
8313  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8314  poly16x8_t __ret; \
8315  __ret = (poly16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 37); \
8316  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8317  __ret; \
8318})
8319#endif
8320
8321#ifdef __LITTLE_ENDIAN__
8322#define vld1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8323  uint8x16_t __s1 = __p1; \
8324  uint8x16_t __ret; \
8325  __ret = (uint8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 48); \
8326  __ret; \
8327})
8328#else
8329#define vld1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8330  uint8x16_t __s1 = __p1; \
8331  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8332  uint8x16_t __ret; \
8333  __ret = (uint8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 48); \
8334  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8335  __ret; \
8336})
8337#endif
8338
8339#ifdef __LITTLE_ENDIAN__
8340#define vld1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8341  uint32x4_t __s1 = __p1; \
8342  uint32x4_t __ret; \
8343  __ret = (uint32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 50); \
8344  __ret; \
8345})
8346#else
8347#define vld1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8348  uint32x4_t __s1 = __p1; \
8349  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8350  uint32x4_t __ret; \
8351  __ret = (uint32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 50); \
8352  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8353  __ret; \
8354})
8355#endif
8356
8357#ifdef __LITTLE_ENDIAN__
8358#define vld1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8359  uint64x2_t __s1 = __p1; \
8360  uint64x2_t __ret; \
8361  __ret = (uint64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 51); \
8362  __ret; \
8363})
8364#else
8365#define vld1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8366  uint64x2_t __s1 = __p1; \
8367  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8368  uint64x2_t __ret; \
8369  __ret = (uint64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 51); \
8370  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8371  __ret; \
8372})
8373#endif
8374
8375#ifdef __LITTLE_ENDIAN__
8376#define vld1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8377  uint16x8_t __s1 = __p1; \
8378  uint16x8_t __ret; \
8379  __ret = (uint16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 49); \
8380  __ret; \
8381})
8382#else
8383#define vld1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8384  uint16x8_t __s1 = __p1; \
8385  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8386  uint16x8_t __ret; \
8387  __ret = (uint16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 49); \
8388  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8389  __ret; \
8390})
8391#endif
8392
8393#ifdef __LITTLE_ENDIAN__
8394#define vld1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8395  int8x16_t __s1 = __p1; \
8396  int8x16_t __ret; \
8397  __ret = (int8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 32); \
8398  __ret; \
8399})
8400#else
8401#define vld1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8402  int8x16_t __s1 = __p1; \
8403  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8404  int8x16_t __ret; \
8405  __ret = (int8x16_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 32); \
8406  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8407  __ret; \
8408})
8409#endif
8410
8411#ifdef __LITTLE_ENDIAN__
8412#define vld1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8413  float32x4_t __s1 = __p1; \
8414  float32x4_t __ret; \
8415  __ret = (float32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 41); \
8416  __ret; \
8417})
8418#else
8419#define vld1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8420  float32x4_t __s1 = __p1; \
8421  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8422  float32x4_t __ret; \
8423  __ret = (float32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 41); \
8424  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8425  __ret; \
8426})
8427#endif
8428
8429#ifdef __LITTLE_ENDIAN__
8430#define vld1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8431  float16x8_t __s1 = __p1; \
8432  float16x8_t __ret; \
8433  __ret = (float16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 40); \
8434  __ret; \
8435})
8436#else
8437#define vld1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8438  float16x8_t __s1 = __p1; \
8439  float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8440  float16x8_t __ret; \
8441  __ret = (float16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 40); \
8442  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8443  __ret; \
8444})
8445#endif
8446
8447#ifdef __LITTLE_ENDIAN__
8448#define vld1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8449  int32x4_t __s1 = __p1; \
8450  int32x4_t __ret; \
8451  __ret = (int32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 34); \
8452  __ret; \
8453})
8454#else
8455#define vld1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8456  int32x4_t __s1 = __p1; \
8457  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8458  int32x4_t __ret; \
8459  __ret = (int32x4_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 34); \
8460  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8461  __ret; \
8462})
8463#endif
8464
8465#ifdef __LITTLE_ENDIAN__
8466#define vld1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8467  int64x2_t __s1 = __p1; \
8468  int64x2_t __ret; \
8469  __ret = (int64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 35); \
8470  __ret; \
8471})
8472#else
8473#define vld1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8474  int64x2_t __s1 = __p1; \
8475  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8476  int64x2_t __ret; \
8477  __ret = (int64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 35); \
8478  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8479  __ret; \
8480})
8481#endif
8482
8483#ifdef __LITTLE_ENDIAN__
8484#define vld1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8485  int16x8_t __s1 = __p1; \
8486  int16x8_t __ret; \
8487  __ret = (int16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 33); \
8488  __ret; \
8489})
8490#else
8491#define vld1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8492  int16x8_t __s1 = __p1; \
8493  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8494  int16x8_t __ret; \
8495  __ret = (int16x8_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 33); \
8496  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8497  __ret; \
8498})
8499#endif
8500
8501#ifdef __LITTLE_ENDIAN__
8502#define vld1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8503  uint8x8_t __s1 = __p1; \
8504  uint8x8_t __ret; \
8505  __ret = (uint8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 16); \
8506  __ret; \
8507})
8508#else
8509#define vld1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
8510  uint8x8_t __s1 = __p1; \
8511  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8512  uint8x8_t __ret; \
8513  __ret = (uint8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 16); \
8514  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8515  __ret; \
8516})
8517#endif
8518
8519#ifdef __LITTLE_ENDIAN__
8520#define vld1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8521  uint32x2_t __s1 = __p1; \
8522  uint32x2_t __ret; \
8523  __ret = (uint32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 18); \
8524  __ret; \
8525})
8526#else
8527#define vld1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
8528  uint32x2_t __s1 = __p1; \
8529  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8530  uint32x2_t __ret; \
8531  __ret = (uint32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 18); \
8532  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8533  __ret; \
8534})
8535#endif
8536
8537#ifdef __LITTLE_ENDIAN__
8538#define vld1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8539  uint64x1_t __s1 = __p1; \
8540  uint64x1_t __ret; \
8541  __ret = (uint64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
8542  __ret; \
8543})
8544#else
8545#define vld1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
8546  uint64x1_t __s1 = __p1; \
8547  uint64x1_t __ret; \
8548  __ret = (uint64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
8549  __ret; \
8550})
8551#endif
8552
8553#ifdef __LITTLE_ENDIAN__
8554#define vld1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8555  uint16x4_t __s1 = __p1; \
8556  uint16x4_t __ret; \
8557  __ret = (uint16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 17); \
8558  __ret; \
8559})
8560#else
8561#define vld1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
8562  uint16x4_t __s1 = __p1; \
8563  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8564  uint16x4_t __ret; \
8565  __ret = (uint16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 17); \
8566  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8567  __ret; \
8568})
8569#endif
8570
8571#ifdef __LITTLE_ENDIAN__
8572#define vld1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8573  int8x8_t __s1 = __p1; \
8574  int8x8_t __ret; \
8575  __ret = (int8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 0); \
8576  __ret; \
8577})
8578#else
8579#define vld1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
8580  int8x8_t __s1 = __p1; \
8581  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
8582  int8x8_t __ret; \
8583  __ret = (int8x8_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 0); \
8584  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
8585  __ret; \
8586})
8587#endif
8588
8589#ifdef __LITTLE_ENDIAN__
8590#define vld1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8591  float32x2_t __s1 = __p1; \
8592  float32x2_t __ret; \
8593  __ret = (float32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 9); \
8594  __ret; \
8595})
8596#else
8597#define vld1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
8598  float32x2_t __s1 = __p1; \
8599  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8600  float32x2_t __ret; \
8601  __ret = (float32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 9); \
8602  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8603  __ret; \
8604})
8605#endif
8606
8607#ifdef __LITTLE_ENDIAN__
8608#define vld1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8609  float16x4_t __s1 = __p1; \
8610  float16x4_t __ret; \
8611  __ret = (float16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 8); \
8612  __ret; \
8613})
8614#else
8615#define vld1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
8616  float16x4_t __s1 = __p1; \
8617  float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8618  float16x4_t __ret; \
8619  __ret = (float16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 8); \
8620  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8621  __ret; \
8622})
8623#endif
8624
8625#ifdef __LITTLE_ENDIAN__
8626#define vld1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8627  int32x2_t __s1 = __p1; \
8628  int32x2_t __ret; \
8629  __ret = (int32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 2); \
8630  __ret; \
8631})
8632#else
8633#define vld1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
8634  int32x2_t __s1 = __p1; \
8635  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
8636  int32x2_t __ret; \
8637  __ret = (int32x2_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 2); \
8638  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
8639  __ret; \
8640})
8641#endif
8642
8643#ifdef __LITTLE_ENDIAN__
8644#define vld1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8645  int64x1_t __s1 = __p1; \
8646  int64x1_t __ret; \
8647  __ret = (int64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
8648  __ret; \
8649})
8650#else
8651#define vld1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
8652  int64x1_t __s1 = __p1; \
8653  int64x1_t __ret; \
8654  __ret = (int64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
8655  __ret; \
8656})
8657#endif
8658
8659#ifdef __LITTLE_ENDIAN__
8660#define vld1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8661  int16x4_t __s1 = __p1; \
8662  int16x4_t __ret; \
8663  __ret = (int16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 1); \
8664  __ret; \
8665})
8666#else
8667#define vld1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
8668  int16x4_t __s1 = __p1; \
8669  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
8670  int16x4_t __ret; \
8671  __ret = (int16x4_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__rev1, __p2, 1); \
8672  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
8673  __ret; \
8674})
8675#endif
8676
8677#ifdef __LITTLE_ENDIAN__
8678#define vld2_p8(__p0) __extension__ ({ \
8679  poly8x8x2_t __ret; \
8680  __builtin_neon_vld2_v(&__ret, __p0, 4); \
8681  __ret; \
8682})
8683#else
8684#define vld2_p8(__p0) __extension__ ({ \
8685  poly8x8x2_t __ret; \
8686  __builtin_neon_vld2_v(&__ret, __p0, 4); \
8687 \
8688  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8689  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8690  __ret; \
8691})
8692#endif
8693
8694#ifdef __LITTLE_ENDIAN__
8695#define vld2_p16(__p0) __extension__ ({ \
8696  poly16x4x2_t __ret; \
8697  __builtin_neon_vld2_v(&__ret, __p0, 5); \
8698  __ret; \
8699})
8700#else
8701#define vld2_p16(__p0) __extension__ ({ \
8702  poly16x4x2_t __ret; \
8703  __builtin_neon_vld2_v(&__ret, __p0, 5); \
8704 \
8705  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8706  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8707  __ret; \
8708})
8709#endif
8710
8711#ifdef __LITTLE_ENDIAN__
8712#define vld2q_p8(__p0) __extension__ ({ \
8713  poly8x16x2_t __ret; \
8714  __builtin_neon_vld2q_v(&__ret, __p0, 36); \
8715  __ret; \
8716})
8717#else
8718#define vld2q_p8(__p0) __extension__ ({ \
8719  poly8x16x2_t __ret; \
8720  __builtin_neon_vld2q_v(&__ret, __p0, 36); \
8721 \
8722  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8723  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8724  __ret; \
8725})
8726#endif
8727
8728#ifdef __LITTLE_ENDIAN__
8729#define vld2q_p16(__p0) __extension__ ({ \
8730  poly16x8x2_t __ret; \
8731  __builtin_neon_vld2q_v(&__ret, __p0, 37); \
8732  __ret; \
8733})
8734#else
8735#define vld2q_p16(__p0) __extension__ ({ \
8736  poly16x8x2_t __ret; \
8737  __builtin_neon_vld2q_v(&__ret, __p0, 37); \
8738 \
8739  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8740  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8741  __ret; \
8742})
8743#endif
8744
8745#ifdef __LITTLE_ENDIAN__
8746#define vld2q_u8(__p0) __extension__ ({ \
8747  uint8x16x2_t __ret; \
8748  __builtin_neon_vld2q_v(&__ret, __p0, 48); \
8749  __ret; \
8750})
8751#else
8752#define vld2q_u8(__p0) __extension__ ({ \
8753  uint8x16x2_t __ret; \
8754  __builtin_neon_vld2q_v(&__ret, __p0, 48); \
8755 \
8756  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8757  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8758  __ret; \
8759})
8760#endif
8761
8762#ifdef __LITTLE_ENDIAN__
8763#define vld2q_u32(__p0) __extension__ ({ \
8764  uint32x4x2_t __ret; \
8765  __builtin_neon_vld2q_v(&__ret, __p0, 50); \
8766  __ret; \
8767})
8768#else
8769#define vld2q_u32(__p0) __extension__ ({ \
8770  uint32x4x2_t __ret; \
8771  __builtin_neon_vld2q_v(&__ret, __p0, 50); \
8772 \
8773  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8774  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8775  __ret; \
8776})
8777#endif
8778
8779#ifdef __LITTLE_ENDIAN__
8780#define vld2q_u16(__p0) __extension__ ({ \
8781  uint16x8x2_t __ret; \
8782  __builtin_neon_vld2q_v(&__ret, __p0, 49); \
8783  __ret; \
8784})
8785#else
8786#define vld2q_u16(__p0) __extension__ ({ \
8787  uint16x8x2_t __ret; \
8788  __builtin_neon_vld2q_v(&__ret, __p0, 49); \
8789 \
8790  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8791  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8792  __ret; \
8793})
8794#endif
8795
8796#ifdef __LITTLE_ENDIAN__
8797#define vld2q_s8(__p0) __extension__ ({ \
8798  int8x16x2_t __ret; \
8799  __builtin_neon_vld2q_v(&__ret, __p0, 32); \
8800  __ret; \
8801})
8802#else
8803#define vld2q_s8(__p0) __extension__ ({ \
8804  int8x16x2_t __ret; \
8805  __builtin_neon_vld2q_v(&__ret, __p0, 32); \
8806 \
8807  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8808  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
8809  __ret; \
8810})
8811#endif
8812
8813#ifdef __LITTLE_ENDIAN__
8814#define vld2q_f32(__p0) __extension__ ({ \
8815  float32x4x2_t __ret; \
8816  __builtin_neon_vld2q_v(&__ret, __p0, 41); \
8817  __ret; \
8818})
8819#else
8820#define vld2q_f32(__p0) __extension__ ({ \
8821  float32x4x2_t __ret; \
8822  __builtin_neon_vld2q_v(&__ret, __p0, 41); \
8823 \
8824  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8825  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8826  __ret; \
8827})
8828#endif
8829
8830#ifdef __LITTLE_ENDIAN__
8831#define vld2q_f16(__p0) __extension__ ({ \
8832  float16x8x2_t __ret; \
8833  __builtin_neon_vld2q_v(&__ret, __p0, 40); \
8834  __ret; \
8835})
8836#else
8837#define vld2q_f16(__p0) __extension__ ({ \
8838  float16x8x2_t __ret; \
8839  __builtin_neon_vld2q_v(&__ret, __p0, 40); \
8840 \
8841  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8842  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8843  __ret; \
8844})
8845#endif
8846
8847#ifdef __LITTLE_ENDIAN__
8848#define vld2q_s32(__p0) __extension__ ({ \
8849  int32x4x2_t __ret; \
8850  __builtin_neon_vld2q_v(&__ret, __p0, 34); \
8851  __ret; \
8852})
8853#else
8854#define vld2q_s32(__p0) __extension__ ({ \
8855  int32x4x2_t __ret; \
8856  __builtin_neon_vld2q_v(&__ret, __p0, 34); \
8857 \
8858  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8859  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8860  __ret; \
8861})
8862#endif
8863
8864#ifdef __LITTLE_ENDIAN__
8865#define vld2q_s16(__p0) __extension__ ({ \
8866  int16x8x2_t __ret; \
8867  __builtin_neon_vld2q_v(&__ret, __p0, 33); \
8868  __ret; \
8869})
8870#else
8871#define vld2q_s16(__p0) __extension__ ({ \
8872  int16x8x2_t __ret; \
8873  __builtin_neon_vld2q_v(&__ret, __p0, 33); \
8874 \
8875  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8876  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8877  __ret; \
8878})
8879#endif
8880
8881#ifdef __LITTLE_ENDIAN__
8882#define vld2_u8(__p0) __extension__ ({ \
8883  uint8x8x2_t __ret; \
8884  __builtin_neon_vld2_v(&__ret, __p0, 16); \
8885  __ret; \
8886})
8887#else
8888#define vld2_u8(__p0) __extension__ ({ \
8889  uint8x8x2_t __ret; \
8890  __builtin_neon_vld2_v(&__ret, __p0, 16); \
8891 \
8892  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8893  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8894  __ret; \
8895})
8896#endif
8897
8898#ifdef __LITTLE_ENDIAN__
8899#define vld2_u32(__p0) __extension__ ({ \
8900  uint32x2x2_t __ret; \
8901  __builtin_neon_vld2_v(&__ret, __p0, 18); \
8902  __ret; \
8903})
8904#else
8905#define vld2_u32(__p0) __extension__ ({ \
8906  uint32x2x2_t __ret; \
8907  __builtin_neon_vld2_v(&__ret, __p0, 18); \
8908 \
8909  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
8910  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
8911  __ret; \
8912})
8913#endif
8914
8915#ifdef __LITTLE_ENDIAN__
8916#define vld2_u64(__p0) __extension__ ({ \
8917  uint64x1x2_t __ret; \
8918  __builtin_neon_vld2_v(&__ret, __p0, 19); \
8919  __ret; \
8920})
8921#else
8922#define vld2_u64(__p0) __extension__ ({ \
8923  uint64x1x2_t __ret; \
8924  __builtin_neon_vld2_v(&__ret, __p0, 19); \
8925  __ret; \
8926})
8927#endif
8928
8929#ifdef __LITTLE_ENDIAN__
8930#define vld2_u16(__p0) __extension__ ({ \
8931  uint16x4x2_t __ret; \
8932  __builtin_neon_vld2_v(&__ret, __p0, 17); \
8933  __ret; \
8934})
8935#else
8936#define vld2_u16(__p0) __extension__ ({ \
8937  uint16x4x2_t __ret; \
8938  __builtin_neon_vld2_v(&__ret, __p0, 17); \
8939 \
8940  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8941  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8942  __ret; \
8943})
8944#endif
8945
8946#ifdef __LITTLE_ENDIAN__
8947#define vld2_s8(__p0) __extension__ ({ \
8948  int8x8x2_t __ret; \
8949  __builtin_neon_vld2_v(&__ret, __p0, 0); \
8950  __ret; \
8951})
8952#else
8953#define vld2_s8(__p0) __extension__ ({ \
8954  int8x8x2_t __ret; \
8955  __builtin_neon_vld2_v(&__ret, __p0, 0); \
8956 \
8957  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
8958  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
8959  __ret; \
8960})
8961#endif
8962
8963#ifdef __LITTLE_ENDIAN__
8964#define vld2_f32(__p0) __extension__ ({ \
8965  float32x2x2_t __ret; \
8966  __builtin_neon_vld2_v(&__ret, __p0, 9); \
8967  __ret; \
8968})
8969#else
8970#define vld2_f32(__p0) __extension__ ({ \
8971  float32x2x2_t __ret; \
8972  __builtin_neon_vld2_v(&__ret, __p0, 9); \
8973 \
8974  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
8975  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
8976  __ret; \
8977})
8978#endif
8979
8980#ifdef __LITTLE_ENDIAN__
8981#define vld2_f16(__p0) __extension__ ({ \
8982  float16x4x2_t __ret; \
8983  __builtin_neon_vld2_v(&__ret, __p0, 8); \
8984  __ret; \
8985})
8986#else
8987#define vld2_f16(__p0) __extension__ ({ \
8988  float16x4x2_t __ret; \
8989  __builtin_neon_vld2_v(&__ret, __p0, 8); \
8990 \
8991  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
8992  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
8993  __ret; \
8994})
8995#endif
8996
8997#ifdef __LITTLE_ENDIAN__
8998#define vld2_s32(__p0) __extension__ ({ \
8999  int32x2x2_t __ret; \
9000  __builtin_neon_vld2_v(&__ret, __p0, 2); \
9001  __ret; \
9002})
9003#else
9004#define vld2_s32(__p0) __extension__ ({ \
9005  int32x2x2_t __ret; \
9006  __builtin_neon_vld2_v(&__ret, __p0, 2); \
9007 \
9008  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9009  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9010  __ret; \
9011})
9012#endif
9013
9014#ifdef __LITTLE_ENDIAN__
9015#define vld2_s64(__p0) __extension__ ({ \
9016  int64x1x2_t __ret; \
9017  __builtin_neon_vld2_v(&__ret, __p0, 3); \
9018  __ret; \
9019})
9020#else
9021#define vld2_s64(__p0) __extension__ ({ \
9022  int64x1x2_t __ret; \
9023  __builtin_neon_vld2_v(&__ret, __p0, 3); \
9024  __ret; \
9025})
9026#endif
9027
9028#ifdef __LITTLE_ENDIAN__
9029#define vld2_s16(__p0) __extension__ ({ \
9030  int16x4x2_t __ret; \
9031  __builtin_neon_vld2_v(&__ret, __p0, 1); \
9032  __ret; \
9033})
9034#else
9035#define vld2_s16(__p0) __extension__ ({ \
9036  int16x4x2_t __ret; \
9037  __builtin_neon_vld2_v(&__ret, __p0, 1); \
9038 \
9039  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9040  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9041  __ret; \
9042})
9043#endif
9044
9045#ifdef __LITTLE_ENDIAN__
9046#define vld2_dup_p8(__p0) __extension__ ({ \
9047  poly8x8x2_t __ret; \
9048  __builtin_neon_vld2_dup_v(&__ret, __p0, 4); \
9049  __ret; \
9050})
9051#else
9052#define vld2_dup_p8(__p0) __extension__ ({ \
9053  poly8x8x2_t __ret; \
9054  __builtin_neon_vld2_dup_v(&__ret, __p0, 4); \
9055 \
9056  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9057  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9058  __ret; \
9059})
9060#endif
9061
9062#ifdef __LITTLE_ENDIAN__
9063#define vld2_dup_p16(__p0) __extension__ ({ \
9064  poly16x4x2_t __ret; \
9065  __builtin_neon_vld2_dup_v(&__ret, __p0, 5); \
9066  __ret; \
9067})
9068#else
9069#define vld2_dup_p16(__p0) __extension__ ({ \
9070  poly16x4x2_t __ret; \
9071  __builtin_neon_vld2_dup_v(&__ret, __p0, 5); \
9072 \
9073  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9074  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9075  __ret; \
9076})
9077#endif
9078
9079#ifdef __LITTLE_ENDIAN__
9080#define vld2_dup_u8(__p0) __extension__ ({ \
9081  uint8x8x2_t __ret; \
9082  __builtin_neon_vld2_dup_v(&__ret, __p0, 16); \
9083  __ret; \
9084})
9085#else
9086#define vld2_dup_u8(__p0) __extension__ ({ \
9087  uint8x8x2_t __ret; \
9088  __builtin_neon_vld2_dup_v(&__ret, __p0, 16); \
9089 \
9090  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9091  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9092  __ret; \
9093})
9094#endif
9095
9096#ifdef __LITTLE_ENDIAN__
9097#define vld2_dup_u32(__p0) __extension__ ({ \
9098  uint32x2x2_t __ret; \
9099  __builtin_neon_vld2_dup_v(&__ret, __p0, 18); \
9100  __ret; \
9101})
9102#else
9103#define vld2_dup_u32(__p0) __extension__ ({ \
9104  uint32x2x2_t __ret; \
9105  __builtin_neon_vld2_dup_v(&__ret, __p0, 18); \
9106 \
9107  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9108  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9109  __ret; \
9110})
9111#endif
9112
9113#ifdef __LITTLE_ENDIAN__
9114#define vld2_dup_u64(__p0) __extension__ ({ \
9115  uint64x1x2_t __ret; \
9116  __builtin_neon_vld2_dup_v(&__ret, __p0, 19); \
9117  __ret; \
9118})
9119#else
9120#define vld2_dup_u64(__p0) __extension__ ({ \
9121  uint64x1x2_t __ret; \
9122  __builtin_neon_vld2_dup_v(&__ret, __p0, 19); \
9123  __ret; \
9124})
9125#endif
9126
9127#ifdef __LITTLE_ENDIAN__
9128#define vld2_dup_u16(__p0) __extension__ ({ \
9129  uint16x4x2_t __ret; \
9130  __builtin_neon_vld2_dup_v(&__ret, __p0, 17); \
9131  __ret; \
9132})
9133#else
9134#define vld2_dup_u16(__p0) __extension__ ({ \
9135  uint16x4x2_t __ret; \
9136  __builtin_neon_vld2_dup_v(&__ret, __p0, 17); \
9137 \
9138  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9139  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9140  __ret; \
9141})
9142#endif
9143
9144#ifdef __LITTLE_ENDIAN__
9145#define vld2_dup_s8(__p0) __extension__ ({ \
9146  int8x8x2_t __ret; \
9147  __builtin_neon_vld2_dup_v(&__ret, __p0, 0); \
9148  __ret; \
9149})
9150#else
9151#define vld2_dup_s8(__p0) __extension__ ({ \
9152  int8x8x2_t __ret; \
9153  __builtin_neon_vld2_dup_v(&__ret, __p0, 0); \
9154 \
9155  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9156  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9157  __ret; \
9158})
9159#endif
9160
9161#ifdef __LITTLE_ENDIAN__
9162#define vld2_dup_f32(__p0) __extension__ ({ \
9163  float32x2x2_t __ret; \
9164  __builtin_neon_vld2_dup_v(&__ret, __p0, 9); \
9165  __ret; \
9166})
9167#else
9168#define vld2_dup_f32(__p0) __extension__ ({ \
9169  float32x2x2_t __ret; \
9170  __builtin_neon_vld2_dup_v(&__ret, __p0, 9); \
9171 \
9172  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9173  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9174  __ret; \
9175})
9176#endif
9177
9178#ifdef __LITTLE_ENDIAN__
9179#define vld2_dup_f16(__p0) __extension__ ({ \
9180  float16x4x2_t __ret; \
9181  __builtin_neon_vld2_dup_v(&__ret, __p0, 8); \
9182  __ret; \
9183})
9184#else
9185#define vld2_dup_f16(__p0) __extension__ ({ \
9186  float16x4x2_t __ret; \
9187  __builtin_neon_vld2_dup_v(&__ret, __p0, 8); \
9188 \
9189  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9190  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9191  __ret; \
9192})
9193#endif
9194
9195#ifdef __LITTLE_ENDIAN__
9196#define vld2_dup_s32(__p0) __extension__ ({ \
9197  int32x2x2_t __ret; \
9198  __builtin_neon_vld2_dup_v(&__ret, __p0, 2); \
9199  __ret; \
9200})
9201#else
9202#define vld2_dup_s32(__p0) __extension__ ({ \
9203  int32x2x2_t __ret; \
9204  __builtin_neon_vld2_dup_v(&__ret, __p0, 2); \
9205 \
9206  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9207  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9208  __ret; \
9209})
9210#endif
9211
9212#ifdef __LITTLE_ENDIAN__
9213#define vld2_dup_s64(__p0) __extension__ ({ \
9214  int64x1x2_t __ret; \
9215  __builtin_neon_vld2_dup_v(&__ret, __p0, 3); \
9216  __ret; \
9217})
9218#else
9219#define vld2_dup_s64(__p0) __extension__ ({ \
9220  int64x1x2_t __ret; \
9221  __builtin_neon_vld2_dup_v(&__ret, __p0, 3); \
9222  __ret; \
9223})
9224#endif
9225
9226#ifdef __LITTLE_ENDIAN__
9227#define vld2_dup_s16(__p0) __extension__ ({ \
9228  int16x4x2_t __ret; \
9229  __builtin_neon_vld2_dup_v(&__ret, __p0, 1); \
9230  __ret; \
9231})
9232#else
9233#define vld2_dup_s16(__p0) __extension__ ({ \
9234  int16x4x2_t __ret; \
9235  __builtin_neon_vld2_dup_v(&__ret, __p0, 1); \
9236 \
9237  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9238  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9239  __ret; \
9240})
9241#endif
9242
9243#ifdef __LITTLE_ENDIAN__
9244#define vld2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
9245  poly8x8x2_t __s1 = __p1; \
9246  poly8x8x2_t __ret; \
9247  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 4); \
9248  __ret; \
9249})
9250#else
9251#define vld2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
9252  poly8x8x2_t __s1 = __p1; \
9253  poly8x8x2_t __rev1; \
9254  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9255  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9256  poly8x8x2_t __ret; \
9257  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 4); \
9258 \
9259  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9260  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9261  __ret; \
9262})
9263#endif
9264
9265#ifdef __LITTLE_ENDIAN__
9266#define vld2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9267  poly16x4x2_t __s1 = __p1; \
9268  poly16x4x2_t __ret; \
9269  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 5); \
9270  __ret; \
9271})
9272#else
9273#define vld2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9274  poly16x4x2_t __s1 = __p1; \
9275  poly16x4x2_t __rev1; \
9276  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9277  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9278  poly16x4x2_t __ret; \
9279  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 5); \
9280 \
9281  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9282  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9283  __ret; \
9284})
9285#endif
9286
9287#ifdef __LITTLE_ENDIAN__
9288#define vld2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9289  poly16x8x2_t __s1 = __p1; \
9290  poly16x8x2_t __ret; \
9291  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 37); \
9292  __ret; \
9293})
9294#else
9295#define vld2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
9296  poly16x8x2_t __s1 = __p1; \
9297  poly16x8x2_t __rev1; \
9298  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9299  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9300  poly16x8x2_t __ret; \
9301  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 37); \
9302 \
9303  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9304  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9305  __ret; \
9306})
9307#endif
9308
9309#ifdef __LITTLE_ENDIAN__
9310#define vld2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9311  uint32x4x2_t __s1 = __p1; \
9312  uint32x4x2_t __ret; \
9313  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 50); \
9314  __ret; \
9315})
9316#else
9317#define vld2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9318  uint32x4x2_t __s1 = __p1; \
9319  uint32x4x2_t __rev1; \
9320  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9321  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9322  uint32x4x2_t __ret; \
9323  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 50); \
9324 \
9325  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9326  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9327  __ret; \
9328})
9329#endif
9330
9331#ifdef __LITTLE_ENDIAN__
9332#define vld2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9333  uint16x8x2_t __s1 = __p1; \
9334  uint16x8x2_t __ret; \
9335  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 49); \
9336  __ret; \
9337})
9338#else
9339#define vld2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9340  uint16x8x2_t __s1 = __p1; \
9341  uint16x8x2_t __rev1; \
9342  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9343  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9344  uint16x8x2_t __ret; \
9345  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 49); \
9346 \
9347  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9348  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9349  __ret; \
9350})
9351#endif
9352
9353#ifdef __LITTLE_ENDIAN__
9354#define vld2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9355  float32x4x2_t __s1 = __p1; \
9356  float32x4x2_t __ret; \
9357  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 41); \
9358  __ret; \
9359})
9360#else
9361#define vld2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9362  float32x4x2_t __s1 = __p1; \
9363  float32x4x2_t __rev1; \
9364  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9365  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9366  float32x4x2_t __ret; \
9367  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 41); \
9368 \
9369  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9370  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9371  __ret; \
9372})
9373#endif
9374
9375#ifdef __LITTLE_ENDIAN__
9376#define vld2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9377  float16x8x2_t __s1 = __p1; \
9378  float16x8x2_t __ret; \
9379  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 40); \
9380  __ret; \
9381})
9382#else
9383#define vld2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9384  float16x8x2_t __s1 = __p1; \
9385  float16x8x2_t __rev1; \
9386  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9387  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9388  float16x8x2_t __ret; \
9389  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 40); \
9390 \
9391  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9392  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9393  __ret; \
9394})
9395#endif
9396
9397#ifdef __LITTLE_ENDIAN__
9398#define vld2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9399  int32x4x2_t __s1 = __p1; \
9400  int32x4x2_t __ret; \
9401  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 34); \
9402  __ret; \
9403})
9404#else
9405#define vld2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9406  int32x4x2_t __s1 = __p1; \
9407  int32x4x2_t __rev1; \
9408  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9409  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9410  int32x4x2_t __ret; \
9411  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 34); \
9412 \
9413  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9414  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9415  __ret; \
9416})
9417#endif
9418
9419#ifdef __LITTLE_ENDIAN__
9420#define vld2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9421  int16x8x2_t __s1 = __p1; \
9422  int16x8x2_t __ret; \
9423  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 33); \
9424  __ret; \
9425})
9426#else
9427#define vld2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9428  int16x8x2_t __s1 = __p1; \
9429  int16x8x2_t __rev1; \
9430  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9431  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9432  int16x8x2_t __ret; \
9433  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 33); \
9434 \
9435  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9436  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9437  __ret; \
9438})
9439#endif
9440
9441#ifdef __LITTLE_ENDIAN__
9442#define vld2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9443  uint8x8x2_t __s1 = __p1; \
9444  uint8x8x2_t __ret; \
9445  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 16); \
9446  __ret; \
9447})
9448#else
9449#define vld2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
9450  uint8x8x2_t __s1 = __p1; \
9451  uint8x8x2_t __rev1; \
9452  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9453  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9454  uint8x8x2_t __ret; \
9455  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 16); \
9456 \
9457  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9458  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9459  __ret; \
9460})
9461#endif
9462
9463#ifdef __LITTLE_ENDIAN__
9464#define vld2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9465  uint32x2x2_t __s1 = __p1; \
9466  uint32x2x2_t __ret; \
9467  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 18); \
9468  __ret; \
9469})
9470#else
9471#define vld2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
9472  uint32x2x2_t __s1 = __p1; \
9473  uint32x2x2_t __rev1; \
9474  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
9475  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
9476  uint32x2x2_t __ret; \
9477  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 18); \
9478 \
9479  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9480  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9481  __ret; \
9482})
9483#endif
9484
9485#ifdef __LITTLE_ENDIAN__
9486#define vld2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9487  uint16x4x2_t __s1 = __p1; \
9488  uint16x4x2_t __ret; \
9489  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 17); \
9490  __ret; \
9491})
9492#else
9493#define vld2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
9494  uint16x4x2_t __s1 = __p1; \
9495  uint16x4x2_t __rev1; \
9496  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9497  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9498  uint16x4x2_t __ret; \
9499  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 17); \
9500 \
9501  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9502  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9503  __ret; \
9504})
9505#endif
9506
9507#ifdef __LITTLE_ENDIAN__
9508#define vld2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9509  int8x8x2_t __s1 = __p1; \
9510  int8x8x2_t __ret; \
9511  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 0); \
9512  __ret; \
9513})
9514#else
9515#define vld2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
9516  int8x8x2_t __s1 = __p1; \
9517  int8x8x2_t __rev1; \
9518  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9519  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9520  int8x8x2_t __ret; \
9521  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 0); \
9522 \
9523  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9524  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9525  __ret; \
9526})
9527#endif
9528
9529#ifdef __LITTLE_ENDIAN__
9530#define vld2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9531  float32x2x2_t __s1 = __p1; \
9532  float32x2x2_t __ret; \
9533  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 9); \
9534  __ret; \
9535})
9536#else
9537#define vld2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
9538  float32x2x2_t __s1 = __p1; \
9539  float32x2x2_t __rev1; \
9540  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
9541  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
9542  float32x2x2_t __ret; \
9543  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 9); \
9544 \
9545  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9546  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9547  __ret; \
9548})
9549#endif
9550
9551#ifdef __LITTLE_ENDIAN__
9552#define vld2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9553  float16x4x2_t __s1 = __p1; \
9554  float16x4x2_t __ret; \
9555  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 8); \
9556  __ret; \
9557})
9558#else
9559#define vld2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
9560  float16x4x2_t __s1 = __p1; \
9561  float16x4x2_t __rev1; \
9562  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9563  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9564  float16x4x2_t __ret; \
9565  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 8); \
9566 \
9567  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9568  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9569  __ret; \
9570})
9571#endif
9572
9573#ifdef __LITTLE_ENDIAN__
9574#define vld2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9575  int32x2x2_t __s1 = __p1; \
9576  int32x2x2_t __ret; \
9577  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 2); \
9578  __ret; \
9579})
9580#else
9581#define vld2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
9582  int32x2x2_t __s1 = __p1; \
9583  int32x2x2_t __rev1; \
9584  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
9585  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
9586  int32x2x2_t __ret; \
9587  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 2); \
9588 \
9589  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9590  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9591  __ret; \
9592})
9593#endif
9594
9595#ifdef __LITTLE_ENDIAN__
9596#define vld2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9597  int16x4x2_t __s1 = __p1; \
9598  int16x4x2_t __ret; \
9599  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 1); \
9600  __ret; \
9601})
9602#else
9603#define vld2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
9604  int16x4x2_t __s1 = __p1; \
9605  int16x4x2_t __rev1; \
9606  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
9607  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
9608  int16x4x2_t __ret; \
9609  __builtin_neon_vld2_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 1); \
9610 \
9611  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9612  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9613  __ret; \
9614})
9615#endif
9616
9617#ifdef __LITTLE_ENDIAN__
9618#define vld3_p8(__p0) __extension__ ({ \
9619  poly8x8x3_t __ret; \
9620  __builtin_neon_vld3_v(&__ret, __p0, 4); \
9621  __ret; \
9622})
9623#else
9624#define vld3_p8(__p0) __extension__ ({ \
9625  poly8x8x3_t __ret; \
9626  __builtin_neon_vld3_v(&__ret, __p0, 4); \
9627 \
9628  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9629  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9630  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9631  __ret; \
9632})
9633#endif
9634
9635#ifdef __LITTLE_ENDIAN__
9636#define vld3_p16(__p0) __extension__ ({ \
9637  poly16x4x3_t __ret; \
9638  __builtin_neon_vld3_v(&__ret, __p0, 5); \
9639  __ret; \
9640})
9641#else
9642#define vld3_p16(__p0) __extension__ ({ \
9643  poly16x4x3_t __ret; \
9644  __builtin_neon_vld3_v(&__ret, __p0, 5); \
9645 \
9646  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9647  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9648  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9649  __ret; \
9650})
9651#endif
9652
9653#ifdef __LITTLE_ENDIAN__
9654#define vld3q_p8(__p0) __extension__ ({ \
9655  poly8x16x3_t __ret; \
9656  __builtin_neon_vld3q_v(&__ret, __p0, 36); \
9657  __ret; \
9658})
9659#else
9660#define vld3q_p8(__p0) __extension__ ({ \
9661  poly8x16x3_t __ret; \
9662  __builtin_neon_vld3q_v(&__ret, __p0, 36); \
9663 \
9664  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9665  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9666  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9667  __ret; \
9668})
9669#endif
9670
9671#ifdef __LITTLE_ENDIAN__
9672#define vld3q_p16(__p0) __extension__ ({ \
9673  poly16x8x3_t __ret; \
9674  __builtin_neon_vld3q_v(&__ret, __p0, 37); \
9675  __ret; \
9676})
9677#else
9678#define vld3q_p16(__p0) __extension__ ({ \
9679  poly16x8x3_t __ret; \
9680  __builtin_neon_vld3q_v(&__ret, __p0, 37); \
9681 \
9682  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9683  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9684  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9685  __ret; \
9686})
9687#endif
9688
9689#ifdef __LITTLE_ENDIAN__
9690#define vld3q_u8(__p0) __extension__ ({ \
9691  uint8x16x3_t __ret; \
9692  __builtin_neon_vld3q_v(&__ret, __p0, 48); \
9693  __ret; \
9694})
9695#else
9696#define vld3q_u8(__p0) __extension__ ({ \
9697  uint8x16x3_t __ret; \
9698  __builtin_neon_vld3q_v(&__ret, __p0, 48); \
9699 \
9700  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9701  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9702  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9703  __ret; \
9704})
9705#endif
9706
9707#ifdef __LITTLE_ENDIAN__
9708#define vld3q_u32(__p0) __extension__ ({ \
9709  uint32x4x3_t __ret; \
9710  __builtin_neon_vld3q_v(&__ret, __p0, 50); \
9711  __ret; \
9712})
9713#else
9714#define vld3q_u32(__p0) __extension__ ({ \
9715  uint32x4x3_t __ret; \
9716  __builtin_neon_vld3q_v(&__ret, __p0, 50); \
9717 \
9718  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9719  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9720  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9721  __ret; \
9722})
9723#endif
9724
9725#ifdef __LITTLE_ENDIAN__
9726#define vld3q_u16(__p0) __extension__ ({ \
9727  uint16x8x3_t __ret; \
9728  __builtin_neon_vld3q_v(&__ret, __p0, 49); \
9729  __ret; \
9730})
9731#else
9732#define vld3q_u16(__p0) __extension__ ({ \
9733  uint16x8x3_t __ret; \
9734  __builtin_neon_vld3q_v(&__ret, __p0, 49); \
9735 \
9736  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9737  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9738  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9739  __ret; \
9740})
9741#endif
9742
9743#ifdef __LITTLE_ENDIAN__
9744#define vld3q_s8(__p0) __extension__ ({ \
9745  int8x16x3_t __ret; \
9746  __builtin_neon_vld3q_v(&__ret, __p0, 32); \
9747  __ret; \
9748})
9749#else
9750#define vld3q_s8(__p0) __extension__ ({ \
9751  int8x16x3_t __ret; \
9752  __builtin_neon_vld3q_v(&__ret, __p0, 32); \
9753 \
9754  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9755  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9756  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
9757  __ret; \
9758})
9759#endif
9760
9761#ifdef __LITTLE_ENDIAN__
9762#define vld3q_f32(__p0) __extension__ ({ \
9763  float32x4x3_t __ret; \
9764  __builtin_neon_vld3q_v(&__ret, __p0, 41); \
9765  __ret; \
9766})
9767#else
9768#define vld3q_f32(__p0) __extension__ ({ \
9769  float32x4x3_t __ret; \
9770  __builtin_neon_vld3q_v(&__ret, __p0, 41); \
9771 \
9772  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9773  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9774  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9775  __ret; \
9776})
9777#endif
9778
9779#ifdef __LITTLE_ENDIAN__
9780#define vld3q_f16(__p0) __extension__ ({ \
9781  float16x8x3_t __ret; \
9782  __builtin_neon_vld3q_v(&__ret, __p0, 40); \
9783  __ret; \
9784})
9785#else
9786#define vld3q_f16(__p0) __extension__ ({ \
9787  float16x8x3_t __ret; \
9788  __builtin_neon_vld3q_v(&__ret, __p0, 40); \
9789 \
9790  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9791  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9792  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9793  __ret; \
9794})
9795#endif
9796
9797#ifdef __LITTLE_ENDIAN__
9798#define vld3q_s32(__p0) __extension__ ({ \
9799  int32x4x3_t __ret; \
9800  __builtin_neon_vld3q_v(&__ret, __p0, 34); \
9801  __ret; \
9802})
9803#else
9804#define vld3q_s32(__p0) __extension__ ({ \
9805  int32x4x3_t __ret; \
9806  __builtin_neon_vld3q_v(&__ret, __p0, 34); \
9807 \
9808  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9809  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9810  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9811  __ret; \
9812})
9813#endif
9814
9815#ifdef __LITTLE_ENDIAN__
9816#define vld3q_s16(__p0) __extension__ ({ \
9817  int16x8x3_t __ret; \
9818  __builtin_neon_vld3q_v(&__ret, __p0, 33); \
9819  __ret; \
9820})
9821#else
9822#define vld3q_s16(__p0) __extension__ ({ \
9823  int16x8x3_t __ret; \
9824  __builtin_neon_vld3q_v(&__ret, __p0, 33); \
9825 \
9826  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9827  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9828  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9829  __ret; \
9830})
9831#endif
9832
9833#ifdef __LITTLE_ENDIAN__
9834#define vld3_u8(__p0) __extension__ ({ \
9835  uint8x8x3_t __ret; \
9836  __builtin_neon_vld3_v(&__ret, __p0, 16); \
9837  __ret; \
9838})
9839#else
9840#define vld3_u8(__p0) __extension__ ({ \
9841  uint8x8x3_t __ret; \
9842  __builtin_neon_vld3_v(&__ret, __p0, 16); \
9843 \
9844  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9845  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9846  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9847  __ret; \
9848})
9849#endif
9850
9851#ifdef __LITTLE_ENDIAN__
9852#define vld3_u32(__p0) __extension__ ({ \
9853  uint32x2x3_t __ret; \
9854  __builtin_neon_vld3_v(&__ret, __p0, 18); \
9855  __ret; \
9856})
9857#else
9858#define vld3_u32(__p0) __extension__ ({ \
9859  uint32x2x3_t __ret; \
9860  __builtin_neon_vld3_v(&__ret, __p0, 18); \
9861 \
9862  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9863  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9864  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9865  __ret; \
9866})
9867#endif
9868
9869#ifdef __LITTLE_ENDIAN__
9870#define vld3_u64(__p0) __extension__ ({ \
9871  uint64x1x3_t __ret; \
9872  __builtin_neon_vld3_v(&__ret, __p0, 19); \
9873  __ret; \
9874})
9875#else
9876#define vld3_u64(__p0) __extension__ ({ \
9877  uint64x1x3_t __ret; \
9878  __builtin_neon_vld3_v(&__ret, __p0, 19); \
9879  __ret; \
9880})
9881#endif
9882
9883#ifdef __LITTLE_ENDIAN__
9884#define vld3_u16(__p0) __extension__ ({ \
9885  uint16x4x3_t __ret; \
9886  __builtin_neon_vld3_v(&__ret, __p0, 17); \
9887  __ret; \
9888})
9889#else
9890#define vld3_u16(__p0) __extension__ ({ \
9891  uint16x4x3_t __ret; \
9892  __builtin_neon_vld3_v(&__ret, __p0, 17); \
9893 \
9894  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9895  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9896  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9897  __ret; \
9898})
9899#endif
9900
9901#ifdef __LITTLE_ENDIAN__
9902#define vld3_s8(__p0) __extension__ ({ \
9903  int8x8x3_t __ret; \
9904  __builtin_neon_vld3_v(&__ret, __p0, 0); \
9905  __ret; \
9906})
9907#else
9908#define vld3_s8(__p0) __extension__ ({ \
9909  int8x8x3_t __ret; \
9910  __builtin_neon_vld3_v(&__ret, __p0, 0); \
9911 \
9912  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
9913  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
9914  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
9915  __ret; \
9916})
9917#endif
9918
9919#ifdef __LITTLE_ENDIAN__
9920#define vld3_f32(__p0) __extension__ ({ \
9921  float32x2x3_t __ret; \
9922  __builtin_neon_vld3_v(&__ret, __p0, 9); \
9923  __ret; \
9924})
9925#else
9926#define vld3_f32(__p0) __extension__ ({ \
9927  float32x2x3_t __ret; \
9928  __builtin_neon_vld3_v(&__ret, __p0, 9); \
9929 \
9930  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9931  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9932  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9933  __ret; \
9934})
9935#endif
9936
9937#ifdef __LITTLE_ENDIAN__
9938#define vld3_f16(__p0) __extension__ ({ \
9939  float16x4x3_t __ret; \
9940  __builtin_neon_vld3_v(&__ret, __p0, 8); \
9941  __ret; \
9942})
9943#else
9944#define vld3_f16(__p0) __extension__ ({ \
9945  float16x4x3_t __ret; \
9946  __builtin_neon_vld3_v(&__ret, __p0, 8); \
9947 \
9948  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9949  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
9950  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
9951  __ret; \
9952})
9953#endif
9954
9955#ifdef __LITTLE_ENDIAN__
9956#define vld3_s32(__p0) __extension__ ({ \
9957  int32x2x3_t __ret; \
9958  __builtin_neon_vld3_v(&__ret, __p0, 2); \
9959  __ret; \
9960})
9961#else
9962#define vld3_s32(__p0) __extension__ ({ \
9963  int32x2x3_t __ret; \
9964  __builtin_neon_vld3_v(&__ret, __p0, 2); \
9965 \
9966  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
9967  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
9968  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
9969  __ret; \
9970})
9971#endif
9972
9973#ifdef __LITTLE_ENDIAN__
9974#define vld3_s64(__p0) __extension__ ({ \
9975  int64x1x3_t __ret; \
9976  __builtin_neon_vld3_v(&__ret, __p0, 3); \
9977  __ret; \
9978})
9979#else
9980#define vld3_s64(__p0) __extension__ ({ \
9981  int64x1x3_t __ret; \
9982  __builtin_neon_vld3_v(&__ret, __p0, 3); \
9983  __ret; \
9984})
9985#endif
9986
9987#ifdef __LITTLE_ENDIAN__
9988#define vld3_s16(__p0) __extension__ ({ \
9989  int16x4x3_t __ret; \
9990  __builtin_neon_vld3_v(&__ret, __p0, 1); \
9991  __ret; \
9992})
9993#else
9994#define vld3_s16(__p0) __extension__ ({ \
9995  int16x4x3_t __ret; \
9996  __builtin_neon_vld3_v(&__ret, __p0, 1); \
9997 \
9998  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
9999  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10000  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10001  __ret; \
10002})
10003#endif
10004
10005#ifdef __LITTLE_ENDIAN__
10006#define vld3_dup_p8(__p0) __extension__ ({ \
10007  poly8x8x3_t __ret; \
10008  __builtin_neon_vld3_dup_v(&__ret, __p0, 4); \
10009  __ret; \
10010})
10011#else
10012#define vld3_dup_p8(__p0) __extension__ ({ \
10013  poly8x8x3_t __ret; \
10014  __builtin_neon_vld3_dup_v(&__ret, __p0, 4); \
10015 \
10016  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10017  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10018  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10019  __ret; \
10020})
10021#endif
10022
10023#ifdef __LITTLE_ENDIAN__
10024#define vld3_dup_p16(__p0) __extension__ ({ \
10025  poly16x4x3_t __ret; \
10026  __builtin_neon_vld3_dup_v(&__ret, __p0, 5); \
10027  __ret; \
10028})
10029#else
10030#define vld3_dup_p16(__p0) __extension__ ({ \
10031  poly16x4x3_t __ret; \
10032  __builtin_neon_vld3_dup_v(&__ret, __p0, 5); \
10033 \
10034  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10035  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10036  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10037  __ret; \
10038})
10039#endif
10040
10041#ifdef __LITTLE_ENDIAN__
10042#define vld3_dup_u8(__p0) __extension__ ({ \
10043  uint8x8x3_t __ret; \
10044  __builtin_neon_vld3_dup_v(&__ret, __p0, 16); \
10045  __ret; \
10046})
10047#else
10048#define vld3_dup_u8(__p0) __extension__ ({ \
10049  uint8x8x3_t __ret; \
10050  __builtin_neon_vld3_dup_v(&__ret, __p0, 16); \
10051 \
10052  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10053  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10054  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10055  __ret; \
10056})
10057#endif
10058
10059#ifdef __LITTLE_ENDIAN__
10060#define vld3_dup_u32(__p0) __extension__ ({ \
10061  uint32x2x3_t __ret; \
10062  __builtin_neon_vld3_dup_v(&__ret, __p0, 18); \
10063  __ret; \
10064})
10065#else
10066#define vld3_dup_u32(__p0) __extension__ ({ \
10067  uint32x2x3_t __ret; \
10068  __builtin_neon_vld3_dup_v(&__ret, __p0, 18); \
10069 \
10070  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10071  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10072  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10073  __ret; \
10074})
10075#endif
10076
10077#ifdef __LITTLE_ENDIAN__
10078#define vld3_dup_u64(__p0) __extension__ ({ \
10079  uint64x1x3_t __ret; \
10080  __builtin_neon_vld3_dup_v(&__ret, __p0, 19); \
10081  __ret; \
10082})
10083#else
10084#define vld3_dup_u64(__p0) __extension__ ({ \
10085  uint64x1x3_t __ret; \
10086  __builtin_neon_vld3_dup_v(&__ret, __p0, 19); \
10087  __ret; \
10088})
10089#endif
10090
10091#ifdef __LITTLE_ENDIAN__
10092#define vld3_dup_u16(__p0) __extension__ ({ \
10093  uint16x4x3_t __ret; \
10094  __builtin_neon_vld3_dup_v(&__ret, __p0, 17); \
10095  __ret; \
10096})
10097#else
10098#define vld3_dup_u16(__p0) __extension__ ({ \
10099  uint16x4x3_t __ret; \
10100  __builtin_neon_vld3_dup_v(&__ret, __p0, 17); \
10101 \
10102  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10103  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10104  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10105  __ret; \
10106})
10107#endif
10108
10109#ifdef __LITTLE_ENDIAN__
10110#define vld3_dup_s8(__p0) __extension__ ({ \
10111  int8x8x3_t __ret; \
10112  __builtin_neon_vld3_dup_v(&__ret, __p0, 0); \
10113  __ret; \
10114})
10115#else
10116#define vld3_dup_s8(__p0) __extension__ ({ \
10117  int8x8x3_t __ret; \
10118  __builtin_neon_vld3_dup_v(&__ret, __p0, 0); \
10119 \
10120  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10121  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10122  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10123  __ret; \
10124})
10125#endif
10126
10127#ifdef __LITTLE_ENDIAN__
10128#define vld3_dup_f32(__p0) __extension__ ({ \
10129  float32x2x3_t __ret; \
10130  __builtin_neon_vld3_dup_v(&__ret, __p0, 9); \
10131  __ret; \
10132})
10133#else
10134#define vld3_dup_f32(__p0) __extension__ ({ \
10135  float32x2x3_t __ret; \
10136  __builtin_neon_vld3_dup_v(&__ret, __p0, 9); \
10137 \
10138  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10139  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10140  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10141  __ret; \
10142})
10143#endif
10144
10145#ifdef __LITTLE_ENDIAN__
10146#define vld3_dup_f16(__p0) __extension__ ({ \
10147  float16x4x3_t __ret; \
10148  __builtin_neon_vld3_dup_v(&__ret, __p0, 8); \
10149  __ret; \
10150})
10151#else
10152#define vld3_dup_f16(__p0) __extension__ ({ \
10153  float16x4x3_t __ret; \
10154  __builtin_neon_vld3_dup_v(&__ret, __p0, 8); \
10155 \
10156  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10157  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10158  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10159  __ret; \
10160})
10161#endif
10162
10163#ifdef __LITTLE_ENDIAN__
10164#define vld3_dup_s32(__p0) __extension__ ({ \
10165  int32x2x3_t __ret; \
10166  __builtin_neon_vld3_dup_v(&__ret, __p0, 2); \
10167  __ret; \
10168})
10169#else
10170#define vld3_dup_s32(__p0) __extension__ ({ \
10171  int32x2x3_t __ret; \
10172  __builtin_neon_vld3_dup_v(&__ret, __p0, 2); \
10173 \
10174  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10175  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10176  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10177  __ret; \
10178})
10179#endif
10180
10181#ifdef __LITTLE_ENDIAN__
10182#define vld3_dup_s64(__p0) __extension__ ({ \
10183  int64x1x3_t __ret; \
10184  __builtin_neon_vld3_dup_v(&__ret, __p0, 3); \
10185  __ret; \
10186})
10187#else
10188#define vld3_dup_s64(__p0) __extension__ ({ \
10189  int64x1x3_t __ret; \
10190  __builtin_neon_vld3_dup_v(&__ret, __p0, 3); \
10191  __ret; \
10192})
10193#endif
10194
10195#ifdef __LITTLE_ENDIAN__
10196#define vld3_dup_s16(__p0) __extension__ ({ \
10197  int16x4x3_t __ret; \
10198  __builtin_neon_vld3_dup_v(&__ret, __p0, 1); \
10199  __ret; \
10200})
10201#else
10202#define vld3_dup_s16(__p0) __extension__ ({ \
10203  int16x4x3_t __ret; \
10204  __builtin_neon_vld3_dup_v(&__ret, __p0, 1); \
10205 \
10206  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10207  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10208  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10209  __ret; \
10210})
10211#endif
10212
10213#ifdef __LITTLE_ENDIAN__
10214#define vld3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
10215  poly8x8x3_t __s1 = __p1; \
10216  poly8x8x3_t __ret; \
10217  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 4); \
10218  __ret; \
10219})
10220#else
10221#define vld3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
10222  poly8x8x3_t __s1 = __p1; \
10223  poly8x8x3_t __rev1; \
10224  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10225  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10226  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10227  poly8x8x3_t __ret; \
10228  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 4); \
10229 \
10230  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10231  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10232  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10233  __ret; \
10234})
10235#endif
10236
10237#ifdef __LITTLE_ENDIAN__
10238#define vld3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10239  poly16x4x3_t __s1 = __p1; \
10240  poly16x4x3_t __ret; \
10241  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 5); \
10242  __ret; \
10243})
10244#else
10245#define vld3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10246  poly16x4x3_t __s1 = __p1; \
10247  poly16x4x3_t __rev1; \
10248  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10249  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10250  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10251  poly16x4x3_t __ret; \
10252  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 5); \
10253 \
10254  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10255  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10256  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10257  __ret; \
10258})
10259#endif
10260
10261#ifdef __LITTLE_ENDIAN__
10262#define vld3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10263  poly16x8x3_t __s1 = __p1; \
10264  poly16x8x3_t __ret; \
10265  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 37); \
10266  __ret; \
10267})
10268#else
10269#define vld3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
10270  poly16x8x3_t __s1 = __p1; \
10271  poly16x8x3_t __rev1; \
10272  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10273  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10274  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10275  poly16x8x3_t __ret; \
10276  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 37); \
10277 \
10278  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10279  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10280  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10281  __ret; \
10282})
10283#endif
10284
10285#ifdef __LITTLE_ENDIAN__
10286#define vld3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10287  uint32x4x3_t __s1 = __p1; \
10288  uint32x4x3_t __ret; \
10289  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 50); \
10290  __ret; \
10291})
10292#else
10293#define vld3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10294  uint32x4x3_t __s1 = __p1; \
10295  uint32x4x3_t __rev1; \
10296  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10297  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10298  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10299  uint32x4x3_t __ret; \
10300  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 50); \
10301 \
10302  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10303  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10304  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10305  __ret; \
10306})
10307#endif
10308
10309#ifdef __LITTLE_ENDIAN__
10310#define vld3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10311  uint16x8x3_t __s1 = __p1; \
10312  uint16x8x3_t __ret; \
10313  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 49); \
10314  __ret; \
10315})
10316#else
10317#define vld3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10318  uint16x8x3_t __s1 = __p1; \
10319  uint16x8x3_t __rev1; \
10320  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10321  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10322  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10323  uint16x8x3_t __ret; \
10324  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 49); \
10325 \
10326  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10327  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10328  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10329  __ret; \
10330})
10331#endif
10332
10333#ifdef __LITTLE_ENDIAN__
10334#define vld3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10335  float32x4x3_t __s1 = __p1; \
10336  float32x4x3_t __ret; \
10337  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 41); \
10338  __ret; \
10339})
10340#else
10341#define vld3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10342  float32x4x3_t __s1 = __p1; \
10343  float32x4x3_t __rev1; \
10344  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10345  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10346  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10347  float32x4x3_t __ret; \
10348  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 41); \
10349 \
10350  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10351  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10352  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10353  __ret; \
10354})
10355#endif
10356
10357#ifdef __LITTLE_ENDIAN__
10358#define vld3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10359  float16x8x3_t __s1 = __p1; \
10360  float16x8x3_t __ret; \
10361  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 40); \
10362  __ret; \
10363})
10364#else
10365#define vld3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10366  float16x8x3_t __s1 = __p1; \
10367  float16x8x3_t __rev1; \
10368  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10369  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10370  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10371  float16x8x3_t __ret; \
10372  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 40); \
10373 \
10374  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10375  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10376  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10377  __ret; \
10378})
10379#endif
10380
10381#ifdef __LITTLE_ENDIAN__
10382#define vld3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10383  int32x4x3_t __s1 = __p1; \
10384  int32x4x3_t __ret; \
10385  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 34); \
10386  __ret; \
10387})
10388#else
10389#define vld3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10390  int32x4x3_t __s1 = __p1; \
10391  int32x4x3_t __rev1; \
10392  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10393  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10394  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10395  int32x4x3_t __ret; \
10396  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 34); \
10397 \
10398  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10399  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10400  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10401  __ret; \
10402})
10403#endif
10404
10405#ifdef __LITTLE_ENDIAN__
10406#define vld3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10407  int16x8x3_t __s1 = __p1; \
10408  int16x8x3_t __ret; \
10409  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 33); \
10410  __ret; \
10411})
10412#else
10413#define vld3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10414  int16x8x3_t __s1 = __p1; \
10415  int16x8x3_t __rev1; \
10416  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10417  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10418  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10419  int16x8x3_t __ret; \
10420  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 33); \
10421 \
10422  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10423  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10424  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10425  __ret; \
10426})
10427#endif
10428
10429#ifdef __LITTLE_ENDIAN__
10430#define vld3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
10431  uint8x8x3_t __s1 = __p1; \
10432  uint8x8x3_t __ret; \
10433  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 16); \
10434  __ret; \
10435})
10436#else
10437#define vld3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
10438  uint8x8x3_t __s1 = __p1; \
10439  uint8x8x3_t __rev1; \
10440  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10441  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10442  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10443  uint8x8x3_t __ret; \
10444  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 16); \
10445 \
10446  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10447  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10448  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10449  __ret; \
10450})
10451#endif
10452
10453#ifdef __LITTLE_ENDIAN__
10454#define vld3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10455  uint32x2x3_t __s1 = __p1; \
10456  uint32x2x3_t __ret; \
10457  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 18); \
10458  __ret; \
10459})
10460#else
10461#define vld3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
10462  uint32x2x3_t __s1 = __p1; \
10463  uint32x2x3_t __rev1; \
10464  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
10465  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
10466  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
10467  uint32x2x3_t __ret; \
10468  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 18); \
10469 \
10470  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10471  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10472  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10473  __ret; \
10474})
10475#endif
10476
10477#ifdef __LITTLE_ENDIAN__
10478#define vld3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10479  uint16x4x3_t __s1 = __p1; \
10480  uint16x4x3_t __ret; \
10481  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 17); \
10482  __ret; \
10483})
10484#else
10485#define vld3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
10486  uint16x4x3_t __s1 = __p1; \
10487  uint16x4x3_t __rev1; \
10488  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10489  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10490  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10491  uint16x4x3_t __ret; \
10492  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 17); \
10493 \
10494  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10495  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10496  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10497  __ret; \
10498})
10499#endif
10500
10501#ifdef __LITTLE_ENDIAN__
10502#define vld3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
10503  int8x8x3_t __s1 = __p1; \
10504  int8x8x3_t __ret; \
10505  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 0); \
10506  __ret; \
10507})
10508#else
10509#define vld3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
10510  int8x8x3_t __s1 = __p1; \
10511  int8x8x3_t __rev1; \
10512  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10513  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10514  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10515  int8x8x3_t __ret; \
10516  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 0); \
10517 \
10518  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10519  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10520  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10521  __ret; \
10522})
10523#endif
10524
10525#ifdef __LITTLE_ENDIAN__
10526#define vld3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10527  float32x2x3_t __s1 = __p1; \
10528  float32x2x3_t __ret; \
10529  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 9); \
10530  __ret; \
10531})
10532#else
10533#define vld3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
10534  float32x2x3_t __s1 = __p1; \
10535  float32x2x3_t __rev1; \
10536  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
10537  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
10538  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
10539  float32x2x3_t __ret; \
10540  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 9); \
10541 \
10542  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10543  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10544  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10545  __ret; \
10546})
10547#endif
10548
10549#ifdef __LITTLE_ENDIAN__
10550#define vld3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10551  float16x4x3_t __s1 = __p1; \
10552  float16x4x3_t __ret; \
10553  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 8); \
10554  __ret; \
10555})
10556#else
10557#define vld3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
10558  float16x4x3_t __s1 = __p1; \
10559  float16x4x3_t __rev1; \
10560  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10561  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10562  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10563  float16x4x3_t __ret; \
10564  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 8); \
10565 \
10566  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10567  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10568  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10569  __ret; \
10570})
10571#endif
10572
10573#ifdef __LITTLE_ENDIAN__
10574#define vld3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10575  int32x2x3_t __s1 = __p1; \
10576  int32x2x3_t __ret; \
10577  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 2); \
10578  __ret; \
10579})
10580#else
10581#define vld3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
10582  int32x2x3_t __s1 = __p1; \
10583  int32x2x3_t __rev1; \
10584  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
10585  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
10586  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
10587  int32x2x3_t __ret; \
10588  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 2); \
10589 \
10590  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10591  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10592  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10593  __ret; \
10594})
10595#endif
10596
10597#ifdef __LITTLE_ENDIAN__
10598#define vld3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10599  int16x4x3_t __s1 = __p1; \
10600  int16x4x3_t __ret; \
10601  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 1); \
10602  __ret; \
10603})
10604#else
10605#define vld3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
10606  int16x4x3_t __s1 = __p1; \
10607  int16x4x3_t __rev1; \
10608  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
10609  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
10610  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
10611  int16x4x3_t __ret; \
10612  __builtin_neon_vld3_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 1); \
10613 \
10614  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10615  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10616  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10617  __ret; \
10618})
10619#endif
10620
10621#ifdef __LITTLE_ENDIAN__
10622#define vld4_p8(__p0) __extension__ ({ \
10623  poly8x8x4_t __ret; \
10624  __builtin_neon_vld4_v(&__ret, __p0, 4); \
10625  __ret; \
10626})
10627#else
10628#define vld4_p8(__p0) __extension__ ({ \
10629  poly8x8x4_t __ret; \
10630  __builtin_neon_vld4_v(&__ret, __p0, 4); \
10631 \
10632  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10633  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10634  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10635  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10636  __ret; \
10637})
10638#endif
10639
10640#ifdef __LITTLE_ENDIAN__
10641#define vld4_p16(__p0) __extension__ ({ \
10642  poly16x4x4_t __ret; \
10643  __builtin_neon_vld4_v(&__ret, __p0, 5); \
10644  __ret; \
10645})
10646#else
10647#define vld4_p16(__p0) __extension__ ({ \
10648  poly16x4x4_t __ret; \
10649  __builtin_neon_vld4_v(&__ret, __p0, 5); \
10650 \
10651  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10652  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10653  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10654  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10655  __ret; \
10656})
10657#endif
10658
10659#ifdef __LITTLE_ENDIAN__
10660#define vld4q_p8(__p0) __extension__ ({ \
10661  poly8x16x4_t __ret; \
10662  __builtin_neon_vld4q_v(&__ret, __p0, 36); \
10663  __ret; \
10664})
10665#else
10666#define vld4q_p8(__p0) __extension__ ({ \
10667  poly8x16x4_t __ret; \
10668  __builtin_neon_vld4q_v(&__ret, __p0, 36); \
10669 \
10670  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10671  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10672  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10673  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10674  __ret; \
10675})
10676#endif
10677
10678#ifdef __LITTLE_ENDIAN__
10679#define vld4q_p16(__p0) __extension__ ({ \
10680  poly16x8x4_t __ret; \
10681  __builtin_neon_vld4q_v(&__ret, __p0, 37); \
10682  __ret; \
10683})
10684#else
10685#define vld4q_p16(__p0) __extension__ ({ \
10686  poly16x8x4_t __ret; \
10687  __builtin_neon_vld4q_v(&__ret, __p0, 37); \
10688 \
10689  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10690  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10691  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10692  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10693  __ret; \
10694})
10695#endif
10696
10697#ifdef __LITTLE_ENDIAN__
10698#define vld4q_u8(__p0) __extension__ ({ \
10699  uint8x16x4_t __ret; \
10700  __builtin_neon_vld4q_v(&__ret, __p0, 48); \
10701  __ret; \
10702})
10703#else
10704#define vld4q_u8(__p0) __extension__ ({ \
10705  uint8x16x4_t __ret; \
10706  __builtin_neon_vld4q_v(&__ret, __p0, 48); \
10707 \
10708  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10709  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10710  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10711  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10712  __ret; \
10713})
10714#endif
10715
10716#ifdef __LITTLE_ENDIAN__
10717#define vld4q_u32(__p0) __extension__ ({ \
10718  uint32x4x4_t __ret; \
10719  __builtin_neon_vld4q_v(&__ret, __p0, 50); \
10720  __ret; \
10721})
10722#else
10723#define vld4q_u32(__p0) __extension__ ({ \
10724  uint32x4x4_t __ret; \
10725  __builtin_neon_vld4q_v(&__ret, __p0, 50); \
10726 \
10727  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10728  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10729  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10730  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10731  __ret; \
10732})
10733#endif
10734
10735#ifdef __LITTLE_ENDIAN__
10736#define vld4q_u16(__p0) __extension__ ({ \
10737  uint16x8x4_t __ret; \
10738  __builtin_neon_vld4q_v(&__ret, __p0, 49); \
10739  __ret; \
10740})
10741#else
10742#define vld4q_u16(__p0) __extension__ ({ \
10743  uint16x8x4_t __ret; \
10744  __builtin_neon_vld4q_v(&__ret, __p0, 49); \
10745 \
10746  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10747  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10748  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10749  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10750  __ret; \
10751})
10752#endif
10753
10754#ifdef __LITTLE_ENDIAN__
10755#define vld4q_s8(__p0) __extension__ ({ \
10756  int8x16x4_t __ret; \
10757  __builtin_neon_vld4q_v(&__ret, __p0, 32); \
10758  __ret; \
10759})
10760#else
10761#define vld4q_s8(__p0) __extension__ ({ \
10762  int8x16x4_t __ret; \
10763  __builtin_neon_vld4q_v(&__ret, __p0, 32); \
10764 \
10765  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10766  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10767  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10768  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
10769  __ret; \
10770})
10771#endif
10772
10773#ifdef __LITTLE_ENDIAN__
10774#define vld4q_f32(__p0) __extension__ ({ \
10775  float32x4x4_t __ret; \
10776  __builtin_neon_vld4q_v(&__ret, __p0, 41); \
10777  __ret; \
10778})
10779#else
10780#define vld4q_f32(__p0) __extension__ ({ \
10781  float32x4x4_t __ret; \
10782  __builtin_neon_vld4q_v(&__ret, __p0, 41); \
10783 \
10784  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10785  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10786  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10787  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10788  __ret; \
10789})
10790#endif
10791
10792#ifdef __LITTLE_ENDIAN__
10793#define vld4q_f16(__p0) __extension__ ({ \
10794  float16x8x4_t __ret; \
10795  __builtin_neon_vld4q_v(&__ret, __p0, 40); \
10796  __ret; \
10797})
10798#else
10799#define vld4q_f16(__p0) __extension__ ({ \
10800  float16x8x4_t __ret; \
10801  __builtin_neon_vld4q_v(&__ret, __p0, 40); \
10802 \
10803  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10804  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10805  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10806  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10807  __ret; \
10808})
10809#endif
10810
10811#ifdef __LITTLE_ENDIAN__
10812#define vld4q_s32(__p0) __extension__ ({ \
10813  int32x4x4_t __ret; \
10814  __builtin_neon_vld4q_v(&__ret, __p0, 34); \
10815  __ret; \
10816})
10817#else
10818#define vld4q_s32(__p0) __extension__ ({ \
10819  int32x4x4_t __ret; \
10820  __builtin_neon_vld4q_v(&__ret, __p0, 34); \
10821 \
10822  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10823  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10824  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10825  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10826  __ret; \
10827})
10828#endif
10829
10830#ifdef __LITTLE_ENDIAN__
10831#define vld4q_s16(__p0) __extension__ ({ \
10832  int16x8x4_t __ret; \
10833  __builtin_neon_vld4q_v(&__ret, __p0, 33); \
10834  __ret; \
10835})
10836#else
10837#define vld4q_s16(__p0) __extension__ ({ \
10838  int16x8x4_t __ret; \
10839  __builtin_neon_vld4q_v(&__ret, __p0, 33); \
10840 \
10841  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10842  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10843  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10844  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10845  __ret; \
10846})
10847#endif
10848
10849#ifdef __LITTLE_ENDIAN__
10850#define vld4_u8(__p0) __extension__ ({ \
10851  uint8x8x4_t __ret; \
10852  __builtin_neon_vld4_v(&__ret, __p0, 16); \
10853  __ret; \
10854})
10855#else
10856#define vld4_u8(__p0) __extension__ ({ \
10857  uint8x8x4_t __ret; \
10858  __builtin_neon_vld4_v(&__ret, __p0, 16); \
10859 \
10860  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10861  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10862  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10863  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10864  __ret; \
10865})
10866#endif
10867
10868#ifdef __LITTLE_ENDIAN__
10869#define vld4_u32(__p0) __extension__ ({ \
10870  uint32x2x4_t __ret; \
10871  __builtin_neon_vld4_v(&__ret, __p0, 18); \
10872  __ret; \
10873})
10874#else
10875#define vld4_u32(__p0) __extension__ ({ \
10876  uint32x2x4_t __ret; \
10877  __builtin_neon_vld4_v(&__ret, __p0, 18); \
10878 \
10879  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10880  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10881  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10882  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10883  __ret; \
10884})
10885#endif
10886
10887#ifdef __LITTLE_ENDIAN__
10888#define vld4_u64(__p0) __extension__ ({ \
10889  uint64x1x4_t __ret; \
10890  __builtin_neon_vld4_v(&__ret, __p0, 19); \
10891  __ret; \
10892})
10893#else
10894#define vld4_u64(__p0) __extension__ ({ \
10895  uint64x1x4_t __ret; \
10896  __builtin_neon_vld4_v(&__ret, __p0, 19); \
10897  __ret; \
10898})
10899#endif
10900
10901#ifdef __LITTLE_ENDIAN__
10902#define vld4_u16(__p0) __extension__ ({ \
10903  uint16x4x4_t __ret; \
10904  __builtin_neon_vld4_v(&__ret, __p0, 17); \
10905  __ret; \
10906})
10907#else
10908#define vld4_u16(__p0) __extension__ ({ \
10909  uint16x4x4_t __ret; \
10910  __builtin_neon_vld4_v(&__ret, __p0, 17); \
10911 \
10912  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10913  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10914  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10915  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10916  __ret; \
10917})
10918#endif
10919
10920#ifdef __LITTLE_ENDIAN__
10921#define vld4_s8(__p0) __extension__ ({ \
10922  int8x8x4_t __ret; \
10923  __builtin_neon_vld4_v(&__ret, __p0, 0); \
10924  __ret; \
10925})
10926#else
10927#define vld4_s8(__p0) __extension__ ({ \
10928  int8x8x4_t __ret; \
10929  __builtin_neon_vld4_v(&__ret, __p0, 0); \
10930 \
10931  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
10932  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
10933  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
10934  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
10935  __ret; \
10936})
10937#endif
10938
10939#ifdef __LITTLE_ENDIAN__
10940#define vld4_f32(__p0) __extension__ ({ \
10941  float32x2x4_t __ret; \
10942  __builtin_neon_vld4_v(&__ret, __p0, 9); \
10943  __ret; \
10944})
10945#else
10946#define vld4_f32(__p0) __extension__ ({ \
10947  float32x2x4_t __ret; \
10948  __builtin_neon_vld4_v(&__ret, __p0, 9); \
10949 \
10950  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10951  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10952  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10953  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10954  __ret; \
10955})
10956#endif
10957
10958#ifdef __LITTLE_ENDIAN__
10959#define vld4_f16(__p0) __extension__ ({ \
10960  float16x4x4_t __ret; \
10961  __builtin_neon_vld4_v(&__ret, __p0, 8); \
10962  __ret; \
10963})
10964#else
10965#define vld4_f16(__p0) __extension__ ({ \
10966  float16x4x4_t __ret; \
10967  __builtin_neon_vld4_v(&__ret, __p0, 8); \
10968 \
10969  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
10970  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
10971  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
10972  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
10973  __ret; \
10974})
10975#endif
10976
10977#ifdef __LITTLE_ENDIAN__
10978#define vld4_s32(__p0) __extension__ ({ \
10979  int32x2x4_t __ret; \
10980  __builtin_neon_vld4_v(&__ret, __p0, 2); \
10981  __ret; \
10982})
10983#else
10984#define vld4_s32(__p0) __extension__ ({ \
10985  int32x2x4_t __ret; \
10986  __builtin_neon_vld4_v(&__ret, __p0, 2); \
10987 \
10988  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
10989  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
10990  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
10991  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
10992  __ret; \
10993})
10994#endif
10995
10996#ifdef __LITTLE_ENDIAN__
10997#define vld4_s64(__p0) __extension__ ({ \
10998  int64x1x4_t __ret; \
10999  __builtin_neon_vld4_v(&__ret, __p0, 3); \
11000  __ret; \
11001})
11002#else
11003#define vld4_s64(__p0) __extension__ ({ \
11004  int64x1x4_t __ret; \
11005  __builtin_neon_vld4_v(&__ret, __p0, 3); \
11006  __ret; \
11007})
11008#endif
11009
11010#ifdef __LITTLE_ENDIAN__
11011#define vld4_s16(__p0) __extension__ ({ \
11012  int16x4x4_t __ret; \
11013  __builtin_neon_vld4_v(&__ret, __p0, 1); \
11014  __ret; \
11015})
11016#else
11017#define vld4_s16(__p0) __extension__ ({ \
11018  int16x4x4_t __ret; \
11019  __builtin_neon_vld4_v(&__ret, __p0, 1); \
11020 \
11021  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11022  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11023  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11024  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11025  __ret; \
11026})
11027#endif
11028
11029#ifdef __LITTLE_ENDIAN__
11030#define vld4_dup_p8(__p0) __extension__ ({ \
11031  poly8x8x4_t __ret; \
11032  __builtin_neon_vld4_dup_v(&__ret, __p0, 4); \
11033  __ret; \
11034})
11035#else
11036#define vld4_dup_p8(__p0) __extension__ ({ \
11037  poly8x8x4_t __ret; \
11038  __builtin_neon_vld4_dup_v(&__ret, __p0, 4); \
11039 \
11040  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11041  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11042  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11043  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11044  __ret; \
11045})
11046#endif
11047
11048#ifdef __LITTLE_ENDIAN__
11049#define vld4_dup_p16(__p0) __extension__ ({ \
11050  poly16x4x4_t __ret; \
11051  __builtin_neon_vld4_dup_v(&__ret, __p0, 5); \
11052  __ret; \
11053})
11054#else
11055#define vld4_dup_p16(__p0) __extension__ ({ \
11056  poly16x4x4_t __ret; \
11057  __builtin_neon_vld4_dup_v(&__ret, __p0, 5); \
11058 \
11059  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11060  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11061  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11062  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11063  __ret; \
11064})
11065#endif
11066
11067#ifdef __LITTLE_ENDIAN__
11068#define vld4_dup_u8(__p0) __extension__ ({ \
11069  uint8x8x4_t __ret; \
11070  __builtin_neon_vld4_dup_v(&__ret, __p0, 16); \
11071  __ret; \
11072})
11073#else
11074#define vld4_dup_u8(__p0) __extension__ ({ \
11075  uint8x8x4_t __ret; \
11076  __builtin_neon_vld4_dup_v(&__ret, __p0, 16); \
11077 \
11078  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11079  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11080  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11081  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11082  __ret; \
11083})
11084#endif
11085
11086#ifdef __LITTLE_ENDIAN__
11087#define vld4_dup_u32(__p0) __extension__ ({ \
11088  uint32x2x4_t __ret; \
11089  __builtin_neon_vld4_dup_v(&__ret, __p0, 18); \
11090  __ret; \
11091})
11092#else
11093#define vld4_dup_u32(__p0) __extension__ ({ \
11094  uint32x2x4_t __ret; \
11095  __builtin_neon_vld4_dup_v(&__ret, __p0, 18); \
11096 \
11097  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11098  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11099  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11100  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11101  __ret; \
11102})
11103#endif
11104
11105#ifdef __LITTLE_ENDIAN__
11106#define vld4_dup_u64(__p0) __extension__ ({ \
11107  uint64x1x4_t __ret; \
11108  __builtin_neon_vld4_dup_v(&__ret, __p0, 19); \
11109  __ret; \
11110})
11111#else
11112#define vld4_dup_u64(__p0) __extension__ ({ \
11113  uint64x1x4_t __ret; \
11114  __builtin_neon_vld4_dup_v(&__ret, __p0, 19); \
11115  __ret; \
11116})
11117#endif
11118
11119#ifdef __LITTLE_ENDIAN__
11120#define vld4_dup_u16(__p0) __extension__ ({ \
11121  uint16x4x4_t __ret; \
11122  __builtin_neon_vld4_dup_v(&__ret, __p0, 17); \
11123  __ret; \
11124})
11125#else
11126#define vld4_dup_u16(__p0) __extension__ ({ \
11127  uint16x4x4_t __ret; \
11128  __builtin_neon_vld4_dup_v(&__ret, __p0, 17); \
11129 \
11130  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11131  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11132  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11133  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11134  __ret; \
11135})
11136#endif
11137
11138#ifdef __LITTLE_ENDIAN__
11139#define vld4_dup_s8(__p0) __extension__ ({ \
11140  int8x8x4_t __ret; \
11141  __builtin_neon_vld4_dup_v(&__ret, __p0, 0); \
11142  __ret; \
11143})
11144#else
11145#define vld4_dup_s8(__p0) __extension__ ({ \
11146  int8x8x4_t __ret; \
11147  __builtin_neon_vld4_dup_v(&__ret, __p0, 0); \
11148 \
11149  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11150  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11151  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11152  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11153  __ret; \
11154})
11155#endif
11156
11157#ifdef __LITTLE_ENDIAN__
11158#define vld4_dup_f32(__p0) __extension__ ({ \
11159  float32x2x4_t __ret; \
11160  __builtin_neon_vld4_dup_v(&__ret, __p0, 9); \
11161  __ret; \
11162})
11163#else
11164#define vld4_dup_f32(__p0) __extension__ ({ \
11165  float32x2x4_t __ret; \
11166  __builtin_neon_vld4_dup_v(&__ret, __p0, 9); \
11167 \
11168  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11169  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11170  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11171  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11172  __ret; \
11173})
11174#endif
11175
11176#ifdef __LITTLE_ENDIAN__
11177#define vld4_dup_f16(__p0) __extension__ ({ \
11178  float16x4x4_t __ret; \
11179  __builtin_neon_vld4_dup_v(&__ret, __p0, 8); \
11180  __ret; \
11181})
11182#else
11183#define vld4_dup_f16(__p0) __extension__ ({ \
11184  float16x4x4_t __ret; \
11185  __builtin_neon_vld4_dup_v(&__ret, __p0, 8); \
11186 \
11187  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11188  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11189  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11190  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11191  __ret; \
11192})
11193#endif
11194
11195#ifdef __LITTLE_ENDIAN__
11196#define vld4_dup_s32(__p0) __extension__ ({ \
11197  int32x2x4_t __ret; \
11198  __builtin_neon_vld4_dup_v(&__ret, __p0, 2); \
11199  __ret; \
11200})
11201#else
11202#define vld4_dup_s32(__p0) __extension__ ({ \
11203  int32x2x4_t __ret; \
11204  __builtin_neon_vld4_dup_v(&__ret, __p0, 2); \
11205 \
11206  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11207  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11208  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11209  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11210  __ret; \
11211})
11212#endif
11213
11214#ifdef __LITTLE_ENDIAN__
11215#define vld4_dup_s64(__p0) __extension__ ({ \
11216  int64x1x4_t __ret; \
11217  __builtin_neon_vld4_dup_v(&__ret, __p0, 3); \
11218  __ret; \
11219})
11220#else
11221#define vld4_dup_s64(__p0) __extension__ ({ \
11222  int64x1x4_t __ret; \
11223  __builtin_neon_vld4_dup_v(&__ret, __p0, 3); \
11224  __ret; \
11225})
11226#endif
11227
11228#ifdef __LITTLE_ENDIAN__
11229#define vld4_dup_s16(__p0) __extension__ ({ \
11230  int16x4x4_t __ret; \
11231  __builtin_neon_vld4_dup_v(&__ret, __p0, 1); \
11232  __ret; \
11233})
11234#else
11235#define vld4_dup_s16(__p0) __extension__ ({ \
11236  int16x4x4_t __ret; \
11237  __builtin_neon_vld4_dup_v(&__ret, __p0, 1); \
11238 \
11239  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11240  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11241  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11242  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11243  __ret; \
11244})
11245#endif
11246
11247#ifdef __LITTLE_ENDIAN__
11248#define vld4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
11249  poly8x8x4_t __s1 = __p1; \
11250  poly8x8x4_t __ret; \
11251  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 4); \
11252  __ret; \
11253})
11254#else
11255#define vld4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
11256  poly8x8x4_t __s1 = __p1; \
11257  poly8x8x4_t __rev1; \
11258  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11259  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11260  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11261  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11262  poly8x8x4_t __ret; \
11263  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 4); \
11264 \
11265  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11266  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11267  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11268  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11269  __ret; \
11270})
11271#endif
11272
11273#ifdef __LITTLE_ENDIAN__
11274#define vld4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11275  poly16x4x4_t __s1 = __p1; \
11276  poly16x4x4_t __ret; \
11277  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 5); \
11278  __ret; \
11279})
11280#else
11281#define vld4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11282  poly16x4x4_t __s1 = __p1; \
11283  poly16x4x4_t __rev1; \
11284  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11285  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11286  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11287  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11288  poly16x4x4_t __ret; \
11289  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 5); \
11290 \
11291  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11292  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11293  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11294  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11295  __ret; \
11296})
11297#endif
11298
11299#ifdef __LITTLE_ENDIAN__
11300#define vld4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11301  poly16x8x4_t __s1 = __p1; \
11302  poly16x8x4_t __ret; \
11303  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 37); \
11304  __ret; \
11305})
11306#else
11307#define vld4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
11308  poly16x8x4_t __s1 = __p1; \
11309  poly16x8x4_t __rev1; \
11310  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11311  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11312  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11313  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11314  poly16x8x4_t __ret; \
11315  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 37); \
11316 \
11317  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11318  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11319  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11320  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11321  __ret; \
11322})
11323#endif
11324
11325#ifdef __LITTLE_ENDIAN__
11326#define vld4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11327  uint32x4x4_t __s1 = __p1; \
11328  uint32x4x4_t __ret; \
11329  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 50); \
11330  __ret; \
11331})
11332#else
11333#define vld4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11334  uint32x4x4_t __s1 = __p1; \
11335  uint32x4x4_t __rev1; \
11336  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11337  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11338  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11339  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11340  uint32x4x4_t __ret; \
11341  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 50); \
11342 \
11343  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11344  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11345  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11346  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11347  __ret; \
11348})
11349#endif
11350
11351#ifdef __LITTLE_ENDIAN__
11352#define vld4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11353  uint16x8x4_t __s1 = __p1; \
11354  uint16x8x4_t __ret; \
11355  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 49); \
11356  __ret; \
11357})
11358#else
11359#define vld4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11360  uint16x8x4_t __s1 = __p1; \
11361  uint16x8x4_t __rev1; \
11362  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11363  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11364  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11365  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11366  uint16x8x4_t __ret; \
11367  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 49); \
11368 \
11369  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11370  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11371  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11372  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11373  __ret; \
11374})
11375#endif
11376
11377#ifdef __LITTLE_ENDIAN__
11378#define vld4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11379  float32x4x4_t __s1 = __p1; \
11380  float32x4x4_t __ret; \
11381  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 41); \
11382  __ret; \
11383})
11384#else
11385#define vld4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11386  float32x4x4_t __s1 = __p1; \
11387  float32x4x4_t __rev1; \
11388  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11389  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11390  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11391  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11392  float32x4x4_t __ret; \
11393  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 41); \
11394 \
11395  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11396  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11397  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11398  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11399  __ret; \
11400})
11401#endif
11402
11403#ifdef __LITTLE_ENDIAN__
11404#define vld4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11405  float16x8x4_t __s1 = __p1; \
11406  float16x8x4_t __ret; \
11407  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 40); \
11408  __ret; \
11409})
11410#else
11411#define vld4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11412  float16x8x4_t __s1 = __p1; \
11413  float16x8x4_t __rev1; \
11414  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11415  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11416  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11417  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11418  float16x8x4_t __ret; \
11419  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 40); \
11420 \
11421  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11422  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11423  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11424  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11425  __ret; \
11426})
11427#endif
11428
11429#ifdef __LITTLE_ENDIAN__
11430#define vld4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11431  int32x4x4_t __s1 = __p1; \
11432  int32x4x4_t __ret; \
11433  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 34); \
11434  __ret; \
11435})
11436#else
11437#define vld4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11438  int32x4x4_t __s1 = __p1; \
11439  int32x4x4_t __rev1; \
11440  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11441  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11442  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11443  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11444  int32x4x4_t __ret; \
11445  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 34); \
11446 \
11447  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11448  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11449  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11450  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11451  __ret; \
11452})
11453#endif
11454
11455#ifdef __LITTLE_ENDIAN__
11456#define vld4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11457  int16x8x4_t __s1 = __p1; \
11458  int16x8x4_t __ret; \
11459  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 33); \
11460  __ret; \
11461})
11462#else
11463#define vld4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11464  int16x8x4_t __s1 = __p1; \
11465  int16x8x4_t __rev1; \
11466  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11467  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11468  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11469  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11470  int16x8x4_t __ret; \
11471  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 33); \
11472 \
11473  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11474  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11475  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11476  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11477  __ret; \
11478})
11479#endif
11480
11481#ifdef __LITTLE_ENDIAN__
11482#define vld4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
11483  uint8x8x4_t __s1 = __p1; \
11484  uint8x8x4_t __ret; \
11485  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 16); \
11486  __ret; \
11487})
11488#else
11489#define vld4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
11490  uint8x8x4_t __s1 = __p1; \
11491  uint8x8x4_t __rev1; \
11492  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11493  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11494  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11495  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11496  uint8x8x4_t __ret; \
11497  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 16); \
11498 \
11499  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11500  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11501  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11502  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11503  __ret; \
11504})
11505#endif
11506
11507#ifdef __LITTLE_ENDIAN__
11508#define vld4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11509  uint32x2x4_t __s1 = __p1; \
11510  uint32x2x4_t __ret; \
11511  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 18); \
11512  __ret; \
11513})
11514#else
11515#define vld4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
11516  uint32x2x4_t __s1 = __p1; \
11517  uint32x2x4_t __rev1; \
11518  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11519  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11520  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
11521  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
11522  uint32x2x4_t __ret; \
11523  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 18); \
11524 \
11525  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11526  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11527  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11528  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11529  __ret; \
11530})
11531#endif
11532
11533#ifdef __LITTLE_ENDIAN__
11534#define vld4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11535  uint16x4x4_t __s1 = __p1; \
11536  uint16x4x4_t __ret; \
11537  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 17); \
11538  __ret; \
11539})
11540#else
11541#define vld4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
11542  uint16x4x4_t __s1 = __p1; \
11543  uint16x4x4_t __rev1; \
11544  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11545  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11546  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11547  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11548  uint16x4x4_t __ret; \
11549  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 17); \
11550 \
11551  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11552  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11553  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11554  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11555  __ret; \
11556})
11557#endif
11558
11559#ifdef __LITTLE_ENDIAN__
11560#define vld4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
11561  int8x8x4_t __s1 = __p1; \
11562  int8x8x4_t __ret; \
11563  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 0); \
11564  __ret; \
11565})
11566#else
11567#define vld4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
11568  int8x8x4_t __s1 = __p1; \
11569  int8x8x4_t __rev1; \
11570  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11571  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11572  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11573  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11574  int8x8x4_t __ret; \
11575  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 0); \
11576 \
11577  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
11578  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
11579  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
11580  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
11581  __ret; \
11582})
11583#endif
11584
11585#ifdef __LITTLE_ENDIAN__
11586#define vld4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11587  float32x2x4_t __s1 = __p1; \
11588  float32x2x4_t __ret; \
11589  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 9); \
11590  __ret; \
11591})
11592#else
11593#define vld4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
11594  float32x2x4_t __s1 = __p1; \
11595  float32x2x4_t __rev1; \
11596  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11597  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11598  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
11599  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
11600  float32x2x4_t __ret; \
11601  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 9); \
11602 \
11603  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11604  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11605  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11606  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11607  __ret; \
11608})
11609#endif
11610
11611#ifdef __LITTLE_ENDIAN__
11612#define vld4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11613  float16x4x4_t __s1 = __p1; \
11614  float16x4x4_t __ret; \
11615  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 8); \
11616  __ret; \
11617})
11618#else
11619#define vld4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
11620  float16x4x4_t __s1 = __p1; \
11621  float16x4x4_t __rev1; \
11622  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11623  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11624  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11625  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11626  float16x4x4_t __ret; \
11627  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 8); \
11628 \
11629  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11630  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11631  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11632  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11633  __ret; \
11634})
11635#endif
11636
11637#ifdef __LITTLE_ENDIAN__
11638#define vld4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11639  int32x2x4_t __s1 = __p1; \
11640  int32x2x4_t __ret; \
11641  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 2); \
11642  __ret; \
11643})
11644#else
11645#define vld4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
11646  int32x2x4_t __s1 = __p1; \
11647  int32x2x4_t __rev1; \
11648  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
11649  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
11650  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
11651  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
11652  int32x2x4_t __ret; \
11653  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 2); \
11654 \
11655  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
11656  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
11657  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
11658  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
11659  __ret; \
11660})
11661#endif
11662
11663#ifdef __LITTLE_ENDIAN__
11664#define vld4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11665  int16x4x4_t __s1 = __p1; \
11666  int16x4x4_t __ret; \
11667  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 1); \
11668  __ret; \
11669})
11670#else
11671#define vld4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
11672  int16x4x4_t __s1 = __p1; \
11673  int16x4x4_t __rev1; \
11674  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
11675  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
11676  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
11677  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
11678  int16x4x4_t __ret; \
11679  __builtin_neon_vld4_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 1); \
11680 \
11681  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
11682  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
11683  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
11684  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
11685  __ret; \
11686})
11687#endif
11688
11689#ifdef __LITTLE_ENDIAN__
11690__ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11691  uint8x16_t __ret;
11692  __ret = (uint8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
11693  return __ret;
11694}
11695#else
11696__ai uint8x16_t vmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11697  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11698  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11699  uint8x16_t __ret;
11700  __ret = (uint8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
11701  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11702  return __ret;
11703}
11704#endif
11705
11706#ifdef __LITTLE_ENDIAN__
11707__ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11708  uint32x4_t __ret;
11709  __ret = (uint32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
11710  return __ret;
11711}
11712#else
11713__ai uint32x4_t vmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11714  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11715  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11716  uint32x4_t __ret;
11717  __ret = (uint32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
11718  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11719  return __ret;
11720}
11721#endif
11722
11723#ifdef __LITTLE_ENDIAN__
11724__ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11725  uint16x8_t __ret;
11726  __ret = (uint16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
11727  return __ret;
11728}
11729#else
11730__ai uint16x8_t vmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11731  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11732  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11733  uint16x8_t __ret;
11734  __ret = (uint16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
11735  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11736  return __ret;
11737}
11738#endif
11739
11740#ifdef __LITTLE_ENDIAN__
11741__ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
11742  int8x16_t __ret;
11743  __ret = (int8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
11744  return __ret;
11745}
11746#else
11747__ai int8x16_t vmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
11748  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11749  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11750  int8x16_t __ret;
11751  __ret = (int8x16_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
11752  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11753  return __ret;
11754}
11755#endif
11756
11757#ifdef __LITTLE_ENDIAN__
11758__ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
11759  float32x4_t __ret;
11760  __ret = (float32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
11761  return __ret;
11762}
11763#else
11764__ai float32x4_t vmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
11765  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11766  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11767  float32x4_t __ret;
11768  __ret = (float32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
11769  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11770  return __ret;
11771}
11772#endif
11773
11774#ifdef __LITTLE_ENDIAN__
11775__ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
11776  int32x4_t __ret;
11777  __ret = (int32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
11778  return __ret;
11779}
11780#else
11781__ai int32x4_t vmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
11782  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11783  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11784  int32x4_t __ret;
11785  __ret = (int32x4_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
11786  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11787  return __ret;
11788}
11789#endif
11790
11791#ifdef __LITTLE_ENDIAN__
11792__ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
11793  int16x8_t __ret;
11794  __ret = (int16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
11795  return __ret;
11796}
11797#else
11798__ai int16x8_t vmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
11799  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11800  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11801  int16x8_t __ret;
11802  __ret = (int16x8_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
11803  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11804  return __ret;
11805}
11806#endif
11807
11808#ifdef __LITTLE_ENDIAN__
11809__ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
11810  uint8x8_t __ret;
11811  __ret = (uint8x8_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
11812  return __ret;
11813}
11814#else
11815__ai uint8x8_t vmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
11816  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11817  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11818  uint8x8_t __ret;
11819  __ret = (uint8x8_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
11820  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11821  return __ret;
11822}
11823#endif
11824
11825#ifdef __LITTLE_ENDIAN__
11826__ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
11827  uint32x2_t __ret;
11828  __ret = (uint32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
11829  return __ret;
11830}
11831#else
11832__ai uint32x2_t vmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
11833  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
11834  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
11835  uint32x2_t __ret;
11836  __ret = (uint32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
11837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
11838  return __ret;
11839}
11840#endif
11841
11842#ifdef __LITTLE_ENDIAN__
11843__ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
11844  uint16x4_t __ret;
11845  __ret = (uint16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
11846  return __ret;
11847}
11848#else
11849__ai uint16x4_t vmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
11850  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11851  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11852  uint16x4_t __ret;
11853  __ret = (uint16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
11854  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11855  return __ret;
11856}
11857#endif
11858
11859#ifdef __LITTLE_ENDIAN__
11860__ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1) {
11861  int8x8_t __ret;
11862  __ret = (int8x8_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
11863  return __ret;
11864}
11865#else
11866__ai int8x8_t vmax_s8(int8x8_t __p0, int8x8_t __p1) {
11867  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11868  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11869  int8x8_t __ret;
11870  __ret = (int8x8_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
11871  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11872  return __ret;
11873}
11874#endif
11875
11876#ifdef __LITTLE_ENDIAN__
11877__ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1) {
11878  float32x2_t __ret;
11879  __ret = (float32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
11880  return __ret;
11881}
11882#else
11883__ai float32x2_t vmax_f32(float32x2_t __p0, float32x2_t __p1) {
11884  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
11885  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
11886  float32x2_t __ret;
11887  __ret = (float32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
11888  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
11889  return __ret;
11890}
11891#endif
11892
11893#ifdef __LITTLE_ENDIAN__
11894__ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1) {
11895  int32x2_t __ret;
11896  __ret = (int32x2_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
11897  return __ret;
11898}
11899#else
11900__ai int32x2_t vmax_s32(int32x2_t __p0, int32x2_t __p1) {
11901  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
11902  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
11903  int32x2_t __ret;
11904  __ret = (int32x2_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
11905  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
11906  return __ret;
11907}
11908#endif
11909
11910#ifdef __LITTLE_ENDIAN__
11911__ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1) {
11912  int16x4_t __ret;
11913  __ret = (int16x4_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
11914  return __ret;
11915}
11916#else
11917__ai int16x4_t vmax_s16(int16x4_t __p0, int16x4_t __p1) {
11918  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11919  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11920  int16x4_t __ret;
11921  __ret = (int16x4_t) __builtin_neon_vmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
11922  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11923  return __ret;
11924}
11925#endif
11926
11927#ifdef __LITTLE_ENDIAN__
11928__ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11929  uint8x16_t __ret;
11930  __ret = (uint8x16_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
11931  return __ret;
11932}
11933#else
11934__ai uint8x16_t vminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
11935  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11936  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11937  uint8x16_t __ret;
11938  __ret = (uint8x16_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
11939  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11940  return __ret;
11941}
11942#endif
11943
11944#ifdef __LITTLE_ENDIAN__
11945__ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11946  uint32x4_t __ret;
11947  __ret = (uint32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
11948  return __ret;
11949}
11950#else
11951__ai uint32x4_t vminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
11952  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
11953  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
11954  uint32x4_t __ret;
11955  __ret = (uint32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
11956  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
11957  return __ret;
11958}
11959#endif
11960
11961#ifdef __LITTLE_ENDIAN__
11962__ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11963  uint16x8_t __ret;
11964  __ret = (uint16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
11965  return __ret;
11966}
11967#else
11968__ai uint16x8_t vminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
11969  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
11970  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
11971  uint16x8_t __ret;
11972  __ret = (uint16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
11973  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
11974  return __ret;
11975}
11976#endif
11977
11978#ifdef __LITTLE_ENDIAN__
11979__ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1) {
11980  int8x16_t __ret;
11981  __ret = (int8x16_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
11982  return __ret;
11983}
11984#else
11985__ai int8x16_t vminq_s8(int8x16_t __p0, int8x16_t __p1) {
11986  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11987  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11988  int8x16_t __ret;
11989  __ret = (int8x16_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
11990  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
11991  return __ret;
11992}
11993#endif
11994
11995#ifdef __LITTLE_ENDIAN__
11996__ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1) {
11997  float32x4_t __ret;
11998  __ret = (float32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
11999  return __ret;
12000}
12001#else
12002__ai float32x4_t vminq_f32(float32x4_t __p0, float32x4_t __p1) {
12003  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12004  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12005  float32x4_t __ret;
12006  __ret = (float32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
12007  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12008  return __ret;
12009}
12010#endif
12011
12012#ifdef __LITTLE_ENDIAN__
12013__ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1) {
12014  int32x4_t __ret;
12015  __ret = (int32x4_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
12016  return __ret;
12017}
12018#else
12019__ai int32x4_t vminq_s32(int32x4_t __p0, int32x4_t __p1) {
12020  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12021  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12022  int32x4_t __ret;
12023  __ret = (int32x4_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
12024  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12025  return __ret;
12026}
12027#endif
12028
12029#ifdef __LITTLE_ENDIAN__
12030__ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1) {
12031  int16x8_t __ret;
12032  __ret = (int16x8_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
12033  return __ret;
12034}
12035#else
12036__ai int16x8_t vminq_s16(int16x8_t __p0, int16x8_t __p1) {
12037  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12038  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12039  int16x8_t __ret;
12040  __ret = (int16x8_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
12041  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12042  return __ret;
12043}
12044#endif
12045
12046#ifdef __LITTLE_ENDIAN__
12047__ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
12048  uint8x8_t __ret;
12049  __ret = (uint8x8_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
12050  return __ret;
12051}
12052#else
12053__ai uint8x8_t vmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
12054  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12055  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12056  uint8x8_t __ret;
12057  __ret = (uint8x8_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
12058  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12059  return __ret;
12060}
12061#endif
12062
12063#ifdef __LITTLE_ENDIAN__
12064__ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
12065  uint32x2_t __ret;
12066  __ret = (uint32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
12067  return __ret;
12068}
12069#else
12070__ai uint32x2_t vmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
12071  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12072  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12073  uint32x2_t __ret;
12074  __ret = (uint32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
12075  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12076  return __ret;
12077}
12078#endif
12079
12080#ifdef __LITTLE_ENDIAN__
12081__ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
12082  uint16x4_t __ret;
12083  __ret = (uint16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
12084  return __ret;
12085}
12086#else
12087__ai uint16x4_t vmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
12088  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12089  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12090  uint16x4_t __ret;
12091  __ret = (uint16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
12092  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12093  return __ret;
12094}
12095#endif
12096
12097#ifdef __LITTLE_ENDIAN__
12098__ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1) {
12099  int8x8_t __ret;
12100  __ret = (int8x8_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
12101  return __ret;
12102}
12103#else
12104__ai int8x8_t vmin_s8(int8x8_t __p0, int8x8_t __p1) {
12105  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12106  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12107  int8x8_t __ret;
12108  __ret = (int8x8_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
12109  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12110  return __ret;
12111}
12112#endif
12113
12114#ifdef __LITTLE_ENDIAN__
12115__ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1) {
12116  float32x2_t __ret;
12117  __ret = (float32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
12118  return __ret;
12119}
12120#else
12121__ai float32x2_t vmin_f32(float32x2_t __p0, float32x2_t __p1) {
12122  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12123  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12124  float32x2_t __ret;
12125  __ret = (float32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
12126  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12127  return __ret;
12128}
12129#endif
12130
12131#ifdef __LITTLE_ENDIAN__
12132__ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1) {
12133  int32x2_t __ret;
12134  __ret = (int32x2_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
12135  return __ret;
12136}
12137#else
12138__ai int32x2_t vmin_s32(int32x2_t __p0, int32x2_t __p1) {
12139  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12140  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12141  int32x2_t __ret;
12142  __ret = (int32x2_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
12143  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12144  return __ret;
12145}
12146#endif
12147
12148#ifdef __LITTLE_ENDIAN__
12149__ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1) {
12150  int16x4_t __ret;
12151  __ret = (int16x4_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
12152  return __ret;
12153}
12154#else
12155__ai int16x4_t vmin_s16(int16x4_t __p0, int16x4_t __p1) {
12156  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12157  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12158  int16x4_t __ret;
12159  __ret = (int16x4_t) __builtin_neon_vmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
12160  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12161  return __ret;
12162}
12163#endif
12164
12165#ifdef __LITTLE_ENDIAN__
12166__ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12167  uint8x16_t __ret;
12168  __ret = __p0 + __p1 * __p2;
12169  return __ret;
12170}
12171#else
12172__ai uint8x16_t vmlaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12173  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12174  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12175  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12176  uint8x16_t __ret;
12177  __ret = __rev0 + __rev1 * __rev2;
12178  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12179  return __ret;
12180}
12181#endif
12182
12183#ifdef __LITTLE_ENDIAN__
12184__ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12185  uint32x4_t __ret;
12186  __ret = __p0 + __p1 * __p2;
12187  return __ret;
12188}
12189#else
12190__ai uint32x4_t vmlaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12191  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12192  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12193  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12194  uint32x4_t __ret;
12195  __ret = __rev0 + __rev1 * __rev2;
12196  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12197  return __ret;
12198}
12199#endif
12200
12201#ifdef __LITTLE_ENDIAN__
12202__ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12203  uint16x8_t __ret;
12204  __ret = __p0 + __p1 * __p2;
12205  return __ret;
12206}
12207#else
12208__ai uint16x8_t vmlaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12209  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12210  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12211  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12212  uint16x8_t __ret;
12213  __ret = __rev0 + __rev1 * __rev2;
12214  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12215  return __ret;
12216}
12217#endif
12218
12219#ifdef __LITTLE_ENDIAN__
12220__ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12221  int8x16_t __ret;
12222  __ret = __p0 + __p1 * __p2;
12223  return __ret;
12224}
12225#else
12226__ai int8x16_t vmlaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12227  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12228  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12229  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12230  int8x16_t __ret;
12231  __ret = __rev0 + __rev1 * __rev2;
12232  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12233  return __ret;
12234}
12235#endif
12236
12237#ifdef __LITTLE_ENDIAN__
12238__ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12239  float32x4_t __ret;
12240  __ret = __p0 + __p1 * __p2;
12241  return __ret;
12242}
12243#else
12244__ai float32x4_t vmlaq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12245  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12246  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12247  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12248  float32x4_t __ret;
12249  __ret = __rev0 + __rev1 * __rev2;
12250  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12251  return __ret;
12252}
12253#endif
12254
12255#ifdef __LITTLE_ENDIAN__
12256__ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12257  int32x4_t __ret;
12258  __ret = __p0 + __p1 * __p2;
12259  return __ret;
12260}
12261#else
12262__ai int32x4_t vmlaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12263  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12264  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12265  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12266  int32x4_t __ret;
12267  __ret = __rev0 + __rev1 * __rev2;
12268  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12269  return __ret;
12270}
12271#endif
12272
12273#ifdef __LITTLE_ENDIAN__
12274__ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12275  int16x8_t __ret;
12276  __ret = __p0 + __p1 * __p2;
12277  return __ret;
12278}
12279#else
12280__ai int16x8_t vmlaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12281  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12282  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12283  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12284  int16x8_t __ret;
12285  __ret = __rev0 + __rev1 * __rev2;
12286  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12287  return __ret;
12288}
12289#endif
12290
12291#ifdef __LITTLE_ENDIAN__
12292__ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12293  uint8x8_t __ret;
12294  __ret = __p0 + __p1 * __p2;
12295  return __ret;
12296}
12297#else
12298__ai uint8x8_t vmla_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12299  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12300  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12301  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12302  uint8x8_t __ret;
12303  __ret = __rev0 + __rev1 * __rev2;
12304  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12305  return __ret;
12306}
12307#endif
12308
12309#ifdef __LITTLE_ENDIAN__
12310__ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12311  uint32x2_t __ret;
12312  __ret = __p0 + __p1 * __p2;
12313  return __ret;
12314}
12315#else
12316__ai uint32x2_t vmla_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12317  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12318  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12319  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12320  uint32x2_t __ret;
12321  __ret = __rev0 + __rev1 * __rev2;
12322  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12323  return __ret;
12324}
12325#endif
12326
12327#ifdef __LITTLE_ENDIAN__
12328__ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12329  uint16x4_t __ret;
12330  __ret = __p0 + __p1 * __p2;
12331  return __ret;
12332}
12333#else
12334__ai uint16x4_t vmla_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12335  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12336  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12337  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12338  uint16x4_t __ret;
12339  __ret = __rev0 + __rev1 * __rev2;
12340  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12341  return __ret;
12342}
12343#endif
12344
12345#ifdef __LITTLE_ENDIAN__
12346__ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
12347  int8x8_t __ret;
12348  __ret = __p0 + __p1 * __p2;
12349  return __ret;
12350}
12351#else
12352__ai int8x8_t vmla_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
12353  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12354  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12355  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12356  int8x8_t __ret;
12357  __ret = __rev0 + __rev1 * __rev2;
12358  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12359  return __ret;
12360}
12361#endif
12362
12363#ifdef __LITTLE_ENDIAN__
12364__ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
12365  float32x2_t __ret;
12366  __ret = __p0 + __p1 * __p2;
12367  return __ret;
12368}
12369#else
12370__ai float32x2_t vmla_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
12371  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12372  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12373  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12374  float32x2_t __ret;
12375  __ret = __rev0 + __rev1 * __rev2;
12376  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12377  return __ret;
12378}
12379#endif
12380
12381#ifdef __LITTLE_ENDIAN__
12382__ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
12383  int32x2_t __ret;
12384  __ret = __p0 + __p1 * __p2;
12385  return __ret;
12386}
12387#else
12388__ai int32x2_t vmla_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
12389  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12390  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12391  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12392  int32x2_t __ret;
12393  __ret = __rev0 + __rev1 * __rev2;
12394  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12395  return __ret;
12396}
12397#endif
12398
12399#ifdef __LITTLE_ENDIAN__
12400__ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
12401  int16x4_t __ret;
12402  __ret = __p0 + __p1 * __p2;
12403  return __ret;
12404}
12405#else
12406__ai int16x4_t vmla_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
12407  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12408  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12409  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12410  int16x4_t __ret;
12411  __ret = __rev0 + __rev1 * __rev2;
12412  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12413  return __ret;
12414}
12415#endif
12416
12417#ifdef __LITTLE_ENDIAN__
12418#define vmlaq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12419  uint32x4_t __s0 = __p0; \
12420  uint32x4_t __s1 = __p1; \
12421  uint32x2_t __s2 = __p2; \
12422  uint32x4_t __ret; \
12423  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12424  __ret; \
12425})
12426#else
12427#define vmlaq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12428  uint32x4_t __s0 = __p0; \
12429  uint32x4_t __s1 = __p1; \
12430  uint32x2_t __s2 = __p2; \
12431  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12432  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12433  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12434  uint32x4_t __ret; \
12435  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12436  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12437  __ret; \
12438})
12439#endif
12440
12441#ifdef __LITTLE_ENDIAN__
12442#define vmlaq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12443  uint16x8_t __s0 = __p0; \
12444  uint16x8_t __s1 = __p1; \
12445  uint16x4_t __s2 = __p2; \
12446  uint16x8_t __ret; \
12447  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12448  __ret; \
12449})
12450#else
12451#define vmlaq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12452  uint16x8_t __s0 = __p0; \
12453  uint16x8_t __s1 = __p1; \
12454  uint16x4_t __s2 = __p2; \
12455  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
12456  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
12457  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12458  uint16x8_t __ret; \
12459  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12460  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
12461  __ret; \
12462})
12463#endif
12464
12465#ifdef __LITTLE_ENDIAN__
12466#define vmlaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12467  float32x4_t __s0 = __p0; \
12468  float32x4_t __s1 = __p1; \
12469  float32x2_t __s2 = __p2; \
12470  float32x4_t __ret; \
12471  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12472  __ret; \
12473})
12474#else
12475#define vmlaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12476  float32x4_t __s0 = __p0; \
12477  float32x4_t __s1 = __p1; \
12478  float32x2_t __s2 = __p2; \
12479  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12480  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12481  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12482  float32x4_t __ret; \
12483  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12484  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12485  __ret; \
12486})
12487#endif
12488
12489#ifdef __LITTLE_ENDIAN__
12490#define vmlaq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12491  int32x4_t __s0 = __p0; \
12492  int32x4_t __s1 = __p1; \
12493  int32x2_t __s2 = __p2; \
12494  int32x4_t __ret; \
12495  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12496  __ret; \
12497})
12498#else
12499#define vmlaq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12500  int32x4_t __s0 = __p0; \
12501  int32x4_t __s1 = __p1; \
12502  int32x2_t __s2 = __p2; \
12503  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12504  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12505  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12506  int32x4_t __ret; \
12507  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12508  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12509  __ret; \
12510})
12511#endif
12512
12513#ifdef __LITTLE_ENDIAN__
12514#define vmlaq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12515  int16x8_t __s0 = __p0; \
12516  int16x8_t __s1 = __p1; \
12517  int16x4_t __s2 = __p2; \
12518  int16x8_t __ret; \
12519  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12520  __ret; \
12521})
12522#else
12523#define vmlaq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12524  int16x8_t __s0 = __p0; \
12525  int16x8_t __s1 = __p1; \
12526  int16x4_t __s2 = __p2; \
12527  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
12528  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
12529  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12530  int16x8_t __ret; \
12531  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
12532  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
12533  __ret; \
12534})
12535#endif
12536
12537#ifdef __LITTLE_ENDIAN__
12538#define vmla_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12539  uint32x2_t __s0 = __p0; \
12540  uint32x2_t __s1 = __p1; \
12541  uint32x2_t __s2 = __p2; \
12542  uint32x2_t __ret; \
12543  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
12544  __ret; \
12545})
12546#else
12547#define vmla_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
12548  uint32x2_t __s0 = __p0; \
12549  uint32x2_t __s1 = __p1; \
12550  uint32x2_t __s2 = __p2; \
12551  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
12552  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
12553  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12554  uint32x2_t __ret; \
12555  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
12556  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
12557  __ret; \
12558})
12559#endif
12560
12561#ifdef __LITTLE_ENDIAN__
12562#define vmla_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12563  uint16x4_t __s0 = __p0; \
12564  uint16x4_t __s1 = __p1; \
12565  uint16x4_t __s2 = __p2; \
12566  uint16x4_t __ret; \
12567  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12568  __ret; \
12569})
12570#else
12571#define vmla_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
12572  uint16x4_t __s0 = __p0; \
12573  uint16x4_t __s1 = __p1; \
12574  uint16x4_t __s2 = __p2; \
12575  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12576  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12577  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12578  uint16x4_t __ret; \
12579  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12580  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12581  __ret; \
12582})
12583#endif
12584
12585#ifdef __LITTLE_ENDIAN__
12586#define vmla_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12587  float32x2_t __s0 = __p0; \
12588  float32x2_t __s1 = __p1; \
12589  float32x2_t __s2 = __p2; \
12590  float32x2_t __ret; \
12591  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
12592  __ret; \
12593})
12594#else
12595#define vmla_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
12596  float32x2_t __s0 = __p0; \
12597  float32x2_t __s1 = __p1; \
12598  float32x2_t __s2 = __p2; \
12599  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
12600  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
12601  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12602  float32x2_t __ret; \
12603  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
12604  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
12605  __ret; \
12606})
12607#endif
12608
12609#ifdef __LITTLE_ENDIAN__
12610#define vmla_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12611  int32x2_t __s0 = __p0; \
12612  int32x2_t __s1 = __p1; \
12613  int32x2_t __s2 = __p2; \
12614  int32x2_t __ret; \
12615  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
12616  __ret; \
12617})
12618#else
12619#define vmla_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
12620  int32x2_t __s0 = __p0; \
12621  int32x2_t __s1 = __p1; \
12622  int32x2_t __s2 = __p2; \
12623  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
12624  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
12625  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
12626  int32x2_t __ret; \
12627  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
12628  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
12629  __ret; \
12630})
12631#endif
12632
12633#ifdef __LITTLE_ENDIAN__
12634#define vmla_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12635  int16x4_t __s0 = __p0; \
12636  int16x4_t __s1 = __p1; \
12637  int16x4_t __s2 = __p2; \
12638  int16x4_t __ret; \
12639  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
12640  __ret; \
12641})
12642#else
12643#define vmla_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
12644  int16x4_t __s0 = __p0; \
12645  int16x4_t __s1 = __p1; \
12646  int16x4_t __s2 = __p2; \
12647  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
12648  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
12649  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
12650  int16x4_t __ret; \
12651  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
12652  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
12653  __ret; \
12654})
12655#endif
12656
12657#ifdef __LITTLE_ENDIAN__
12658__ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
12659  uint32x4_t __ret;
12660  __ret = __p0 + __p1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
12661  return __ret;
12662}
12663#else
12664__ai uint32x4_t vmlaq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
12665  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12666  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12667  uint32x4_t __ret;
12668  __ret = __rev0 + __rev1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
12669  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12670  return __ret;
12671}
12672#endif
12673
12674#ifdef __LITTLE_ENDIAN__
12675__ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
12676  uint16x8_t __ret;
12677  __ret = __p0 + __p1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12678  return __ret;
12679}
12680#else
12681__ai uint16x8_t vmlaq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
12682  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12683  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12684  uint16x8_t __ret;
12685  __ret = __rev0 + __rev1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12686  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12687  return __ret;
12688}
12689#endif
12690
12691#ifdef __LITTLE_ENDIAN__
12692__ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
12693  float32x4_t __ret;
12694  __ret = __p0 + __p1 * (float32x4_t) {__p2, __p2, __p2, __p2};
12695  return __ret;
12696}
12697#else
12698__ai float32x4_t vmlaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
12699  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12700  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12701  float32x4_t __ret;
12702  __ret = __rev0 + __rev1 * (float32x4_t) {__p2, __p2, __p2, __p2};
12703  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12704  return __ret;
12705}
12706#endif
12707
12708#ifdef __LITTLE_ENDIAN__
12709__ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
12710  int32x4_t __ret;
12711  __ret = __p0 + __p1 * (int32x4_t) {__p2, __p2, __p2, __p2};
12712  return __ret;
12713}
12714#else
12715__ai int32x4_t vmlaq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
12716  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12717  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12718  int32x4_t __ret;
12719  __ret = __rev0 + __rev1 * (int32x4_t) {__p2, __p2, __p2, __p2};
12720  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12721  return __ret;
12722}
12723#endif
12724
12725#ifdef __LITTLE_ENDIAN__
12726__ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
12727  int16x8_t __ret;
12728  __ret = __p0 + __p1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12729  return __ret;
12730}
12731#else
12732__ai int16x8_t vmlaq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
12733  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12734  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12735  int16x8_t __ret;
12736  __ret = __rev0 + __rev1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
12737  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12738  return __ret;
12739}
12740#endif
12741
12742#ifdef __LITTLE_ENDIAN__
12743__ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
12744  uint32x2_t __ret;
12745  __ret = __p0 + __p1 * (uint32x2_t) {__p2, __p2};
12746  return __ret;
12747}
12748#else
12749__ai uint32x2_t vmla_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
12750  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12751  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12752  uint32x2_t __ret;
12753  __ret = __rev0 + __rev1 * (uint32x2_t) {__p2, __p2};
12754  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12755  return __ret;
12756}
12757#endif
12758
12759#ifdef __LITTLE_ENDIAN__
12760__ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
12761  uint16x4_t __ret;
12762  __ret = __p0 + __p1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
12763  return __ret;
12764}
12765#else
12766__ai uint16x4_t vmla_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
12767  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12768  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12769  uint16x4_t __ret;
12770  __ret = __rev0 + __rev1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
12771  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12772  return __ret;
12773}
12774#endif
12775
12776#ifdef __LITTLE_ENDIAN__
12777__ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
12778  float32x2_t __ret;
12779  __ret = __p0 + __p1 * (float32x2_t) {__p2, __p2};
12780  return __ret;
12781}
12782#else
12783__ai float32x2_t vmla_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
12784  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12785  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12786  float32x2_t __ret;
12787  __ret = __rev0 + __rev1 * (float32x2_t) {__p2, __p2};
12788  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12789  return __ret;
12790}
12791#endif
12792
12793#ifdef __LITTLE_ENDIAN__
12794__ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
12795  int32x2_t __ret;
12796  __ret = __p0 + __p1 * (int32x2_t) {__p2, __p2};
12797  return __ret;
12798}
12799#else
12800__ai int32x2_t vmla_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
12801  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12802  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12803  int32x2_t __ret;
12804  __ret = __rev0 + __rev1 * (int32x2_t) {__p2, __p2};
12805  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12806  return __ret;
12807}
12808#endif
12809
12810#ifdef __LITTLE_ENDIAN__
12811__ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
12812  int16x4_t __ret;
12813  __ret = __p0 + __p1 * (int16x4_t) {__p2, __p2, __p2, __p2};
12814  return __ret;
12815}
12816#else
12817__ai int16x4_t vmla_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
12818  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12819  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12820  int16x4_t __ret;
12821  __ret = __rev0 + __rev1 * (int16x4_t) {__p2, __p2, __p2, __p2};
12822  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12823  return __ret;
12824}
12825#endif
12826
12827#ifdef __LITTLE_ENDIAN__
12828__ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12829  uint8x16_t __ret;
12830  __ret = __p0 - __p1 * __p2;
12831  return __ret;
12832}
12833#else
12834__ai uint8x16_t vmlsq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
12835  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12836  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12837  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12838  uint8x16_t __ret;
12839  __ret = __rev0 - __rev1 * __rev2;
12840  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12841  return __ret;
12842}
12843#endif
12844
12845#ifdef __LITTLE_ENDIAN__
12846__ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12847  uint32x4_t __ret;
12848  __ret = __p0 - __p1 * __p2;
12849  return __ret;
12850}
12851#else
12852__ai uint32x4_t vmlsq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
12853  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12854  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12855  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12856  uint32x4_t __ret;
12857  __ret = __rev0 - __rev1 * __rev2;
12858  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12859  return __ret;
12860}
12861#endif
12862
12863#ifdef __LITTLE_ENDIAN__
12864__ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12865  uint16x8_t __ret;
12866  __ret = __p0 - __p1 * __p2;
12867  return __ret;
12868}
12869#else
12870__ai uint16x8_t vmlsq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
12871  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12872  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12873  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12874  uint16x8_t __ret;
12875  __ret = __rev0 - __rev1 * __rev2;
12876  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12877  return __ret;
12878}
12879#endif
12880
12881#ifdef __LITTLE_ENDIAN__
12882__ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12883  int8x16_t __ret;
12884  __ret = __p0 - __p1 * __p2;
12885  return __ret;
12886}
12887#else
12888__ai int8x16_t vmlsq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
12889  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12890  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12891  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12892  int8x16_t __ret;
12893  __ret = __rev0 - __rev1 * __rev2;
12894  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
12895  return __ret;
12896}
12897#endif
12898
12899#ifdef __LITTLE_ENDIAN__
12900__ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12901  float32x4_t __ret;
12902  __ret = __p0 - __p1 * __p2;
12903  return __ret;
12904}
12905#else
12906__ai float32x4_t vmlsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
12907  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12908  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12909  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12910  float32x4_t __ret;
12911  __ret = __rev0 - __rev1 * __rev2;
12912  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12913  return __ret;
12914}
12915#endif
12916
12917#ifdef __LITTLE_ENDIAN__
12918__ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12919  int32x4_t __ret;
12920  __ret = __p0 - __p1 * __p2;
12921  return __ret;
12922}
12923#else
12924__ai int32x4_t vmlsq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
12925  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12926  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12927  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
12928  int32x4_t __ret;
12929  __ret = __rev0 - __rev1 * __rev2;
12930  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
12931  return __ret;
12932}
12933#endif
12934
12935#ifdef __LITTLE_ENDIAN__
12936__ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12937  int16x8_t __ret;
12938  __ret = __p0 - __p1 * __p2;
12939  return __ret;
12940}
12941#else
12942__ai int16x8_t vmlsq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
12943  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12944  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12945  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12946  int16x8_t __ret;
12947  __ret = __rev0 - __rev1 * __rev2;
12948  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12949  return __ret;
12950}
12951#endif
12952
12953#ifdef __LITTLE_ENDIAN__
12954__ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12955  uint8x8_t __ret;
12956  __ret = __p0 - __p1 * __p2;
12957  return __ret;
12958}
12959#else
12960__ai uint8x8_t vmls_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
12961  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
12962  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
12963  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
12964  uint8x8_t __ret;
12965  __ret = __rev0 - __rev1 * __rev2;
12966  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
12967  return __ret;
12968}
12969#endif
12970
12971#ifdef __LITTLE_ENDIAN__
12972__ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12973  uint32x2_t __ret;
12974  __ret = __p0 - __p1 * __p2;
12975  return __ret;
12976}
12977#else
12978__ai uint32x2_t vmls_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
12979  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
12980  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
12981  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
12982  uint32x2_t __ret;
12983  __ret = __rev0 - __rev1 * __rev2;
12984  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
12985  return __ret;
12986}
12987#endif
12988
12989#ifdef __LITTLE_ENDIAN__
12990__ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12991  uint16x4_t __ret;
12992  __ret = __p0 - __p1 * __p2;
12993  return __ret;
12994}
12995#else
12996__ai uint16x4_t vmls_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
12997  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
12998  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
12999  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
13000  uint16x4_t __ret;
13001  __ret = __rev0 - __rev1 * __rev2;
13002  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13003  return __ret;
13004}
13005#endif
13006
13007#ifdef __LITTLE_ENDIAN__
13008__ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
13009  int8x8_t __ret;
13010  __ret = __p0 - __p1 * __p2;
13011  return __ret;
13012}
13013#else
13014__ai int8x8_t vmls_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
13015  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13016  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13017  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
13018  int8x8_t __ret;
13019  __ret = __rev0 - __rev1 * __rev2;
13020  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13021  return __ret;
13022}
13023#endif
13024
13025#ifdef __LITTLE_ENDIAN__
13026__ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
13027  float32x2_t __ret;
13028  __ret = __p0 - __p1 * __p2;
13029  return __ret;
13030}
13031#else
13032__ai float32x2_t vmls_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
13033  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13034  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13035  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
13036  float32x2_t __ret;
13037  __ret = __rev0 - __rev1 * __rev2;
13038  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13039  return __ret;
13040}
13041#endif
13042
13043#ifdef __LITTLE_ENDIAN__
13044__ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
13045  int32x2_t __ret;
13046  __ret = __p0 - __p1 * __p2;
13047  return __ret;
13048}
13049#else
13050__ai int32x2_t vmls_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
13051  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13052  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13053  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
13054  int32x2_t __ret;
13055  __ret = __rev0 - __rev1 * __rev2;
13056  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13057  return __ret;
13058}
13059#endif
13060
13061#ifdef __LITTLE_ENDIAN__
13062__ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
13063  int16x4_t __ret;
13064  __ret = __p0 - __p1 * __p2;
13065  return __ret;
13066}
13067#else
13068__ai int16x4_t vmls_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
13069  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13070  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13071  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
13072  int16x4_t __ret;
13073  __ret = __rev0 - __rev1 * __rev2;
13074  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13075  return __ret;
13076}
13077#endif
13078
13079#ifdef __LITTLE_ENDIAN__
13080#define vmlsq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13081  uint32x4_t __s0 = __p0; \
13082  uint32x4_t __s1 = __p1; \
13083  uint32x2_t __s2 = __p2; \
13084  uint32x4_t __ret; \
13085  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13086  __ret; \
13087})
13088#else
13089#define vmlsq_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13090  uint32x4_t __s0 = __p0; \
13091  uint32x4_t __s1 = __p1; \
13092  uint32x2_t __s2 = __p2; \
13093  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13094  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13095  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13096  uint32x4_t __ret; \
13097  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13098  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13099  __ret; \
13100})
13101#endif
13102
13103#ifdef __LITTLE_ENDIAN__
13104#define vmlsq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13105  uint16x8_t __s0 = __p0; \
13106  uint16x8_t __s1 = __p1; \
13107  uint16x4_t __s2 = __p2; \
13108  uint16x8_t __ret; \
13109  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13110  __ret; \
13111})
13112#else
13113#define vmlsq_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13114  uint16x8_t __s0 = __p0; \
13115  uint16x8_t __s1 = __p1; \
13116  uint16x4_t __s2 = __p2; \
13117  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
13118  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
13119  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13120  uint16x8_t __ret; \
13121  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13122  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
13123  __ret; \
13124})
13125#endif
13126
13127#ifdef __LITTLE_ENDIAN__
13128#define vmlsq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13129  float32x4_t __s0 = __p0; \
13130  float32x4_t __s1 = __p1; \
13131  float32x2_t __s2 = __p2; \
13132  float32x4_t __ret; \
13133  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13134  __ret; \
13135})
13136#else
13137#define vmlsq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13138  float32x4_t __s0 = __p0; \
13139  float32x4_t __s1 = __p1; \
13140  float32x2_t __s2 = __p2; \
13141  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13142  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13143  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13144  float32x4_t __ret; \
13145  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13146  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13147  __ret; \
13148})
13149#endif
13150
13151#ifdef __LITTLE_ENDIAN__
13152#define vmlsq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13153  int32x4_t __s0 = __p0; \
13154  int32x4_t __s1 = __p1; \
13155  int32x2_t __s2 = __p2; \
13156  int32x4_t __ret; \
13157  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13158  __ret; \
13159})
13160#else
13161#define vmlsq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13162  int32x4_t __s0 = __p0; \
13163  int32x4_t __s1 = __p1; \
13164  int32x2_t __s2 = __p2; \
13165  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13166  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13167  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13168  int32x4_t __ret; \
13169  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13170  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13171  __ret; \
13172})
13173#endif
13174
13175#ifdef __LITTLE_ENDIAN__
13176#define vmlsq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13177  int16x8_t __s0 = __p0; \
13178  int16x8_t __s1 = __p1; \
13179  int16x4_t __s2 = __p2; \
13180  int16x8_t __ret; \
13181  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13182  __ret; \
13183})
13184#else
13185#define vmlsq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13186  int16x8_t __s0 = __p0; \
13187  int16x8_t __s1 = __p1; \
13188  int16x4_t __s2 = __p2; \
13189  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
13190  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
13191  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13192  int16x8_t __ret; \
13193  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
13194  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
13195  __ret; \
13196})
13197#endif
13198
13199#ifdef __LITTLE_ENDIAN__
13200#define vmls_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13201  uint32x2_t __s0 = __p0; \
13202  uint32x2_t __s1 = __p1; \
13203  uint32x2_t __s2 = __p2; \
13204  uint32x2_t __ret; \
13205  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
13206  __ret; \
13207})
13208#else
13209#define vmls_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
13210  uint32x2_t __s0 = __p0; \
13211  uint32x2_t __s1 = __p1; \
13212  uint32x2_t __s2 = __p2; \
13213  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
13214  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
13215  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13216  uint32x2_t __ret; \
13217  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
13218  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
13219  __ret; \
13220})
13221#endif
13222
13223#ifdef __LITTLE_ENDIAN__
13224#define vmls_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13225  uint16x4_t __s0 = __p0; \
13226  uint16x4_t __s1 = __p1; \
13227  uint16x4_t __s2 = __p2; \
13228  uint16x4_t __ret; \
13229  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13230  __ret; \
13231})
13232#else
13233#define vmls_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
13234  uint16x4_t __s0 = __p0; \
13235  uint16x4_t __s1 = __p1; \
13236  uint16x4_t __s2 = __p2; \
13237  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13238  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13239  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13240  uint16x4_t __ret; \
13241  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13242  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13243  __ret; \
13244})
13245#endif
13246
13247#ifdef __LITTLE_ENDIAN__
13248#define vmls_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13249  float32x2_t __s0 = __p0; \
13250  float32x2_t __s1 = __p1; \
13251  float32x2_t __s2 = __p2; \
13252  float32x2_t __ret; \
13253  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
13254  __ret; \
13255})
13256#else
13257#define vmls_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
13258  float32x2_t __s0 = __p0; \
13259  float32x2_t __s1 = __p1; \
13260  float32x2_t __s2 = __p2; \
13261  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
13262  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
13263  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13264  float32x2_t __ret; \
13265  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
13266  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
13267  __ret; \
13268})
13269#endif
13270
13271#ifdef __LITTLE_ENDIAN__
13272#define vmls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13273  int32x2_t __s0 = __p0; \
13274  int32x2_t __s1 = __p1; \
13275  int32x2_t __s2 = __p2; \
13276  int32x2_t __ret; \
13277  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
13278  __ret; \
13279})
13280#else
13281#define vmls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
13282  int32x2_t __s0 = __p0; \
13283  int32x2_t __s1 = __p1; \
13284  int32x2_t __s2 = __p2; \
13285  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
13286  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
13287  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
13288  int32x2_t __ret; \
13289  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
13290  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
13291  __ret; \
13292})
13293#endif
13294
13295#ifdef __LITTLE_ENDIAN__
13296#define vmls_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13297  int16x4_t __s0 = __p0; \
13298  int16x4_t __s1 = __p1; \
13299  int16x4_t __s2 = __p2; \
13300  int16x4_t __ret; \
13301  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
13302  __ret; \
13303})
13304#else
13305#define vmls_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
13306  int16x4_t __s0 = __p0; \
13307  int16x4_t __s1 = __p1; \
13308  int16x4_t __s2 = __p2; \
13309  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
13310  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
13311  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
13312  int16x4_t __ret; \
13313  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
13314  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13315  __ret; \
13316})
13317#endif
13318
13319#ifdef __LITTLE_ENDIAN__
13320__ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
13321  uint32x4_t __ret;
13322  __ret = __p0 - __p1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
13323  return __ret;
13324}
13325#else
13326__ai uint32x4_t vmlsq_n_u32(uint32x4_t __p0, uint32x4_t __p1, uint32_t __p2) {
13327  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13328  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13329  uint32x4_t __ret;
13330  __ret = __rev0 - __rev1 * (uint32x4_t) {__p2, __p2, __p2, __p2};
13331  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13332  return __ret;
13333}
13334#endif
13335
13336#ifdef __LITTLE_ENDIAN__
13337__ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
13338  uint16x8_t __ret;
13339  __ret = __p0 - __p1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13340  return __ret;
13341}
13342#else
13343__ai uint16x8_t vmlsq_n_u16(uint16x8_t __p0, uint16x8_t __p1, uint16_t __p2) {
13344  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13345  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13346  uint16x8_t __ret;
13347  __ret = __rev0 - __rev1 * (uint16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13348  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13349  return __ret;
13350}
13351#endif
13352
13353#ifdef __LITTLE_ENDIAN__
13354__ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
13355  float32x4_t __ret;
13356  __ret = __p0 - __p1 * (float32x4_t) {__p2, __p2, __p2, __p2};
13357  return __ret;
13358}
13359#else
13360__ai float32x4_t vmlsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
13361  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13362  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13363  float32x4_t __ret;
13364  __ret = __rev0 - __rev1 * (float32x4_t) {__p2, __p2, __p2, __p2};
13365  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13366  return __ret;
13367}
13368#endif
13369
13370#ifdef __LITTLE_ENDIAN__
13371__ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
13372  int32x4_t __ret;
13373  __ret = __p0 - __p1 * (int32x4_t) {__p2, __p2, __p2, __p2};
13374  return __ret;
13375}
13376#else
13377__ai int32x4_t vmlsq_n_s32(int32x4_t __p0, int32x4_t __p1, int32_t __p2) {
13378  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13379  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13380  int32x4_t __ret;
13381  __ret = __rev0 - __rev1 * (int32x4_t) {__p2, __p2, __p2, __p2};
13382  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13383  return __ret;
13384}
13385#endif
13386
13387#ifdef __LITTLE_ENDIAN__
13388__ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
13389  int16x8_t __ret;
13390  __ret = __p0 - __p1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13391  return __ret;
13392}
13393#else
13394__ai int16x8_t vmlsq_n_s16(int16x8_t __p0, int16x8_t __p1, int16_t __p2) {
13395  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13396  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
13397  int16x8_t __ret;
13398  __ret = __rev0 - __rev1 * (int16x8_t) {__p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2};
13399  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13400  return __ret;
13401}
13402#endif
13403
13404#ifdef __LITTLE_ENDIAN__
13405__ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
13406  uint32x2_t __ret;
13407  __ret = __p0 - __p1 * (uint32x2_t) {__p2, __p2};
13408  return __ret;
13409}
13410#else
13411__ai uint32x2_t vmls_n_u32(uint32x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
13412  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13413  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13414  uint32x2_t __ret;
13415  __ret = __rev0 - __rev1 * (uint32x2_t) {__p2, __p2};
13416  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13417  return __ret;
13418}
13419#endif
13420
13421#ifdef __LITTLE_ENDIAN__
13422__ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
13423  uint16x4_t __ret;
13424  __ret = __p0 - __p1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
13425  return __ret;
13426}
13427#else
13428__ai uint16x4_t vmls_n_u16(uint16x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
13429  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13430  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13431  uint16x4_t __ret;
13432  __ret = __rev0 - __rev1 * (uint16x4_t) {__p2, __p2, __p2, __p2};
13433  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13434  return __ret;
13435}
13436#endif
13437
13438#ifdef __LITTLE_ENDIAN__
13439__ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
13440  float32x2_t __ret;
13441  __ret = __p0 - __p1 * (float32x2_t) {__p2, __p2};
13442  return __ret;
13443}
13444#else
13445__ai float32x2_t vmls_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
13446  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13447  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13448  float32x2_t __ret;
13449  __ret = __rev0 - __rev1 * (float32x2_t) {__p2, __p2};
13450  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13451  return __ret;
13452}
13453#endif
13454
13455#ifdef __LITTLE_ENDIAN__
13456__ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
13457  int32x2_t __ret;
13458  __ret = __p0 - __p1 * (int32x2_t) {__p2, __p2};
13459  return __ret;
13460}
13461#else
13462__ai int32x2_t vmls_n_s32(int32x2_t __p0, int32x2_t __p1, int32_t __p2) {
13463  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13464  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
13465  int32x2_t __ret;
13466  __ret = __rev0 - __rev1 * (int32x2_t) {__p2, __p2};
13467  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13468  return __ret;
13469}
13470#endif
13471
13472#ifdef __LITTLE_ENDIAN__
13473__ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
13474  int16x4_t __ret;
13475  __ret = __p0 - __p1 * (int16x4_t) {__p2, __p2, __p2, __p2};
13476  return __ret;
13477}
13478#else
13479__ai int16x4_t vmls_n_s16(int16x4_t __p0, int16x4_t __p1, int16_t __p2) {
13480  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13481  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
13482  int16x4_t __ret;
13483  __ret = __rev0 - __rev1 * (int16x4_t) {__p2, __p2, __p2, __p2};
13484  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13485  return __ret;
13486}
13487#endif
13488
13489#ifdef __LITTLE_ENDIAN__
13490__ai poly8x8_t vmov_n_p8(poly8_t __p0) {
13491  poly8x8_t __ret;
13492  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13493  return __ret;
13494}
13495#else
13496__ai poly8x8_t vmov_n_p8(poly8_t __p0) {
13497  poly8x8_t __ret;
13498  __ret = (poly8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13499  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13500  return __ret;
13501}
13502#endif
13503
13504#ifdef __LITTLE_ENDIAN__
13505__ai poly16x4_t vmov_n_p16(poly16_t __p0) {
13506  poly16x4_t __ret;
13507  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
13508  return __ret;
13509}
13510#else
13511__ai poly16x4_t vmov_n_p16(poly16_t __p0) {
13512  poly16x4_t __ret;
13513  __ret = (poly16x4_t) {__p0, __p0, __p0, __p0};
13514  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13515  return __ret;
13516}
13517#endif
13518
13519#ifdef __LITTLE_ENDIAN__
13520__ai poly8x16_t vmovq_n_p8(poly8_t __p0) {
13521  poly8x16_t __ret;
13522  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13523  return __ret;
13524}
13525#else
13526__ai poly8x16_t vmovq_n_p8(poly8_t __p0) {
13527  poly8x16_t __ret;
13528  __ret = (poly8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13529  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13530  return __ret;
13531}
13532#endif
13533
13534#ifdef __LITTLE_ENDIAN__
13535__ai poly16x8_t vmovq_n_p16(poly16_t __p0) {
13536  poly16x8_t __ret;
13537  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13538  return __ret;
13539}
13540#else
13541__ai poly16x8_t vmovq_n_p16(poly16_t __p0) {
13542  poly16x8_t __ret;
13543  __ret = (poly16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13544  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13545  return __ret;
13546}
13547#endif
13548
13549#ifdef __LITTLE_ENDIAN__
13550__ai uint8x16_t vmovq_n_u8(uint8_t __p0) {
13551  uint8x16_t __ret;
13552  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13553  return __ret;
13554}
13555#else
13556__ai uint8x16_t vmovq_n_u8(uint8_t __p0) {
13557  uint8x16_t __ret;
13558  __ret = (uint8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13559  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13560  return __ret;
13561}
13562#endif
13563
13564#ifdef __LITTLE_ENDIAN__
13565__ai uint32x4_t vmovq_n_u32(uint32_t __p0) {
13566  uint32x4_t __ret;
13567  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
13568  return __ret;
13569}
13570#else
13571__ai uint32x4_t vmovq_n_u32(uint32_t __p0) {
13572  uint32x4_t __ret;
13573  __ret = (uint32x4_t) {__p0, __p0, __p0, __p0};
13574  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13575  return __ret;
13576}
13577#endif
13578
13579#ifdef __LITTLE_ENDIAN__
13580__ai uint64x2_t vmovq_n_u64(uint64_t __p0) {
13581  uint64x2_t __ret;
13582  __ret = (uint64x2_t) {__p0, __p0};
13583  return __ret;
13584}
13585#else
13586__ai uint64x2_t vmovq_n_u64(uint64_t __p0) {
13587  uint64x2_t __ret;
13588  __ret = (uint64x2_t) {__p0, __p0};
13589  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13590  return __ret;
13591}
13592#endif
13593
13594#ifdef __LITTLE_ENDIAN__
13595__ai uint16x8_t vmovq_n_u16(uint16_t __p0) {
13596  uint16x8_t __ret;
13597  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13598  return __ret;
13599}
13600#else
13601__ai uint16x8_t vmovq_n_u16(uint16_t __p0) {
13602  uint16x8_t __ret;
13603  __ret = (uint16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13604  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13605  return __ret;
13606}
13607#endif
13608
13609#ifdef __LITTLE_ENDIAN__
13610__ai int8x16_t vmovq_n_s8(int8_t __p0) {
13611  int8x16_t __ret;
13612  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13613  return __ret;
13614}
13615#else
13616__ai int8x16_t vmovq_n_s8(int8_t __p0) {
13617  int8x16_t __ret;
13618  __ret = (int8x16_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13619  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
13620  return __ret;
13621}
13622#endif
13623
13624#ifdef __LITTLE_ENDIAN__
13625__ai float32x4_t vmovq_n_f32(float32_t __p0) {
13626  float32x4_t __ret;
13627  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
13628  return __ret;
13629}
13630#else
13631__ai float32x4_t vmovq_n_f32(float32_t __p0) {
13632  float32x4_t __ret;
13633  __ret = (float32x4_t) {__p0, __p0, __p0, __p0};
13634  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13635  return __ret;
13636}
13637#endif
13638
13639#ifdef __LITTLE_ENDIAN__
13640#define vmovq_n_f16(__p0) __extension__ ({ \
13641  float16_t __s0 = __p0; \
13642  float16x8_t __ret; \
13643  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
13644  __ret; \
13645})
13646#else
13647#define vmovq_n_f16(__p0) __extension__ ({ \
13648  float16_t __s0 = __p0; \
13649  float16x8_t __ret; \
13650  __ret = (float16x8_t) {__s0, __s0, __s0, __s0, __s0, __s0, __s0, __s0}; \
13651  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
13652  __ret; \
13653})
13654#endif
13655
13656#ifdef __LITTLE_ENDIAN__
13657__ai int32x4_t vmovq_n_s32(int32_t __p0) {
13658  int32x4_t __ret;
13659  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
13660  return __ret;
13661}
13662#else
13663__ai int32x4_t vmovq_n_s32(int32_t __p0) {
13664  int32x4_t __ret;
13665  __ret = (int32x4_t) {__p0, __p0, __p0, __p0};
13666  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13667  return __ret;
13668}
13669#endif
13670
13671#ifdef __LITTLE_ENDIAN__
13672__ai int64x2_t vmovq_n_s64(int64_t __p0) {
13673  int64x2_t __ret;
13674  __ret = (int64x2_t) {__p0, __p0};
13675  return __ret;
13676}
13677#else
13678__ai int64x2_t vmovq_n_s64(int64_t __p0) {
13679  int64x2_t __ret;
13680  __ret = (int64x2_t) {__p0, __p0};
13681  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13682  return __ret;
13683}
13684#endif
13685
13686#ifdef __LITTLE_ENDIAN__
13687__ai int16x8_t vmovq_n_s16(int16_t __p0) {
13688  int16x8_t __ret;
13689  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13690  return __ret;
13691}
13692#else
13693__ai int16x8_t vmovq_n_s16(int16_t __p0) {
13694  int16x8_t __ret;
13695  __ret = (int16x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13696  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13697  return __ret;
13698}
13699#endif
13700
13701#ifdef __LITTLE_ENDIAN__
13702__ai uint8x8_t vmov_n_u8(uint8_t __p0) {
13703  uint8x8_t __ret;
13704  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13705  return __ret;
13706}
13707#else
13708__ai uint8x8_t vmov_n_u8(uint8_t __p0) {
13709  uint8x8_t __ret;
13710  __ret = (uint8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13711  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13712  return __ret;
13713}
13714#endif
13715
13716#ifdef __LITTLE_ENDIAN__
13717__ai uint32x2_t vmov_n_u32(uint32_t __p0) {
13718  uint32x2_t __ret;
13719  __ret = (uint32x2_t) {__p0, __p0};
13720  return __ret;
13721}
13722#else
13723__ai uint32x2_t vmov_n_u32(uint32_t __p0) {
13724  uint32x2_t __ret;
13725  __ret = (uint32x2_t) {__p0, __p0};
13726  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13727  return __ret;
13728}
13729#endif
13730
13731#ifdef __LITTLE_ENDIAN__
13732__ai uint64x1_t vmov_n_u64(uint64_t __p0) {
13733  uint64x1_t __ret;
13734  __ret = (uint64x1_t) {__p0};
13735  return __ret;
13736}
13737#else
13738__ai uint64x1_t vmov_n_u64(uint64_t __p0) {
13739  uint64x1_t __ret;
13740  __ret = (uint64x1_t) {__p0};
13741  return __ret;
13742}
13743#endif
13744
13745#ifdef __LITTLE_ENDIAN__
13746__ai uint16x4_t vmov_n_u16(uint16_t __p0) {
13747  uint16x4_t __ret;
13748  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
13749  return __ret;
13750}
13751#else
13752__ai uint16x4_t vmov_n_u16(uint16_t __p0) {
13753  uint16x4_t __ret;
13754  __ret = (uint16x4_t) {__p0, __p0, __p0, __p0};
13755  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13756  return __ret;
13757}
13758#endif
13759
13760#ifdef __LITTLE_ENDIAN__
13761__ai int8x8_t vmov_n_s8(int8_t __p0) {
13762  int8x8_t __ret;
13763  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13764  return __ret;
13765}
13766#else
13767__ai int8x8_t vmov_n_s8(int8_t __p0) {
13768  int8x8_t __ret;
13769  __ret = (int8x8_t) {__p0, __p0, __p0, __p0, __p0, __p0, __p0, __p0};
13770  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13771  return __ret;
13772}
13773#endif
13774
13775#ifdef __LITTLE_ENDIAN__
13776__ai float32x2_t vmov_n_f32(float32_t __p0) {
13777  float32x2_t __ret;
13778  __ret = (float32x2_t) {__p0, __p0};
13779  return __ret;
13780}
13781#else
13782__ai float32x2_t vmov_n_f32(float32_t __p0) {
13783  float32x2_t __ret;
13784  __ret = (float32x2_t) {__p0, __p0};
13785  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13786  return __ret;
13787}
13788#endif
13789
13790#ifdef __LITTLE_ENDIAN__
13791#define vmov_n_f16(__p0) __extension__ ({ \
13792  float16_t __s0 = __p0; \
13793  float16x4_t __ret; \
13794  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
13795  __ret; \
13796})
13797#else
13798#define vmov_n_f16(__p0) __extension__ ({ \
13799  float16_t __s0 = __p0; \
13800  float16x4_t __ret; \
13801  __ret = (float16x4_t) {__s0, __s0, __s0, __s0}; \
13802  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
13803  __ret; \
13804})
13805#endif
13806
13807#ifdef __LITTLE_ENDIAN__
13808__ai int32x2_t vmov_n_s32(int32_t __p0) {
13809  int32x2_t __ret;
13810  __ret = (int32x2_t) {__p0, __p0};
13811  return __ret;
13812}
13813#else
13814__ai int32x2_t vmov_n_s32(int32_t __p0) {
13815  int32x2_t __ret;
13816  __ret = (int32x2_t) {__p0, __p0};
13817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13818  return __ret;
13819}
13820#endif
13821
13822#ifdef __LITTLE_ENDIAN__
13823__ai int64x1_t vmov_n_s64(int64_t __p0) {
13824  int64x1_t __ret;
13825  __ret = (int64x1_t) {__p0};
13826  return __ret;
13827}
13828#else
13829__ai int64x1_t vmov_n_s64(int64_t __p0) {
13830  int64x1_t __ret;
13831  __ret = (int64x1_t) {__p0};
13832  return __ret;
13833}
13834#endif
13835
13836#ifdef __LITTLE_ENDIAN__
13837__ai int16x4_t vmov_n_s16(int16_t __p0) {
13838  int16x4_t __ret;
13839  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
13840  return __ret;
13841}
13842#else
13843__ai int16x4_t vmov_n_s16(int16_t __p0) {
13844  int16x4_t __ret;
13845  __ret = (int16x4_t) {__p0, __p0, __p0, __p0};
13846  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13847  return __ret;
13848}
13849#endif
13850
13851#ifdef __LITTLE_ENDIAN__
13852__ai uint16x8_t vmovl_u8(uint8x8_t __p0) {
13853  uint16x8_t __ret;
13854  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 49);
13855  return __ret;
13856}
13857#else
13858__ai uint16x8_t vmovl_u8(uint8x8_t __p0) {
13859  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13860  uint16x8_t __ret;
13861  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 49);
13862  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13863  return __ret;
13864}
13865__ai uint16x8_t __noswap_vmovl_u8(uint8x8_t __p0) {
13866  uint16x8_t __ret;
13867  __ret = (uint16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 49);
13868  return __ret;
13869}
13870#endif
13871
13872#ifdef __LITTLE_ENDIAN__
13873__ai uint64x2_t vmovl_u32(uint32x2_t __p0) {
13874  uint64x2_t __ret;
13875  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 51);
13876  return __ret;
13877}
13878#else
13879__ai uint64x2_t vmovl_u32(uint32x2_t __p0) {
13880  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13881  uint64x2_t __ret;
13882  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 51);
13883  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13884  return __ret;
13885}
13886__ai uint64x2_t __noswap_vmovl_u32(uint32x2_t __p0) {
13887  uint64x2_t __ret;
13888  __ret = (uint64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 51);
13889  return __ret;
13890}
13891#endif
13892
13893#ifdef __LITTLE_ENDIAN__
13894__ai uint32x4_t vmovl_u16(uint16x4_t __p0) {
13895  uint32x4_t __ret;
13896  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 50);
13897  return __ret;
13898}
13899#else
13900__ai uint32x4_t vmovl_u16(uint16x4_t __p0) {
13901  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13902  uint32x4_t __ret;
13903  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 50);
13904  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13905  return __ret;
13906}
13907__ai uint32x4_t __noswap_vmovl_u16(uint16x4_t __p0) {
13908  uint32x4_t __ret;
13909  __ret = (uint32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 50);
13910  return __ret;
13911}
13912#endif
13913
13914#ifdef __LITTLE_ENDIAN__
13915__ai int16x8_t vmovl_s8(int8x8_t __p0) {
13916  int16x8_t __ret;
13917  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 33);
13918  return __ret;
13919}
13920#else
13921__ai int16x8_t vmovl_s8(int8x8_t __p0) {
13922  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
13923  int16x8_t __ret;
13924  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 33);
13925  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
13926  return __ret;
13927}
13928__ai int16x8_t __noswap_vmovl_s8(int8x8_t __p0) {
13929  int16x8_t __ret;
13930  __ret = (int16x8_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 33);
13931  return __ret;
13932}
13933#endif
13934
13935#ifdef __LITTLE_ENDIAN__
13936__ai int64x2_t vmovl_s32(int32x2_t __p0) {
13937  int64x2_t __ret;
13938  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 35);
13939  return __ret;
13940}
13941#else
13942__ai int64x2_t vmovl_s32(int32x2_t __p0) {
13943  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
13944  int64x2_t __ret;
13945  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 35);
13946  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
13947  return __ret;
13948}
13949__ai int64x2_t __noswap_vmovl_s32(int32x2_t __p0) {
13950  int64x2_t __ret;
13951  __ret = (int64x2_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 35);
13952  return __ret;
13953}
13954#endif
13955
13956#ifdef __LITTLE_ENDIAN__
13957__ai int32x4_t vmovl_s16(int16x4_t __p0) {
13958  int32x4_t __ret;
13959  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 34);
13960  return __ret;
13961}
13962#else
13963__ai int32x4_t vmovl_s16(int16x4_t __p0) {
13964  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13965  int32x4_t __ret;
13966  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__rev0, 34);
13967  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13968  return __ret;
13969}
13970__ai int32x4_t __noswap_vmovl_s16(int16x4_t __p0) {
13971  int32x4_t __ret;
13972  __ret = (int32x4_t) __builtin_neon_vmovl_v((int8x8_t)__p0, 34);
13973  return __ret;
13974}
13975#endif
13976
13977#ifdef __LITTLE_ENDIAN__
13978__ai uint16x4_t vmovn_u32(uint32x4_t __p0) {
13979  uint16x4_t __ret;
13980  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 17);
13981  return __ret;
13982}
13983#else
13984__ai uint16x4_t vmovn_u32(uint32x4_t __p0) {
13985  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
13986  uint16x4_t __ret;
13987  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 17);
13988  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
13989  return __ret;
13990}
13991__ai uint16x4_t __noswap_vmovn_u32(uint32x4_t __p0) {
13992  uint16x4_t __ret;
13993  __ret = (uint16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 17);
13994  return __ret;
13995}
13996#endif
13997
13998#ifdef __LITTLE_ENDIAN__
13999__ai uint32x2_t vmovn_u64(uint64x2_t __p0) {
14000  uint32x2_t __ret;
14001  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 18);
14002  return __ret;
14003}
14004#else
14005__ai uint32x2_t vmovn_u64(uint64x2_t __p0) {
14006  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14007  uint32x2_t __ret;
14008  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 18);
14009  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14010  return __ret;
14011}
14012__ai uint32x2_t __noswap_vmovn_u64(uint64x2_t __p0) {
14013  uint32x2_t __ret;
14014  __ret = (uint32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 18);
14015  return __ret;
14016}
14017#endif
14018
14019#ifdef __LITTLE_ENDIAN__
14020__ai uint8x8_t vmovn_u16(uint16x8_t __p0) {
14021  uint8x8_t __ret;
14022  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 16);
14023  return __ret;
14024}
14025#else
14026__ai uint8x8_t vmovn_u16(uint16x8_t __p0) {
14027  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14028  uint8x8_t __ret;
14029  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 16);
14030  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14031  return __ret;
14032}
14033__ai uint8x8_t __noswap_vmovn_u16(uint16x8_t __p0) {
14034  uint8x8_t __ret;
14035  __ret = (uint8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 16);
14036  return __ret;
14037}
14038#endif
14039
14040#ifdef __LITTLE_ENDIAN__
14041__ai int16x4_t vmovn_s32(int32x4_t __p0) {
14042  int16x4_t __ret;
14043  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 1);
14044  return __ret;
14045}
14046#else
14047__ai int16x4_t vmovn_s32(int32x4_t __p0) {
14048  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14049  int16x4_t __ret;
14050  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 1);
14051  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14052  return __ret;
14053}
14054__ai int16x4_t __noswap_vmovn_s32(int32x4_t __p0) {
14055  int16x4_t __ret;
14056  __ret = (int16x4_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 1);
14057  return __ret;
14058}
14059#endif
14060
14061#ifdef __LITTLE_ENDIAN__
14062__ai int32x2_t vmovn_s64(int64x2_t __p0) {
14063  int32x2_t __ret;
14064  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 2);
14065  return __ret;
14066}
14067#else
14068__ai int32x2_t vmovn_s64(int64x2_t __p0) {
14069  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14070  int32x2_t __ret;
14071  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 2);
14072  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14073  return __ret;
14074}
14075__ai int32x2_t __noswap_vmovn_s64(int64x2_t __p0) {
14076  int32x2_t __ret;
14077  __ret = (int32x2_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 2);
14078  return __ret;
14079}
14080#endif
14081
14082#ifdef __LITTLE_ENDIAN__
14083__ai int8x8_t vmovn_s16(int16x8_t __p0) {
14084  int8x8_t __ret;
14085  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 0);
14086  return __ret;
14087}
14088#else
14089__ai int8x8_t vmovn_s16(int16x8_t __p0) {
14090  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14091  int8x8_t __ret;
14092  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__rev0, 0);
14093  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14094  return __ret;
14095}
14096__ai int8x8_t __noswap_vmovn_s16(int16x8_t __p0) {
14097  int8x8_t __ret;
14098  __ret = (int8x8_t) __builtin_neon_vmovn_v((int8x16_t)__p0, 0);
14099  return __ret;
14100}
14101#endif
14102
14103#ifdef __LITTLE_ENDIAN__
14104__ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1) {
14105  uint8x16_t __ret;
14106  __ret = __p0 * __p1;
14107  return __ret;
14108}
14109#else
14110__ai uint8x16_t vmulq_u8(uint8x16_t __p0, uint8x16_t __p1) {
14111  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14112  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14113  uint8x16_t __ret;
14114  __ret = __rev0 * __rev1;
14115  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14116  return __ret;
14117}
14118#endif
14119
14120#ifdef __LITTLE_ENDIAN__
14121__ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1) {
14122  uint32x4_t __ret;
14123  __ret = __p0 * __p1;
14124  return __ret;
14125}
14126#else
14127__ai uint32x4_t vmulq_u32(uint32x4_t __p0, uint32x4_t __p1) {
14128  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14129  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14130  uint32x4_t __ret;
14131  __ret = __rev0 * __rev1;
14132  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14133  return __ret;
14134}
14135#endif
14136
14137#ifdef __LITTLE_ENDIAN__
14138__ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1) {
14139  uint16x8_t __ret;
14140  __ret = __p0 * __p1;
14141  return __ret;
14142}
14143#else
14144__ai uint16x8_t vmulq_u16(uint16x8_t __p0, uint16x8_t __p1) {
14145  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14146  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14147  uint16x8_t __ret;
14148  __ret = __rev0 * __rev1;
14149  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14150  return __ret;
14151}
14152#endif
14153
14154#ifdef __LITTLE_ENDIAN__
14155__ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1) {
14156  int8x16_t __ret;
14157  __ret = __p0 * __p1;
14158  return __ret;
14159}
14160#else
14161__ai int8x16_t vmulq_s8(int8x16_t __p0, int8x16_t __p1) {
14162  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14163  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14164  int8x16_t __ret;
14165  __ret = __rev0 * __rev1;
14166  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14167  return __ret;
14168}
14169#endif
14170
14171#ifdef __LITTLE_ENDIAN__
14172__ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1) {
14173  float32x4_t __ret;
14174  __ret = __p0 * __p1;
14175  return __ret;
14176}
14177#else
14178__ai float32x4_t vmulq_f32(float32x4_t __p0, float32x4_t __p1) {
14179  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14180  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14181  float32x4_t __ret;
14182  __ret = __rev0 * __rev1;
14183  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14184  return __ret;
14185}
14186#endif
14187
14188#ifdef __LITTLE_ENDIAN__
14189__ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1) {
14190  int32x4_t __ret;
14191  __ret = __p0 * __p1;
14192  return __ret;
14193}
14194#else
14195__ai int32x4_t vmulq_s32(int32x4_t __p0, int32x4_t __p1) {
14196  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14197  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14198  int32x4_t __ret;
14199  __ret = __rev0 * __rev1;
14200  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14201  return __ret;
14202}
14203#endif
14204
14205#ifdef __LITTLE_ENDIAN__
14206__ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1) {
14207  int16x8_t __ret;
14208  __ret = __p0 * __p1;
14209  return __ret;
14210}
14211#else
14212__ai int16x8_t vmulq_s16(int16x8_t __p0, int16x8_t __p1) {
14213  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14214  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14215  int16x8_t __ret;
14216  __ret = __rev0 * __rev1;
14217  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14218  return __ret;
14219}
14220#endif
14221
14222#ifdef __LITTLE_ENDIAN__
14223__ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1) {
14224  uint8x8_t __ret;
14225  __ret = __p0 * __p1;
14226  return __ret;
14227}
14228#else
14229__ai uint8x8_t vmul_u8(uint8x8_t __p0, uint8x8_t __p1) {
14230  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14231  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14232  uint8x8_t __ret;
14233  __ret = __rev0 * __rev1;
14234  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14235  return __ret;
14236}
14237#endif
14238
14239#ifdef __LITTLE_ENDIAN__
14240__ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1) {
14241  uint32x2_t __ret;
14242  __ret = __p0 * __p1;
14243  return __ret;
14244}
14245#else
14246__ai uint32x2_t vmul_u32(uint32x2_t __p0, uint32x2_t __p1) {
14247  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14248  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14249  uint32x2_t __ret;
14250  __ret = __rev0 * __rev1;
14251  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14252  return __ret;
14253}
14254#endif
14255
14256#ifdef __LITTLE_ENDIAN__
14257__ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1) {
14258  uint16x4_t __ret;
14259  __ret = __p0 * __p1;
14260  return __ret;
14261}
14262#else
14263__ai uint16x4_t vmul_u16(uint16x4_t __p0, uint16x4_t __p1) {
14264  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14265  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14266  uint16x4_t __ret;
14267  __ret = __rev0 * __rev1;
14268  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14269  return __ret;
14270}
14271#endif
14272
14273#ifdef __LITTLE_ENDIAN__
14274__ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1) {
14275  int8x8_t __ret;
14276  __ret = __p0 * __p1;
14277  return __ret;
14278}
14279#else
14280__ai int8x8_t vmul_s8(int8x8_t __p0, int8x8_t __p1) {
14281  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14282  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14283  int8x8_t __ret;
14284  __ret = __rev0 * __rev1;
14285  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14286  return __ret;
14287}
14288#endif
14289
14290#ifdef __LITTLE_ENDIAN__
14291__ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1) {
14292  float32x2_t __ret;
14293  __ret = __p0 * __p1;
14294  return __ret;
14295}
14296#else
14297__ai float32x2_t vmul_f32(float32x2_t __p0, float32x2_t __p1) {
14298  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14299  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14300  float32x2_t __ret;
14301  __ret = __rev0 * __rev1;
14302  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14303  return __ret;
14304}
14305#endif
14306
14307#ifdef __LITTLE_ENDIAN__
14308__ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1) {
14309  int32x2_t __ret;
14310  __ret = __p0 * __p1;
14311  return __ret;
14312}
14313#else
14314__ai int32x2_t vmul_s32(int32x2_t __p0, int32x2_t __p1) {
14315  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14316  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14317  int32x2_t __ret;
14318  __ret = __rev0 * __rev1;
14319  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14320  return __ret;
14321}
14322#endif
14323
14324#ifdef __LITTLE_ENDIAN__
14325__ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1) {
14326  int16x4_t __ret;
14327  __ret = __p0 * __p1;
14328  return __ret;
14329}
14330#else
14331__ai int16x4_t vmul_s16(int16x4_t __p0, int16x4_t __p1) {
14332  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14333  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14334  int16x4_t __ret;
14335  __ret = __rev0 * __rev1;
14336  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14337  return __ret;
14338}
14339#endif
14340
14341#ifdef __LITTLE_ENDIAN__
14342__ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1) {
14343  poly8x8_t __ret;
14344  __ret = (poly8x8_t) __builtin_neon_vmul_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
14345  return __ret;
14346}
14347#else
14348__ai poly8x8_t vmul_p8(poly8x8_t __p0, poly8x8_t __p1) {
14349  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14350  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14351  poly8x8_t __ret;
14352  __ret = (poly8x8_t) __builtin_neon_vmul_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
14353  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14354  return __ret;
14355}
14356#endif
14357
14358#ifdef __LITTLE_ENDIAN__
14359__ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1) {
14360  poly8x16_t __ret;
14361  __ret = (poly8x16_t) __builtin_neon_vmulq_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
14362  return __ret;
14363}
14364#else
14365__ai poly8x16_t vmulq_p8(poly8x16_t __p0, poly8x16_t __p1) {
14366  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14367  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14368  poly8x16_t __ret;
14369  __ret = (poly8x16_t) __builtin_neon_vmulq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
14370  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
14371  return __ret;
14372}
14373#endif
14374
14375#ifdef __LITTLE_ENDIAN__
14376#define vmulq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14377  uint32x4_t __s0 = __p0; \
14378  uint32x2_t __s1 = __p1; \
14379  uint32x4_t __ret; \
14380  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14381  __ret; \
14382})
14383#else
14384#define vmulq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14385  uint32x4_t __s0 = __p0; \
14386  uint32x2_t __s1 = __p1; \
14387  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14388  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14389  uint32x4_t __ret; \
14390  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14391  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14392  __ret; \
14393})
14394#endif
14395
14396#ifdef __LITTLE_ENDIAN__
14397#define vmulq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14398  uint16x8_t __s0 = __p0; \
14399  uint16x4_t __s1 = __p1; \
14400  uint16x8_t __ret; \
14401  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14402  __ret; \
14403})
14404#else
14405#define vmulq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14406  uint16x8_t __s0 = __p0; \
14407  uint16x4_t __s1 = __p1; \
14408  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
14409  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14410  uint16x8_t __ret; \
14411  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14412  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
14413  __ret; \
14414})
14415#endif
14416
14417#ifdef __LITTLE_ENDIAN__
14418#define vmulq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14419  float32x4_t __s0 = __p0; \
14420  float32x2_t __s1 = __p1; \
14421  float32x4_t __ret; \
14422  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14423  __ret; \
14424})
14425#else
14426#define vmulq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14427  float32x4_t __s0 = __p0; \
14428  float32x2_t __s1 = __p1; \
14429  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14430  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14431  float32x4_t __ret; \
14432  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14433  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14434  __ret; \
14435})
14436#endif
14437
14438#ifdef __LITTLE_ENDIAN__
14439#define vmulq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14440  int32x4_t __s0 = __p0; \
14441  int32x2_t __s1 = __p1; \
14442  int32x4_t __ret; \
14443  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14444  __ret; \
14445})
14446#else
14447#define vmulq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14448  int32x4_t __s0 = __p0; \
14449  int32x2_t __s1 = __p1; \
14450  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14451  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14452  int32x4_t __ret; \
14453  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14454  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14455  __ret; \
14456})
14457#endif
14458
14459#ifdef __LITTLE_ENDIAN__
14460#define vmulq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14461  int16x8_t __s0 = __p0; \
14462  int16x4_t __s1 = __p1; \
14463  int16x8_t __ret; \
14464  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14465  __ret; \
14466})
14467#else
14468#define vmulq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14469  int16x8_t __s0 = __p0; \
14470  int16x4_t __s1 = __p1; \
14471  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
14472  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14473  int16x8_t __ret; \
14474  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
14475  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
14476  __ret; \
14477})
14478#endif
14479
14480#ifdef __LITTLE_ENDIAN__
14481#define vmul_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14482  uint32x2_t __s0 = __p0; \
14483  uint32x2_t __s1 = __p1; \
14484  uint32x2_t __ret; \
14485  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
14486  __ret; \
14487})
14488#else
14489#define vmul_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14490  uint32x2_t __s0 = __p0; \
14491  uint32x2_t __s1 = __p1; \
14492  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14493  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14494  uint32x2_t __ret; \
14495  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
14496  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14497  __ret; \
14498})
14499#endif
14500
14501#ifdef __LITTLE_ENDIAN__
14502#define vmul_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14503  uint16x4_t __s0 = __p0; \
14504  uint16x4_t __s1 = __p1; \
14505  uint16x4_t __ret; \
14506  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14507  __ret; \
14508})
14509#else
14510#define vmul_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14511  uint16x4_t __s0 = __p0; \
14512  uint16x4_t __s1 = __p1; \
14513  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14514  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14515  uint16x4_t __ret; \
14516  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14517  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14518  __ret; \
14519})
14520#endif
14521
14522#ifdef __LITTLE_ENDIAN__
14523#define vmul_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14524  float32x2_t __s0 = __p0; \
14525  float32x2_t __s1 = __p1; \
14526  float32x2_t __ret; \
14527  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
14528  __ret; \
14529})
14530#else
14531#define vmul_lane_f32(__p0, __p1, __p2) __extension__ ({ \
14532  float32x2_t __s0 = __p0; \
14533  float32x2_t __s1 = __p1; \
14534  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14535  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14536  float32x2_t __ret; \
14537  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
14538  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14539  __ret; \
14540})
14541#endif
14542
14543#ifdef __LITTLE_ENDIAN__
14544#define vmul_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14545  int32x2_t __s0 = __p0; \
14546  int32x2_t __s1 = __p1; \
14547  int32x2_t __ret; \
14548  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
14549  __ret; \
14550})
14551#else
14552#define vmul_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14553  int32x2_t __s0 = __p0; \
14554  int32x2_t __s1 = __p1; \
14555  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14556  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14557  int32x2_t __ret; \
14558  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
14559  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14560  __ret; \
14561})
14562#endif
14563
14564#ifdef __LITTLE_ENDIAN__
14565#define vmul_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14566  int16x4_t __s0 = __p0; \
14567  int16x4_t __s1 = __p1; \
14568  int16x4_t __ret; \
14569  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
14570  __ret; \
14571})
14572#else
14573#define vmul_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14574  int16x4_t __s0 = __p0; \
14575  int16x4_t __s1 = __p1; \
14576  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14577  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14578  int16x4_t __ret; \
14579  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
14580  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14581  __ret; \
14582})
14583#endif
14584
14585#ifdef __LITTLE_ENDIAN__
14586__ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1) {
14587  uint32x4_t __ret;
14588  __ret = __p0 * (uint32x4_t) {__p1, __p1, __p1, __p1};
14589  return __ret;
14590}
14591#else
14592__ai uint32x4_t vmulq_n_u32(uint32x4_t __p0, uint32_t __p1) {
14593  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14594  uint32x4_t __ret;
14595  __ret = __rev0 * (uint32x4_t) {__p1, __p1, __p1, __p1};
14596  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14597  return __ret;
14598}
14599#endif
14600
14601#ifdef __LITTLE_ENDIAN__
14602__ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1) {
14603  uint16x8_t __ret;
14604  __ret = __p0 * (uint16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14605  return __ret;
14606}
14607#else
14608__ai uint16x8_t vmulq_n_u16(uint16x8_t __p0, uint16_t __p1) {
14609  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14610  uint16x8_t __ret;
14611  __ret = __rev0 * (uint16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14612  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14613  return __ret;
14614}
14615#endif
14616
14617#ifdef __LITTLE_ENDIAN__
14618__ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1) {
14619  float32x4_t __ret;
14620  __ret = __p0 * (float32x4_t) {__p1, __p1, __p1, __p1};
14621  return __ret;
14622}
14623#else
14624__ai float32x4_t vmulq_n_f32(float32x4_t __p0, float32_t __p1) {
14625  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14626  float32x4_t __ret;
14627  __ret = __rev0 * (float32x4_t) {__p1, __p1, __p1, __p1};
14628  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14629  return __ret;
14630}
14631#endif
14632
14633#ifdef __LITTLE_ENDIAN__
14634__ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1) {
14635  int32x4_t __ret;
14636  __ret = __p0 * (int32x4_t) {__p1, __p1, __p1, __p1};
14637  return __ret;
14638}
14639#else
14640__ai int32x4_t vmulq_n_s32(int32x4_t __p0, int32_t __p1) {
14641  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14642  int32x4_t __ret;
14643  __ret = __rev0 * (int32x4_t) {__p1, __p1, __p1, __p1};
14644  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14645  return __ret;
14646}
14647#endif
14648
14649#ifdef __LITTLE_ENDIAN__
14650__ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1) {
14651  int16x8_t __ret;
14652  __ret = __p0 * (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14653  return __ret;
14654}
14655#else
14656__ai int16x8_t vmulq_n_s16(int16x8_t __p0, int16_t __p1) {
14657  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14658  int16x8_t __ret;
14659  __ret = __rev0 * (int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1};
14660  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14661  return __ret;
14662}
14663#endif
14664
14665#ifdef __LITTLE_ENDIAN__
14666__ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1) {
14667  uint32x2_t __ret;
14668  __ret = __p0 * (uint32x2_t) {__p1, __p1};
14669  return __ret;
14670}
14671#else
14672__ai uint32x2_t vmul_n_u32(uint32x2_t __p0, uint32_t __p1) {
14673  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14674  uint32x2_t __ret;
14675  __ret = __rev0 * (uint32x2_t) {__p1, __p1};
14676  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14677  return __ret;
14678}
14679#endif
14680
14681#ifdef __LITTLE_ENDIAN__
14682__ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1) {
14683  uint16x4_t __ret;
14684  __ret = __p0 * (uint16x4_t) {__p1, __p1, __p1, __p1};
14685  return __ret;
14686}
14687#else
14688__ai uint16x4_t vmul_n_u16(uint16x4_t __p0, uint16_t __p1) {
14689  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14690  uint16x4_t __ret;
14691  __ret = __rev0 * (uint16x4_t) {__p1, __p1, __p1, __p1};
14692  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14693  return __ret;
14694}
14695#endif
14696
14697#ifdef __LITTLE_ENDIAN__
14698__ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1) {
14699  float32x2_t __ret;
14700  __ret = __p0 * (float32x2_t) {__p1, __p1};
14701  return __ret;
14702}
14703#else
14704__ai float32x2_t vmul_n_f32(float32x2_t __p0, float32_t __p1) {
14705  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14706  float32x2_t __ret;
14707  __ret = __rev0 * (float32x2_t) {__p1, __p1};
14708  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14709  return __ret;
14710}
14711#endif
14712
14713#ifdef __LITTLE_ENDIAN__
14714__ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1) {
14715  int32x2_t __ret;
14716  __ret = __p0 * (int32x2_t) {__p1, __p1};
14717  return __ret;
14718}
14719#else
14720__ai int32x2_t vmul_n_s32(int32x2_t __p0, int32_t __p1) {
14721  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14722  int32x2_t __ret;
14723  __ret = __rev0 * (int32x2_t) {__p1, __p1};
14724  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14725  return __ret;
14726}
14727#endif
14728
14729#ifdef __LITTLE_ENDIAN__
14730__ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1) {
14731  int16x4_t __ret;
14732  __ret = __p0 * (int16x4_t) {__p1, __p1, __p1, __p1};
14733  return __ret;
14734}
14735#else
14736__ai int16x4_t vmul_n_s16(int16x4_t __p0, int16_t __p1) {
14737  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14738  int16x4_t __ret;
14739  __ret = __rev0 * (int16x4_t) {__p1, __p1, __p1, __p1};
14740  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14741  return __ret;
14742}
14743#endif
14744
14745#ifdef __LITTLE_ENDIAN__
14746__ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
14747  poly16x8_t __ret;
14748  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 37);
14749  return __ret;
14750}
14751#else
14752__ai poly16x8_t vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
14753  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14754  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14755  poly16x8_t __ret;
14756  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 37);
14757  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14758  return __ret;
14759}
14760__ai poly16x8_t __noswap_vmull_p8(poly8x8_t __p0, poly8x8_t __p1) {
14761  poly16x8_t __ret;
14762  __ret = (poly16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 37);
14763  return __ret;
14764}
14765#endif
14766
14767#ifdef __LITTLE_ENDIAN__
14768__ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
14769  uint16x8_t __ret;
14770  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 49);
14771  return __ret;
14772}
14773#else
14774__ai uint16x8_t vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
14775  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14776  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14777  uint16x8_t __ret;
14778  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 49);
14779  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14780  return __ret;
14781}
14782__ai uint16x8_t __noswap_vmull_u8(uint8x8_t __p0, uint8x8_t __p1) {
14783  uint16x8_t __ret;
14784  __ret = (uint16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 49);
14785  return __ret;
14786}
14787#endif
14788
14789#ifdef __LITTLE_ENDIAN__
14790__ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
14791  uint64x2_t __ret;
14792  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 51);
14793  return __ret;
14794}
14795#else
14796__ai uint64x2_t vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
14797  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14798  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14799  uint64x2_t __ret;
14800  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 51);
14801  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14802  return __ret;
14803}
14804__ai uint64x2_t __noswap_vmull_u32(uint32x2_t __p0, uint32x2_t __p1) {
14805  uint64x2_t __ret;
14806  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 51);
14807  return __ret;
14808}
14809#endif
14810
14811#ifdef __LITTLE_ENDIAN__
14812__ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
14813  uint32x4_t __ret;
14814  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 50);
14815  return __ret;
14816}
14817#else
14818__ai uint32x4_t vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
14819  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14820  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14821  uint32x4_t __ret;
14822  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 50);
14823  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14824  return __ret;
14825}
14826__ai uint32x4_t __noswap_vmull_u16(uint16x4_t __p0, uint16x4_t __p1) {
14827  uint32x4_t __ret;
14828  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 50);
14829  return __ret;
14830}
14831#endif
14832
14833#ifdef __LITTLE_ENDIAN__
14834__ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1) {
14835  int16x8_t __ret;
14836  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 33);
14837  return __ret;
14838}
14839#else
14840__ai int16x8_t vmull_s8(int8x8_t __p0, int8x8_t __p1) {
14841  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
14842  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
14843  int16x8_t __ret;
14844  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 33);
14845  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
14846  return __ret;
14847}
14848__ai int16x8_t __noswap_vmull_s8(int8x8_t __p0, int8x8_t __p1) {
14849  int16x8_t __ret;
14850  __ret = (int16x8_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 33);
14851  return __ret;
14852}
14853#endif
14854
14855#ifdef __LITTLE_ENDIAN__
14856__ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1) {
14857  int64x2_t __ret;
14858  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
14859  return __ret;
14860}
14861#else
14862__ai int64x2_t vmull_s32(int32x2_t __p0, int32x2_t __p1) {
14863  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14864  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
14865  int64x2_t __ret;
14866  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 35);
14867  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14868  return __ret;
14869}
14870__ai int64x2_t __noswap_vmull_s32(int32x2_t __p0, int32x2_t __p1) {
14871  int64x2_t __ret;
14872  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
14873  return __ret;
14874}
14875#endif
14876
14877#ifdef __LITTLE_ENDIAN__
14878__ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1) {
14879  int32x4_t __ret;
14880  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
14881  return __ret;
14882}
14883#else
14884__ai int32x4_t vmull_s16(int16x4_t __p0, int16x4_t __p1) {
14885  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
14886  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
14887  int32x4_t __ret;
14888  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 34);
14889  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
14890  return __ret;
14891}
14892__ai int32x4_t __noswap_vmull_s16(int16x4_t __p0, int16x4_t __p1) {
14893  int32x4_t __ret;
14894  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
14895  return __ret;
14896}
14897#endif
14898
14899#ifdef __LITTLE_ENDIAN__
14900#define vmull_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14901  uint32x2_t __s0 = __p0; \
14902  uint32x2_t __s1 = __p1; \
14903  uint64x2_t __ret; \
14904  __ret = vmull_u32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
14905  __ret; \
14906})
14907#else
14908#define vmull_lane_u32(__p0, __p1, __p2) __extension__ ({ \
14909  uint32x2_t __s0 = __p0; \
14910  uint32x2_t __s1 = __p1; \
14911  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14912  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14913  uint64x2_t __ret; \
14914  __ret = __noswap_vmull_u32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
14915  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14916  __ret; \
14917})
14918#endif
14919
14920#ifdef __LITTLE_ENDIAN__
14921#define vmull_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14922  uint16x4_t __s0 = __p0; \
14923  uint16x4_t __s1 = __p1; \
14924  uint32x4_t __ret; \
14925  __ret = vmull_u16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
14926  __ret; \
14927})
14928#else
14929#define vmull_lane_u16(__p0, __p1, __p2) __extension__ ({ \
14930  uint16x4_t __s0 = __p0; \
14931  uint16x4_t __s1 = __p1; \
14932  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14933  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14934  uint32x4_t __ret; \
14935  __ret = __noswap_vmull_u16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
14936  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14937  __ret; \
14938})
14939#endif
14940
14941#ifdef __LITTLE_ENDIAN__
14942#define vmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14943  int32x2_t __s0 = __p0; \
14944  int32x2_t __s1 = __p1; \
14945  int64x2_t __ret; \
14946  __ret = vmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
14947  __ret; \
14948})
14949#else
14950#define vmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
14951  int32x2_t __s0 = __p0; \
14952  int32x2_t __s1 = __p1; \
14953  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
14954  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
14955  int64x2_t __ret; \
14956  __ret = __noswap_vmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
14957  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
14958  __ret; \
14959})
14960#endif
14961
14962#ifdef __LITTLE_ENDIAN__
14963#define vmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14964  int16x4_t __s0 = __p0; \
14965  int16x4_t __s1 = __p1; \
14966  int32x4_t __ret; \
14967  __ret = vmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
14968  __ret; \
14969})
14970#else
14971#define vmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
14972  int16x4_t __s0 = __p0; \
14973  int16x4_t __s1 = __p1; \
14974  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
14975  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
14976  int32x4_t __ret; \
14977  __ret = __noswap_vmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
14978  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
14979  __ret; \
14980})
14981#endif
14982
14983#ifdef __LITTLE_ENDIAN__
14984__ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
14985  uint64x2_t __ret;
14986  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
14987  return __ret;
14988}
14989#else
14990__ai uint64x2_t vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
14991  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
14992  uint64x2_t __ret;
14993  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
14994  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
14995  return __ret;
14996}
14997__ai uint64x2_t __noswap_vmull_n_u32(uint32x2_t __p0, uint32_t __p1) {
14998  uint64x2_t __ret;
14999  __ret = (uint64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint32x2_t) {__p1, __p1}, 51);
15000  return __ret;
15001}
15002#endif
15003
15004#ifdef __LITTLE_ENDIAN__
15005__ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
15006  uint32x4_t __ret;
15007  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
15008  return __ret;
15009}
15010#else
15011__ai uint32x4_t vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
15012  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15013  uint32x4_t __ret;
15014  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
15015  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15016  return __ret;
15017}
15018__ai uint32x4_t __noswap_vmull_n_u16(uint16x4_t __p0, uint16_t __p1) {
15019  uint32x4_t __ret;
15020  __ret = (uint32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(uint16x4_t) {__p1, __p1, __p1, __p1}, 50);
15021  return __ret;
15022}
15023#endif
15024
15025#ifdef __LITTLE_ENDIAN__
15026__ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1) {
15027  int64x2_t __ret;
15028  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
15029  return __ret;
15030}
15031#else
15032__ai int64x2_t vmull_n_s32(int32x2_t __p0, int32_t __p1) {
15033  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15034  int64x2_t __ret;
15035  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
15036  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15037  return __ret;
15038}
15039__ai int64x2_t __noswap_vmull_n_s32(int32x2_t __p0, int32_t __p1) {
15040  int64x2_t __ret;
15041  __ret = (int64x2_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
15042  return __ret;
15043}
15044#endif
15045
15046#ifdef __LITTLE_ENDIAN__
15047__ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1) {
15048  int32x4_t __ret;
15049  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
15050  return __ret;
15051}
15052#else
15053__ai int32x4_t vmull_n_s16(int16x4_t __p0, int16_t __p1) {
15054  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15055  int32x4_t __ret;
15056  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
15057  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15058  return __ret;
15059}
15060__ai int32x4_t __noswap_vmull_n_s16(int16x4_t __p0, int16_t __p1) {
15061  int32x4_t __ret;
15062  __ret = (int32x4_t) __builtin_neon_vmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
15063  return __ret;
15064}
15065#endif
15066
15067#ifdef __LITTLE_ENDIAN__
15068__ai poly8x8_t vmvn_p8(poly8x8_t __p0) {
15069  poly8x8_t __ret;
15070  __ret = ~__p0;
15071  return __ret;
15072}
15073#else
15074__ai poly8x8_t vmvn_p8(poly8x8_t __p0) {
15075  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15076  poly8x8_t __ret;
15077  __ret = ~__rev0;
15078  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15079  return __ret;
15080}
15081#endif
15082
15083#ifdef __LITTLE_ENDIAN__
15084__ai poly8x16_t vmvnq_p8(poly8x16_t __p0) {
15085  poly8x16_t __ret;
15086  __ret = ~__p0;
15087  return __ret;
15088}
15089#else
15090__ai poly8x16_t vmvnq_p8(poly8x16_t __p0) {
15091  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15092  poly8x16_t __ret;
15093  __ret = ~__rev0;
15094  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15095  return __ret;
15096}
15097#endif
15098
15099#ifdef __LITTLE_ENDIAN__
15100__ai uint8x16_t vmvnq_u8(uint8x16_t __p0) {
15101  uint8x16_t __ret;
15102  __ret = ~__p0;
15103  return __ret;
15104}
15105#else
15106__ai uint8x16_t vmvnq_u8(uint8x16_t __p0) {
15107  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15108  uint8x16_t __ret;
15109  __ret = ~__rev0;
15110  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15111  return __ret;
15112}
15113#endif
15114
15115#ifdef __LITTLE_ENDIAN__
15116__ai uint32x4_t vmvnq_u32(uint32x4_t __p0) {
15117  uint32x4_t __ret;
15118  __ret = ~__p0;
15119  return __ret;
15120}
15121#else
15122__ai uint32x4_t vmvnq_u32(uint32x4_t __p0) {
15123  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15124  uint32x4_t __ret;
15125  __ret = ~__rev0;
15126  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15127  return __ret;
15128}
15129#endif
15130
15131#ifdef __LITTLE_ENDIAN__
15132__ai uint16x8_t vmvnq_u16(uint16x8_t __p0) {
15133  uint16x8_t __ret;
15134  __ret = ~__p0;
15135  return __ret;
15136}
15137#else
15138__ai uint16x8_t vmvnq_u16(uint16x8_t __p0) {
15139  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15140  uint16x8_t __ret;
15141  __ret = ~__rev0;
15142  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15143  return __ret;
15144}
15145#endif
15146
15147#ifdef __LITTLE_ENDIAN__
15148__ai int8x16_t vmvnq_s8(int8x16_t __p0) {
15149  int8x16_t __ret;
15150  __ret = ~__p0;
15151  return __ret;
15152}
15153#else
15154__ai int8x16_t vmvnq_s8(int8x16_t __p0) {
15155  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15156  int8x16_t __ret;
15157  __ret = ~__rev0;
15158  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15159  return __ret;
15160}
15161#endif
15162
15163#ifdef __LITTLE_ENDIAN__
15164__ai int32x4_t vmvnq_s32(int32x4_t __p0) {
15165  int32x4_t __ret;
15166  __ret = ~__p0;
15167  return __ret;
15168}
15169#else
15170__ai int32x4_t vmvnq_s32(int32x4_t __p0) {
15171  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15172  int32x4_t __ret;
15173  __ret = ~__rev0;
15174  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15175  return __ret;
15176}
15177#endif
15178
15179#ifdef __LITTLE_ENDIAN__
15180__ai int16x8_t vmvnq_s16(int16x8_t __p0) {
15181  int16x8_t __ret;
15182  __ret = ~__p0;
15183  return __ret;
15184}
15185#else
15186__ai int16x8_t vmvnq_s16(int16x8_t __p0) {
15187  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15188  int16x8_t __ret;
15189  __ret = ~__rev0;
15190  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15191  return __ret;
15192}
15193#endif
15194
15195#ifdef __LITTLE_ENDIAN__
15196__ai uint8x8_t vmvn_u8(uint8x8_t __p0) {
15197  uint8x8_t __ret;
15198  __ret = ~__p0;
15199  return __ret;
15200}
15201#else
15202__ai uint8x8_t vmvn_u8(uint8x8_t __p0) {
15203  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15204  uint8x8_t __ret;
15205  __ret = ~__rev0;
15206  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15207  return __ret;
15208}
15209#endif
15210
15211#ifdef __LITTLE_ENDIAN__
15212__ai uint32x2_t vmvn_u32(uint32x2_t __p0) {
15213  uint32x2_t __ret;
15214  __ret = ~__p0;
15215  return __ret;
15216}
15217#else
15218__ai uint32x2_t vmvn_u32(uint32x2_t __p0) {
15219  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15220  uint32x2_t __ret;
15221  __ret = ~__rev0;
15222  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15223  return __ret;
15224}
15225#endif
15226
15227#ifdef __LITTLE_ENDIAN__
15228__ai uint16x4_t vmvn_u16(uint16x4_t __p0) {
15229  uint16x4_t __ret;
15230  __ret = ~__p0;
15231  return __ret;
15232}
15233#else
15234__ai uint16x4_t vmvn_u16(uint16x4_t __p0) {
15235  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15236  uint16x4_t __ret;
15237  __ret = ~__rev0;
15238  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15239  return __ret;
15240}
15241#endif
15242
15243#ifdef __LITTLE_ENDIAN__
15244__ai int8x8_t vmvn_s8(int8x8_t __p0) {
15245  int8x8_t __ret;
15246  __ret = ~__p0;
15247  return __ret;
15248}
15249#else
15250__ai int8x8_t vmvn_s8(int8x8_t __p0) {
15251  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15252  int8x8_t __ret;
15253  __ret = ~__rev0;
15254  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15255  return __ret;
15256}
15257#endif
15258
15259#ifdef __LITTLE_ENDIAN__
15260__ai int32x2_t vmvn_s32(int32x2_t __p0) {
15261  int32x2_t __ret;
15262  __ret = ~__p0;
15263  return __ret;
15264}
15265#else
15266__ai int32x2_t vmvn_s32(int32x2_t __p0) {
15267  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15268  int32x2_t __ret;
15269  __ret = ~__rev0;
15270  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15271  return __ret;
15272}
15273#endif
15274
15275#ifdef __LITTLE_ENDIAN__
15276__ai int16x4_t vmvn_s16(int16x4_t __p0) {
15277  int16x4_t __ret;
15278  __ret = ~__p0;
15279  return __ret;
15280}
15281#else
15282__ai int16x4_t vmvn_s16(int16x4_t __p0) {
15283  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15284  int16x4_t __ret;
15285  __ret = ~__rev0;
15286  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15287  return __ret;
15288}
15289#endif
15290
15291#ifdef __LITTLE_ENDIAN__
15292__ai int8x16_t vnegq_s8(int8x16_t __p0) {
15293  int8x16_t __ret;
15294  __ret = -__p0;
15295  return __ret;
15296}
15297#else
15298__ai int8x16_t vnegq_s8(int8x16_t __p0) {
15299  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15300  int8x16_t __ret;
15301  __ret = -__rev0;
15302  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15303  return __ret;
15304}
15305#endif
15306
15307#ifdef __LITTLE_ENDIAN__
15308__ai float32x4_t vnegq_f32(float32x4_t __p0) {
15309  float32x4_t __ret;
15310  __ret = -__p0;
15311  return __ret;
15312}
15313#else
15314__ai float32x4_t vnegq_f32(float32x4_t __p0) {
15315  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15316  float32x4_t __ret;
15317  __ret = -__rev0;
15318  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15319  return __ret;
15320}
15321#endif
15322
15323#ifdef __LITTLE_ENDIAN__
15324__ai int32x4_t vnegq_s32(int32x4_t __p0) {
15325  int32x4_t __ret;
15326  __ret = -__p0;
15327  return __ret;
15328}
15329#else
15330__ai int32x4_t vnegq_s32(int32x4_t __p0) {
15331  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15332  int32x4_t __ret;
15333  __ret = -__rev0;
15334  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15335  return __ret;
15336}
15337#endif
15338
15339#ifdef __LITTLE_ENDIAN__
15340__ai int16x8_t vnegq_s16(int16x8_t __p0) {
15341  int16x8_t __ret;
15342  __ret = -__p0;
15343  return __ret;
15344}
15345#else
15346__ai int16x8_t vnegq_s16(int16x8_t __p0) {
15347  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15348  int16x8_t __ret;
15349  __ret = -__rev0;
15350  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15351  return __ret;
15352}
15353#endif
15354
15355#ifdef __LITTLE_ENDIAN__
15356__ai int8x8_t vneg_s8(int8x8_t __p0) {
15357  int8x8_t __ret;
15358  __ret = -__p0;
15359  return __ret;
15360}
15361#else
15362__ai int8x8_t vneg_s8(int8x8_t __p0) {
15363  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15364  int8x8_t __ret;
15365  __ret = -__rev0;
15366  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15367  return __ret;
15368}
15369#endif
15370
15371#ifdef __LITTLE_ENDIAN__
15372__ai float32x2_t vneg_f32(float32x2_t __p0) {
15373  float32x2_t __ret;
15374  __ret = -__p0;
15375  return __ret;
15376}
15377#else
15378__ai float32x2_t vneg_f32(float32x2_t __p0) {
15379  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15380  float32x2_t __ret;
15381  __ret = -__rev0;
15382  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15383  return __ret;
15384}
15385#endif
15386
15387#ifdef __LITTLE_ENDIAN__
15388__ai int32x2_t vneg_s32(int32x2_t __p0) {
15389  int32x2_t __ret;
15390  __ret = -__p0;
15391  return __ret;
15392}
15393#else
15394__ai int32x2_t vneg_s32(int32x2_t __p0) {
15395  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15396  int32x2_t __ret;
15397  __ret = -__rev0;
15398  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15399  return __ret;
15400}
15401#endif
15402
15403#ifdef __LITTLE_ENDIAN__
15404__ai int16x4_t vneg_s16(int16x4_t __p0) {
15405  int16x4_t __ret;
15406  __ret = -__p0;
15407  return __ret;
15408}
15409#else
15410__ai int16x4_t vneg_s16(int16x4_t __p0) {
15411  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15412  int16x4_t __ret;
15413  __ret = -__rev0;
15414  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15415  return __ret;
15416}
15417#endif
15418
15419#ifdef __LITTLE_ENDIAN__
15420__ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15421  uint8x16_t __ret;
15422  __ret = __p0 | ~__p1;
15423  return __ret;
15424}
15425#else
15426__ai uint8x16_t vornq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15427  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15428  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15429  uint8x16_t __ret;
15430  __ret = __rev0 | ~__rev1;
15431  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15432  return __ret;
15433}
15434#endif
15435
15436#ifdef __LITTLE_ENDIAN__
15437__ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15438  uint32x4_t __ret;
15439  __ret = __p0 | ~__p1;
15440  return __ret;
15441}
15442#else
15443__ai uint32x4_t vornq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15444  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15445  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15446  uint32x4_t __ret;
15447  __ret = __rev0 | ~__rev1;
15448  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15449  return __ret;
15450}
15451#endif
15452
15453#ifdef __LITTLE_ENDIAN__
15454__ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15455  uint64x2_t __ret;
15456  __ret = __p0 | ~__p1;
15457  return __ret;
15458}
15459#else
15460__ai uint64x2_t vornq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15461  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15462  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15463  uint64x2_t __ret;
15464  __ret = __rev0 | ~__rev1;
15465  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15466  return __ret;
15467}
15468#endif
15469
15470#ifdef __LITTLE_ENDIAN__
15471__ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15472  uint16x8_t __ret;
15473  __ret = __p0 | ~__p1;
15474  return __ret;
15475}
15476#else
15477__ai uint16x8_t vornq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15478  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15479  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15480  uint16x8_t __ret;
15481  __ret = __rev0 | ~__rev1;
15482  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15483  return __ret;
15484}
15485#endif
15486
15487#ifdef __LITTLE_ENDIAN__
15488__ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1) {
15489  int8x16_t __ret;
15490  __ret = __p0 | ~__p1;
15491  return __ret;
15492}
15493#else
15494__ai int8x16_t vornq_s8(int8x16_t __p0, int8x16_t __p1) {
15495  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15496  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15497  int8x16_t __ret;
15498  __ret = __rev0 | ~__rev1;
15499  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15500  return __ret;
15501}
15502#endif
15503
15504#ifdef __LITTLE_ENDIAN__
15505__ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1) {
15506  int32x4_t __ret;
15507  __ret = __p0 | ~__p1;
15508  return __ret;
15509}
15510#else
15511__ai int32x4_t vornq_s32(int32x4_t __p0, int32x4_t __p1) {
15512  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15513  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15514  int32x4_t __ret;
15515  __ret = __rev0 | ~__rev1;
15516  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15517  return __ret;
15518}
15519#endif
15520
15521#ifdef __LITTLE_ENDIAN__
15522__ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1) {
15523  int64x2_t __ret;
15524  __ret = __p0 | ~__p1;
15525  return __ret;
15526}
15527#else
15528__ai int64x2_t vornq_s64(int64x2_t __p0, int64x2_t __p1) {
15529  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15530  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15531  int64x2_t __ret;
15532  __ret = __rev0 | ~__rev1;
15533  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15534  return __ret;
15535}
15536#endif
15537
15538#ifdef __LITTLE_ENDIAN__
15539__ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1) {
15540  int16x8_t __ret;
15541  __ret = __p0 | ~__p1;
15542  return __ret;
15543}
15544#else
15545__ai int16x8_t vornq_s16(int16x8_t __p0, int16x8_t __p1) {
15546  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15547  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15548  int16x8_t __ret;
15549  __ret = __rev0 | ~__rev1;
15550  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15551  return __ret;
15552}
15553#endif
15554
15555#ifdef __LITTLE_ENDIAN__
15556__ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1) {
15557  uint8x8_t __ret;
15558  __ret = __p0 | ~__p1;
15559  return __ret;
15560}
15561#else
15562__ai uint8x8_t vorn_u8(uint8x8_t __p0, uint8x8_t __p1) {
15563  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15564  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15565  uint8x8_t __ret;
15566  __ret = __rev0 | ~__rev1;
15567  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15568  return __ret;
15569}
15570#endif
15571
15572#ifdef __LITTLE_ENDIAN__
15573__ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1) {
15574  uint32x2_t __ret;
15575  __ret = __p0 | ~__p1;
15576  return __ret;
15577}
15578#else
15579__ai uint32x2_t vorn_u32(uint32x2_t __p0, uint32x2_t __p1) {
15580  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15581  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15582  uint32x2_t __ret;
15583  __ret = __rev0 | ~__rev1;
15584  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15585  return __ret;
15586}
15587#endif
15588
15589#ifdef __LITTLE_ENDIAN__
15590__ai uint64x1_t vorn_u64(uint64x1_t __p0, uint64x1_t __p1) {
15591  uint64x1_t __ret;
15592  __ret = __p0 | ~__p1;
15593  return __ret;
15594}
15595#else
15596__ai uint64x1_t vorn_u64(uint64x1_t __p0, uint64x1_t __p1) {
15597  uint64x1_t __ret;
15598  __ret = __p0 | ~__p1;
15599  return __ret;
15600}
15601#endif
15602
15603#ifdef __LITTLE_ENDIAN__
15604__ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1) {
15605  uint16x4_t __ret;
15606  __ret = __p0 | ~__p1;
15607  return __ret;
15608}
15609#else
15610__ai uint16x4_t vorn_u16(uint16x4_t __p0, uint16x4_t __p1) {
15611  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15612  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15613  uint16x4_t __ret;
15614  __ret = __rev0 | ~__rev1;
15615  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15616  return __ret;
15617}
15618#endif
15619
15620#ifdef __LITTLE_ENDIAN__
15621__ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1) {
15622  int8x8_t __ret;
15623  __ret = __p0 | ~__p1;
15624  return __ret;
15625}
15626#else
15627__ai int8x8_t vorn_s8(int8x8_t __p0, int8x8_t __p1) {
15628  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15629  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15630  int8x8_t __ret;
15631  __ret = __rev0 | ~__rev1;
15632  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15633  return __ret;
15634}
15635#endif
15636
15637#ifdef __LITTLE_ENDIAN__
15638__ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1) {
15639  int32x2_t __ret;
15640  __ret = __p0 | ~__p1;
15641  return __ret;
15642}
15643#else
15644__ai int32x2_t vorn_s32(int32x2_t __p0, int32x2_t __p1) {
15645  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15646  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15647  int32x2_t __ret;
15648  __ret = __rev0 | ~__rev1;
15649  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15650  return __ret;
15651}
15652#endif
15653
15654#ifdef __LITTLE_ENDIAN__
15655__ai int64x1_t vorn_s64(int64x1_t __p0, int64x1_t __p1) {
15656  int64x1_t __ret;
15657  __ret = __p0 | ~__p1;
15658  return __ret;
15659}
15660#else
15661__ai int64x1_t vorn_s64(int64x1_t __p0, int64x1_t __p1) {
15662  int64x1_t __ret;
15663  __ret = __p0 | ~__p1;
15664  return __ret;
15665}
15666#endif
15667
15668#ifdef __LITTLE_ENDIAN__
15669__ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1) {
15670  int16x4_t __ret;
15671  __ret = __p0 | ~__p1;
15672  return __ret;
15673}
15674#else
15675__ai int16x4_t vorn_s16(int16x4_t __p0, int16x4_t __p1) {
15676  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15677  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15678  int16x4_t __ret;
15679  __ret = __rev0 | ~__rev1;
15680  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15681  return __ret;
15682}
15683#endif
15684
15685#ifdef __LITTLE_ENDIAN__
15686__ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15687  uint8x16_t __ret;
15688  __ret = __p0 | __p1;
15689  return __ret;
15690}
15691#else
15692__ai uint8x16_t vorrq_u8(uint8x16_t __p0, uint8x16_t __p1) {
15693  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15694  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15695  uint8x16_t __ret;
15696  __ret = __rev0 | __rev1;
15697  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15698  return __ret;
15699}
15700#endif
15701
15702#ifdef __LITTLE_ENDIAN__
15703__ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15704  uint32x4_t __ret;
15705  __ret = __p0 | __p1;
15706  return __ret;
15707}
15708#else
15709__ai uint32x4_t vorrq_u32(uint32x4_t __p0, uint32x4_t __p1) {
15710  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15711  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15712  uint32x4_t __ret;
15713  __ret = __rev0 | __rev1;
15714  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15715  return __ret;
15716}
15717#endif
15718
15719#ifdef __LITTLE_ENDIAN__
15720__ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15721  uint64x2_t __ret;
15722  __ret = __p0 | __p1;
15723  return __ret;
15724}
15725#else
15726__ai uint64x2_t vorrq_u64(uint64x2_t __p0, uint64x2_t __p1) {
15727  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15728  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15729  uint64x2_t __ret;
15730  __ret = __rev0 | __rev1;
15731  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15732  return __ret;
15733}
15734#endif
15735
15736#ifdef __LITTLE_ENDIAN__
15737__ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15738  uint16x8_t __ret;
15739  __ret = __p0 | __p1;
15740  return __ret;
15741}
15742#else
15743__ai uint16x8_t vorrq_u16(uint16x8_t __p0, uint16x8_t __p1) {
15744  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15745  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15746  uint16x8_t __ret;
15747  __ret = __rev0 | __rev1;
15748  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15749  return __ret;
15750}
15751#endif
15752
15753#ifdef __LITTLE_ENDIAN__
15754__ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1) {
15755  int8x16_t __ret;
15756  __ret = __p0 | __p1;
15757  return __ret;
15758}
15759#else
15760__ai int8x16_t vorrq_s8(int8x16_t __p0, int8x16_t __p1) {
15761  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15762  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15763  int8x16_t __ret;
15764  __ret = __rev0 | __rev1;
15765  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15766  return __ret;
15767}
15768#endif
15769
15770#ifdef __LITTLE_ENDIAN__
15771__ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1) {
15772  int32x4_t __ret;
15773  __ret = __p0 | __p1;
15774  return __ret;
15775}
15776#else
15777__ai int32x4_t vorrq_s32(int32x4_t __p0, int32x4_t __p1) {
15778  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15779  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15780  int32x4_t __ret;
15781  __ret = __rev0 | __rev1;
15782  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15783  return __ret;
15784}
15785#endif
15786
15787#ifdef __LITTLE_ENDIAN__
15788__ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1) {
15789  int64x2_t __ret;
15790  __ret = __p0 | __p1;
15791  return __ret;
15792}
15793#else
15794__ai int64x2_t vorrq_s64(int64x2_t __p0, int64x2_t __p1) {
15795  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15796  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15797  int64x2_t __ret;
15798  __ret = __rev0 | __rev1;
15799  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15800  return __ret;
15801}
15802#endif
15803
15804#ifdef __LITTLE_ENDIAN__
15805__ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1) {
15806  int16x8_t __ret;
15807  __ret = __p0 | __p1;
15808  return __ret;
15809}
15810#else
15811__ai int16x8_t vorrq_s16(int16x8_t __p0, int16x8_t __p1) {
15812  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15813  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15814  int16x8_t __ret;
15815  __ret = __rev0 | __rev1;
15816  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15817  return __ret;
15818}
15819#endif
15820
15821#ifdef __LITTLE_ENDIAN__
15822__ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1) {
15823  uint8x8_t __ret;
15824  __ret = __p0 | __p1;
15825  return __ret;
15826}
15827#else
15828__ai uint8x8_t vorr_u8(uint8x8_t __p0, uint8x8_t __p1) {
15829  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15830  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15831  uint8x8_t __ret;
15832  __ret = __rev0 | __rev1;
15833  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15834  return __ret;
15835}
15836#endif
15837
15838#ifdef __LITTLE_ENDIAN__
15839__ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1) {
15840  uint32x2_t __ret;
15841  __ret = __p0 | __p1;
15842  return __ret;
15843}
15844#else
15845__ai uint32x2_t vorr_u32(uint32x2_t __p0, uint32x2_t __p1) {
15846  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15847  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15848  uint32x2_t __ret;
15849  __ret = __rev0 | __rev1;
15850  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15851  return __ret;
15852}
15853#endif
15854
15855#ifdef __LITTLE_ENDIAN__
15856__ai uint64x1_t vorr_u64(uint64x1_t __p0, uint64x1_t __p1) {
15857  uint64x1_t __ret;
15858  __ret = __p0 | __p1;
15859  return __ret;
15860}
15861#else
15862__ai uint64x1_t vorr_u64(uint64x1_t __p0, uint64x1_t __p1) {
15863  uint64x1_t __ret;
15864  __ret = __p0 | __p1;
15865  return __ret;
15866}
15867#endif
15868
15869#ifdef __LITTLE_ENDIAN__
15870__ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1) {
15871  uint16x4_t __ret;
15872  __ret = __p0 | __p1;
15873  return __ret;
15874}
15875#else
15876__ai uint16x4_t vorr_u16(uint16x4_t __p0, uint16x4_t __p1) {
15877  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15878  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15879  uint16x4_t __ret;
15880  __ret = __rev0 | __rev1;
15881  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15882  return __ret;
15883}
15884#endif
15885
15886#ifdef __LITTLE_ENDIAN__
15887__ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1) {
15888  int8x8_t __ret;
15889  __ret = __p0 | __p1;
15890  return __ret;
15891}
15892#else
15893__ai int8x8_t vorr_s8(int8x8_t __p0, int8x8_t __p1) {
15894  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15895  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15896  int8x8_t __ret;
15897  __ret = __rev0 | __rev1;
15898  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15899  return __ret;
15900}
15901#endif
15902
15903#ifdef __LITTLE_ENDIAN__
15904__ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1) {
15905  int32x2_t __ret;
15906  __ret = __p0 | __p1;
15907  return __ret;
15908}
15909#else
15910__ai int32x2_t vorr_s32(int32x2_t __p0, int32x2_t __p1) {
15911  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15912  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
15913  int32x2_t __ret;
15914  __ret = __rev0 | __rev1;
15915  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15916  return __ret;
15917}
15918#endif
15919
15920#ifdef __LITTLE_ENDIAN__
15921__ai int64x1_t vorr_s64(int64x1_t __p0, int64x1_t __p1) {
15922  int64x1_t __ret;
15923  __ret = __p0 | __p1;
15924  return __ret;
15925}
15926#else
15927__ai int64x1_t vorr_s64(int64x1_t __p0, int64x1_t __p1) {
15928  int64x1_t __ret;
15929  __ret = __p0 | __p1;
15930  return __ret;
15931}
15932#endif
15933
15934#ifdef __LITTLE_ENDIAN__
15935__ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1) {
15936  int16x4_t __ret;
15937  __ret = __p0 | __p1;
15938  return __ret;
15939}
15940#else
15941__ai int16x4_t vorr_s16(int16x4_t __p0, int16x4_t __p1) {
15942  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15943  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15944  int16x4_t __ret;
15945  __ret = __rev0 | __rev1;
15946  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15947  return __ret;
15948}
15949#endif
15950
15951#ifdef __LITTLE_ENDIAN__
15952__ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1) {
15953  uint16x8_t __ret;
15954  __ret = (uint16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
15955  return __ret;
15956}
15957#else
15958__ai uint16x8_t vpadalq_u8(uint16x8_t __p0, uint8x16_t __p1) {
15959  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
15960  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
15961  uint16x8_t __ret;
15962  __ret = (uint16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
15963  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
15964  return __ret;
15965}
15966#endif
15967
15968#ifdef __LITTLE_ENDIAN__
15969__ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1) {
15970  uint64x2_t __ret;
15971  __ret = (uint64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
15972  return __ret;
15973}
15974#else
15975__ai uint64x2_t vpadalq_u32(uint64x2_t __p0, uint32x4_t __p1) {
15976  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
15977  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
15978  uint64x2_t __ret;
15979  __ret = (uint64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
15980  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
15981  return __ret;
15982}
15983#endif
15984
15985#ifdef __LITTLE_ENDIAN__
15986__ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1) {
15987  uint32x4_t __ret;
15988  __ret = (uint32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
15989  return __ret;
15990}
15991#else
15992__ai uint32x4_t vpadalq_u16(uint32x4_t __p0, uint16x8_t __p1) {
15993  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
15994  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
15995  uint32x4_t __ret;
15996  __ret = (uint32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
15997  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
15998  return __ret;
15999}
16000#endif
16001
16002#ifdef __LITTLE_ENDIAN__
16003__ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1) {
16004  int16x8_t __ret;
16005  __ret = (int16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
16006  return __ret;
16007}
16008#else
16009__ai int16x8_t vpadalq_s8(int16x8_t __p0, int8x16_t __p1) {
16010  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16011  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16012  int16x8_t __ret;
16013  __ret = (int16x8_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
16014  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16015  return __ret;
16016}
16017#endif
16018
16019#ifdef __LITTLE_ENDIAN__
16020__ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1) {
16021  int64x2_t __ret;
16022  __ret = (int64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
16023  return __ret;
16024}
16025#else
16026__ai int64x2_t vpadalq_s32(int64x2_t __p0, int32x4_t __p1) {
16027  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16028  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16029  int64x2_t __ret;
16030  __ret = (int64x2_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
16031  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16032  return __ret;
16033}
16034#endif
16035
16036#ifdef __LITTLE_ENDIAN__
16037__ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1) {
16038  int32x4_t __ret;
16039  __ret = (int32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
16040  return __ret;
16041}
16042#else
16043__ai int32x4_t vpadalq_s16(int32x4_t __p0, int16x8_t __p1) {
16044  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16045  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16046  int32x4_t __ret;
16047  __ret = (int32x4_t) __builtin_neon_vpadalq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
16048  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16049  return __ret;
16050}
16051#endif
16052
16053#ifdef __LITTLE_ENDIAN__
16054__ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1) {
16055  uint16x4_t __ret;
16056  __ret = (uint16x4_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16057  return __ret;
16058}
16059#else
16060__ai uint16x4_t vpadal_u8(uint16x4_t __p0, uint8x8_t __p1) {
16061  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16062  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16063  uint16x4_t __ret;
16064  __ret = (uint16x4_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16065  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16066  return __ret;
16067}
16068#endif
16069
16070#ifdef __LITTLE_ENDIAN__
16071__ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1) {
16072  uint64x1_t __ret;
16073  __ret = (uint64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
16074  return __ret;
16075}
16076#else
16077__ai uint64x1_t vpadal_u32(uint64x1_t __p0, uint32x2_t __p1) {
16078  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16079  uint64x1_t __ret;
16080  __ret = (uint64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__rev1, 19);
16081  return __ret;
16082}
16083#endif
16084
16085#ifdef __LITTLE_ENDIAN__
16086__ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1) {
16087  uint32x2_t __ret;
16088  __ret = (uint32x2_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16089  return __ret;
16090}
16091#else
16092__ai uint32x2_t vpadal_u16(uint32x2_t __p0, uint16x4_t __p1) {
16093  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16094  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16095  uint32x2_t __ret;
16096  __ret = (uint32x2_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16097  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16098  return __ret;
16099}
16100#endif
16101
16102#ifdef __LITTLE_ENDIAN__
16103__ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1) {
16104  int16x4_t __ret;
16105  __ret = (int16x4_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16106  return __ret;
16107}
16108#else
16109__ai int16x4_t vpadal_s8(int16x4_t __p0, int8x8_t __p1) {
16110  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16111  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16112  int16x4_t __ret;
16113  __ret = (int16x4_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16114  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16115  return __ret;
16116}
16117#endif
16118
16119#ifdef __LITTLE_ENDIAN__
16120__ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1) {
16121  int64x1_t __ret;
16122  __ret = (int64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
16123  return __ret;
16124}
16125#else
16126__ai int64x1_t vpadal_s32(int64x1_t __p0, int32x2_t __p1) {
16127  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16128  int64x1_t __ret;
16129  __ret = (int64x1_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__rev1, 3);
16130  return __ret;
16131}
16132#endif
16133
16134#ifdef __LITTLE_ENDIAN__
16135__ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1) {
16136  int32x2_t __ret;
16137  __ret = (int32x2_t) __builtin_neon_vpadal_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16138  return __ret;
16139}
16140#else
16141__ai int32x2_t vpadal_s16(int32x2_t __p0, int16x4_t __p1) {
16142  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16143  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16144  int32x2_t __ret;
16145  __ret = (int32x2_t) __builtin_neon_vpadal_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16146  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16147  return __ret;
16148}
16149#endif
16150
16151#ifdef __LITTLE_ENDIAN__
16152__ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16153  uint8x8_t __ret;
16154  __ret = (uint8x8_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16155  return __ret;
16156}
16157#else
16158__ai uint8x8_t vpadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16159  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16160  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16161  uint8x8_t __ret;
16162  __ret = (uint8x8_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16163  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16164  return __ret;
16165}
16166#endif
16167
16168#ifdef __LITTLE_ENDIAN__
16169__ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16170  uint32x2_t __ret;
16171  __ret = (uint32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16172  return __ret;
16173}
16174#else
16175__ai uint32x2_t vpadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16176  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16177  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16178  uint32x2_t __ret;
16179  __ret = (uint32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16180  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16181  return __ret;
16182}
16183#endif
16184
16185#ifdef __LITTLE_ENDIAN__
16186__ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16187  uint16x4_t __ret;
16188  __ret = (uint16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16189  return __ret;
16190}
16191#else
16192__ai uint16x4_t vpadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16193  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16194  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16195  uint16x4_t __ret;
16196  __ret = (uint16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16197  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16198  return __ret;
16199}
16200#endif
16201
16202#ifdef __LITTLE_ENDIAN__
16203__ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1) {
16204  int8x8_t __ret;
16205  __ret = (int8x8_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
16206  return __ret;
16207}
16208#else
16209__ai int8x8_t vpadd_s8(int8x8_t __p0, int8x8_t __p1) {
16210  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16211  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16212  int8x8_t __ret;
16213  __ret = (int8x8_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
16214  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16215  return __ret;
16216}
16217#endif
16218
16219#ifdef __LITTLE_ENDIAN__
16220__ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1) {
16221  float32x2_t __ret;
16222  __ret = (float32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
16223  return __ret;
16224}
16225#else
16226__ai float32x2_t vpadd_f32(float32x2_t __p0, float32x2_t __p1) {
16227  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16228  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16229  float32x2_t __ret;
16230  __ret = (float32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
16231  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16232  return __ret;
16233}
16234#endif
16235
16236#ifdef __LITTLE_ENDIAN__
16237__ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1) {
16238  int32x2_t __ret;
16239  __ret = (int32x2_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16240  return __ret;
16241}
16242#else
16243__ai int32x2_t vpadd_s32(int32x2_t __p0, int32x2_t __p1) {
16244  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16245  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16246  int32x2_t __ret;
16247  __ret = (int32x2_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16248  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16249  return __ret;
16250}
16251#endif
16252
16253#ifdef __LITTLE_ENDIAN__
16254__ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1) {
16255  int16x4_t __ret;
16256  __ret = (int16x4_t) __builtin_neon_vpadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16257  return __ret;
16258}
16259#else
16260__ai int16x4_t vpadd_s16(int16x4_t __p0, int16x4_t __p1) {
16261  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16262  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16263  int16x4_t __ret;
16264  __ret = (int16x4_t) __builtin_neon_vpadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16265  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16266  return __ret;
16267}
16268#endif
16269
16270#ifdef __LITTLE_ENDIAN__
16271__ai uint16x8_t vpaddlq_u8(uint8x16_t __p0) {
16272  uint16x8_t __ret;
16273  __ret = (uint16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 49);
16274  return __ret;
16275}
16276#else
16277__ai uint16x8_t vpaddlq_u8(uint8x16_t __p0) {
16278  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16279  uint16x8_t __ret;
16280  __ret = (uint16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 49);
16281  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16282  return __ret;
16283}
16284#endif
16285
16286#ifdef __LITTLE_ENDIAN__
16287__ai uint64x2_t vpaddlq_u32(uint32x4_t __p0) {
16288  uint64x2_t __ret;
16289  __ret = (uint64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 51);
16290  return __ret;
16291}
16292#else
16293__ai uint64x2_t vpaddlq_u32(uint32x4_t __p0) {
16294  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16295  uint64x2_t __ret;
16296  __ret = (uint64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 51);
16297  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16298  return __ret;
16299}
16300#endif
16301
16302#ifdef __LITTLE_ENDIAN__
16303__ai uint32x4_t vpaddlq_u16(uint16x8_t __p0) {
16304  uint32x4_t __ret;
16305  __ret = (uint32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 50);
16306  return __ret;
16307}
16308#else
16309__ai uint32x4_t vpaddlq_u16(uint16x8_t __p0) {
16310  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16311  uint32x4_t __ret;
16312  __ret = (uint32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 50);
16313  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16314  return __ret;
16315}
16316#endif
16317
16318#ifdef __LITTLE_ENDIAN__
16319__ai int16x8_t vpaddlq_s8(int8x16_t __p0) {
16320  int16x8_t __ret;
16321  __ret = (int16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 33);
16322  return __ret;
16323}
16324#else
16325__ai int16x8_t vpaddlq_s8(int8x16_t __p0) {
16326  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16327  int16x8_t __ret;
16328  __ret = (int16x8_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 33);
16329  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16330  return __ret;
16331}
16332#endif
16333
16334#ifdef __LITTLE_ENDIAN__
16335__ai int64x2_t vpaddlq_s32(int32x4_t __p0) {
16336  int64x2_t __ret;
16337  __ret = (int64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 35);
16338  return __ret;
16339}
16340#else
16341__ai int64x2_t vpaddlq_s32(int32x4_t __p0) {
16342  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16343  int64x2_t __ret;
16344  __ret = (int64x2_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 35);
16345  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16346  return __ret;
16347}
16348#endif
16349
16350#ifdef __LITTLE_ENDIAN__
16351__ai int32x4_t vpaddlq_s16(int16x8_t __p0) {
16352  int32x4_t __ret;
16353  __ret = (int32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__p0, 34);
16354  return __ret;
16355}
16356#else
16357__ai int32x4_t vpaddlq_s16(int16x8_t __p0) {
16358  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16359  int32x4_t __ret;
16360  __ret = (int32x4_t) __builtin_neon_vpaddlq_v((int8x16_t)__rev0, 34);
16361  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16362  return __ret;
16363}
16364#endif
16365
16366#ifdef __LITTLE_ENDIAN__
16367__ai uint16x4_t vpaddl_u8(uint8x8_t __p0) {
16368  uint16x4_t __ret;
16369  __ret = (uint16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 17);
16370  return __ret;
16371}
16372#else
16373__ai uint16x4_t vpaddl_u8(uint8x8_t __p0) {
16374  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16375  uint16x4_t __ret;
16376  __ret = (uint16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 17);
16377  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16378  return __ret;
16379}
16380#endif
16381
16382#ifdef __LITTLE_ENDIAN__
16383__ai uint64x1_t vpaddl_u32(uint32x2_t __p0) {
16384  uint64x1_t __ret;
16385  __ret = (uint64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 19);
16386  return __ret;
16387}
16388#else
16389__ai uint64x1_t vpaddl_u32(uint32x2_t __p0) {
16390  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16391  uint64x1_t __ret;
16392  __ret = (uint64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 19);
16393  return __ret;
16394}
16395#endif
16396
16397#ifdef __LITTLE_ENDIAN__
16398__ai uint32x2_t vpaddl_u16(uint16x4_t __p0) {
16399  uint32x2_t __ret;
16400  __ret = (uint32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 18);
16401  return __ret;
16402}
16403#else
16404__ai uint32x2_t vpaddl_u16(uint16x4_t __p0) {
16405  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16406  uint32x2_t __ret;
16407  __ret = (uint32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 18);
16408  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16409  return __ret;
16410}
16411#endif
16412
16413#ifdef __LITTLE_ENDIAN__
16414__ai int16x4_t vpaddl_s8(int8x8_t __p0) {
16415  int16x4_t __ret;
16416  __ret = (int16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 1);
16417  return __ret;
16418}
16419#else
16420__ai int16x4_t vpaddl_s8(int8x8_t __p0) {
16421  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16422  int16x4_t __ret;
16423  __ret = (int16x4_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 1);
16424  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16425  return __ret;
16426}
16427#endif
16428
16429#ifdef __LITTLE_ENDIAN__
16430__ai int64x1_t vpaddl_s32(int32x2_t __p0) {
16431  int64x1_t __ret;
16432  __ret = (int64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 3);
16433  return __ret;
16434}
16435#else
16436__ai int64x1_t vpaddl_s32(int32x2_t __p0) {
16437  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16438  int64x1_t __ret;
16439  __ret = (int64x1_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 3);
16440  return __ret;
16441}
16442#endif
16443
16444#ifdef __LITTLE_ENDIAN__
16445__ai int32x2_t vpaddl_s16(int16x4_t __p0) {
16446  int32x2_t __ret;
16447  __ret = (int32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__p0, 2);
16448  return __ret;
16449}
16450#else
16451__ai int32x2_t vpaddl_s16(int16x4_t __p0) {
16452  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16453  int32x2_t __ret;
16454  __ret = (int32x2_t) __builtin_neon_vpaddl_v((int8x8_t)__rev0, 2);
16455  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16456  return __ret;
16457}
16458#endif
16459
16460#ifdef __LITTLE_ENDIAN__
16461__ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
16462  uint8x8_t __ret;
16463  __ret = (uint8x8_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16464  return __ret;
16465}
16466#else
16467__ai uint8x8_t vpmax_u8(uint8x8_t __p0, uint8x8_t __p1) {
16468  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16469  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16470  uint8x8_t __ret;
16471  __ret = (uint8x8_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16472  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16473  return __ret;
16474}
16475#endif
16476
16477#ifdef __LITTLE_ENDIAN__
16478__ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
16479  uint32x2_t __ret;
16480  __ret = (uint32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16481  return __ret;
16482}
16483#else
16484__ai uint32x2_t vpmax_u32(uint32x2_t __p0, uint32x2_t __p1) {
16485  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16486  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16487  uint32x2_t __ret;
16488  __ret = (uint32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16489  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16490  return __ret;
16491}
16492#endif
16493
16494#ifdef __LITTLE_ENDIAN__
16495__ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
16496  uint16x4_t __ret;
16497  __ret = (uint16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16498  return __ret;
16499}
16500#else
16501__ai uint16x4_t vpmax_u16(uint16x4_t __p0, uint16x4_t __p1) {
16502  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16503  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16504  uint16x4_t __ret;
16505  __ret = (uint16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16506  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16507  return __ret;
16508}
16509#endif
16510
16511#ifdef __LITTLE_ENDIAN__
16512__ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1) {
16513  int8x8_t __ret;
16514  __ret = (int8x8_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
16515  return __ret;
16516}
16517#else
16518__ai int8x8_t vpmax_s8(int8x8_t __p0, int8x8_t __p1) {
16519  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16520  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16521  int8x8_t __ret;
16522  __ret = (int8x8_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
16523  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16524  return __ret;
16525}
16526#endif
16527
16528#ifdef __LITTLE_ENDIAN__
16529__ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1) {
16530  float32x2_t __ret;
16531  __ret = (float32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
16532  return __ret;
16533}
16534#else
16535__ai float32x2_t vpmax_f32(float32x2_t __p0, float32x2_t __p1) {
16536  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16537  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16538  float32x2_t __ret;
16539  __ret = (float32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
16540  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16541  return __ret;
16542}
16543#endif
16544
16545#ifdef __LITTLE_ENDIAN__
16546__ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1) {
16547  int32x2_t __ret;
16548  __ret = (int32x2_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16549  return __ret;
16550}
16551#else
16552__ai int32x2_t vpmax_s32(int32x2_t __p0, int32x2_t __p1) {
16553  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16554  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16555  int32x2_t __ret;
16556  __ret = (int32x2_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16557  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16558  return __ret;
16559}
16560#endif
16561
16562#ifdef __LITTLE_ENDIAN__
16563__ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1) {
16564  int16x4_t __ret;
16565  __ret = (int16x4_t) __builtin_neon_vpmax_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16566  return __ret;
16567}
16568#else
16569__ai int16x4_t vpmax_s16(int16x4_t __p0, int16x4_t __p1) {
16570  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16571  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16572  int16x4_t __ret;
16573  __ret = (int16x4_t) __builtin_neon_vpmax_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16574  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16575  return __ret;
16576}
16577#endif
16578
16579#ifdef __LITTLE_ENDIAN__
16580__ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
16581  uint8x8_t __ret;
16582  __ret = (uint8x8_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16583  return __ret;
16584}
16585#else
16586__ai uint8x8_t vpmin_u8(uint8x8_t __p0, uint8x8_t __p1) {
16587  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16588  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16589  uint8x8_t __ret;
16590  __ret = (uint8x8_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16591  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16592  return __ret;
16593}
16594#endif
16595
16596#ifdef __LITTLE_ENDIAN__
16597__ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
16598  uint32x2_t __ret;
16599  __ret = (uint32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16600  return __ret;
16601}
16602#else
16603__ai uint32x2_t vpmin_u32(uint32x2_t __p0, uint32x2_t __p1) {
16604  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16605  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16606  uint32x2_t __ret;
16607  __ret = (uint32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16608  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16609  return __ret;
16610}
16611#endif
16612
16613#ifdef __LITTLE_ENDIAN__
16614__ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
16615  uint16x4_t __ret;
16616  __ret = (uint16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16617  return __ret;
16618}
16619#else
16620__ai uint16x4_t vpmin_u16(uint16x4_t __p0, uint16x4_t __p1) {
16621  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16622  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16623  uint16x4_t __ret;
16624  __ret = (uint16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
16625  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16626  return __ret;
16627}
16628#endif
16629
16630#ifdef __LITTLE_ENDIAN__
16631__ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1) {
16632  int8x8_t __ret;
16633  __ret = (int8x8_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
16634  return __ret;
16635}
16636#else
16637__ai int8x8_t vpmin_s8(int8x8_t __p0, int8x8_t __p1) {
16638  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16639  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16640  int8x8_t __ret;
16641  __ret = (int8x8_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
16642  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16643  return __ret;
16644}
16645#endif
16646
16647#ifdef __LITTLE_ENDIAN__
16648__ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1) {
16649  float32x2_t __ret;
16650  __ret = (float32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
16651  return __ret;
16652}
16653#else
16654__ai float32x2_t vpmin_f32(float32x2_t __p0, float32x2_t __p1) {
16655  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16656  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16657  float32x2_t __ret;
16658  __ret = (float32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
16659  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16660  return __ret;
16661}
16662#endif
16663
16664#ifdef __LITTLE_ENDIAN__
16665__ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1) {
16666  int32x2_t __ret;
16667  __ret = (int32x2_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
16668  return __ret;
16669}
16670#else
16671__ai int32x2_t vpmin_s32(int32x2_t __p0, int32x2_t __p1) {
16672  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16673  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16674  int32x2_t __ret;
16675  __ret = (int32x2_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
16676  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16677  return __ret;
16678}
16679#endif
16680
16681#ifdef __LITTLE_ENDIAN__
16682__ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1) {
16683  int16x4_t __ret;
16684  __ret = (int16x4_t) __builtin_neon_vpmin_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
16685  return __ret;
16686}
16687#else
16688__ai int16x4_t vpmin_s16(int16x4_t __p0, int16x4_t __p1) {
16689  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16690  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16691  int16x4_t __ret;
16692  __ret = (int16x4_t) __builtin_neon_vpmin_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
16693  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16694  return __ret;
16695}
16696#endif
16697
16698#ifdef __LITTLE_ENDIAN__
16699__ai int8x16_t vqabsq_s8(int8x16_t __p0) {
16700  int8x16_t __ret;
16701  __ret = (int8x16_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 32);
16702  return __ret;
16703}
16704#else
16705__ai int8x16_t vqabsq_s8(int8x16_t __p0) {
16706  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16707  int8x16_t __ret;
16708  __ret = (int8x16_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 32);
16709  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16710  return __ret;
16711}
16712#endif
16713
16714#ifdef __LITTLE_ENDIAN__
16715__ai int32x4_t vqabsq_s32(int32x4_t __p0) {
16716  int32x4_t __ret;
16717  __ret = (int32x4_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 34);
16718  return __ret;
16719}
16720#else
16721__ai int32x4_t vqabsq_s32(int32x4_t __p0) {
16722  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16723  int32x4_t __ret;
16724  __ret = (int32x4_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 34);
16725  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16726  return __ret;
16727}
16728#endif
16729
16730#ifdef __LITTLE_ENDIAN__
16731__ai int16x8_t vqabsq_s16(int16x8_t __p0) {
16732  int16x8_t __ret;
16733  __ret = (int16x8_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 33);
16734  return __ret;
16735}
16736#else
16737__ai int16x8_t vqabsq_s16(int16x8_t __p0) {
16738  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16739  int16x8_t __ret;
16740  __ret = (int16x8_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 33);
16741  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16742  return __ret;
16743}
16744#endif
16745
16746#ifdef __LITTLE_ENDIAN__
16747__ai int8x8_t vqabs_s8(int8x8_t __p0) {
16748  int8x8_t __ret;
16749  __ret = (int8x8_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 0);
16750  return __ret;
16751}
16752#else
16753__ai int8x8_t vqabs_s8(int8x8_t __p0) {
16754  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16755  int8x8_t __ret;
16756  __ret = (int8x8_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 0);
16757  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16758  return __ret;
16759}
16760#endif
16761
16762#ifdef __LITTLE_ENDIAN__
16763__ai int32x2_t vqabs_s32(int32x2_t __p0) {
16764  int32x2_t __ret;
16765  __ret = (int32x2_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 2);
16766  return __ret;
16767}
16768#else
16769__ai int32x2_t vqabs_s32(int32x2_t __p0) {
16770  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16771  int32x2_t __ret;
16772  __ret = (int32x2_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 2);
16773  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16774  return __ret;
16775}
16776#endif
16777
16778#ifdef __LITTLE_ENDIAN__
16779__ai int16x4_t vqabs_s16(int16x4_t __p0) {
16780  int16x4_t __ret;
16781  __ret = (int16x4_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 1);
16782  return __ret;
16783}
16784#else
16785__ai int16x4_t vqabs_s16(int16x4_t __p0) {
16786  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16787  int16x4_t __ret;
16788  __ret = (int16x4_t) __builtin_neon_vqabs_v((int8x8_t)__rev0, 1);
16789  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16790  return __ret;
16791}
16792#endif
16793
16794#ifdef __LITTLE_ENDIAN__
16795__ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
16796  uint8x16_t __ret;
16797  __ret = (uint8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
16798  return __ret;
16799}
16800#else
16801__ai uint8x16_t vqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
16802  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16803  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16804  uint8x16_t __ret;
16805  __ret = (uint8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
16806  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16807  return __ret;
16808}
16809#endif
16810
16811#ifdef __LITTLE_ENDIAN__
16812__ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
16813  uint32x4_t __ret;
16814  __ret = (uint32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
16815  return __ret;
16816}
16817#else
16818__ai uint32x4_t vqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
16819  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16820  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16821  uint32x4_t __ret;
16822  __ret = (uint32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
16823  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16824  return __ret;
16825}
16826#endif
16827
16828#ifdef __LITTLE_ENDIAN__
16829__ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
16830  uint64x2_t __ret;
16831  __ret = (uint64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
16832  return __ret;
16833}
16834#else
16835__ai uint64x2_t vqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
16836  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16837  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16838  uint64x2_t __ret;
16839  __ret = (uint64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
16840  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16841  return __ret;
16842}
16843#endif
16844
16845#ifdef __LITTLE_ENDIAN__
16846__ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
16847  uint16x8_t __ret;
16848  __ret = (uint16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
16849  return __ret;
16850}
16851#else
16852__ai uint16x8_t vqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
16853  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16854  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16855  uint16x8_t __ret;
16856  __ret = (uint16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
16857  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16858  return __ret;
16859}
16860#endif
16861
16862#ifdef __LITTLE_ENDIAN__
16863__ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
16864  int8x16_t __ret;
16865  __ret = (int8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
16866  return __ret;
16867}
16868#else
16869__ai int8x16_t vqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
16870  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16871  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16872  int8x16_t __ret;
16873  __ret = (int8x16_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
16874  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
16875  return __ret;
16876}
16877#endif
16878
16879#ifdef __LITTLE_ENDIAN__
16880__ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
16881  int32x4_t __ret;
16882  __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
16883  return __ret;
16884}
16885#else
16886__ai int32x4_t vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
16887  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16888  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16889  int32x4_t __ret;
16890  __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
16891  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
16892  return __ret;
16893}
16894__ai int32x4_t __noswap_vqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
16895  int32x4_t __ret;
16896  __ret = (int32x4_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
16897  return __ret;
16898}
16899#endif
16900
16901#ifdef __LITTLE_ENDIAN__
16902__ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
16903  int64x2_t __ret;
16904  __ret = (int64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
16905  return __ret;
16906}
16907#else
16908__ai int64x2_t vqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
16909  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16910  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16911  int64x2_t __ret;
16912  __ret = (int64x2_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
16913  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16914  return __ret;
16915}
16916#endif
16917
16918#ifdef __LITTLE_ENDIAN__
16919__ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
16920  int16x8_t __ret;
16921  __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
16922  return __ret;
16923}
16924#else
16925__ai int16x8_t vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
16926  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16927  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16928  int16x8_t __ret;
16929  __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
16930  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16931  return __ret;
16932}
16933__ai int16x8_t __noswap_vqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
16934  int16x8_t __ret;
16935  __ret = (int16x8_t) __builtin_neon_vqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
16936  return __ret;
16937}
16938#endif
16939
16940#ifdef __LITTLE_ENDIAN__
16941__ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16942  uint8x8_t __ret;
16943  __ret = (uint8x8_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
16944  return __ret;
16945}
16946#else
16947__ai uint8x8_t vqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
16948  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
16949  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
16950  uint8x8_t __ret;
16951  __ret = (uint8x8_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
16952  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
16953  return __ret;
16954}
16955#endif
16956
16957#ifdef __LITTLE_ENDIAN__
16958__ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16959  uint32x2_t __ret;
16960  __ret = (uint32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
16961  return __ret;
16962}
16963#else
16964__ai uint32x2_t vqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
16965  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
16966  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
16967  uint32x2_t __ret;
16968  __ret = (uint32x2_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
16969  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
16970  return __ret;
16971}
16972#endif
16973
16974#ifdef __LITTLE_ENDIAN__
16975__ai uint64x1_t vqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
16976  uint64x1_t __ret;
16977  __ret = (uint64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
16978  return __ret;
16979}
16980#else
16981__ai uint64x1_t vqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
16982  uint64x1_t __ret;
16983  __ret = (uint64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
16984  return __ret;
16985}
16986#endif
16987
16988#ifdef __LITTLE_ENDIAN__
16989__ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16990  uint16x4_t __ret;
16991  __ret = (uint16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
16992  return __ret;
16993}
16994#else
16995__ai uint16x4_t vqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
16996  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
16997  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
16998  uint16x4_t __ret;
16999  __ret = (uint16x4_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
17000  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17001  return __ret;
17002}
17003#endif
17004
17005#ifdef __LITTLE_ENDIAN__
17006__ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1) {
17007  int8x8_t __ret;
17008  __ret = (int8x8_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
17009  return __ret;
17010}
17011#else
17012__ai int8x8_t vqadd_s8(int8x8_t __p0, int8x8_t __p1) {
17013  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17014  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17015  int8x8_t __ret;
17016  __ret = (int8x8_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
17017  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17018  return __ret;
17019}
17020#endif
17021
17022#ifdef __LITTLE_ENDIAN__
17023__ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
17024  int32x2_t __ret;
17025  __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17026  return __ret;
17027}
17028#else
17029__ai int32x2_t vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
17030  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17031  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17032  int32x2_t __ret;
17033  __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
17034  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17035  return __ret;
17036}
17037__ai int32x2_t __noswap_vqadd_s32(int32x2_t __p0, int32x2_t __p1) {
17038  int32x2_t __ret;
17039  __ret = (int32x2_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17040  return __ret;
17041}
17042#endif
17043
17044#ifdef __LITTLE_ENDIAN__
17045__ai int64x1_t vqadd_s64(int64x1_t __p0, int64x1_t __p1) {
17046  int64x1_t __ret;
17047  __ret = (int64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
17048  return __ret;
17049}
17050#else
17051__ai int64x1_t vqadd_s64(int64x1_t __p0, int64x1_t __p1) {
17052  int64x1_t __ret;
17053  __ret = (int64x1_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
17054  return __ret;
17055}
17056#endif
17057
17058#ifdef __LITTLE_ENDIAN__
17059__ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
17060  int16x4_t __ret;
17061  __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
17062  return __ret;
17063}
17064#else
17065__ai int16x4_t vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
17066  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17067  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17068  int16x4_t __ret;
17069  __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
17070  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17071  return __ret;
17072}
17073__ai int16x4_t __noswap_vqadd_s16(int16x4_t __p0, int16x4_t __p1) {
17074  int16x4_t __ret;
17075  __ret = (int16x4_t) __builtin_neon_vqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
17076  return __ret;
17077}
17078#endif
17079
17080#ifdef __LITTLE_ENDIAN__
17081__ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17082  int64x2_t __ret;
17083  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17084  return __ret;
17085}
17086#else
17087__ai int64x2_t vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17088  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17089  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17090  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
17091  int64x2_t __ret;
17092  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 35);
17093  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17094  return __ret;
17095}
17096__ai int64x2_t __noswap_vqdmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17097  int64x2_t __ret;
17098  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17099  return __ret;
17100}
17101#endif
17102
17103#ifdef __LITTLE_ENDIAN__
17104__ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17105  int32x4_t __ret;
17106  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17107  return __ret;
17108}
17109#else
17110__ai int32x4_t vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17111  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17112  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17113  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
17114  int32x4_t __ret;
17115  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 34);
17116  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17117  return __ret;
17118}
17119__ai int32x4_t __noswap_vqdmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17120  int32x4_t __ret;
17121  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17122  return __ret;
17123}
17124#endif
17125
17126#ifdef __LITTLE_ENDIAN__
17127#define vqdmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17128  int64x2_t __s0 = __p0; \
17129  int32x2_t __s1 = __p1; \
17130  int32x2_t __s2 = __p2; \
17131  int64x2_t __ret; \
17132  __ret = vqdmlal_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
17133  __ret; \
17134})
17135#else
17136#define vqdmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17137  int64x2_t __s0 = __p0; \
17138  int32x2_t __s1 = __p1; \
17139  int32x2_t __s2 = __p2; \
17140  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17141  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17142  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
17143  int64x2_t __ret; \
17144  __ret = __noswap_vqdmlal_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
17145  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17146  __ret; \
17147})
17148#endif
17149
17150#ifdef __LITTLE_ENDIAN__
17151#define vqdmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17152  int32x4_t __s0 = __p0; \
17153  int16x4_t __s1 = __p1; \
17154  int16x4_t __s2 = __p2; \
17155  int32x4_t __ret; \
17156  __ret = vqdmlal_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
17157  __ret; \
17158})
17159#else
17160#define vqdmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17161  int32x4_t __s0 = __p0; \
17162  int16x4_t __s1 = __p1; \
17163  int16x4_t __s2 = __p2; \
17164  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17165  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17166  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
17167  int32x4_t __ret; \
17168  __ret = __noswap_vqdmlal_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
17169  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17170  __ret; \
17171})
17172#endif
17173
17174#ifdef __LITTLE_ENDIAN__
17175__ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17176  int64x2_t __ret;
17177  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17178  return __ret;
17179}
17180#else
17181__ai int64x2_t vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17182  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17183  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17184  int64x2_t __ret;
17185  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17186  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17187  return __ret;
17188}
17189__ai int64x2_t __noswap_vqdmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17190  int64x2_t __ret;
17191  __ret = (int64x2_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17192  return __ret;
17193}
17194#endif
17195
17196#ifdef __LITTLE_ENDIAN__
17197__ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17198  int32x4_t __ret;
17199  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17200  return __ret;
17201}
17202#else
17203__ai int32x4_t vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17204  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17205  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17206  int32x4_t __ret;
17207  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17208  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17209  return __ret;
17210}
17211__ai int32x4_t __noswap_vqdmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17212  int32x4_t __ret;
17213  __ret = (int32x4_t) __builtin_neon_vqdmlal_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17214  return __ret;
17215}
17216#endif
17217
17218#ifdef __LITTLE_ENDIAN__
17219__ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17220  int64x2_t __ret;
17221  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17222  return __ret;
17223}
17224#else
17225__ai int64x2_t vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17226  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17227  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17228  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
17229  int64x2_t __ret;
17230  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 35);
17231  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17232  return __ret;
17233}
17234__ai int64x2_t __noswap_vqdmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
17235  int64x2_t __ret;
17236  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 35);
17237  return __ret;
17238}
17239#endif
17240
17241#ifdef __LITTLE_ENDIAN__
17242__ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17243  int32x4_t __ret;
17244  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17245  return __ret;
17246}
17247#else
17248__ai int32x4_t vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17249  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17250  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17251  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
17252  int32x4_t __ret;
17253  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 34);
17254  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17255  return __ret;
17256}
17257__ai int32x4_t __noswap_vqdmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
17258  int32x4_t __ret;
17259  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 34);
17260  return __ret;
17261}
17262#endif
17263
17264#ifdef __LITTLE_ENDIAN__
17265#define vqdmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17266  int64x2_t __s0 = __p0; \
17267  int32x2_t __s1 = __p1; \
17268  int32x2_t __s2 = __p2; \
17269  int64x2_t __ret; \
17270  __ret = vqdmlsl_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
17271  __ret; \
17272})
17273#else
17274#define vqdmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
17275  int64x2_t __s0 = __p0; \
17276  int32x2_t __s1 = __p1; \
17277  int32x2_t __s2 = __p2; \
17278  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17279  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17280  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
17281  int64x2_t __ret; \
17282  __ret = __noswap_vqdmlsl_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
17283  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17284  __ret; \
17285})
17286#endif
17287
17288#ifdef __LITTLE_ENDIAN__
17289#define vqdmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17290  int32x4_t __s0 = __p0; \
17291  int16x4_t __s1 = __p1; \
17292  int16x4_t __s2 = __p2; \
17293  int32x4_t __ret; \
17294  __ret = vqdmlsl_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
17295  __ret; \
17296})
17297#else
17298#define vqdmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
17299  int32x4_t __s0 = __p0; \
17300  int16x4_t __s1 = __p1; \
17301  int16x4_t __s2 = __p2; \
17302  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17303  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17304  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
17305  int32x4_t __ret; \
17306  __ret = __noswap_vqdmlsl_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
17307  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17308  __ret; \
17309})
17310#endif
17311
17312#ifdef __LITTLE_ENDIAN__
17313__ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17314  int64x2_t __ret;
17315  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17316  return __ret;
17317}
17318#else
17319__ai int64x2_t vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17320  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17321  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17322  int64x2_t __ret;
17323  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17324  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17325  return __ret;
17326}
17327__ai int64x2_t __noswap_vqdmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
17328  int64x2_t __ret;
17329  __ret = (int64x2_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int32x2_t) {__p2, __p2}, 35);
17330  return __ret;
17331}
17332#endif
17333
17334#ifdef __LITTLE_ENDIAN__
17335__ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17336  int32x4_t __ret;
17337  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17338  return __ret;
17339}
17340#else
17341__ai int32x4_t vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17342  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17343  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17344  int32x4_t __ret;
17345  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__rev0, (int8x8_t)__rev1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17346  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17347  return __ret;
17348}
17349__ai int32x4_t __noswap_vqdmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
17350  int32x4_t __ret;
17351  __ret = (int32x4_t) __builtin_neon_vqdmlsl_v((int8x16_t)__p0, (int8x8_t)__p1, (int8x8_t)(int16x4_t) {__p2, __p2, __p2, __p2}, 34);
17352  return __ret;
17353}
17354#endif
17355
17356#ifdef __LITTLE_ENDIAN__
17357__ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17358  int32x4_t __ret;
17359  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
17360  return __ret;
17361}
17362#else
17363__ai int32x4_t vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17364  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17365  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17366  int32x4_t __ret;
17367  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
17368  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17369  return __ret;
17370}
17371__ai int32x4_t __noswap_vqdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
17372  int32x4_t __ret;
17373  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
17374  return __ret;
17375}
17376#endif
17377
17378#ifdef __LITTLE_ENDIAN__
17379__ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17380  int16x8_t __ret;
17381  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
17382  return __ret;
17383}
17384#else
17385__ai int16x8_t vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17386  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17387  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
17388  int16x8_t __ret;
17389  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
17390  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17391  return __ret;
17392}
17393__ai int16x8_t __noswap_vqdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
17394  int16x8_t __ret;
17395  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
17396  return __ret;
17397}
17398#endif
17399
17400#ifdef __LITTLE_ENDIAN__
17401__ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17402  int32x2_t __ret;
17403  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17404  return __ret;
17405}
17406#else
17407__ai int32x2_t vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17408  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17409  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17410  int32x2_t __ret;
17411  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
17412  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17413  return __ret;
17414}
17415__ai int32x2_t __noswap_vqdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
17416  int32x2_t __ret;
17417  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
17418  return __ret;
17419}
17420#endif
17421
17422#ifdef __LITTLE_ENDIAN__
17423__ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17424  int16x4_t __ret;
17425  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
17426  return __ret;
17427}
17428#else
17429__ai int16x4_t vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17430  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17431  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17432  int16x4_t __ret;
17433  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
17434  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17435  return __ret;
17436}
17437__ai int16x4_t __noswap_vqdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
17438  int16x4_t __ret;
17439  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
17440  return __ret;
17441}
17442#endif
17443
17444#ifdef __LITTLE_ENDIAN__
17445#define vqdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17446  int32x4_t __s0 = __p0; \
17447  int32x2_t __s1 = __p1; \
17448  int32x4_t __ret; \
17449  __ret = vqdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
17450  __ret; \
17451})
17452#else
17453#define vqdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17454  int32x4_t __s0 = __p0; \
17455  int32x2_t __s1 = __p1; \
17456  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17457  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17458  int32x4_t __ret; \
17459  __ret = __noswap_vqdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
17460  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17461  __ret; \
17462})
17463#endif
17464
17465#ifdef __LITTLE_ENDIAN__
17466#define vqdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17467  int16x8_t __s0 = __p0; \
17468  int16x4_t __s1 = __p1; \
17469  int16x8_t __ret; \
17470  __ret = vqdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
17471  __ret; \
17472})
17473#else
17474#define vqdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17475  int16x8_t __s0 = __p0; \
17476  int16x4_t __s1 = __p1; \
17477  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
17478  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17479  int16x8_t __ret; \
17480  __ret = __noswap_vqdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
17481  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
17482  __ret; \
17483})
17484#endif
17485
17486#ifdef __LITTLE_ENDIAN__
17487#define vqdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17488  int32x2_t __s0 = __p0; \
17489  int32x2_t __s1 = __p1; \
17490  int32x2_t __ret; \
17491  __ret = vqdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
17492  __ret; \
17493})
17494#else
17495#define vqdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17496  int32x2_t __s0 = __p0; \
17497  int32x2_t __s1 = __p1; \
17498  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17499  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17500  int32x2_t __ret; \
17501  __ret = __noswap_vqdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
17502  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17503  __ret; \
17504})
17505#endif
17506
17507#ifdef __LITTLE_ENDIAN__
17508#define vqdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17509  int16x4_t __s0 = __p0; \
17510  int16x4_t __s1 = __p1; \
17511  int16x4_t __ret; \
17512  __ret = vqdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
17513  __ret; \
17514})
17515#else
17516#define vqdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17517  int16x4_t __s0 = __p0; \
17518  int16x4_t __s1 = __p1; \
17519  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17520  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17521  int16x4_t __ret; \
17522  __ret = __noswap_vqdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
17523  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17524  __ret; \
17525})
17526#endif
17527
17528#ifdef __LITTLE_ENDIAN__
17529__ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
17530  int32x4_t __ret;
17531  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
17532  return __ret;
17533}
17534#else
17535__ai int32x4_t vqdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
17536  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17537  int32x4_t __ret;
17538  __ret = (int32x4_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
17539  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17540  return __ret;
17541}
17542#endif
17543
17544#ifdef __LITTLE_ENDIAN__
17545__ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
17546  int16x8_t __ret;
17547  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__p0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
17548  return __ret;
17549}
17550#else
17551__ai int16x8_t vqdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
17552  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17553  int16x8_t __ret;
17554  __ret = (int16x8_t) __builtin_neon_vqdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
17555  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17556  return __ret;
17557}
17558#endif
17559
17560#ifdef __LITTLE_ENDIAN__
17561__ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
17562  int32x2_t __ret;
17563  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
17564  return __ret;
17565}
17566#else
17567__ai int32x2_t vqdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
17568  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17569  int32x2_t __ret;
17570  __ret = (int32x2_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
17571  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17572  return __ret;
17573}
17574#endif
17575
17576#ifdef __LITTLE_ENDIAN__
17577__ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
17578  int16x4_t __ret;
17579  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
17580  return __ret;
17581}
17582#else
17583__ai int16x4_t vqdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
17584  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17585  int16x4_t __ret;
17586  __ret = (int16x4_t) __builtin_neon_vqdmulh_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
17587  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17588  return __ret;
17589}
17590#endif
17591
17592#ifdef __LITTLE_ENDIAN__
17593__ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
17594  int64x2_t __ret;
17595  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
17596  return __ret;
17597}
17598#else
17599__ai int64x2_t vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
17600  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17601  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
17602  int64x2_t __ret;
17603  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 35);
17604  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17605  return __ret;
17606}
17607__ai int64x2_t __noswap_vqdmull_s32(int32x2_t __p0, int32x2_t __p1) {
17608  int64x2_t __ret;
17609  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 35);
17610  return __ret;
17611}
17612#endif
17613
17614#ifdef __LITTLE_ENDIAN__
17615__ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
17616  int32x4_t __ret;
17617  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
17618  return __ret;
17619}
17620#else
17621__ai int32x4_t vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
17622  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17623  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
17624  int32x4_t __ret;
17625  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)__rev1, 34);
17626  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17627  return __ret;
17628}
17629__ai int32x4_t __noswap_vqdmull_s16(int16x4_t __p0, int16x4_t __p1) {
17630  int32x4_t __ret;
17631  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)__p1, 34);
17632  return __ret;
17633}
17634#endif
17635
17636#ifdef __LITTLE_ENDIAN__
17637#define vqdmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17638  int32x2_t __s0 = __p0; \
17639  int32x2_t __s1 = __p1; \
17640  int64x2_t __ret; \
17641  __ret = vqdmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
17642  __ret; \
17643})
17644#else
17645#define vqdmull_lane_s32(__p0, __p1, __p2) __extension__ ({ \
17646  int32x2_t __s0 = __p0; \
17647  int32x2_t __s1 = __p1; \
17648  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
17649  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
17650  int64x2_t __ret; \
17651  __ret = __noswap_vqdmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
17652  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
17653  __ret; \
17654})
17655#endif
17656
17657#ifdef __LITTLE_ENDIAN__
17658#define vqdmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17659  int16x4_t __s0 = __p0; \
17660  int16x4_t __s1 = __p1; \
17661  int32x4_t __ret; \
17662  __ret = vqdmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
17663  __ret; \
17664})
17665#else
17666#define vqdmull_lane_s16(__p0, __p1, __p2) __extension__ ({ \
17667  int16x4_t __s0 = __p0; \
17668  int16x4_t __s1 = __p1; \
17669  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
17670  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
17671  int32x4_t __ret; \
17672  __ret = __noswap_vqdmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
17673  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
17674  __ret; \
17675})
17676#endif
17677
17678#ifdef __LITTLE_ENDIAN__
17679__ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
17680  int64x2_t __ret;
17681  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
17682  return __ret;
17683}
17684#else
17685__ai int64x2_t vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
17686  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17687  int64x2_t __ret;
17688  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
17689  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17690  return __ret;
17691}
17692__ai int64x2_t __noswap_vqdmull_n_s32(int32x2_t __p0, int32_t __p1) {
17693  int64x2_t __ret;
17694  __ret = (int64x2_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 35);
17695  return __ret;
17696}
17697#endif
17698
17699#ifdef __LITTLE_ENDIAN__
17700__ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
17701  int32x4_t __ret;
17702  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
17703  return __ret;
17704}
17705#else
17706__ai int32x4_t vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
17707  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17708  int32x4_t __ret;
17709  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
17710  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17711  return __ret;
17712}
17713__ai int32x4_t __noswap_vqdmull_n_s16(int16x4_t __p0, int16_t __p1) {
17714  int32x4_t __ret;
17715  __ret = (int32x4_t) __builtin_neon_vqdmull_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 34);
17716  return __ret;
17717}
17718#endif
17719
17720#ifdef __LITTLE_ENDIAN__
17721__ai uint16x4_t vqmovn_u32(uint32x4_t __p0) {
17722  uint16x4_t __ret;
17723  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 17);
17724  return __ret;
17725}
17726#else
17727__ai uint16x4_t vqmovn_u32(uint32x4_t __p0) {
17728  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17729  uint16x4_t __ret;
17730  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 17);
17731  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17732  return __ret;
17733}
17734__ai uint16x4_t __noswap_vqmovn_u32(uint32x4_t __p0) {
17735  uint16x4_t __ret;
17736  __ret = (uint16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 17);
17737  return __ret;
17738}
17739#endif
17740
17741#ifdef __LITTLE_ENDIAN__
17742__ai uint32x2_t vqmovn_u64(uint64x2_t __p0) {
17743  uint32x2_t __ret;
17744  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 18);
17745  return __ret;
17746}
17747#else
17748__ai uint32x2_t vqmovn_u64(uint64x2_t __p0) {
17749  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17750  uint32x2_t __ret;
17751  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 18);
17752  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17753  return __ret;
17754}
17755__ai uint32x2_t __noswap_vqmovn_u64(uint64x2_t __p0) {
17756  uint32x2_t __ret;
17757  __ret = (uint32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 18);
17758  return __ret;
17759}
17760#endif
17761
17762#ifdef __LITTLE_ENDIAN__
17763__ai uint8x8_t vqmovn_u16(uint16x8_t __p0) {
17764  uint8x8_t __ret;
17765  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 16);
17766  return __ret;
17767}
17768#else
17769__ai uint8x8_t vqmovn_u16(uint16x8_t __p0) {
17770  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17771  uint8x8_t __ret;
17772  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 16);
17773  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17774  return __ret;
17775}
17776__ai uint8x8_t __noswap_vqmovn_u16(uint16x8_t __p0) {
17777  uint8x8_t __ret;
17778  __ret = (uint8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 16);
17779  return __ret;
17780}
17781#endif
17782
17783#ifdef __LITTLE_ENDIAN__
17784__ai int16x4_t vqmovn_s32(int32x4_t __p0) {
17785  int16x4_t __ret;
17786  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 1);
17787  return __ret;
17788}
17789#else
17790__ai int16x4_t vqmovn_s32(int32x4_t __p0) {
17791  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17792  int16x4_t __ret;
17793  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 1);
17794  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17795  return __ret;
17796}
17797__ai int16x4_t __noswap_vqmovn_s32(int32x4_t __p0) {
17798  int16x4_t __ret;
17799  __ret = (int16x4_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 1);
17800  return __ret;
17801}
17802#endif
17803
17804#ifdef __LITTLE_ENDIAN__
17805__ai int32x2_t vqmovn_s64(int64x2_t __p0) {
17806  int32x2_t __ret;
17807  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 2);
17808  return __ret;
17809}
17810#else
17811__ai int32x2_t vqmovn_s64(int64x2_t __p0) {
17812  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17813  int32x2_t __ret;
17814  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 2);
17815  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17816  return __ret;
17817}
17818__ai int32x2_t __noswap_vqmovn_s64(int64x2_t __p0) {
17819  int32x2_t __ret;
17820  __ret = (int32x2_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 2);
17821  return __ret;
17822}
17823#endif
17824
17825#ifdef __LITTLE_ENDIAN__
17826__ai int8x8_t vqmovn_s16(int16x8_t __p0) {
17827  int8x8_t __ret;
17828  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 0);
17829  return __ret;
17830}
17831#else
17832__ai int8x8_t vqmovn_s16(int16x8_t __p0) {
17833  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17834  int8x8_t __ret;
17835  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__rev0, 0);
17836  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17837  return __ret;
17838}
17839__ai int8x8_t __noswap_vqmovn_s16(int16x8_t __p0) {
17840  int8x8_t __ret;
17841  __ret = (int8x8_t) __builtin_neon_vqmovn_v((int8x16_t)__p0, 0);
17842  return __ret;
17843}
17844#endif
17845
17846#ifdef __LITTLE_ENDIAN__
17847__ai uint16x4_t vqmovun_s32(int32x4_t __p0) {
17848  uint16x4_t __ret;
17849  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 17);
17850  return __ret;
17851}
17852#else
17853__ai uint16x4_t vqmovun_s32(int32x4_t __p0) {
17854  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17855  uint16x4_t __ret;
17856  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 17);
17857  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17858  return __ret;
17859}
17860__ai uint16x4_t __noswap_vqmovun_s32(int32x4_t __p0) {
17861  uint16x4_t __ret;
17862  __ret = (uint16x4_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 17);
17863  return __ret;
17864}
17865#endif
17866
17867#ifdef __LITTLE_ENDIAN__
17868__ai uint32x2_t vqmovun_s64(int64x2_t __p0) {
17869  uint32x2_t __ret;
17870  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 18);
17871  return __ret;
17872}
17873#else
17874__ai uint32x2_t vqmovun_s64(int64x2_t __p0) {
17875  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17876  uint32x2_t __ret;
17877  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 18);
17878  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17879  return __ret;
17880}
17881__ai uint32x2_t __noswap_vqmovun_s64(int64x2_t __p0) {
17882  uint32x2_t __ret;
17883  __ret = (uint32x2_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 18);
17884  return __ret;
17885}
17886#endif
17887
17888#ifdef __LITTLE_ENDIAN__
17889__ai uint8x8_t vqmovun_s16(int16x8_t __p0) {
17890  uint8x8_t __ret;
17891  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 16);
17892  return __ret;
17893}
17894#else
17895__ai uint8x8_t vqmovun_s16(int16x8_t __p0) {
17896  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17897  uint8x8_t __ret;
17898  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__rev0, 16);
17899  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17900  return __ret;
17901}
17902__ai uint8x8_t __noswap_vqmovun_s16(int16x8_t __p0) {
17903  uint8x8_t __ret;
17904  __ret = (uint8x8_t) __builtin_neon_vqmovun_v((int8x16_t)__p0, 16);
17905  return __ret;
17906}
17907#endif
17908
17909#ifdef __LITTLE_ENDIAN__
17910__ai int8x16_t vqnegq_s8(int8x16_t __p0) {
17911  int8x16_t __ret;
17912  __ret = (int8x16_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 32);
17913  return __ret;
17914}
17915#else
17916__ai int8x16_t vqnegq_s8(int8x16_t __p0) {
17917  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17918  int8x16_t __ret;
17919  __ret = (int8x16_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 32);
17920  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
17921  return __ret;
17922}
17923#endif
17924
17925#ifdef __LITTLE_ENDIAN__
17926__ai int32x4_t vqnegq_s32(int32x4_t __p0) {
17927  int32x4_t __ret;
17928  __ret = (int32x4_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 34);
17929  return __ret;
17930}
17931#else
17932__ai int32x4_t vqnegq_s32(int32x4_t __p0) {
17933  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17934  int32x4_t __ret;
17935  __ret = (int32x4_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 34);
17936  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
17937  return __ret;
17938}
17939#endif
17940
17941#ifdef __LITTLE_ENDIAN__
17942__ai int16x8_t vqnegq_s16(int16x8_t __p0) {
17943  int16x8_t __ret;
17944  __ret = (int16x8_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 33);
17945  return __ret;
17946}
17947#else
17948__ai int16x8_t vqnegq_s16(int16x8_t __p0) {
17949  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17950  int16x8_t __ret;
17951  __ret = (int16x8_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 33);
17952  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17953  return __ret;
17954}
17955#endif
17956
17957#ifdef __LITTLE_ENDIAN__
17958__ai int8x8_t vqneg_s8(int8x8_t __p0) {
17959  int8x8_t __ret;
17960  __ret = (int8x8_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 0);
17961  return __ret;
17962}
17963#else
17964__ai int8x8_t vqneg_s8(int8x8_t __p0) {
17965  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
17966  int8x8_t __ret;
17967  __ret = (int8x8_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 0);
17968  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
17969  return __ret;
17970}
17971#endif
17972
17973#ifdef __LITTLE_ENDIAN__
17974__ai int32x2_t vqneg_s32(int32x2_t __p0) {
17975  int32x2_t __ret;
17976  __ret = (int32x2_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 2);
17977  return __ret;
17978}
17979#else
17980__ai int32x2_t vqneg_s32(int32x2_t __p0) {
17981  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
17982  int32x2_t __ret;
17983  __ret = (int32x2_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 2);
17984  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
17985  return __ret;
17986}
17987#endif
17988
17989#ifdef __LITTLE_ENDIAN__
17990__ai int16x4_t vqneg_s16(int16x4_t __p0) {
17991  int16x4_t __ret;
17992  __ret = (int16x4_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 1);
17993  return __ret;
17994}
17995#else
17996__ai int16x4_t vqneg_s16(int16x4_t __p0) {
17997  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
17998  int16x4_t __ret;
17999  __ret = (int16x4_t) __builtin_neon_vqneg_v((int8x8_t)__rev0, 1);
18000  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18001  return __ret;
18002}
18003#endif
18004
18005#ifdef __LITTLE_ENDIAN__
18006__ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
18007  int32x4_t __ret;
18008  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18009  return __ret;
18010}
18011#else
18012__ai int32x4_t vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
18013  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18014  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18015  int32x4_t __ret;
18016  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
18017  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18018  return __ret;
18019}
18020__ai int32x4_t __noswap_vqrdmulhq_s32(int32x4_t __p0, int32x4_t __p1) {
18021  int32x4_t __ret;
18022  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18023  return __ret;
18024}
18025#endif
18026
18027#ifdef __LITTLE_ENDIAN__
18028__ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
18029  int16x8_t __ret;
18030  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18031  return __ret;
18032}
18033#else
18034__ai int16x8_t vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
18035  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18036  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18037  int16x8_t __ret;
18038  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
18039  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18040  return __ret;
18041}
18042__ai int16x8_t __noswap_vqrdmulhq_s16(int16x8_t __p0, int16x8_t __p1) {
18043  int16x8_t __ret;
18044  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18045  return __ret;
18046}
18047#endif
18048
18049#ifdef __LITTLE_ENDIAN__
18050__ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
18051  int32x2_t __ret;
18052  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18053  return __ret;
18054}
18055#else
18056__ai int32x2_t vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
18057  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18058  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18059  int32x2_t __ret;
18060  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18061  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18062  return __ret;
18063}
18064__ai int32x2_t __noswap_vqrdmulh_s32(int32x2_t __p0, int32x2_t __p1) {
18065  int32x2_t __ret;
18066  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18067  return __ret;
18068}
18069#endif
18070
18071#ifdef __LITTLE_ENDIAN__
18072__ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
18073  int16x4_t __ret;
18074  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18075  return __ret;
18076}
18077#else
18078__ai int16x4_t vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
18079  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18080  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18081  int16x4_t __ret;
18082  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18083  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18084  return __ret;
18085}
18086__ai int16x4_t __noswap_vqrdmulh_s16(int16x4_t __p0, int16x4_t __p1) {
18087  int16x4_t __ret;
18088  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18089  return __ret;
18090}
18091#endif
18092
18093#ifdef __LITTLE_ENDIAN__
18094#define vqrdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18095  int32x4_t __s0 = __p0; \
18096  int32x2_t __s1 = __p1; \
18097  int32x4_t __ret; \
18098  __ret = vqrdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
18099  __ret; \
18100})
18101#else
18102#define vqrdmulhq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18103  int32x4_t __s0 = __p0; \
18104  int32x2_t __s1 = __p1; \
18105  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18106  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
18107  int32x4_t __ret; \
18108  __ret = __noswap_vqrdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
18109  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18110  __ret; \
18111})
18112#endif
18113
18114#ifdef __LITTLE_ENDIAN__
18115#define vqrdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18116  int16x8_t __s0 = __p0; \
18117  int16x4_t __s1 = __p1; \
18118  int16x8_t __ret; \
18119  __ret = vqrdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
18120  __ret; \
18121})
18122#else
18123#define vqrdmulhq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18124  int16x8_t __s0 = __p0; \
18125  int16x4_t __s1 = __p1; \
18126  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18127  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
18128  int16x8_t __ret; \
18129  __ret = __noswap_vqrdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
18130  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18131  __ret; \
18132})
18133#endif
18134
18135#ifdef __LITTLE_ENDIAN__
18136#define vqrdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18137  int32x2_t __s0 = __p0; \
18138  int32x2_t __s1 = __p1; \
18139  int32x2_t __ret; \
18140  __ret = vqrdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
18141  __ret; \
18142})
18143#else
18144#define vqrdmulh_lane_s32(__p0, __p1, __p2) __extension__ ({ \
18145  int32x2_t __s0 = __p0; \
18146  int32x2_t __s1 = __p1; \
18147  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18148  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
18149  int32x2_t __ret; \
18150  __ret = __noswap_vqrdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
18151  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18152  __ret; \
18153})
18154#endif
18155
18156#ifdef __LITTLE_ENDIAN__
18157#define vqrdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18158  int16x4_t __s0 = __p0; \
18159  int16x4_t __s1 = __p1; \
18160  int16x4_t __ret; \
18161  __ret = vqrdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
18162  __ret; \
18163})
18164#else
18165#define vqrdmulh_lane_s16(__p0, __p1, __p2) __extension__ ({ \
18166  int16x4_t __s0 = __p0; \
18167  int16x4_t __s1 = __p1; \
18168  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18169  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
18170  int16x4_t __ret; \
18171  __ret = __noswap_vqrdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
18172  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18173  __ret; \
18174})
18175#endif
18176
18177#ifdef __LITTLE_ENDIAN__
18178__ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
18179  int32x4_t __ret;
18180  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
18181  return __ret;
18182}
18183#else
18184__ai int32x4_t vqrdmulhq_n_s32(int32x4_t __p0, int32_t __p1) {
18185  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18186  int32x4_t __ret;
18187  __ret = (int32x4_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int32x4_t) {__p1, __p1, __p1, __p1}, 34);
18188  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18189  return __ret;
18190}
18191#endif
18192
18193#ifdef __LITTLE_ENDIAN__
18194__ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
18195  int16x8_t __ret;
18196  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__p0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
18197  return __ret;
18198}
18199#else
18200__ai int16x8_t vqrdmulhq_n_s16(int16x8_t __p0, int16_t __p1) {
18201  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18202  int16x8_t __ret;
18203  __ret = (int16x8_t) __builtin_neon_vqrdmulhq_v((int8x16_t)__rev0, (int8x16_t)(int16x8_t) {__p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1}, 33);
18204  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18205  return __ret;
18206}
18207#endif
18208
18209#ifdef __LITTLE_ENDIAN__
18210__ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
18211  int32x2_t __ret;
18212  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
18213  return __ret;
18214}
18215#else
18216__ai int32x2_t vqrdmulh_n_s32(int32x2_t __p0, int32_t __p1) {
18217  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18218  int32x2_t __ret;
18219  __ret = (int32x2_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)(int32x2_t) {__p1, __p1}, 2);
18220  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18221  return __ret;
18222}
18223#endif
18224
18225#ifdef __LITTLE_ENDIAN__
18226__ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
18227  int16x4_t __ret;
18228  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__p0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
18229  return __ret;
18230}
18231#else
18232__ai int16x4_t vqrdmulh_n_s16(int16x4_t __p0, int16_t __p1) {
18233  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18234  int16x4_t __ret;
18235  __ret = (int16x4_t) __builtin_neon_vqrdmulh_v((int8x8_t)__rev0, (int8x8_t)(int16x4_t) {__p1, __p1, __p1, __p1}, 1);
18236  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18237  return __ret;
18238}
18239#endif
18240
18241#ifdef __LITTLE_ENDIAN__
18242__ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18243  uint8x16_t __ret;
18244  __ret = (uint8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
18245  return __ret;
18246}
18247#else
18248__ai uint8x16_t vqrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18249  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18250  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18251  uint8x16_t __ret;
18252  __ret = (uint8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
18253  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18254  return __ret;
18255}
18256#endif
18257
18258#ifdef __LITTLE_ENDIAN__
18259__ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18260  uint32x4_t __ret;
18261  __ret = (uint32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
18262  return __ret;
18263}
18264#else
18265__ai uint32x4_t vqrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18266  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18267  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18268  uint32x4_t __ret;
18269  __ret = (uint32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
18270  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18271  return __ret;
18272}
18273#endif
18274
18275#ifdef __LITTLE_ENDIAN__
18276__ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18277  uint64x2_t __ret;
18278  __ret = (uint64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
18279  return __ret;
18280}
18281#else
18282__ai uint64x2_t vqrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18283  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18284  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18285  uint64x2_t __ret;
18286  __ret = (uint64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
18287  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18288  return __ret;
18289}
18290#endif
18291
18292#ifdef __LITTLE_ENDIAN__
18293__ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18294  uint16x8_t __ret;
18295  __ret = (uint16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
18296  return __ret;
18297}
18298#else
18299__ai uint16x8_t vqrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18300  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18301  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18302  uint16x8_t __ret;
18303  __ret = (uint16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
18304  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18305  return __ret;
18306}
18307#endif
18308
18309#ifdef __LITTLE_ENDIAN__
18310__ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18311  int8x16_t __ret;
18312  __ret = (int8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
18313  return __ret;
18314}
18315#else
18316__ai int8x16_t vqrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18317  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18318  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18319  int8x16_t __ret;
18320  __ret = (int8x16_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
18321  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18322  return __ret;
18323}
18324#endif
18325
18326#ifdef __LITTLE_ENDIAN__
18327__ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18328  int32x4_t __ret;
18329  __ret = (int32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18330  return __ret;
18331}
18332#else
18333__ai int32x4_t vqrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18334  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18335  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18336  int32x4_t __ret;
18337  __ret = (int32x4_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
18338  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18339  return __ret;
18340}
18341#endif
18342
18343#ifdef __LITTLE_ENDIAN__
18344__ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18345  int64x2_t __ret;
18346  __ret = (int64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
18347  return __ret;
18348}
18349#else
18350__ai int64x2_t vqrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18351  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18352  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18353  int64x2_t __ret;
18354  __ret = (int64x2_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
18355  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18356  return __ret;
18357}
18358#endif
18359
18360#ifdef __LITTLE_ENDIAN__
18361__ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18362  int16x8_t __ret;
18363  __ret = (int16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18364  return __ret;
18365}
18366#else
18367__ai int16x8_t vqrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18368  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18369  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18370  int16x8_t __ret;
18371  __ret = (int16x8_t) __builtin_neon_vqrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
18372  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18373  return __ret;
18374}
18375#endif
18376
18377#ifdef __LITTLE_ENDIAN__
18378__ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18379  uint8x8_t __ret;
18380  __ret = (uint8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18381  return __ret;
18382}
18383#else
18384__ai uint8x8_t vqrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18385  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18386  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18387  uint8x8_t __ret;
18388  __ret = (uint8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18389  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18390  return __ret;
18391}
18392#endif
18393
18394#ifdef __LITTLE_ENDIAN__
18395__ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18396  uint32x2_t __ret;
18397  __ret = (uint32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18398  return __ret;
18399}
18400#else
18401__ai uint32x2_t vqrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18402  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18403  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18404  uint32x2_t __ret;
18405  __ret = (uint32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18406  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18407  return __ret;
18408}
18409#endif
18410
18411#ifdef __LITTLE_ENDIAN__
18412__ai uint64x1_t vqrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18413  uint64x1_t __ret;
18414  __ret = (uint64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18415  return __ret;
18416}
18417#else
18418__ai uint64x1_t vqrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18419  uint64x1_t __ret;
18420  __ret = (uint64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18421  return __ret;
18422}
18423#endif
18424
18425#ifdef __LITTLE_ENDIAN__
18426__ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18427  uint16x4_t __ret;
18428  __ret = (uint16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18429  return __ret;
18430}
18431#else
18432__ai uint16x4_t vqrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18433  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18434  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18435  uint16x4_t __ret;
18436  __ret = (uint16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18437  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18438  return __ret;
18439}
18440#endif
18441
18442#ifdef __LITTLE_ENDIAN__
18443__ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1) {
18444  int8x8_t __ret;
18445  __ret = (int8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18446  return __ret;
18447}
18448#else
18449__ai int8x8_t vqrshl_s8(int8x8_t __p0, int8x8_t __p1) {
18450  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18451  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18452  int8x8_t __ret;
18453  __ret = (int8x8_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18454  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18455  return __ret;
18456}
18457#endif
18458
18459#ifdef __LITTLE_ENDIAN__
18460__ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1) {
18461  int32x2_t __ret;
18462  __ret = (int32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18463  return __ret;
18464}
18465#else
18466__ai int32x2_t vqrshl_s32(int32x2_t __p0, int32x2_t __p1) {
18467  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18468  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18469  int32x2_t __ret;
18470  __ret = (int32x2_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18471  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18472  return __ret;
18473}
18474#endif
18475
18476#ifdef __LITTLE_ENDIAN__
18477__ai int64x1_t vqrshl_s64(int64x1_t __p0, int64x1_t __p1) {
18478  int64x1_t __ret;
18479  __ret = (int64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18480  return __ret;
18481}
18482#else
18483__ai int64x1_t vqrshl_s64(int64x1_t __p0, int64x1_t __p1) {
18484  int64x1_t __ret;
18485  __ret = (int64x1_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18486  return __ret;
18487}
18488#endif
18489
18490#ifdef __LITTLE_ENDIAN__
18491__ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1) {
18492  int16x4_t __ret;
18493  __ret = (int16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18494  return __ret;
18495}
18496#else
18497__ai int16x4_t vqrshl_s16(int16x4_t __p0, int16x4_t __p1) {
18498  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18499  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18500  int16x4_t __ret;
18501  __ret = (int16x4_t) __builtin_neon_vqrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18502  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18503  return __ret;
18504}
18505#endif
18506
18507#ifdef __LITTLE_ENDIAN__
18508#define vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
18509  uint32x4_t __s0 = __p0; \
18510  uint16x4_t __ret; \
18511  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 17); \
18512  __ret; \
18513})
18514#else
18515#define vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
18516  uint32x4_t __s0 = __p0; \
18517  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18518  uint16x4_t __ret; \
18519  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 17); \
18520  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18521  __ret; \
18522})
18523#define __noswap_vqrshrn_n_u32(__p0, __p1) __extension__ ({ \
18524  uint32x4_t __s0 = __p0; \
18525  uint16x4_t __ret; \
18526  __ret = (uint16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 17); \
18527  __ret; \
18528})
18529#endif
18530
18531#ifdef __LITTLE_ENDIAN__
18532#define vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
18533  uint64x2_t __s0 = __p0; \
18534  uint32x2_t __ret; \
18535  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 18); \
18536  __ret; \
18537})
18538#else
18539#define vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
18540  uint64x2_t __s0 = __p0; \
18541  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18542  uint32x2_t __ret; \
18543  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 18); \
18544  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18545  __ret; \
18546})
18547#define __noswap_vqrshrn_n_u64(__p0, __p1) __extension__ ({ \
18548  uint64x2_t __s0 = __p0; \
18549  uint32x2_t __ret; \
18550  __ret = (uint32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 18); \
18551  __ret; \
18552})
18553#endif
18554
18555#ifdef __LITTLE_ENDIAN__
18556#define vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
18557  uint16x8_t __s0 = __p0; \
18558  uint8x8_t __ret; \
18559  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 16); \
18560  __ret; \
18561})
18562#else
18563#define vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
18564  uint16x8_t __s0 = __p0; \
18565  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18566  uint8x8_t __ret; \
18567  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 16); \
18568  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18569  __ret; \
18570})
18571#define __noswap_vqrshrn_n_u16(__p0, __p1) __extension__ ({ \
18572  uint16x8_t __s0 = __p0; \
18573  uint8x8_t __ret; \
18574  __ret = (uint8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 16); \
18575  __ret; \
18576})
18577#endif
18578
18579#ifdef __LITTLE_ENDIAN__
18580#define vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
18581  int32x4_t __s0 = __p0; \
18582  int16x4_t __ret; \
18583  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 1); \
18584  __ret; \
18585})
18586#else
18587#define vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
18588  int32x4_t __s0 = __p0; \
18589  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18590  int16x4_t __ret; \
18591  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 1); \
18592  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18593  __ret; \
18594})
18595#define __noswap_vqrshrn_n_s32(__p0, __p1) __extension__ ({ \
18596  int32x4_t __s0 = __p0; \
18597  int16x4_t __ret; \
18598  __ret = (int16x4_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 1); \
18599  __ret; \
18600})
18601#endif
18602
18603#ifdef __LITTLE_ENDIAN__
18604#define vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
18605  int64x2_t __s0 = __p0; \
18606  int32x2_t __ret; \
18607  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 2); \
18608  __ret; \
18609})
18610#else
18611#define vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
18612  int64x2_t __s0 = __p0; \
18613  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18614  int32x2_t __ret; \
18615  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 2); \
18616  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18617  __ret; \
18618})
18619#define __noswap_vqrshrn_n_s64(__p0, __p1) __extension__ ({ \
18620  int64x2_t __s0 = __p0; \
18621  int32x2_t __ret; \
18622  __ret = (int32x2_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 2); \
18623  __ret; \
18624})
18625#endif
18626
18627#ifdef __LITTLE_ENDIAN__
18628#define vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
18629  int16x8_t __s0 = __p0; \
18630  int8x8_t __ret; \
18631  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 0); \
18632  __ret; \
18633})
18634#else
18635#define vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
18636  int16x8_t __s0 = __p0; \
18637  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18638  int8x8_t __ret; \
18639  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__rev0, __p1, 0); \
18640  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18641  __ret; \
18642})
18643#define __noswap_vqrshrn_n_s16(__p0, __p1) __extension__ ({ \
18644  int16x8_t __s0 = __p0; \
18645  int8x8_t __ret; \
18646  __ret = (int8x8_t) __builtin_neon_vqrshrn_n_v((int8x16_t)__s0, __p1, 0); \
18647  __ret; \
18648})
18649#endif
18650
18651#ifdef __LITTLE_ENDIAN__
18652#define vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
18653  int32x4_t __s0 = __p0; \
18654  uint16x4_t __ret; \
18655  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 17); \
18656  __ret; \
18657})
18658#else
18659#define vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
18660  int32x4_t __s0 = __p0; \
18661  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
18662  uint16x4_t __ret; \
18663  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 17); \
18664  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
18665  __ret; \
18666})
18667#define __noswap_vqrshrun_n_s32(__p0, __p1) __extension__ ({ \
18668  int32x4_t __s0 = __p0; \
18669  uint16x4_t __ret; \
18670  __ret = (uint16x4_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 17); \
18671  __ret; \
18672})
18673#endif
18674
18675#ifdef __LITTLE_ENDIAN__
18676#define vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
18677  int64x2_t __s0 = __p0; \
18678  uint32x2_t __ret; \
18679  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 18); \
18680  __ret; \
18681})
18682#else
18683#define vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
18684  int64x2_t __s0 = __p0; \
18685  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
18686  uint32x2_t __ret; \
18687  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 18); \
18688  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
18689  __ret; \
18690})
18691#define __noswap_vqrshrun_n_s64(__p0, __p1) __extension__ ({ \
18692  int64x2_t __s0 = __p0; \
18693  uint32x2_t __ret; \
18694  __ret = (uint32x2_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 18); \
18695  __ret; \
18696})
18697#endif
18698
18699#ifdef __LITTLE_ENDIAN__
18700#define vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
18701  int16x8_t __s0 = __p0; \
18702  uint8x8_t __ret; \
18703  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 16); \
18704  __ret; \
18705})
18706#else
18707#define vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
18708  int16x8_t __s0 = __p0; \
18709  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
18710  uint8x8_t __ret; \
18711  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__rev0, __p1, 16); \
18712  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
18713  __ret; \
18714})
18715#define __noswap_vqrshrun_n_s16(__p0, __p1) __extension__ ({ \
18716  int16x8_t __s0 = __p0; \
18717  uint8x8_t __ret; \
18718  __ret = (uint8x8_t) __builtin_neon_vqrshrun_n_v((int8x16_t)__s0, __p1, 16); \
18719  __ret; \
18720})
18721#endif
18722
18723#ifdef __LITTLE_ENDIAN__
18724__ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18725  uint8x16_t __ret;
18726  __ret = (uint8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
18727  return __ret;
18728}
18729#else
18730__ai uint8x16_t vqshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
18731  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18732  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18733  uint8x16_t __ret;
18734  __ret = (uint8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
18735  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18736  return __ret;
18737}
18738#endif
18739
18740#ifdef __LITTLE_ENDIAN__
18741__ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18742  uint32x4_t __ret;
18743  __ret = (uint32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
18744  return __ret;
18745}
18746#else
18747__ai uint32x4_t vqshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
18748  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18749  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18750  uint32x4_t __ret;
18751  __ret = (uint32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
18752  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18753  return __ret;
18754}
18755#endif
18756
18757#ifdef __LITTLE_ENDIAN__
18758__ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18759  uint64x2_t __ret;
18760  __ret = (uint64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
18761  return __ret;
18762}
18763#else
18764__ai uint64x2_t vqshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
18765  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18766  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18767  uint64x2_t __ret;
18768  __ret = (uint64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
18769  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18770  return __ret;
18771}
18772#endif
18773
18774#ifdef __LITTLE_ENDIAN__
18775__ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18776  uint16x8_t __ret;
18777  __ret = (uint16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
18778  return __ret;
18779}
18780#else
18781__ai uint16x8_t vqshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
18782  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18783  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18784  uint16x8_t __ret;
18785  __ret = (uint16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
18786  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18787  return __ret;
18788}
18789#endif
18790
18791#ifdef __LITTLE_ENDIAN__
18792__ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18793  int8x16_t __ret;
18794  __ret = (int8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
18795  return __ret;
18796}
18797#else
18798__ai int8x16_t vqshlq_s8(int8x16_t __p0, int8x16_t __p1) {
18799  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18800  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18801  int8x16_t __ret;
18802  __ret = (int8x16_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
18803  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
18804  return __ret;
18805}
18806#endif
18807
18808#ifdef __LITTLE_ENDIAN__
18809__ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18810  int32x4_t __ret;
18811  __ret = (int32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
18812  return __ret;
18813}
18814#else
18815__ai int32x4_t vqshlq_s32(int32x4_t __p0, int32x4_t __p1) {
18816  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18817  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18818  int32x4_t __ret;
18819  __ret = (int32x4_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
18820  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18821  return __ret;
18822}
18823#endif
18824
18825#ifdef __LITTLE_ENDIAN__
18826__ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18827  int64x2_t __ret;
18828  __ret = (int64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
18829  return __ret;
18830}
18831#else
18832__ai int64x2_t vqshlq_s64(int64x2_t __p0, int64x2_t __p1) {
18833  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18834  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18835  int64x2_t __ret;
18836  __ret = (int64x2_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
18837  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18838  return __ret;
18839}
18840#endif
18841
18842#ifdef __LITTLE_ENDIAN__
18843__ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18844  int16x8_t __ret;
18845  __ret = (int16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
18846  return __ret;
18847}
18848#else
18849__ai int16x8_t vqshlq_s16(int16x8_t __p0, int16x8_t __p1) {
18850  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18851  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18852  int16x8_t __ret;
18853  __ret = (int16x8_t) __builtin_neon_vqshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
18854  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18855  return __ret;
18856}
18857#endif
18858
18859#ifdef __LITTLE_ENDIAN__
18860__ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18861  uint8x8_t __ret;
18862  __ret = (uint8x8_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
18863  return __ret;
18864}
18865#else
18866__ai uint8x8_t vqshl_u8(uint8x8_t __p0, int8x8_t __p1) {
18867  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18868  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18869  uint8x8_t __ret;
18870  __ret = (uint8x8_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
18871  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18872  return __ret;
18873}
18874#endif
18875
18876#ifdef __LITTLE_ENDIAN__
18877__ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18878  uint32x2_t __ret;
18879  __ret = (uint32x2_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
18880  return __ret;
18881}
18882#else
18883__ai uint32x2_t vqshl_u32(uint32x2_t __p0, int32x2_t __p1) {
18884  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18885  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18886  uint32x2_t __ret;
18887  __ret = (uint32x2_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
18888  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18889  return __ret;
18890}
18891#endif
18892
18893#ifdef __LITTLE_ENDIAN__
18894__ai uint64x1_t vqshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18895  uint64x1_t __ret;
18896  __ret = (uint64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18897  return __ret;
18898}
18899#else
18900__ai uint64x1_t vqshl_u64(uint64x1_t __p0, int64x1_t __p1) {
18901  uint64x1_t __ret;
18902  __ret = (uint64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
18903  return __ret;
18904}
18905#endif
18906
18907#ifdef __LITTLE_ENDIAN__
18908__ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18909  uint16x4_t __ret;
18910  __ret = (uint16x4_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
18911  return __ret;
18912}
18913#else
18914__ai uint16x4_t vqshl_u16(uint16x4_t __p0, int16x4_t __p1) {
18915  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18916  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18917  uint16x4_t __ret;
18918  __ret = (uint16x4_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
18919  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18920  return __ret;
18921}
18922#endif
18923
18924#ifdef __LITTLE_ENDIAN__
18925__ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1) {
18926  int8x8_t __ret;
18927  __ret = (int8x8_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
18928  return __ret;
18929}
18930#else
18931__ai int8x8_t vqshl_s8(int8x8_t __p0, int8x8_t __p1) {
18932  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
18933  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
18934  int8x8_t __ret;
18935  __ret = (int8x8_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
18936  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
18937  return __ret;
18938}
18939#endif
18940
18941#ifdef __LITTLE_ENDIAN__
18942__ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1) {
18943  int32x2_t __ret;
18944  __ret = (int32x2_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
18945  return __ret;
18946}
18947#else
18948__ai int32x2_t vqshl_s32(int32x2_t __p0, int32x2_t __p1) {
18949  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
18950  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
18951  int32x2_t __ret;
18952  __ret = (int32x2_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
18953  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
18954  return __ret;
18955}
18956#endif
18957
18958#ifdef __LITTLE_ENDIAN__
18959__ai int64x1_t vqshl_s64(int64x1_t __p0, int64x1_t __p1) {
18960  int64x1_t __ret;
18961  __ret = (int64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18962  return __ret;
18963}
18964#else
18965__ai int64x1_t vqshl_s64(int64x1_t __p0, int64x1_t __p1) {
18966  int64x1_t __ret;
18967  __ret = (int64x1_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
18968  return __ret;
18969}
18970#endif
18971
18972#ifdef __LITTLE_ENDIAN__
18973__ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1) {
18974  int16x4_t __ret;
18975  __ret = (int16x4_t) __builtin_neon_vqshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
18976  return __ret;
18977}
18978#else
18979__ai int16x4_t vqshl_s16(int16x4_t __p0, int16x4_t __p1) {
18980  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
18981  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
18982  int16x4_t __ret;
18983  __ret = (int16x4_t) __builtin_neon_vqshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
18984  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
18985  return __ret;
18986}
18987#endif
18988
18989#ifdef __LITTLE_ENDIAN__
18990#define vqshlq_n_u8(__p0, __p1) __extension__ ({ \
18991  uint8x16_t __s0 = __p0; \
18992  uint8x16_t __ret; \
18993  __ret = (uint8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 48); \
18994  __ret; \
18995})
18996#else
18997#define vqshlq_n_u8(__p0, __p1) __extension__ ({ \
18998  uint8x16_t __s0 = __p0; \
18999  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19000  uint8x16_t __ret; \
19001  __ret = (uint8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 48); \
19002  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19003  __ret; \
19004})
19005#endif
19006
19007#ifdef __LITTLE_ENDIAN__
19008#define vqshlq_n_u32(__p0, __p1) __extension__ ({ \
19009  uint32x4_t __s0 = __p0; \
19010  uint32x4_t __ret; \
19011  __ret = (uint32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 50); \
19012  __ret; \
19013})
19014#else
19015#define vqshlq_n_u32(__p0, __p1) __extension__ ({ \
19016  uint32x4_t __s0 = __p0; \
19017  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19018  uint32x4_t __ret; \
19019  __ret = (uint32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 50); \
19020  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19021  __ret; \
19022})
19023#endif
19024
19025#ifdef __LITTLE_ENDIAN__
19026#define vqshlq_n_u64(__p0, __p1) __extension__ ({ \
19027  uint64x2_t __s0 = __p0; \
19028  uint64x2_t __ret; \
19029  __ret = (uint64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 51); \
19030  __ret; \
19031})
19032#else
19033#define vqshlq_n_u64(__p0, __p1) __extension__ ({ \
19034  uint64x2_t __s0 = __p0; \
19035  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19036  uint64x2_t __ret; \
19037  __ret = (uint64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 51); \
19038  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19039  __ret; \
19040})
19041#endif
19042
19043#ifdef __LITTLE_ENDIAN__
19044#define vqshlq_n_u16(__p0, __p1) __extension__ ({ \
19045  uint16x8_t __s0 = __p0; \
19046  uint16x8_t __ret; \
19047  __ret = (uint16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 49); \
19048  __ret; \
19049})
19050#else
19051#define vqshlq_n_u16(__p0, __p1) __extension__ ({ \
19052  uint16x8_t __s0 = __p0; \
19053  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19054  uint16x8_t __ret; \
19055  __ret = (uint16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 49); \
19056  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19057  __ret; \
19058})
19059#endif
19060
19061#ifdef __LITTLE_ENDIAN__
19062#define vqshlq_n_s8(__p0, __p1) __extension__ ({ \
19063  int8x16_t __s0 = __p0; \
19064  int8x16_t __ret; \
19065  __ret = (int8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 32); \
19066  __ret; \
19067})
19068#else
19069#define vqshlq_n_s8(__p0, __p1) __extension__ ({ \
19070  int8x16_t __s0 = __p0; \
19071  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19072  int8x16_t __ret; \
19073  __ret = (int8x16_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 32); \
19074  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19075  __ret; \
19076})
19077#endif
19078
19079#ifdef __LITTLE_ENDIAN__
19080#define vqshlq_n_s32(__p0, __p1) __extension__ ({ \
19081  int32x4_t __s0 = __p0; \
19082  int32x4_t __ret; \
19083  __ret = (int32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 34); \
19084  __ret; \
19085})
19086#else
19087#define vqshlq_n_s32(__p0, __p1) __extension__ ({ \
19088  int32x4_t __s0 = __p0; \
19089  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19090  int32x4_t __ret; \
19091  __ret = (int32x4_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 34); \
19092  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19093  __ret; \
19094})
19095#endif
19096
19097#ifdef __LITTLE_ENDIAN__
19098#define vqshlq_n_s64(__p0, __p1) __extension__ ({ \
19099  int64x2_t __s0 = __p0; \
19100  int64x2_t __ret; \
19101  __ret = (int64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 35); \
19102  __ret; \
19103})
19104#else
19105#define vqshlq_n_s64(__p0, __p1) __extension__ ({ \
19106  int64x2_t __s0 = __p0; \
19107  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19108  int64x2_t __ret; \
19109  __ret = (int64x2_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 35); \
19110  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19111  __ret; \
19112})
19113#endif
19114
19115#ifdef __LITTLE_ENDIAN__
19116#define vqshlq_n_s16(__p0, __p1) __extension__ ({ \
19117  int16x8_t __s0 = __p0; \
19118  int16x8_t __ret; \
19119  __ret = (int16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__s0, __p1, 33); \
19120  __ret; \
19121})
19122#else
19123#define vqshlq_n_s16(__p0, __p1) __extension__ ({ \
19124  int16x8_t __s0 = __p0; \
19125  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19126  int16x8_t __ret; \
19127  __ret = (int16x8_t) __builtin_neon_vqshlq_n_v((int8x16_t)__rev0, __p1, 33); \
19128  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19129  __ret; \
19130})
19131#endif
19132
19133#ifdef __LITTLE_ENDIAN__
19134#define vqshl_n_u8(__p0, __p1) __extension__ ({ \
19135  uint8x8_t __s0 = __p0; \
19136  uint8x8_t __ret; \
19137  __ret = (uint8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 16); \
19138  __ret; \
19139})
19140#else
19141#define vqshl_n_u8(__p0, __p1) __extension__ ({ \
19142  uint8x8_t __s0 = __p0; \
19143  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19144  uint8x8_t __ret; \
19145  __ret = (uint8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 16); \
19146  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19147  __ret; \
19148})
19149#endif
19150
19151#ifdef __LITTLE_ENDIAN__
19152#define vqshl_n_u32(__p0, __p1) __extension__ ({ \
19153  uint32x2_t __s0 = __p0; \
19154  uint32x2_t __ret; \
19155  __ret = (uint32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 18); \
19156  __ret; \
19157})
19158#else
19159#define vqshl_n_u32(__p0, __p1) __extension__ ({ \
19160  uint32x2_t __s0 = __p0; \
19161  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19162  uint32x2_t __ret; \
19163  __ret = (uint32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 18); \
19164  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19165  __ret; \
19166})
19167#endif
19168
19169#ifdef __LITTLE_ENDIAN__
19170#define vqshl_n_u64(__p0, __p1) __extension__ ({ \
19171  uint64x1_t __s0 = __p0; \
19172  uint64x1_t __ret; \
19173  __ret = (uint64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 19); \
19174  __ret; \
19175})
19176#else
19177#define vqshl_n_u64(__p0, __p1) __extension__ ({ \
19178  uint64x1_t __s0 = __p0; \
19179  uint64x1_t __ret; \
19180  __ret = (uint64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 19); \
19181  __ret; \
19182})
19183#endif
19184
19185#ifdef __LITTLE_ENDIAN__
19186#define vqshl_n_u16(__p0, __p1) __extension__ ({ \
19187  uint16x4_t __s0 = __p0; \
19188  uint16x4_t __ret; \
19189  __ret = (uint16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 17); \
19190  __ret; \
19191})
19192#else
19193#define vqshl_n_u16(__p0, __p1) __extension__ ({ \
19194  uint16x4_t __s0 = __p0; \
19195  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19196  uint16x4_t __ret; \
19197  __ret = (uint16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 17); \
19198  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19199  __ret; \
19200})
19201#endif
19202
19203#ifdef __LITTLE_ENDIAN__
19204#define vqshl_n_s8(__p0, __p1) __extension__ ({ \
19205  int8x8_t __s0 = __p0; \
19206  int8x8_t __ret; \
19207  __ret = (int8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 0); \
19208  __ret; \
19209})
19210#else
19211#define vqshl_n_s8(__p0, __p1) __extension__ ({ \
19212  int8x8_t __s0 = __p0; \
19213  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19214  int8x8_t __ret; \
19215  __ret = (int8x8_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 0); \
19216  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19217  __ret; \
19218})
19219#endif
19220
19221#ifdef __LITTLE_ENDIAN__
19222#define vqshl_n_s32(__p0, __p1) __extension__ ({ \
19223  int32x2_t __s0 = __p0; \
19224  int32x2_t __ret; \
19225  __ret = (int32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 2); \
19226  __ret; \
19227})
19228#else
19229#define vqshl_n_s32(__p0, __p1) __extension__ ({ \
19230  int32x2_t __s0 = __p0; \
19231  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19232  int32x2_t __ret; \
19233  __ret = (int32x2_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 2); \
19234  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19235  __ret; \
19236})
19237#endif
19238
19239#ifdef __LITTLE_ENDIAN__
19240#define vqshl_n_s64(__p0, __p1) __extension__ ({ \
19241  int64x1_t __s0 = __p0; \
19242  int64x1_t __ret; \
19243  __ret = (int64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 3); \
19244  __ret; \
19245})
19246#else
19247#define vqshl_n_s64(__p0, __p1) __extension__ ({ \
19248  int64x1_t __s0 = __p0; \
19249  int64x1_t __ret; \
19250  __ret = (int64x1_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 3); \
19251  __ret; \
19252})
19253#endif
19254
19255#ifdef __LITTLE_ENDIAN__
19256#define vqshl_n_s16(__p0, __p1) __extension__ ({ \
19257  int16x4_t __s0 = __p0; \
19258  int16x4_t __ret; \
19259  __ret = (int16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__s0, __p1, 1); \
19260  __ret; \
19261})
19262#else
19263#define vqshl_n_s16(__p0, __p1) __extension__ ({ \
19264  int16x4_t __s0 = __p0; \
19265  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19266  int16x4_t __ret; \
19267  __ret = (int16x4_t) __builtin_neon_vqshl_n_v((int8x8_t)__rev0, __p1, 1); \
19268  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19269  __ret; \
19270})
19271#endif
19272
19273#ifdef __LITTLE_ENDIAN__
19274#define vqshluq_n_s8(__p0, __p1) __extension__ ({ \
19275  int8x16_t __s0 = __p0; \
19276  uint8x16_t __ret; \
19277  __ret = (uint8x16_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 48); \
19278  __ret; \
19279})
19280#else
19281#define vqshluq_n_s8(__p0, __p1) __extension__ ({ \
19282  int8x16_t __s0 = __p0; \
19283  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19284  uint8x16_t __ret; \
19285  __ret = (uint8x16_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 48); \
19286  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
19287  __ret; \
19288})
19289#endif
19290
19291#ifdef __LITTLE_ENDIAN__
19292#define vqshluq_n_s32(__p0, __p1) __extension__ ({ \
19293  int32x4_t __s0 = __p0; \
19294  uint32x4_t __ret; \
19295  __ret = (uint32x4_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 50); \
19296  __ret; \
19297})
19298#else
19299#define vqshluq_n_s32(__p0, __p1) __extension__ ({ \
19300  int32x4_t __s0 = __p0; \
19301  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19302  uint32x4_t __ret; \
19303  __ret = (uint32x4_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 50); \
19304  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19305  __ret; \
19306})
19307#endif
19308
19309#ifdef __LITTLE_ENDIAN__
19310#define vqshluq_n_s64(__p0, __p1) __extension__ ({ \
19311  int64x2_t __s0 = __p0; \
19312  uint64x2_t __ret; \
19313  __ret = (uint64x2_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 51); \
19314  __ret; \
19315})
19316#else
19317#define vqshluq_n_s64(__p0, __p1) __extension__ ({ \
19318  int64x2_t __s0 = __p0; \
19319  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19320  uint64x2_t __ret; \
19321  __ret = (uint64x2_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 51); \
19322  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19323  __ret; \
19324})
19325#endif
19326
19327#ifdef __LITTLE_ENDIAN__
19328#define vqshluq_n_s16(__p0, __p1) __extension__ ({ \
19329  int16x8_t __s0 = __p0; \
19330  uint16x8_t __ret; \
19331  __ret = (uint16x8_t) __builtin_neon_vqshluq_n_v((int8x16_t)__s0, __p1, 49); \
19332  __ret; \
19333})
19334#else
19335#define vqshluq_n_s16(__p0, __p1) __extension__ ({ \
19336  int16x8_t __s0 = __p0; \
19337  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19338  uint16x8_t __ret; \
19339  __ret = (uint16x8_t) __builtin_neon_vqshluq_n_v((int8x16_t)__rev0, __p1, 49); \
19340  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19341  __ret; \
19342})
19343#endif
19344
19345#ifdef __LITTLE_ENDIAN__
19346#define vqshlu_n_s8(__p0, __p1) __extension__ ({ \
19347  int8x8_t __s0 = __p0; \
19348  uint8x8_t __ret; \
19349  __ret = (uint8x8_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 16); \
19350  __ret; \
19351})
19352#else
19353#define vqshlu_n_s8(__p0, __p1) __extension__ ({ \
19354  int8x8_t __s0 = __p0; \
19355  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19356  uint8x8_t __ret; \
19357  __ret = (uint8x8_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 16); \
19358  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19359  __ret; \
19360})
19361#endif
19362
19363#ifdef __LITTLE_ENDIAN__
19364#define vqshlu_n_s32(__p0, __p1) __extension__ ({ \
19365  int32x2_t __s0 = __p0; \
19366  uint32x2_t __ret; \
19367  __ret = (uint32x2_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 18); \
19368  __ret; \
19369})
19370#else
19371#define vqshlu_n_s32(__p0, __p1) __extension__ ({ \
19372  int32x2_t __s0 = __p0; \
19373  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19374  uint32x2_t __ret; \
19375  __ret = (uint32x2_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 18); \
19376  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19377  __ret; \
19378})
19379#endif
19380
19381#ifdef __LITTLE_ENDIAN__
19382#define vqshlu_n_s64(__p0, __p1) __extension__ ({ \
19383  int64x1_t __s0 = __p0; \
19384  uint64x1_t __ret; \
19385  __ret = (uint64x1_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 19); \
19386  __ret; \
19387})
19388#else
19389#define vqshlu_n_s64(__p0, __p1) __extension__ ({ \
19390  int64x1_t __s0 = __p0; \
19391  uint64x1_t __ret; \
19392  __ret = (uint64x1_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 19); \
19393  __ret; \
19394})
19395#endif
19396
19397#ifdef __LITTLE_ENDIAN__
19398#define vqshlu_n_s16(__p0, __p1) __extension__ ({ \
19399  int16x4_t __s0 = __p0; \
19400  uint16x4_t __ret; \
19401  __ret = (uint16x4_t) __builtin_neon_vqshlu_n_v((int8x8_t)__s0, __p1, 17); \
19402  __ret; \
19403})
19404#else
19405#define vqshlu_n_s16(__p0, __p1) __extension__ ({ \
19406  int16x4_t __s0 = __p0; \
19407  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19408  uint16x4_t __ret; \
19409  __ret = (uint16x4_t) __builtin_neon_vqshlu_n_v((int8x8_t)__rev0, __p1, 17); \
19410  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19411  __ret; \
19412})
19413#endif
19414
19415#ifdef __LITTLE_ENDIAN__
19416#define vqshrn_n_u32(__p0, __p1) __extension__ ({ \
19417  uint32x4_t __s0 = __p0; \
19418  uint16x4_t __ret; \
19419  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 17); \
19420  __ret; \
19421})
19422#else
19423#define vqshrn_n_u32(__p0, __p1) __extension__ ({ \
19424  uint32x4_t __s0 = __p0; \
19425  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19426  uint16x4_t __ret; \
19427  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 17); \
19428  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19429  __ret; \
19430})
19431#define __noswap_vqshrn_n_u32(__p0, __p1) __extension__ ({ \
19432  uint32x4_t __s0 = __p0; \
19433  uint16x4_t __ret; \
19434  __ret = (uint16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 17); \
19435  __ret; \
19436})
19437#endif
19438
19439#ifdef __LITTLE_ENDIAN__
19440#define vqshrn_n_u64(__p0, __p1) __extension__ ({ \
19441  uint64x2_t __s0 = __p0; \
19442  uint32x2_t __ret; \
19443  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 18); \
19444  __ret; \
19445})
19446#else
19447#define vqshrn_n_u64(__p0, __p1) __extension__ ({ \
19448  uint64x2_t __s0 = __p0; \
19449  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19450  uint32x2_t __ret; \
19451  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 18); \
19452  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19453  __ret; \
19454})
19455#define __noswap_vqshrn_n_u64(__p0, __p1) __extension__ ({ \
19456  uint64x2_t __s0 = __p0; \
19457  uint32x2_t __ret; \
19458  __ret = (uint32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 18); \
19459  __ret; \
19460})
19461#endif
19462
19463#ifdef __LITTLE_ENDIAN__
19464#define vqshrn_n_u16(__p0, __p1) __extension__ ({ \
19465  uint16x8_t __s0 = __p0; \
19466  uint8x8_t __ret; \
19467  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 16); \
19468  __ret; \
19469})
19470#else
19471#define vqshrn_n_u16(__p0, __p1) __extension__ ({ \
19472  uint16x8_t __s0 = __p0; \
19473  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19474  uint8x8_t __ret; \
19475  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 16); \
19476  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19477  __ret; \
19478})
19479#define __noswap_vqshrn_n_u16(__p0, __p1) __extension__ ({ \
19480  uint16x8_t __s0 = __p0; \
19481  uint8x8_t __ret; \
19482  __ret = (uint8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 16); \
19483  __ret; \
19484})
19485#endif
19486
19487#ifdef __LITTLE_ENDIAN__
19488#define vqshrn_n_s32(__p0, __p1) __extension__ ({ \
19489  int32x4_t __s0 = __p0; \
19490  int16x4_t __ret; \
19491  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 1); \
19492  __ret; \
19493})
19494#else
19495#define vqshrn_n_s32(__p0, __p1) __extension__ ({ \
19496  int32x4_t __s0 = __p0; \
19497  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19498  int16x4_t __ret; \
19499  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 1); \
19500  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19501  __ret; \
19502})
19503#define __noswap_vqshrn_n_s32(__p0, __p1) __extension__ ({ \
19504  int32x4_t __s0 = __p0; \
19505  int16x4_t __ret; \
19506  __ret = (int16x4_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 1); \
19507  __ret; \
19508})
19509#endif
19510
19511#ifdef __LITTLE_ENDIAN__
19512#define vqshrn_n_s64(__p0, __p1) __extension__ ({ \
19513  int64x2_t __s0 = __p0; \
19514  int32x2_t __ret; \
19515  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 2); \
19516  __ret; \
19517})
19518#else
19519#define vqshrn_n_s64(__p0, __p1) __extension__ ({ \
19520  int64x2_t __s0 = __p0; \
19521  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19522  int32x2_t __ret; \
19523  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 2); \
19524  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19525  __ret; \
19526})
19527#define __noswap_vqshrn_n_s64(__p0, __p1) __extension__ ({ \
19528  int64x2_t __s0 = __p0; \
19529  int32x2_t __ret; \
19530  __ret = (int32x2_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 2); \
19531  __ret; \
19532})
19533#endif
19534
19535#ifdef __LITTLE_ENDIAN__
19536#define vqshrn_n_s16(__p0, __p1) __extension__ ({ \
19537  int16x8_t __s0 = __p0; \
19538  int8x8_t __ret; \
19539  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 0); \
19540  __ret; \
19541})
19542#else
19543#define vqshrn_n_s16(__p0, __p1) __extension__ ({ \
19544  int16x8_t __s0 = __p0; \
19545  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19546  int8x8_t __ret; \
19547  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__rev0, __p1, 0); \
19548  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19549  __ret; \
19550})
19551#define __noswap_vqshrn_n_s16(__p0, __p1) __extension__ ({ \
19552  int16x8_t __s0 = __p0; \
19553  int8x8_t __ret; \
19554  __ret = (int8x8_t) __builtin_neon_vqshrn_n_v((int8x16_t)__s0, __p1, 0); \
19555  __ret; \
19556})
19557#endif
19558
19559#ifdef __LITTLE_ENDIAN__
19560#define vqshrun_n_s32(__p0, __p1) __extension__ ({ \
19561  int32x4_t __s0 = __p0; \
19562  uint16x4_t __ret; \
19563  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 17); \
19564  __ret; \
19565})
19566#else
19567#define vqshrun_n_s32(__p0, __p1) __extension__ ({ \
19568  int32x4_t __s0 = __p0; \
19569  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
19570  uint16x4_t __ret; \
19571  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 17); \
19572  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
19573  __ret; \
19574})
19575#define __noswap_vqshrun_n_s32(__p0, __p1) __extension__ ({ \
19576  int32x4_t __s0 = __p0; \
19577  uint16x4_t __ret; \
19578  __ret = (uint16x4_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 17); \
19579  __ret; \
19580})
19581#endif
19582
19583#ifdef __LITTLE_ENDIAN__
19584#define vqshrun_n_s64(__p0, __p1) __extension__ ({ \
19585  int64x2_t __s0 = __p0; \
19586  uint32x2_t __ret; \
19587  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 18); \
19588  __ret; \
19589})
19590#else
19591#define vqshrun_n_s64(__p0, __p1) __extension__ ({ \
19592  int64x2_t __s0 = __p0; \
19593  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
19594  uint32x2_t __ret; \
19595  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 18); \
19596  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
19597  __ret; \
19598})
19599#define __noswap_vqshrun_n_s64(__p0, __p1) __extension__ ({ \
19600  int64x2_t __s0 = __p0; \
19601  uint32x2_t __ret; \
19602  __ret = (uint32x2_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 18); \
19603  __ret; \
19604})
19605#endif
19606
19607#ifdef __LITTLE_ENDIAN__
19608#define vqshrun_n_s16(__p0, __p1) __extension__ ({ \
19609  int16x8_t __s0 = __p0; \
19610  uint8x8_t __ret; \
19611  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 16); \
19612  __ret; \
19613})
19614#else
19615#define vqshrun_n_s16(__p0, __p1) __extension__ ({ \
19616  int16x8_t __s0 = __p0; \
19617  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
19618  uint8x8_t __ret; \
19619  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__rev0, __p1, 16); \
19620  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
19621  __ret; \
19622})
19623#define __noswap_vqshrun_n_s16(__p0, __p1) __extension__ ({ \
19624  int16x8_t __s0 = __p0; \
19625  uint8x8_t __ret; \
19626  __ret = (uint8x8_t) __builtin_neon_vqshrun_n_v((int8x16_t)__s0, __p1, 16); \
19627  __ret; \
19628})
19629#endif
19630
19631#ifdef __LITTLE_ENDIAN__
19632__ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
19633  uint8x16_t __ret;
19634  __ret = (uint8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
19635  return __ret;
19636}
19637#else
19638__ai uint8x16_t vqsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
19639  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19640  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19641  uint8x16_t __ret;
19642  __ret = (uint8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
19643  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19644  return __ret;
19645}
19646#endif
19647
19648#ifdef __LITTLE_ENDIAN__
19649__ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
19650  uint32x4_t __ret;
19651  __ret = (uint32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
19652  return __ret;
19653}
19654#else
19655__ai uint32x4_t vqsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
19656  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19657  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19658  uint32x4_t __ret;
19659  __ret = (uint32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
19660  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19661  return __ret;
19662}
19663#endif
19664
19665#ifdef __LITTLE_ENDIAN__
19666__ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
19667  uint64x2_t __ret;
19668  __ret = (uint64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
19669  return __ret;
19670}
19671#else
19672__ai uint64x2_t vqsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
19673  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19674  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19675  uint64x2_t __ret;
19676  __ret = (uint64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
19677  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19678  return __ret;
19679}
19680#endif
19681
19682#ifdef __LITTLE_ENDIAN__
19683__ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
19684  uint16x8_t __ret;
19685  __ret = (uint16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
19686  return __ret;
19687}
19688#else
19689__ai uint16x8_t vqsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
19690  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19691  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19692  uint16x8_t __ret;
19693  __ret = (uint16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
19694  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19695  return __ret;
19696}
19697#endif
19698
19699#ifdef __LITTLE_ENDIAN__
19700__ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1) {
19701  int8x16_t __ret;
19702  __ret = (int8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
19703  return __ret;
19704}
19705#else
19706__ai int8x16_t vqsubq_s8(int8x16_t __p0, int8x16_t __p1) {
19707  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19708  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19709  int8x16_t __ret;
19710  __ret = (int8x16_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
19711  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
19712  return __ret;
19713}
19714#endif
19715
19716#ifdef __LITTLE_ENDIAN__
19717__ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
19718  int32x4_t __ret;
19719  __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19720  return __ret;
19721}
19722#else
19723__ai int32x4_t vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
19724  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19725  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19726  int32x4_t __ret;
19727  __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
19728  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19729  return __ret;
19730}
19731__ai int32x4_t __noswap_vqsubq_s32(int32x4_t __p0, int32x4_t __p1) {
19732  int32x4_t __ret;
19733  __ret = (int32x4_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
19734  return __ret;
19735}
19736#endif
19737
19738#ifdef __LITTLE_ENDIAN__
19739__ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1) {
19740  int64x2_t __ret;
19741  __ret = (int64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
19742  return __ret;
19743}
19744#else
19745__ai int64x2_t vqsubq_s64(int64x2_t __p0, int64x2_t __p1) {
19746  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19747  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19748  int64x2_t __ret;
19749  __ret = (int64x2_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
19750  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19751  return __ret;
19752}
19753#endif
19754
19755#ifdef __LITTLE_ENDIAN__
19756__ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
19757  int16x8_t __ret;
19758  __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19759  return __ret;
19760}
19761#else
19762__ai int16x8_t vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
19763  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19764  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19765  int16x8_t __ret;
19766  __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
19767  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19768  return __ret;
19769}
19770__ai int16x8_t __noswap_vqsubq_s16(int16x8_t __p0, int16x8_t __p1) {
19771  int16x8_t __ret;
19772  __ret = (int16x8_t) __builtin_neon_vqsubq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
19773  return __ret;
19774}
19775#endif
19776
19777#ifdef __LITTLE_ENDIAN__
19778__ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
19779  uint8x8_t __ret;
19780  __ret = (uint8x8_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
19781  return __ret;
19782}
19783#else
19784__ai uint8x8_t vqsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
19785  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19786  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19787  uint8x8_t __ret;
19788  __ret = (uint8x8_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
19789  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19790  return __ret;
19791}
19792#endif
19793
19794#ifdef __LITTLE_ENDIAN__
19795__ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
19796  uint32x2_t __ret;
19797  __ret = (uint32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
19798  return __ret;
19799}
19800#else
19801__ai uint32x2_t vqsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
19802  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19803  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19804  uint32x2_t __ret;
19805  __ret = (uint32x2_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
19806  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19807  return __ret;
19808}
19809#endif
19810
19811#ifdef __LITTLE_ENDIAN__
19812__ai uint64x1_t vqsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
19813  uint64x1_t __ret;
19814  __ret = (uint64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
19815  return __ret;
19816}
19817#else
19818__ai uint64x1_t vqsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
19819  uint64x1_t __ret;
19820  __ret = (uint64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
19821  return __ret;
19822}
19823#endif
19824
19825#ifdef __LITTLE_ENDIAN__
19826__ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
19827  uint16x4_t __ret;
19828  __ret = (uint16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
19829  return __ret;
19830}
19831#else
19832__ai uint16x4_t vqsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
19833  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19834  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19835  uint16x4_t __ret;
19836  __ret = (uint16x4_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
19837  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19838  return __ret;
19839}
19840#endif
19841
19842#ifdef __LITTLE_ENDIAN__
19843__ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1) {
19844  int8x8_t __ret;
19845  __ret = (int8x8_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
19846  return __ret;
19847}
19848#else
19849__ai int8x8_t vqsub_s8(int8x8_t __p0, int8x8_t __p1) {
19850  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19851  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19852  int8x8_t __ret;
19853  __ret = (int8x8_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
19854  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19855  return __ret;
19856}
19857#endif
19858
19859#ifdef __LITTLE_ENDIAN__
19860__ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
19861  int32x2_t __ret;
19862  __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19863  return __ret;
19864}
19865#else
19866__ai int32x2_t vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
19867  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19868  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19869  int32x2_t __ret;
19870  __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
19871  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19872  return __ret;
19873}
19874__ai int32x2_t __noswap_vqsub_s32(int32x2_t __p0, int32x2_t __p1) {
19875  int32x2_t __ret;
19876  __ret = (int32x2_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
19877  return __ret;
19878}
19879#endif
19880
19881#ifdef __LITTLE_ENDIAN__
19882__ai int64x1_t vqsub_s64(int64x1_t __p0, int64x1_t __p1) {
19883  int64x1_t __ret;
19884  __ret = (int64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
19885  return __ret;
19886}
19887#else
19888__ai int64x1_t vqsub_s64(int64x1_t __p0, int64x1_t __p1) {
19889  int64x1_t __ret;
19890  __ret = (int64x1_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
19891  return __ret;
19892}
19893#endif
19894
19895#ifdef __LITTLE_ENDIAN__
19896__ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
19897  int16x4_t __ret;
19898  __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19899  return __ret;
19900}
19901#else
19902__ai int16x4_t vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
19903  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19904  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19905  int16x4_t __ret;
19906  __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
19907  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19908  return __ret;
19909}
19910__ai int16x4_t __noswap_vqsub_s16(int16x4_t __p0, int16x4_t __p1) {
19911  int16x4_t __ret;
19912  __ret = (int16x4_t) __builtin_neon_vqsub_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
19913  return __ret;
19914}
19915#endif
19916
19917#ifdef __LITTLE_ENDIAN__
19918__ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
19919  uint16x4_t __ret;
19920  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
19921  return __ret;
19922}
19923#else
19924__ai uint16x4_t vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
19925  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19926  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19927  uint16x4_t __ret;
19928  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
19929  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19930  return __ret;
19931}
19932__ai uint16x4_t __noswap_vraddhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
19933  uint16x4_t __ret;
19934  __ret = (uint16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
19935  return __ret;
19936}
19937#endif
19938
19939#ifdef __LITTLE_ENDIAN__
19940__ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
19941  uint32x2_t __ret;
19942  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
19943  return __ret;
19944}
19945#else
19946__ai uint32x2_t vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
19947  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
19948  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
19949  uint32x2_t __ret;
19950  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
19951  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
19952  return __ret;
19953}
19954__ai uint32x2_t __noswap_vraddhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
19955  uint32x2_t __ret;
19956  __ret = (uint32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
19957  return __ret;
19958}
19959#endif
19960
19961#ifdef __LITTLE_ENDIAN__
19962__ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
19963  uint8x8_t __ret;
19964  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
19965  return __ret;
19966}
19967#else
19968__ai uint8x8_t vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
19969  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
19970  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
19971  uint8x8_t __ret;
19972  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
19973  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
19974  return __ret;
19975}
19976__ai uint8x8_t __noswap_vraddhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
19977  uint8x8_t __ret;
19978  __ret = (uint8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
19979  return __ret;
19980}
19981#endif
19982
19983#ifdef __LITTLE_ENDIAN__
19984__ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
19985  int16x4_t __ret;
19986  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
19987  return __ret;
19988}
19989#else
19990__ai int16x4_t vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
19991  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
19992  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
19993  int16x4_t __ret;
19994  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
19995  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
19996  return __ret;
19997}
19998__ai int16x4_t __noswap_vraddhn_s32(int32x4_t __p0, int32x4_t __p1) {
19999  int16x4_t __ret;
20000  __ret = (int16x4_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
20001  return __ret;
20002}
20003#endif
20004
20005#ifdef __LITTLE_ENDIAN__
20006__ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
20007  int32x2_t __ret;
20008  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
20009  return __ret;
20010}
20011#else
20012__ai int32x2_t vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
20013  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20014  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20015  int32x2_t __ret;
20016  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
20017  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20018  return __ret;
20019}
20020__ai int32x2_t __noswap_vraddhn_s64(int64x2_t __p0, int64x2_t __p1) {
20021  int32x2_t __ret;
20022  __ret = (int32x2_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
20023  return __ret;
20024}
20025#endif
20026
20027#ifdef __LITTLE_ENDIAN__
20028__ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
20029  int8x8_t __ret;
20030  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
20031  return __ret;
20032}
20033#else
20034__ai int8x8_t vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
20035  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20036  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20037  int8x8_t __ret;
20038  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
20039  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20040  return __ret;
20041}
20042__ai int8x8_t __noswap_vraddhn_s16(int16x8_t __p0, int16x8_t __p1) {
20043  int8x8_t __ret;
20044  __ret = (int8x8_t) __builtin_neon_vraddhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
20045  return __ret;
20046}
20047#endif
20048
20049#ifdef __LITTLE_ENDIAN__
20050__ai uint32x4_t vrecpeq_u32(uint32x4_t __p0) {
20051  uint32x4_t __ret;
20052  __ret = (uint32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 50);
20053  return __ret;
20054}
20055#else
20056__ai uint32x4_t vrecpeq_u32(uint32x4_t __p0) {
20057  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20058  uint32x4_t __ret;
20059  __ret = (uint32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 50);
20060  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20061  return __ret;
20062}
20063#endif
20064
20065#ifdef __LITTLE_ENDIAN__
20066__ai float32x4_t vrecpeq_f32(float32x4_t __p0) {
20067  float32x4_t __ret;
20068  __ret = (float32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 41);
20069  return __ret;
20070}
20071#else
20072__ai float32x4_t vrecpeq_f32(float32x4_t __p0) {
20073  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20074  float32x4_t __ret;
20075  __ret = (float32x4_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 41);
20076  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20077  return __ret;
20078}
20079#endif
20080
20081#ifdef __LITTLE_ENDIAN__
20082__ai uint32x2_t vrecpe_u32(uint32x2_t __p0) {
20083  uint32x2_t __ret;
20084  __ret = (uint32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 18);
20085  return __ret;
20086}
20087#else
20088__ai uint32x2_t vrecpe_u32(uint32x2_t __p0) {
20089  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20090  uint32x2_t __ret;
20091  __ret = (uint32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 18);
20092  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20093  return __ret;
20094}
20095#endif
20096
20097#ifdef __LITTLE_ENDIAN__
20098__ai float32x2_t vrecpe_f32(float32x2_t __p0) {
20099  float32x2_t __ret;
20100  __ret = (float32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 9);
20101  return __ret;
20102}
20103#else
20104__ai float32x2_t vrecpe_f32(float32x2_t __p0) {
20105  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20106  float32x2_t __ret;
20107  __ret = (float32x2_t) __builtin_neon_vrecpe_v((int8x8_t)__rev0, 9);
20108  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20109  return __ret;
20110}
20111#endif
20112
20113#ifdef __LITTLE_ENDIAN__
20114__ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1) {
20115  float32x4_t __ret;
20116  __ret = (float32x4_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
20117  return __ret;
20118}
20119#else
20120__ai float32x4_t vrecpsq_f32(float32x4_t __p0, float32x4_t __p1) {
20121  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20122  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20123  float32x4_t __ret;
20124  __ret = (float32x4_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
20125  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20126  return __ret;
20127}
20128#endif
20129
20130#ifdef __LITTLE_ENDIAN__
20131__ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1) {
20132  float32x2_t __ret;
20133  __ret = (float32x2_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
20134  return __ret;
20135}
20136#else
20137__ai float32x2_t vrecps_f32(float32x2_t __p0, float32x2_t __p1) {
20138  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20139  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20140  float32x2_t __ret;
20141  __ret = (float32x2_t) __builtin_neon_vrecps_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
20142  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20143  return __ret;
20144}
20145#endif
20146
20147#ifdef __LITTLE_ENDIAN__
20148__ai poly8x8_t vrev16_p8(poly8x8_t __p0) {
20149  poly8x8_t __ret;
20150  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20151  return __ret;
20152}
20153#else
20154__ai poly8x8_t vrev16_p8(poly8x8_t __p0) {
20155  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20156  poly8x8_t __ret;
20157  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20158  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20159  return __ret;
20160}
20161#endif
20162
20163#ifdef __LITTLE_ENDIAN__
20164__ai poly8x16_t vrev16q_p8(poly8x16_t __p0) {
20165  poly8x16_t __ret;
20166  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20167  return __ret;
20168}
20169#else
20170__ai poly8x16_t vrev16q_p8(poly8x16_t __p0) {
20171  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20172  poly8x16_t __ret;
20173  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20174  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20175  return __ret;
20176}
20177#endif
20178
20179#ifdef __LITTLE_ENDIAN__
20180__ai uint8x16_t vrev16q_u8(uint8x16_t __p0) {
20181  uint8x16_t __ret;
20182  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20183  return __ret;
20184}
20185#else
20186__ai uint8x16_t vrev16q_u8(uint8x16_t __p0) {
20187  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20188  uint8x16_t __ret;
20189  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20190  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20191  return __ret;
20192}
20193#endif
20194
20195#ifdef __LITTLE_ENDIAN__
20196__ai int8x16_t vrev16q_s8(int8x16_t __p0) {
20197  int8x16_t __ret;
20198  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20199  return __ret;
20200}
20201#else
20202__ai int8x16_t vrev16q_s8(int8x16_t __p0) {
20203  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20204  int8x16_t __ret;
20205  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14);
20206  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20207  return __ret;
20208}
20209#endif
20210
20211#ifdef __LITTLE_ENDIAN__
20212__ai uint8x8_t vrev16_u8(uint8x8_t __p0) {
20213  uint8x8_t __ret;
20214  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20215  return __ret;
20216}
20217#else
20218__ai uint8x8_t vrev16_u8(uint8x8_t __p0) {
20219  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20220  uint8x8_t __ret;
20221  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20222  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20223  return __ret;
20224}
20225#endif
20226
20227#ifdef __LITTLE_ENDIAN__
20228__ai int8x8_t vrev16_s8(int8x8_t __p0) {
20229  int8x8_t __ret;
20230  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20231  return __ret;
20232}
20233#else
20234__ai int8x8_t vrev16_s8(int8x8_t __p0) {
20235  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20236  int8x8_t __ret;
20237  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20238  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20239  return __ret;
20240}
20241#endif
20242
20243#ifdef __LITTLE_ENDIAN__
20244__ai poly8x8_t vrev32_p8(poly8x8_t __p0) {
20245  poly8x8_t __ret;
20246  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20247  return __ret;
20248}
20249#else
20250__ai poly8x8_t vrev32_p8(poly8x8_t __p0) {
20251  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20252  poly8x8_t __ret;
20253  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20254  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20255  return __ret;
20256}
20257#endif
20258
20259#ifdef __LITTLE_ENDIAN__
20260__ai poly16x4_t vrev32_p16(poly16x4_t __p0) {
20261  poly16x4_t __ret;
20262  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20263  return __ret;
20264}
20265#else
20266__ai poly16x4_t vrev32_p16(poly16x4_t __p0) {
20267  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20268  poly16x4_t __ret;
20269  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20270  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20271  return __ret;
20272}
20273#endif
20274
20275#ifdef __LITTLE_ENDIAN__
20276__ai poly8x16_t vrev32q_p8(poly8x16_t __p0) {
20277  poly8x16_t __ret;
20278  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20279  return __ret;
20280}
20281#else
20282__ai poly8x16_t vrev32q_p8(poly8x16_t __p0) {
20283  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20284  poly8x16_t __ret;
20285  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20286  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20287  return __ret;
20288}
20289#endif
20290
20291#ifdef __LITTLE_ENDIAN__
20292__ai poly16x8_t vrev32q_p16(poly16x8_t __p0) {
20293  poly16x8_t __ret;
20294  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20295  return __ret;
20296}
20297#else
20298__ai poly16x8_t vrev32q_p16(poly16x8_t __p0) {
20299  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20300  poly16x8_t __ret;
20301  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20302  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20303  return __ret;
20304}
20305#endif
20306
20307#ifdef __LITTLE_ENDIAN__
20308__ai uint8x16_t vrev32q_u8(uint8x16_t __p0) {
20309  uint8x16_t __ret;
20310  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20311  return __ret;
20312}
20313#else
20314__ai uint8x16_t vrev32q_u8(uint8x16_t __p0) {
20315  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20316  uint8x16_t __ret;
20317  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20318  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20319  return __ret;
20320}
20321#endif
20322
20323#ifdef __LITTLE_ENDIAN__
20324__ai uint16x8_t vrev32q_u16(uint16x8_t __p0) {
20325  uint16x8_t __ret;
20326  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20327  return __ret;
20328}
20329#else
20330__ai uint16x8_t vrev32q_u16(uint16x8_t __p0) {
20331  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20332  uint16x8_t __ret;
20333  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20334  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20335  return __ret;
20336}
20337#endif
20338
20339#ifdef __LITTLE_ENDIAN__
20340__ai int8x16_t vrev32q_s8(int8x16_t __p0) {
20341  int8x16_t __ret;
20342  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20343  return __ret;
20344}
20345#else
20346__ai int8x16_t vrev32q_s8(int8x16_t __p0) {
20347  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20348  int8x16_t __ret;
20349  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12);
20350  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20351  return __ret;
20352}
20353#endif
20354
20355#ifdef __LITTLE_ENDIAN__
20356__ai int16x8_t vrev32q_s16(int16x8_t __p0) {
20357  int16x8_t __ret;
20358  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2, 5, 4, 7, 6);
20359  return __ret;
20360}
20361#else
20362__ai int16x8_t vrev32q_s16(int16x8_t __p0) {
20363  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20364  int16x8_t __ret;
20365  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2, 5, 4, 7, 6);
20366  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20367  return __ret;
20368}
20369#endif
20370
20371#ifdef __LITTLE_ENDIAN__
20372__ai uint8x8_t vrev32_u8(uint8x8_t __p0) {
20373  uint8x8_t __ret;
20374  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20375  return __ret;
20376}
20377#else
20378__ai uint8x8_t vrev32_u8(uint8x8_t __p0) {
20379  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20380  uint8x8_t __ret;
20381  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20382  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20383  return __ret;
20384}
20385#endif
20386
20387#ifdef __LITTLE_ENDIAN__
20388__ai uint16x4_t vrev32_u16(uint16x4_t __p0) {
20389  uint16x4_t __ret;
20390  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20391  return __ret;
20392}
20393#else
20394__ai uint16x4_t vrev32_u16(uint16x4_t __p0) {
20395  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20396  uint16x4_t __ret;
20397  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20398  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20399  return __ret;
20400}
20401#endif
20402
20403#ifdef __LITTLE_ENDIAN__
20404__ai int8x8_t vrev32_s8(int8x8_t __p0) {
20405  int8x8_t __ret;
20406  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20407  return __ret;
20408}
20409#else
20410__ai int8x8_t vrev32_s8(int8x8_t __p0) {
20411  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20412  int8x8_t __ret;
20413  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20414  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20415  return __ret;
20416}
20417#endif
20418
20419#ifdef __LITTLE_ENDIAN__
20420__ai int16x4_t vrev32_s16(int16x4_t __p0) {
20421  int16x4_t __ret;
20422  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20423  return __ret;
20424}
20425#else
20426__ai int16x4_t vrev32_s16(int16x4_t __p0) {
20427  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20428  int16x4_t __ret;
20429  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20430  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20431  return __ret;
20432}
20433#endif
20434
20435#ifdef __LITTLE_ENDIAN__
20436__ai poly8x8_t vrev64_p8(poly8x8_t __p0) {
20437  poly8x8_t __ret;
20438  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20439  return __ret;
20440}
20441#else
20442__ai poly8x8_t vrev64_p8(poly8x8_t __p0) {
20443  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20444  poly8x8_t __ret;
20445  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
20446  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20447  return __ret;
20448}
20449#endif
20450
20451#ifdef __LITTLE_ENDIAN__
20452__ai poly16x4_t vrev64_p16(poly16x4_t __p0) {
20453  poly16x4_t __ret;
20454  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20455  return __ret;
20456}
20457#else
20458__ai poly16x4_t vrev64_p16(poly16x4_t __p0) {
20459  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20460  poly16x4_t __ret;
20461  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
20462  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20463  return __ret;
20464}
20465#endif
20466
20467#ifdef __LITTLE_ENDIAN__
20468__ai poly8x16_t vrev64q_p8(poly8x16_t __p0) {
20469  poly8x16_t __ret;
20470  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20471  return __ret;
20472}
20473#else
20474__ai poly8x16_t vrev64q_p8(poly8x16_t __p0) {
20475  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20476  poly8x16_t __ret;
20477  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20478  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20479  return __ret;
20480}
20481#endif
20482
20483#ifdef __LITTLE_ENDIAN__
20484__ai poly16x8_t vrev64q_p16(poly16x8_t __p0) {
20485  poly16x8_t __ret;
20486  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20487  return __ret;
20488}
20489#else
20490__ai poly16x8_t vrev64q_p16(poly16x8_t __p0) {
20491  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20492  poly16x8_t __ret;
20493  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20494  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20495  return __ret;
20496}
20497#endif
20498
20499#ifdef __LITTLE_ENDIAN__
20500__ai uint8x16_t vrev64q_u8(uint8x16_t __p0) {
20501  uint8x16_t __ret;
20502  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20503  return __ret;
20504}
20505#else
20506__ai uint8x16_t vrev64q_u8(uint8x16_t __p0) {
20507  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20508  uint8x16_t __ret;
20509  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20510  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20511  return __ret;
20512}
20513#endif
20514
20515#ifdef __LITTLE_ENDIAN__
20516__ai uint32x4_t vrev64q_u32(uint32x4_t __p0) {
20517  uint32x4_t __ret;
20518  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20519  return __ret;
20520}
20521#else
20522__ai uint32x4_t vrev64q_u32(uint32x4_t __p0) {
20523  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20524  uint32x4_t __ret;
20525  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20526  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20527  return __ret;
20528}
20529#endif
20530
20531#ifdef __LITTLE_ENDIAN__
20532__ai uint16x8_t vrev64q_u16(uint16x8_t __p0) {
20533  uint16x8_t __ret;
20534  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20535  return __ret;
20536}
20537#else
20538__ai uint16x8_t vrev64q_u16(uint16x8_t __p0) {
20539  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20540  uint16x8_t __ret;
20541  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20542  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20543  return __ret;
20544}
20545#endif
20546
20547#ifdef __LITTLE_ENDIAN__
20548__ai int8x16_t vrev64q_s8(int8x16_t __p0) {
20549  int8x16_t __ret;
20550  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20551  return __ret;
20552}
20553#else
20554__ai int8x16_t vrev64q_s8(int8x16_t __p0) {
20555  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20556  int8x16_t __ret;
20557  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8);
20558  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20559  return __ret;
20560}
20561#endif
20562
20563#ifdef __LITTLE_ENDIAN__
20564__ai float32x4_t vrev64q_f32(float32x4_t __p0) {
20565  float32x4_t __ret;
20566  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20567  return __ret;
20568}
20569#else
20570__ai float32x4_t vrev64q_f32(float32x4_t __p0) {
20571  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20572  float32x4_t __ret;
20573  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20574  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20575  return __ret;
20576}
20577#endif
20578
20579#ifdef __LITTLE_ENDIAN__
20580__ai int32x4_t vrev64q_s32(int32x4_t __p0) {
20581  int32x4_t __ret;
20582  __ret = __builtin_shufflevector(__p0, __p0, 1, 0, 3, 2);
20583  return __ret;
20584}
20585#else
20586__ai int32x4_t vrev64q_s32(int32x4_t __p0) {
20587  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20588  int32x4_t __ret;
20589  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0, 3, 2);
20590  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20591  return __ret;
20592}
20593#endif
20594
20595#ifdef __LITTLE_ENDIAN__
20596__ai int16x8_t vrev64q_s16(int16x8_t __p0) {
20597  int16x8_t __ret;
20598  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0, 7, 6, 5, 4);
20599  return __ret;
20600}
20601#else
20602__ai int16x8_t vrev64q_s16(int16x8_t __p0) {
20603  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20604  int16x8_t __ret;
20605  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0, 7, 6, 5, 4);
20606  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20607  return __ret;
20608}
20609#endif
20610
20611#ifdef __LITTLE_ENDIAN__
20612__ai uint8x8_t vrev64_u8(uint8x8_t __p0) {
20613  uint8x8_t __ret;
20614  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20615  return __ret;
20616}
20617#else
20618__ai uint8x8_t vrev64_u8(uint8x8_t __p0) {
20619  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20620  uint8x8_t __ret;
20621  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
20622  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20623  return __ret;
20624}
20625#endif
20626
20627#ifdef __LITTLE_ENDIAN__
20628__ai uint32x2_t vrev64_u32(uint32x2_t __p0) {
20629  uint32x2_t __ret;
20630  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
20631  return __ret;
20632}
20633#else
20634__ai uint32x2_t vrev64_u32(uint32x2_t __p0) {
20635  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20636  uint32x2_t __ret;
20637  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
20638  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20639  return __ret;
20640}
20641#endif
20642
20643#ifdef __LITTLE_ENDIAN__
20644__ai uint16x4_t vrev64_u16(uint16x4_t __p0) {
20645  uint16x4_t __ret;
20646  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20647  return __ret;
20648}
20649#else
20650__ai uint16x4_t vrev64_u16(uint16x4_t __p0) {
20651  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20652  uint16x4_t __ret;
20653  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
20654  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20655  return __ret;
20656}
20657#endif
20658
20659#ifdef __LITTLE_ENDIAN__
20660__ai int8x8_t vrev64_s8(int8x8_t __p0) {
20661  int8x8_t __ret;
20662  __ret = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20663  return __ret;
20664}
20665#else
20666__ai int8x8_t vrev64_s8(int8x8_t __p0) {
20667  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20668  int8x8_t __ret;
20669  __ret = __builtin_shufflevector(__rev0, __rev0, 7, 6, 5, 4, 3, 2, 1, 0);
20670  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20671  return __ret;
20672}
20673#endif
20674
20675#ifdef __LITTLE_ENDIAN__
20676__ai float32x2_t vrev64_f32(float32x2_t __p0) {
20677  float32x2_t __ret;
20678  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
20679  return __ret;
20680}
20681#else
20682__ai float32x2_t vrev64_f32(float32x2_t __p0) {
20683  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20684  float32x2_t __ret;
20685  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
20686  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20687  return __ret;
20688}
20689#endif
20690
20691#ifdef __LITTLE_ENDIAN__
20692__ai int32x2_t vrev64_s32(int32x2_t __p0) {
20693  int32x2_t __ret;
20694  __ret = __builtin_shufflevector(__p0, __p0, 1, 0);
20695  return __ret;
20696}
20697#else
20698__ai int32x2_t vrev64_s32(int32x2_t __p0) {
20699  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20700  int32x2_t __ret;
20701  __ret = __builtin_shufflevector(__rev0, __rev0, 1, 0);
20702  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20703  return __ret;
20704}
20705#endif
20706
20707#ifdef __LITTLE_ENDIAN__
20708__ai int16x4_t vrev64_s16(int16x4_t __p0) {
20709  int16x4_t __ret;
20710  __ret = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20711  return __ret;
20712}
20713#else
20714__ai int16x4_t vrev64_s16(int16x4_t __p0) {
20715  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20716  int16x4_t __ret;
20717  __ret = __builtin_shufflevector(__rev0, __rev0, 3, 2, 1, 0);
20718  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20719  return __ret;
20720}
20721#endif
20722
20723#ifdef __LITTLE_ENDIAN__
20724__ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
20725  uint8x16_t __ret;
20726  __ret = (uint8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
20727  return __ret;
20728}
20729#else
20730__ai uint8x16_t vrhaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
20731  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20732  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20733  uint8x16_t __ret;
20734  __ret = (uint8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
20735  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20736  return __ret;
20737}
20738#endif
20739
20740#ifdef __LITTLE_ENDIAN__
20741__ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
20742  uint32x4_t __ret;
20743  __ret = (uint32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
20744  return __ret;
20745}
20746#else
20747__ai uint32x4_t vrhaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
20748  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20749  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20750  uint32x4_t __ret;
20751  __ret = (uint32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
20752  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20753  return __ret;
20754}
20755#endif
20756
20757#ifdef __LITTLE_ENDIAN__
20758__ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
20759  uint16x8_t __ret;
20760  __ret = (uint16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
20761  return __ret;
20762}
20763#else
20764__ai uint16x8_t vrhaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
20765  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20766  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20767  uint16x8_t __ret;
20768  __ret = (uint16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
20769  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20770  return __ret;
20771}
20772#endif
20773
20774#ifdef __LITTLE_ENDIAN__
20775__ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
20776  int8x16_t __ret;
20777  __ret = (int8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
20778  return __ret;
20779}
20780#else
20781__ai int8x16_t vrhaddq_s8(int8x16_t __p0, int8x16_t __p1) {
20782  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20783  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20784  int8x16_t __ret;
20785  __ret = (int8x16_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
20786  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20787  return __ret;
20788}
20789#endif
20790
20791#ifdef __LITTLE_ENDIAN__
20792__ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
20793  int32x4_t __ret;
20794  __ret = (int32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
20795  return __ret;
20796}
20797#else
20798__ai int32x4_t vrhaddq_s32(int32x4_t __p0, int32x4_t __p1) {
20799  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20800  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20801  int32x4_t __ret;
20802  __ret = (int32x4_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
20803  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20804  return __ret;
20805}
20806#endif
20807
20808#ifdef __LITTLE_ENDIAN__
20809__ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
20810  int16x8_t __ret;
20811  __ret = (int16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
20812  return __ret;
20813}
20814#else
20815__ai int16x8_t vrhaddq_s16(int16x8_t __p0, int16x8_t __p1) {
20816  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20817  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20818  int16x8_t __ret;
20819  __ret = (int16x8_t) __builtin_neon_vrhaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
20820  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20821  return __ret;
20822}
20823#endif
20824
20825#ifdef __LITTLE_ENDIAN__
20826__ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
20827  uint8x8_t __ret;
20828  __ret = (uint8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
20829  return __ret;
20830}
20831#else
20832__ai uint8x8_t vrhadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
20833  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20834  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20835  uint8x8_t __ret;
20836  __ret = (uint8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
20837  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20838  return __ret;
20839}
20840#endif
20841
20842#ifdef __LITTLE_ENDIAN__
20843__ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
20844  uint32x2_t __ret;
20845  __ret = (uint32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
20846  return __ret;
20847}
20848#else
20849__ai uint32x2_t vrhadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
20850  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20851  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20852  uint32x2_t __ret;
20853  __ret = (uint32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
20854  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20855  return __ret;
20856}
20857#endif
20858
20859#ifdef __LITTLE_ENDIAN__
20860__ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
20861  uint16x4_t __ret;
20862  __ret = (uint16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
20863  return __ret;
20864}
20865#else
20866__ai uint16x4_t vrhadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
20867  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20868  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20869  uint16x4_t __ret;
20870  __ret = (uint16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
20871  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20872  return __ret;
20873}
20874#endif
20875
20876#ifdef __LITTLE_ENDIAN__
20877__ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1) {
20878  int8x8_t __ret;
20879  __ret = (int8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
20880  return __ret;
20881}
20882#else
20883__ai int8x8_t vrhadd_s8(int8x8_t __p0, int8x8_t __p1) {
20884  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20885  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20886  int8x8_t __ret;
20887  __ret = (int8x8_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
20888  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20889  return __ret;
20890}
20891#endif
20892
20893#ifdef __LITTLE_ENDIAN__
20894__ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1) {
20895  int32x2_t __ret;
20896  __ret = (int32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
20897  return __ret;
20898}
20899#else
20900__ai int32x2_t vrhadd_s32(int32x2_t __p0, int32x2_t __p1) {
20901  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20902  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20903  int32x2_t __ret;
20904  __ret = (int32x2_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
20905  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20906  return __ret;
20907}
20908#endif
20909
20910#ifdef __LITTLE_ENDIAN__
20911__ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1) {
20912  int16x4_t __ret;
20913  __ret = (int16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
20914  return __ret;
20915}
20916#else
20917__ai int16x4_t vrhadd_s16(int16x4_t __p0, int16x4_t __p1) {
20918  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20919  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20920  int16x4_t __ret;
20921  __ret = (int16x4_t) __builtin_neon_vrhadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
20922  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20923  return __ret;
20924}
20925#endif
20926
20927#ifdef __LITTLE_ENDIAN__
20928__ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
20929  uint8x16_t __ret;
20930  __ret = (uint8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
20931  return __ret;
20932}
20933#else
20934__ai uint8x16_t vrshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
20935  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20936  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20937  uint8x16_t __ret;
20938  __ret = (uint8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
20939  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
20940  return __ret;
20941}
20942#endif
20943
20944#ifdef __LITTLE_ENDIAN__
20945__ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
20946  uint32x4_t __ret;
20947  __ret = (uint32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
20948  return __ret;
20949}
20950#else
20951__ai uint32x4_t vrshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
20952  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
20953  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
20954  uint32x4_t __ret;
20955  __ret = (uint32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
20956  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
20957  return __ret;
20958}
20959#endif
20960
20961#ifdef __LITTLE_ENDIAN__
20962__ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
20963  uint64x2_t __ret;
20964  __ret = (uint64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
20965  return __ret;
20966}
20967#else
20968__ai uint64x2_t vrshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
20969  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
20970  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
20971  uint64x2_t __ret;
20972  __ret = (uint64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
20973  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
20974  return __ret;
20975}
20976#endif
20977
20978#ifdef __LITTLE_ENDIAN__
20979__ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
20980  uint16x8_t __ret;
20981  __ret = (uint16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
20982  return __ret;
20983}
20984#else
20985__ai uint16x8_t vrshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
20986  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
20987  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
20988  uint16x8_t __ret;
20989  __ret = (uint16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
20990  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
20991  return __ret;
20992}
20993#endif
20994
20995#ifdef __LITTLE_ENDIAN__
20996__ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
20997  int8x16_t __ret;
20998  __ret = (int8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
20999  return __ret;
21000}
21001#else
21002__ai int8x16_t vrshlq_s8(int8x16_t __p0, int8x16_t __p1) {
21003  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21004  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21005  int8x16_t __ret;
21006  __ret = (int8x16_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
21007  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
21008  return __ret;
21009}
21010#endif
21011
21012#ifdef __LITTLE_ENDIAN__
21013__ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
21014  int32x4_t __ret;
21015  __ret = (int32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
21016  return __ret;
21017}
21018#else
21019__ai int32x4_t vrshlq_s32(int32x4_t __p0, int32x4_t __p1) {
21020  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21021  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21022  int32x4_t __ret;
21023  __ret = (int32x4_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
21024  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21025  return __ret;
21026}
21027#endif
21028
21029#ifdef __LITTLE_ENDIAN__
21030__ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
21031  int64x2_t __ret;
21032  __ret = (int64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
21033  return __ret;
21034}
21035#else
21036__ai int64x2_t vrshlq_s64(int64x2_t __p0, int64x2_t __p1) {
21037  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21038  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21039  int64x2_t __ret;
21040  __ret = (int64x2_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
21041  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21042  return __ret;
21043}
21044#endif
21045
21046#ifdef __LITTLE_ENDIAN__
21047__ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
21048  int16x8_t __ret;
21049  __ret = (int16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
21050  return __ret;
21051}
21052#else
21053__ai int16x8_t vrshlq_s16(int16x8_t __p0, int16x8_t __p1) {
21054  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21055  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21056  int16x8_t __ret;
21057  __ret = (int16x8_t) __builtin_neon_vrshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
21058  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21059  return __ret;
21060}
21061#endif
21062
21063#ifdef __LITTLE_ENDIAN__
21064__ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
21065  uint8x8_t __ret;
21066  __ret = (uint8x8_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
21067  return __ret;
21068}
21069#else
21070__ai uint8x8_t vrshl_u8(uint8x8_t __p0, int8x8_t __p1) {
21071  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21072  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21073  uint8x8_t __ret;
21074  __ret = (uint8x8_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
21075  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21076  return __ret;
21077}
21078#endif
21079
21080#ifdef __LITTLE_ENDIAN__
21081__ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
21082  uint32x2_t __ret;
21083  __ret = (uint32x2_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
21084  return __ret;
21085}
21086#else
21087__ai uint32x2_t vrshl_u32(uint32x2_t __p0, int32x2_t __p1) {
21088  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21089  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21090  uint32x2_t __ret;
21091  __ret = (uint32x2_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
21092  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21093  return __ret;
21094}
21095#endif
21096
21097#ifdef __LITTLE_ENDIAN__
21098__ai uint64x1_t vrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
21099  uint64x1_t __ret;
21100  __ret = (uint64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
21101  return __ret;
21102}
21103#else
21104__ai uint64x1_t vrshl_u64(uint64x1_t __p0, int64x1_t __p1) {
21105  uint64x1_t __ret;
21106  __ret = (uint64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
21107  return __ret;
21108}
21109#endif
21110
21111#ifdef __LITTLE_ENDIAN__
21112__ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
21113  uint16x4_t __ret;
21114  __ret = (uint16x4_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
21115  return __ret;
21116}
21117#else
21118__ai uint16x4_t vrshl_u16(uint16x4_t __p0, int16x4_t __p1) {
21119  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21120  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21121  uint16x4_t __ret;
21122  __ret = (uint16x4_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
21123  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21124  return __ret;
21125}
21126#endif
21127
21128#ifdef __LITTLE_ENDIAN__
21129__ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1) {
21130  int8x8_t __ret;
21131  __ret = (int8x8_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
21132  return __ret;
21133}
21134#else
21135__ai int8x8_t vrshl_s8(int8x8_t __p0, int8x8_t __p1) {
21136  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
21137  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
21138  int8x8_t __ret;
21139  __ret = (int8x8_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
21140  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
21141  return __ret;
21142}
21143#endif
21144
21145#ifdef __LITTLE_ENDIAN__
21146__ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1) {
21147  int32x2_t __ret;
21148  __ret = (int32x2_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
21149  return __ret;
21150}
21151#else
21152__ai int32x2_t vrshl_s32(int32x2_t __p0, int32x2_t __p1) {
21153  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21154  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21155  int32x2_t __ret;
21156  __ret = (int32x2_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
21157  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21158  return __ret;
21159}
21160#endif
21161
21162#ifdef __LITTLE_ENDIAN__
21163__ai int64x1_t vrshl_s64(int64x1_t __p0, int64x1_t __p1) {
21164  int64x1_t __ret;
21165  __ret = (int64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
21166  return __ret;
21167}
21168#else
21169__ai int64x1_t vrshl_s64(int64x1_t __p0, int64x1_t __p1) {
21170  int64x1_t __ret;
21171  __ret = (int64x1_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
21172  return __ret;
21173}
21174#endif
21175
21176#ifdef __LITTLE_ENDIAN__
21177__ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1) {
21178  int16x4_t __ret;
21179  __ret = (int16x4_t) __builtin_neon_vrshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
21180  return __ret;
21181}
21182#else
21183__ai int16x4_t vrshl_s16(int16x4_t __p0, int16x4_t __p1) {
21184  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21185  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21186  int16x4_t __ret;
21187  __ret = (int16x4_t) __builtin_neon_vrshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
21188  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21189  return __ret;
21190}
21191#endif
21192
21193#ifdef __LITTLE_ENDIAN__
21194#define vrshrq_n_u8(__p0, __p1) __extension__ ({ \
21195  uint8x16_t __s0 = __p0; \
21196  uint8x16_t __ret; \
21197  __ret = (uint8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 48); \
21198  __ret; \
21199})
21200#else
21201#define vrshrq_n_u8(__p0, __p1) __extension__ ({ \
21202  uint8x16_t __s0 = __p0; \
21203  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21204  uint8x16_t __ret; \
21205  __ret = (uint8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 48); \
21206  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21207  __ret; \
21208})
21209#endif
21210
21211#ifdef __LITTLE_ENDIAN__
21212#define vrshrq_n_u32(__p0, __p1) __extension__ ({ \
21213  uint32x4_t __s0 = __p0; \
21214  uint32x4_t __ret; \
21215  __ret = (uint32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 50); \
21216  __ret; \
21217})
21218#else
21219#define vrshrq_n_u32(__p0, __p1) __extension__ ({ \
21220  uint32x4_t __s0 = __p0; \
21221  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21222  uint32x4_t __ret; \
21223  __ret = (uint32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 50); \
21224  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21225  __ret; \
21226})
21227#endif
21228
21229#ifdef __LITTLE_ENDIAN__
21230#define vrshrq_n_u64(__p0, __p1) __extension__ ({ \
21231  uint64x2_t __s0 = __p0; \
21232  uint64x2_t __ret; \
21233  __ret = (uint64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 51); \
21234  __ret; \
21235})
21236#else
21237#define vrshrq_n_u64(__p0, __p1) __extension__ ({ \
21238  uint64x2_t __s0 = __p0; \
21239  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21240  uint64x2_t __ret; \
21241  __ret = (uint64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 51); \
21242  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21243  __ret; \
21244})
21245#endif
21246
21247#ifdef __LITTLE_ENDIAN__
21248#define vrshrq_n_u16(__p0, __p1) __extension__ ({ \
21249  uint16x8_t __s0 = __p0; \
21250  uint16x8_t __ret; \
21251  __ret = (uint16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 49); \
21252  __ret; \
21253})
21254#else
21255#define vrshrq_n_u16(__p0, __p1) __extension__ ({ \
21256  uint16x8_t __s0 = __p0; \
21257  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21258  uint16x8_t __ret; \
21259  __ret = (uint16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 49); \
21260  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21261  __ret; \
21262})
21263#endif
21264
21265#ifdef __LITTLE_ENDIAN__
21266#define vrshrq_n_s8(__p0, __p1) __extension__ ({ \
21267  int8x16_t __s0 = __p0; \
21268  int8x16_t __ret; \
21269  __ret = (int8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 32); \
21270  __ret; \
21271})
21272#else
21273#define vrshrq_n_s8(__p0, __p1) __extension__ ({ \
21274  int8x16_t __s0 = __p0; \
21275  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21276  int8x16_t __ret; \
21277  __ret = (int8x16_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 32); \
21278  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21279  __ret; \
21280})
21281#endif
21282
21283#ifdef __LITTLE_ENDIAN__
21284#define vrshrq_n_s32(__p0, __p1) __extension__ ({ \
21285  int32x4_t __s0 = __p0; \
21286  int32x4_t __ret; \
21287  __ret = (int32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 34); \
21288  __ret; \
21289})
21290#else
21291#define vrshrq_n_s32(__p0, __p1) __extension__ ({ \
21292  int32x4_t __s0 = __p0; \
21293  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21294  int32x4_t __ret; \
21295  __ret = (int32x4_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 34); \
21296  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21297  __ret; \
21298})
21299#endif
21300
21301#ifdef __LITTLE_ENDIAN__
21302#define vrshrq_n_s64(__p0, __p1) __extension__ ({ \
21303  int64x2_t __s0 = __p0; \
21304  int64x2_t __ret; \
21305  __ret = (int64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 35); \
21306  __ret; \
21307})
21308#else
21309#define vrshrq_n_s64(__p0, __p1) __extension__ ({ \
21310  int64x2_t __s0 = __p0; \
21311  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21312  int64x2_t __ret; \
21313  __ret = (int64x2_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 35); \
21314  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21315  __ret; \
21316})
21317#endif
21318
21319#ifdef __LITTLE_ENDIAN__
21320#define vrshrq_n_s16(__p0, __p1) __extension__ ({ \
21321  int16x8_t __s0 = __p0; \
21322  int16x8_t __ret; \
21323  __ret = (int16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__s0, __p1, 33); \
21324  __ret; \
21325})
21326#else
21327#define vrshrq_n_s16(__p0, __p1) __extension__ ({ \
21328  int16x8_t __s0 = __p0; \
21329  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21330  int16x8_t __ret; \
21331  __ret = (int16x8_t) __builtin_neon_vrshrq_n_v((int8x16_t)__rev0, __p1, 33); \
21332  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21333  __ret; \
21334})
21335#endif
21336
21337#ifdef __LITTLE_ENDIAN__
21338#define vrshr_n_u8(__p0, __p1) __extension__ ({ \
21339  uint8x8_t __s0 = __p0; \
21340  uint8x8_t __ret; \
21341  __ret = (uint8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 16); \
21342  __ret; \
21343})
21344#else
21345#define vrshr_n_u8(__p0, __p1) __extension__ ({ \
21346  uint8x8_t __s0 = __p0; \
21347  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21348  uint8x8_t __ret; \
21349  __ret = (uint8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 16); \
21350  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21351  __ret; \
21352})
21353#endif
21354
21355#ifdef __LITTLE_ENDIAN__
21356#define vrshr_n_u32(__p0, __p1) __extension__ ({ \
21357  uint32x2_t __s0 = __p0; \
21358  uint32x2_t __ret; \
21359  __ret = (uint32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 18); \
21360  __ret; \
21361})
21362#else
21363#define vrshr_n_u32(__p0, __p1) __extension__ ({ \
21364  uint32x2_t __s0 = __p0; \
21365  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21366  uint32x2_t __ret; \
21367  __ret = (uint32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 18); \
21368  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21369  __ret; \
21370})
21371#endif
21372
21373#ifdef __LITTLE_ENDIAN__
21374#define vrshr_n_u64(__p0, __p1) __extension__ ({ \
21375  uint64x1_t __s0 = __p0; \
21376  uint64x1_t __ret; \
21377  __ret = (uint64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 19); \
21378  __ret; \
21379})
21380#else
21381#define vrshr_n_u64(__p0, __p1) __extension__ ({ \
21382  uint64x1_t __s0 = __p0; \
21383  uint64x1_t __ret; \
21384  __ret = (uint64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 19); \
21385  __ret; \
21386})
21387#endif
21388
21389#ifdef __LITTLE_ENDIAN__
21390#define vrshr_n_u16(__p0, __p1) __extension__ ({ \
21391  uint16x4_t __s0 = __p0; \
21392  uint16x4_t __ret; \
21393  __ret = (uint16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 17); \
21394  __ret; \
21395})
21396#else
21397#define vrshr_n_u16(__p0, __p1) __extension__ ({ \
21398  uint16x4_t __s0 = __p0; \
21399  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21400  uint16x4_t __ret; \
21401  __ret = (uint16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 17); \
21402  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21403  __ret; \
21404})
21405#endif
21406
21407#ifdef __LITTLE_ENDIAN__
21408#define vrshr_n_s8(__p0, __p1) __extension__ ({ \
21409  int8x8_t __s0 = __p0; \
21410  int8x8_t __ret; \
21411  __ret = (int8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 0); \
21412  __ret; \
21413})
21414#else
21415#define vrshr_n_s8(__p0, __p1) __extension__ ({ \
21416  int8x8_t __s0 = __p0; \
21417  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21418  int8x8_t __ret; \
21419  __ret = (int8x8_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 0); \
21420  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21421  __ret; \
21422})
21423#endif
21424
21425#ifdef __LITTLE_ENDIAN__
21426#define vrshr_n_s32(__p0, __p1) __extension__ ({ \
21427  int32x2_t __s0 = __p0; \
21428  int32x2_t __ret; \
21429  __ret = (int32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 2); \
21430  __ret; \
21431})
21432#else
21433#define vrshr_n_s32(__p0, __p1) __extension__ ({ \
21434  int32x2_t __s0 = __p0; \
21435  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21436  int32x2_t __ret; \
21437  __ret = (int32x2_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 2); \
21438  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21439  __ret; \
21440})
21441#endif
21442
21443#ifdef __LITTLE_ENDIAN__
21444#define vrshr_n_s64(__p0, __p1) __extension__ ({ \
21445  int64x1_t __s0 = __p0; \
21446  int64x1_t __ret; \
21447  __ret = (int64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 3); \
21448  __ret; \
21449})
21450#else
21451#define vrshr_n_s64(__p0, __p1) __extension__ ({ \
21452  int64x1_t __s0 = __p0; \
21453  int64x1_t __ret; \
21454  __ret = (int64x1_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 3); \
21455  __ret; \
21456})
21457#endif
21458
21459#ifdef __LITTLE_ENDIAN__
21460#define vrshr_n_s16(__p0, __p1) __extension__ ({ \
21461  int16x4_t __s0 = __p0; \
21462  int16x4_t __ret; \
21463  __ret = (int16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__s0, __p1, 1); \
21464  __ret; \
21465})
21466#else
21467#define vrshr_n_s16(__p0, __p1) __extension__ ({ \
21468  int16x4_t __s0 = __p0; \
21469  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21470  int16x4_t __ret; \
21471  __ret = (int16x4_t) __builtin_neon_vrshr_n_v((int8x8_t)__rev0, __p1, 1); \
21472  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21473  __ret; \
21474})
21475#endif
21476
21477#ifdef __LITTLE_ENDIAN__
21478#define vrshrn_n_u32(__p0, __p1) __extension__ ({ \
21479  uint32x4_t __s0 = __p0; \
21480  uint16x4_t __ret; \
21481  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 17); \
21482  __ret; \
21483})
21484#else
21485#define vrshrn_n_u32(__p0, __p1) __extension__ ({ \
21486  uint32x4_t __s0 = __p0; \
21487  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21488  uint16x4_t __ret; \
21489  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 17); \
21490  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21491  __ret; \
21492})
21493#define __noswap_vrshrn_n_u32(__p0, __p1) __extension__ ({ \
21494  uint32x4_t __s0 = __p0; \
21495  uint16x4_t __ret; \
21496  __ret = (uint16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 17); \
21497  __ret; \
21498})
21499#endif
21500
21501#ifdef __LITTLE_ENDIAN__
21502#define vrshrn_n_u64(__p0, __p1) __extension__ ({ \
21503  uint64x2_t __s0 = __p0; \
21504  uint32x2_t __ret; \
21505  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 18); \
21506  __ret; \
21507})
21508#else
21509#define vrshrn_n_u64(__p0, __p1) __extension__ ({ \
21510  uint64x2_t __s0 = __p0; \
21511  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21512  uint32x2_t __ret; \
21513  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 18); \
21514  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21515  __ret; \
21516})
21517#define __noswap_vrshrn_n_u64(__p0, __p1) __extension__ ({ \
21518  uint64x2_t __s0 = __p0; \
21519  uint32x2_t __ret; \
21520  __ret = (uint32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 18); \
21521  __ret; \
21522})
21523#endif
21524
21525#ifdef __LITTLE_ENDIAN__
21526#define vrshrn_n_u16(__p0, __p1) __extension__ ({ \
21527  uint16x8_t __s0 = __p0; \
21528  uint8x8_t __ret; \
21529  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 16); \
21530  __ret; \
21531})
21532#else
21533#define vrshrn_n_u16(__p0, __p1) __extension__ ({ \
21534  uint16x8_t __s0 = __p0; \
21535  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21536  uint8x8_t __ret; \
21537  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 16); \
21538  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21539  __ret; \
21540})
21541#define __noswap_vrshrn_n_u16(__p0, __p1) __extension__ ({ \
21542  uint16x8_t __s0 = __p0; \
21543  uint8x8_t __ret; \
21544  __ret = (uint8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 16); \
21545  __ret; \
21546})
21547#endif
21548
21549#ifdef __LITTLE_ENDIAN__
21550#define vrshrn_n_s32(__p0, __p1) __extension__ ({ \
21551  int32x4_t __s0 = __p0; \
21552  int16x4_t __ret; \
21553  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 1); \
21554  __ret; \
21555})
21556#else
21557#define vrshrn_n_s32(__p0, __p1) __extension__ ({ \
21558  int32x4_t __s0 = __p0; \
21559  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21560  int16x4_t __ret; \
21561  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 1); \
21562  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21563  __ret; \
21564})
21565#define __noswap_vrshrn_n_s32(__p0, __p1) __extension__ ({ \
21566  int32x4_t __s0 = __p0; \
21567  int16x4_t __ret; \
21568  __ret = (int16x4_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 1); \
21569  __ret; \
21570})
21571#endif
21572
21573#ifdef __LITTLE_ENDIAN__
21574#define vrshrn_n_s64(__p0, __p1) __extension__ ({ \
21575  int64x2_t __s0 = __p0; \
21576  int32x2_t __ret; \
21577  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 2); \
21578  __ret; \
21579})
21580#else
21581#define vrshrn_n_s64(__p0, __p1) __extension__ ({ \
21582  int64x2_t __s0 = __p0; \
21583  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21584  int32x2_t __ret; \
21585  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 2); \
21586  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21587  __ret; \
21588})
21589#define __noswap_vrshrn_n_s64(__p0, __p1) __extension__ ({ \
21590  int64x2_t __s0 = __p0; \
21591  int32x2_t __ret; \
21592  __ret = (int32x2_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 2); \
21593  __ret; \
21594})
21595#endif
21596
21597#ifdef __LITTLE_ENDIAN__
21598#define vrshrn_n_s16(__p0, __p1) __extension__ ({ \
21599  int16x8_t __s0 = __p0; \
21600  int8x8_t __ret; \
21601  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 0); \
21602  __ret; \
21603})
21604#else
21605#define vrshrn_n_s16(__p0, __p1) __extension__ ({ \
21606  int16x8_t __s0 = __p0; \
21607  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21608  int8x8_t __ret; \
21609  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__rev0, __p1, 0); \
21610  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21611  __ret; \
21612})
21613#define __noswap_vrshrn_n_s16(__p0, __p1) __extension__ ({ \
21614  int16x8_t __s0 = __p0; \
21615  int8x8_t __ret; \
21616  __ret = (int8x8_t) __builtin_neon_vrshrn_n_v((int8x16_t)__s0, __p1, 0); \
21617  __ret; \
21618})
21619#endif
21620
21621#ifdef __LITTLE_ENDIAN__
21622__ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0) {
21623  uint32x4_t __ret;
21624  __ret = (uint32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 50);
21625  return __ret;
21626}
21627#else
21628__ai uint32x4_t vrsqrteq_u32(uint32x4_t __p0) {
21629  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21630  uint32x4_t __ret;
21631  __ret = (uint32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 50);
21632  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21633  return __ret;
21634}
21635#endif
21636
21637#ifdef __LITTLE_ENDIAN__
21638__ai float32x4_t vrsqrteq_f32(float32x4_t __p0) {
21639  float32x4_t __ret;
21640  __ret = (float32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 41);
21641  return __ret;
21642}
21643#else
21644__ai float32x4_t vrsqrteq_f32(float32x4_t __p0) {
21645  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21646  float32x4_t __ret;
21647  __ret = (float32x4_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 41);
21648  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21649  return __ret;
21650}
21651#endif
21652
21653#ifdef __LITTLE_ENDIAN__
21654__ai uint32x2_t vrsqrte_u32(uint32x2_t __p0) {
21655  uint32x2_t __ret;
21656  __ret = (uint32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 18);
21657  return __ret;
21658}
21659#else
21660__ai uint32x2_t vrsqrte_u32(uint32x2_t __p0) {
21661  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21662  uint32x2_t __ret;
21663  __ret = (uint32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 18);
21664  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21665  return __ret;
21666}
21667#endif
21668
21669#ifdef __LITTLE_ENDIAN__
21670__ai float32x2_t vrsqrte_f32(float32x2_t __p0) {
21671  float32x2_t __ret;
21672  __ret = (float32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 9);
21673  return __ret;
21674}
21675#else
21676__ai float32x2_t vrsqrte_f32(float32x2_t __p0) {
21677  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21678  float32x2_t __ret;
21679  __ret = (float32x2_t) __builtin_neon_vrsqrte_v((int8x8_t)__rev0, 9);
21680  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21681  return __ret;
21682}
21683#endif
21684
21685#ifdef __LITTLE_ENDIAN__
21686__ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1) {
21687  float32x4_t __ret;
21688  __ret = (float32x4_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
21689  return __ret;
21690}
21691#else
21692__ai float32x4_t vrsqrtsq_f32(float32x4_t __p0, float32x4_t __p1) {
21693  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
21694  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
21695  float32x4_t __ret;
21696  __ret = (float32x4_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
21697  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
21698  return __ret;
21699}
21700#endif
21701
21702#ifdef __LITTLE_ENDIAN__
21703__ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1) {
21704  float32x2_t __ret;
21705  __ret = (float32x2_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
21706  return __ret;
21707}
21708#else
21709__ai float32x2_t vrsqrts_f32(float32x2_t __p0, float32x2_t __p1) {
21710  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
21711  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
21712  float32x2_t __ret;
21713  __ret = (float32x2_t) __builtin_neon_vrsqrts_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
21714  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
21715  return __ret;
21716}
21717#endif
21718
21719#ifdef __LITTLE_ENDIAN__
21720#define vrsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
21721  uint8x16_t __s0 = __p0; \
21722  uint8x16_t __s1 = __p1; \
21723  uint8x16_t __ret; \
21724  __ret = (uint8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
21725  __ret; \
21726})
21727#else
21728#define vrsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
21729  uint8x16_t __s0 = __p0; \
21730  uint8x16_t __s1 = __p1; \
21731  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21732  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21733  uint8x16_t __ret; \
21734  __ret = (uint8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
21735  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21736  __ret; \
21737})
21738#endif
21739
21740#ifdef __LITTLE_ENDIAN__
21741#define vrsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
21742  uint32x4_t __s0 = __p0; \
21743  uint32x4_t __s1 = __p1; \
21744  uint32x4_t __ret; \
21745  __ret = (uint32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
21746  __ret; \
21747})
21748#else
21749#define vrsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
21750  uint32x4_t __s0 = __p0; \
21751  uint32x4_t __s1 = __p1; \
21752  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21753  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
21754  uint32x4_t __ret; \
21755  __ret = (uint32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
21756  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21757  __ret; \
21758})
21759#endif
21760
21761#ifdef __LITTLE_ENDIAN__
21762#define vrsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
21763  uint64x2_t __s0 = __p0; \
21764  uint64x2_t __s1 = __p1; \
21765  uint64x2_t __ret; \
21766  __ret = (uint64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
21767  __ret; \
21768})
21769#else
21770#define vrsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
21771  uint64x2_t __s0 = __p0; \
21772  uint64x2_t __s1 = __p1; \
21773  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21774  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
21775  uint64x2_t __ret; \
21776  __ret = (uint64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
21777  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21778  __ret; \
21779})
21780#endif
21781
21782#ifdef __LITTLE_ENDIAN__
21783#define vrsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
21784  uint16x8_t __s0 = __p0; \
21785  uint16x8_t __s1 = __p1; \
21786  uint16x8_t __ret; \
21787  __ret = (uint16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
21788  __ret; \
21789})
21790#else
21791#define vrsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
21792  uint16x8_t __s0 = __p0; \
21793  uint16x8_t __s1 = __p1; \
21794  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21795  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21796  uint16x8_t __ret; \
21797  __ret = (uint16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
21798  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21799  __ret; \
21800})
21801#endif
21802
21803#ifdef __LITTLE_ENDIAN__
21804#define vrsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
21805  int8x16_t __s0 = __p0; \
21806  int8x16_t __s1 = __p1; \
21807  int8x16_t __ret; \
21808  __ret = (int8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
21809  __ret; \
21810})
21811#else
21812#define vrsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
21813  int8x16_t __s0 = __p0; \
21814  int8x16_t __s1 = __p1; \
21815  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21816  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21817  int8x16_t __ret; \
21818  __ret = (int8x16_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
21819  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
21820  __ret; \
21821})
21822#endif
21823
21824#ifdef __LITTLE_ENDIAN__
21825#define vrsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
21826  int32x4_t __s0 = __p0; \
21827  int32x4_t __s1 = __p1; \
21828  int32x4_t __ret; \
21829  __ret = (int32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
21830  __ret; \
21831})
21832#else
21833#define vrsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
21834  int32x4_t __s0 = __p0; \
21835  int32x4_t __s1 = __p1; \
21836  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21837  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
21838  int32x4_t __ret; \
21839  __ret = (int32x4_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
21840  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21841  __ret; \
21842})
21843#endif
21844
21845#ifdef __LITTLE_ENDIAN__
21846#define vrsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
21847  int64x2_t __s0 = __p0; \
21848  int64x2_t __s1 = __p1; \
21849  int64x2_t __ret; \
21850  __ret = (int64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
21851  __ret; \
21852})
21853#else
21854#define vrsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
21855  int64x2_t __s0 = __p0; \
21856  int64x2_t __s1 = __p1; \
21857  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21858  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
21859  int64x2_t __ret; \
21860  __ret = (int64x2_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
21861  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21862  __ret; \
21863})
21864#endif
21865
21866#ifdef __LITTLE_ENDIAN__
21867#define vrsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
21868  int16x8_t __s0 = __p0; \
21869  int16x8_t __s1 = __p1; \
21870  int16x8_t __ret; \
21871  __ret = (int16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
21872  __ret; \
21873})
21874#else
21875#define vrsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
21876  int16x8_t __s0 = __p0; \
21877  int16x8_t __s1 = __p1; \
21878  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21879  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21880  int16x8_t __ret; \
21881  __ret = (int16x8_t) __builtin_neon_vrsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
21882  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21883  __ret; \
21884})
21885#endif
21886
21887#ifdef __LITTLE_ENDIAN__
21888#define vrsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
21889  uint8x8_t __s0 = __p0; \
21890  uint8x8_t __s1 = __p1; \
21891  uint8x8_t __ret; \
21892  __ret = (uint8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
21893  __ret; \
21894})
21895#else
21896#define vrsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
21897  uint8x8_t __s0 = __p0; \
21898  uint8x8_t __s1 = __p1; \
21899  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21900  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21901  uint8x8_t __ret; \
21902  __ret = (uint8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
21903  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21904  __ret; \
21905})
21906#endif
21907
21908#ifdef __LITTLE_ENDIAN__
21909#define vrsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
21910  uint32x2_t __s0 = __p0; \
21911  uint32x2_t __s1 = __p1; \
21912  uint32x2_t __ret; \
21913  __ret = (uint32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
21914  __ret; \
21915})
21916#else
21917#define vrsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
21918  uint32x2_t __s0 = __p0; \
21919  uint32x2_t __s1 = __p1; \
21920  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
21921  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
21922  uint32x2_t __ret; \
21923  __ret = (uint32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
21924  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
21925  __ret; \
21926})
21927#endif
21928
21929#ifdef __LITTLE_ENDIAN__
21930#define vrsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
21931  uint64x1_t __s0 = __p0; \
21932  uint64x1_t __s1 = __p1; \
21933  uint64x1_t __ret; \
21934  __ret = (uint64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
21935  __ret; \
21936})
21937#else
21938#define vrsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
21939  uint64x1_t __s0 = __p0; \
21940  uint64x1_t __s1 = __p1; \
21941  uint64x1_t __ret; \
21942  __ret = (uint64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
21943  __ret; \
21944})
21945#endif
21946
21947#ifdef __LITTLE_ENDIAN__
21948#define vrsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
21949  uint16x4_t __s0 = __p0; \
21950  uint16x4_t __s1 = __p1; \
21951  uint16x4_t __ret; \
21952  __ret = (uint16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
21953  __ret; \
21954})
21955#else
21956#define vrsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
21957  uint16x4_t __s0 = __p0; \
21958  uint16x4_t __s1 = __p1; \
21959  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
21960  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
21961  uint16x4_t __ret; \
21962  __ret = (uint16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
21963  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
21964  __ret; \
21965})
21966#endif
21967
21968#ifdef __LITTLE_ENDIAN__
21969#define vrsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
21970  int8x8_t __s0 = __p0; \
21971  int8x8_t __s1 = __p1; \
21972  int8x8_t __ret; \
21973  __ret = (int8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
21974  __ret; \
21975})
21976#else
21977#define vrsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
21978  int8x8_t __s0 = __p0; \
21979  int8x8_t __s1 = __p1; \
21980  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
21981  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
21982  int8x8_t __ret; \
21983  __ret = (int8x8_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
21984  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
21985  __ret; \
21986})
21987#endif
21988
21989#ifdef __LITTLE_ENDIAN__
21990#define vrsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
21991  int32x2_t __s0 = __p0; \
21992  int32x2_t __s1 = __p1; \
21993  int32x2_t __ret; \
21994  __ret = (int32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
21995  __ret; \
21996})
21997#else
21998#define vrsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
21999  int32x2_t __s0 = __p0; \
22000  int32x2_t __s1 = __p1; \
22001  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
22002  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22003  int32x2_t __ret; \
22004  __ret = (int32x2_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
22005  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22006  __ret; \
22007})
22008#endif
22009
22010#ifdef __LITTLE_ENDIAN__
22011#define vrsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
22012  int64x1_t __s0 = __p0; \
22013  int64x1_t __s1 = __p1; \
22014  int64x1_t __ret; \
22015  __ret = (int64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
22016  __ret; \
22017})
22018#else
22019#define vrsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
22020  int64x1_t __s0 = __p0; \
22021  int64x1_t __s1 = __p1; \
22022  int64x1_t __ret; \
22023  __ret = (int64x1_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
22024  __ret; \
22025})
22026#endif
22027
22028#ifdef __LITTLE_ENDIAN__
22029#define vrsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
22030  int16x4_t __s0 = __p0; \
22031  int16x4_t __s1 = __p1; \
22032  int16x4_t __ret; \
22033  __ret = (int16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
22034  __ret; \
22035})
22036#else
22037#define vrsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
22038  int16x4_t __s0 = __p0; \
22039  int16x4_t __s1 = __p1; \
22040  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
22041  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22042  int16x4_t __ret; \
22043  __ret = (int16x4_t) __builtin_neon_vrsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
22044  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22045  __ret; \
22046})
22047#endif
22048
22049#ifdef __LITTLE_ENDIAN__
22050__ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
22051  uint16x4_t __ret;
22052  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
22053  return __ret;
22054}
22055#else
22056__ai uint16x4_t vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
22057  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22058  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22059  uint16x4_t __ret;
22060  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
22061  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22062  return __ret;
22063}
22064__ai uint16x4_t __noswap_vrsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
22065  uint16x4_t __ret;
22066  __ret = (uint16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
22067  return __ret;
22068}
22069#endif
22070
22071#ifdef __LITTLE_ENDIAN__
22072__ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
22073  uint32x2_t __ret;
22074  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
22075  return __ret;
22076}
22077#else
22078__ai uint32x2_t vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
22079  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22080  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22081  uint32x2_t __ret;
22082  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
22083  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22084  return __ret;
22085}
22086__ai uint32x2_t __noswap_vrsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
22087  uint32x2_t __ret;
22088  __ret = (uint32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
22089  return __ret;
22090}
22091#endif
22092
22093#ifdef __LITTLE_ENDIAN__
22094__ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
22095  uint8x8_t __ret;
22096  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
22097  return __ret;
22098}
22099#else
22100__ai uint8x8_t vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
22101  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22102  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22103  uint8x8_t __ret;
22104  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
22105  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22106  return __ret;
22107}
22108__ai uint8x8_t __noswap_vrsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
22109  uint8x8_t __ret;
22110  __ret = (uint8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
22111  return __ret;
22112}
22113#endif
22114
22115#ifdef __LITTLE_ENDIAN__
22116__ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
22117  int16x4_t __ret;
22118  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
22119  return __ret;
22120}
22121#else
22122__ai int16x4_t vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
22123  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22124  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22125  int16x4_t __ret;
22126  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
22127  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22128  return __ret;
22129}
22130__ai int16x4_t __noswap_vrsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
22131  int16x4_t __ret;
22132  __ret = (int16x4_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
22133  return __ret;
22134}
22135#endif
22136
22137#ifdef __LITTLE_ENDIAN__
22138__ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
22139  int32x2_t __ret;
22140  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
22141  return __ret;
22142}
22143#else
22144__ai int32x2_t vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
22145  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22146  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22147  int32x2_t __ret;
22148  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
22149  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22150  return __ret;
22151}
22152__ai int32x2_t __noswap_vrsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
22153  int32x2_t __ret;
22154  __ret = (int32x2_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
22155  return __ret;
22156}
22157#endif
22158
22159#ifdef __LITTLE_ENDIAN__
22160__ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
22161  int8x8_t __ret;
22162  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
22163  return __ret;
22164}
22165#else
22166__ai int8x8_t vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
22167  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22168  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22169  int8x8_t __ret;
22170  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
22171  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22172  return __ret;
22173}
22174__ai int8x8_t __noswap_vrsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
22175  int8x8_t __ret;
22176  __ret = (int8x8_t) __builtin_neon_vrsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
22177  return __ret;
22178}
22179#endif
22180
22181#ifdef __LITTLE_ENDIAN__
22182#define vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22183  poly8_t __s0 = __p0; \
22184  poly8x8_t __s1 = __p1; \
22185  poly8x8_t __ret; \
22186  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22187  __ret; \
22188})
22189#else
22190#define vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22191  poly8_t __s0 = __p0; \
22192  poly8x8_t __s1 = __p1; \
22193  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22194  poly8x8_t __ret; \
22195  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
22196  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22197  __ret; \
22198})
22199#define __noswap_vset_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22200  poly8_t __s0 = __p0; \
22201  poly8x8_t __s1 = __p1; \
22202  poly8x8_t __ret; \
22203  __ret = (poly8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22204  __ret; \
22205})
22206#endif
22207
22208#ifdef __LITTLE_ENDIAN__
22209#define vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22210  poly16_t __s0 = __p0; \
22211  poly16x4_t __s1 = __p1; \
22212  poly16x4_t __ret; \
22213  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22214  __ret; \
22215})
22216#else
22217#define vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22218  poly16_t __s0 = __p0; \
22219  poly16x4_t __s1 = __p1; \
22220  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22221  poly16x4_t __ret; \
22222  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
22223  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22224  __ret; \
22225})
22226#define __noswap_vset_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22227  poly16_t __s0 = __p0; \
22228  poly16x4_t __s1 = __p1; \
22229  poly16x4_t __ret; \
22230  __ret = (poly16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22231  __ret; \
22232})
22233#endif
22234
22235#ifdef __LITTLE_ENDIAN__
22236#define vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22237  poly8_t __s0 = __p0; \
22238  poly8x16_t __s1 = __p1; \
22239  poly8x16_t __ret; \
22240  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22241  __ret; \
22242})
22243#else
22244#define vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22245  poly8_t __s0 = __p0; \
22246  poly8x16_t __s1 = __p1; \
22247  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22248  poly8x16_t __ret; \
22249  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
22250  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22251  __ret; \
22252})
22253#define __noswap_vsetq_lane_p8(__p0, __p1, __p2) __extension__ ({ \
22254  poly8_t __s0 = __p0; \
22255  poly8x16_t __s1 = __p1; \
22256  poly8x16_t __ret; \
22257  __ret = (poly8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22258  __ret; \
22259})
22260#endif
22261
22262#ifdef __LITTLE_ENDIAN__
22263#define vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22264  poly16_t __s0 = __p0; \
22265  poly16x8_t __s1 = __p1; \
22266  poly16x8_t __ret; \
22267  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22268  __ret; \
22269})
22270#else
22271#define vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22272  poly16_t __s0 = __p0; \
22273  poly16x8_t __s1 = __p1; \
22274  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22275  poly16x8_t __ret; \
22276  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
22277  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22278  __ret; \
22279})
22280#define __noswap_vsetq_lane_p16(__p0, __p1, __p2) __extension__ ({ \
22281  poly16_t __s0 = __p0; \
22282  poly16x8_t __s1 = __p1; \
22283  poly16x8_t __ret; \
22284  __ret = (poly16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22285  __ret; \
22286})
22287#endif
22288
22289#ifdef __LITTLE_ENDIAN__
22290#define vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22291  uint8_t __s0 = __p0; \
22292  uint8x16_t __s1 = __p1; \
22293  uint8x16_t __ret; \
22294  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22295  __ret; \
22296})
22297#else
22298#define vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22299  uint8_t __s0 = __p0; \
22300  uint8x16_t __s1 = __p1; \
22301  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22302  uint8x16_t __ret; \
22303  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
22304  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22305  __ret; \
22306})
22307#define __noswap_vsetq_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22308  uint8_t __s0 = __p0; \
22309  uint8x16_t __s1 = __p1; \
22310  uint8x16_t __ret; \
22311  __ret = (uint8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22312  __ret; \
22313})
22314#endif
22315
22316#ifdef __LITTLE_ENDIAN__
22317#define vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22318  uint32_t __s0 = __p0; \
22319  uint32x4_t __s1 = __p1; \
22320  uint32x4_t __ret; \
22321  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22322  __ret; \
22323})
22324#else
22325#define vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22326  uint32_t __s0 = __p0; \
22327  uint32x4_t __s1 = __p1; \
22328  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22329  uint32x4_t __ret; \
22330  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__rev1, __p2); \
22331  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22332  __ret; \
22333})
22334#define __noswap_vsetq_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22335  uint32_t __s0 = __p0; \
22336  uint32x4_t __s1 = __p1; \
22337  uint32x4_t __ret; \
22338  __ret = (uint32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22339  __ret; \
22340})
22341#endif
22342
22343#ifdef __LITTLE_ENDIAN__
22344#define vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22345  uint64_t __s0 = __p0; \
22346  uint64x2_t __s1 = __p1; \
22347  uint64x2_t __ret; \
22348  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22349  __ret; \
22350})
22351#else
22352#define vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22353  uint64_t __s0 = __p0; \
22354  uint64x2_t __s1 = __p1; \
22355  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22356  uint64x2_t __ret; \
22357  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
22358  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22359  __ret; \
22360})
22361#define __noswap_vsetq_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22362  uint64_t __s0 = __p0; \
22363  uint64x2_t __s1 = __p1; \
22364  uint64x2_t __ret; \
22365  __ret = (uint64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22366  __ret; \
22367})
22368#endif
22369
22370#ifdef __LITTLE_ENDIAN__
22371#define vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22372  uint16_t __s0 = __p0; \
22373  uint16x8_t __s1 = __p1; \
22374  uint16x8_t __ret; \
22375  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22376  __ret; \
22377})
22378#else
22379#define vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22380  uint16_t __s0 = __p0; \
22381  uint16x8_t __s1 = __p1; \
22382  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22383  uint16x8_t __ret; \
22384  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
22385  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22386  __ret; \
22387})
22388#define __noswap_vsetq_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22389  uint16_t __s0 = __p0; \
22390  uint16x8_t __s1 = __p1; \
22391  uint16x8_t __ret; \
22392  __ret = (uint16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22393  __ret; \
22394})
22395#endif
22396
22397#ifdef __LITTLE_ENDIAN__
22398#define vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22399  int8_t __s0 = __p0; \
22400  int8x16_t __s1 = __p1; \
22401  int8x16_t __ret; \
22402  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22403  __ret; \
22404})
22405#else
22406#define vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22407  int8_t __s0 = __p0; \
22408  int8x16_t __s1 = __p1; \
22409  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22410  int8x16_t __ret; \
22411  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__rev1, __p2); \
22412  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
22413  __ret; \
22414})
22415#define __noswap_vsetq_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22416  int8_t __s0 = __p0; \
22417  int8x16_t __s1 = __p1; \
22418  int8x16_t __ret; \
22419  __ret = (int8x16_t) __builtin_neon_vsetq_lane_i8(__s0, (int8x16_t)__s1, __p2); \
22420  __ret; \
22421})
22422#endif
22423
22424#ifdef __LITTLE_ENDIAN__
22425#define vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22426  float32_t __s0 = __p0; \
22427  float32x4_t __s1 = __p1; \
22428  float32x4_t __ret; \
22429  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__s1, __p2); \
22430  __ret; \
22431})
22432#else
22433#define vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22434  float32_t __s0 = __p0; \
22435  float32x4_t __s1 = __p1; \
22436  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22437  float32x4_t __ret; \
22438  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__rev1, __p2); \
22439  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22440  __ret; \
22441})
22442#define __noswap_vsetq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22443  float32_t __s0 = __p0; \
22444  float32x4_t __s1 = __p1; \
22445  float32x4_t __ret; \
22446  __ret = (float32x4_t) __builtin_neon_vsetq_lane_f32(__s0, (int8x16_t)__s1, __p2); \
22447  __ret; \
22448})
22449#endif
22450
22451#ifdef __LITTLE_ENDIAN__
22452#define vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22453  int32_t __s0 = __p0; \
22454  int32x4_t __s1 = __p1; \
22455  int32x4_t __ret; \
22456  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22457  __ret; \
22458})
22459#else
22460#define vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22461  int32_t __s0 = __p0; \
22462  int32x4_t __s1 = __p1; \
22463  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22464  int32x4_t __ret; \
22465  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__rev1, __p2); \
22466  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22467  __ret; \
22468})
22469#define __noswap_vsetq_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22470  int32_t __s0 = __p0; \
22471  int32x4_t __s1 = __p1; \
22472  int32x4_t __ret; \
22473  __ret = (int32x4_t) __builtin_neon_vsetq_lane_i32(__s0, (int8x16_t)__s1, __p2); \
22474  __ret; \
22475})
22476#endif
22477
22478#ifdef __LITTLE_ENDIAN__
22479#define vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22480  int64_t __s0 = __p0; \
22481  int64x2_t __s1 = __p1; \
22482  int64x2_t __ret; \
22483  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22484  __ret; \
22485})
22486#else
22487#define vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22488  int64_t __s0 = __p0; \
22489  int64x2_t __s1 = __p1; \
22490  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22491  int64x2_t __ret; \
22492  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
22493  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22494  __ret; \
22495})
22496#define __noswap_vsetq_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22497  int64_t __s0 = __p0; \
22498  int64x2_t __s1 = __p1; \
22499  int64x2_t __ret; \
22500  __ret = (int64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
22501  __ret; \
22502})
22503#endif
22504
22505#ifdef __LITTLE_ENDIAN__
22506#define vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22507  int16_t __s0 = __p0; \
22508  int16x8_t __s1 = __p1; \
22509  int16x8_t __ret; \
22510  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22511  __ret; \
22512})
22513#else
22514#define vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22515  int16_t __s0 = __p0; \
22516  int16x8_t __s1 = __p1; \
22517  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22518  int16x8_t __ret; \
22519  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__rev1, __p2); \
22520  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22521  __ret; \
22522})
22523#define __noswap_vsetq_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22524  int16_t __s0 = __p0; \
22525  int16x8_t __s1 = __p1; \
22526  int16x8_t __ret; \
22527  __ret = (int16x8_t) __builtin_neon_vsetq_lane_i16(__s0, (int8x16_t)__s1, __p2); \
22528  __ret; \
22529})
22530#endif
22531
22532#ifdef __LITTLE_ENDIAN__
22533#define vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22534  uint8_t __s0 = __p0; \
22535  uint8x8_t __s1 = __p1; \
22536  uint8x8_t __ret; \
22537  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22538  __ret; \
22539})
22540#else
22541#define vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22542  uint8_t __s0 = __p0; \
22543  uint8x8_t __s1 = __p1; \
22544  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22545  uint8x8_t __ret; \
22546  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
22547  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22548  __ret; \
22549})
22550#define __noswap_vset_lane_u8(__p0, __p1, __p2) __extension__ ({ \
22551  uint8_t __s0 = __p0; \
22552  uint8x8_t __s1 = __p1; \
22553  uint8x8_t __ret; \
22554  __ret = (uint8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22555  __ret; \
22556})
22557#endif
22558
22559#ifdef __LITTLE_ENDIAN__
22560#define vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22561  uint32_t __s0 = __p0; \
22562  uint32x2_t __s1 = __p1; \
22563  uint32x2_t __ret; \
22564  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22565  __ret; \
22566})
22567#else
22568#define vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22569  uint32_t __s0 = __p0; \
22570  uint32x2_t __s1 = __p1; \
22571  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22572  uint32x2_t __ret; \
22573  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__rev1, __p2); \
22574  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22575  __ret; \
22576})
22577#define __noswap_vset_lane_u32(__p0, __p1, __p2) __extension__ ({ \
22578  uint32_t __s0 = __p0; \
22579  uint32x2_t __s1 = __p1; \
22580  uint32x2_t __ret; \
22581  __ret = (uint32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22582  __ret; \
22583})
22584#endif
22585
22586#ifdef __LITTLE_ENDIAN__
22587#define vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22588  uint64_t __s0 = __p0; \
22589  uint64x1_t __s1 = __p1; \
22590  uint64x1_t __ret; \
22591  __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22592  __ret; \
22593})
22594#else
22595#define vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22596  uint64_t __s0 = __p0; \
22597  uint64x1_t __s1 = __p1; \
22598  uint64x1_t __ret; \
22599  __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22600  __ret; \
22601})
22602#define __noswap_vset_lane_u64(__p0, __p1, __p2) __extension__ ({ \
22603  uint64_t __s0 = __p0; \
22604  uint64x1_t __s1 = __p1; \
22605  uint64x1_t __ret; \
22606  __ret = (uint64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22607  __ret; \
22608})
22609#endif
22610
22611#ifdef __LITTLE_ENDIAN__
22612#define vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22613  uint16_t __s0 = __p0; \
22614  uint16x4_t __s1 = __p1; \
22615  uint16x4_t __ret; \
22616  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22617  __ret; \
22618})
22619#else
22620#define vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22621  uint16_t __s0 = __p0; \
22622  uint16x4_t __s1 = __p1; \
22623  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22624  uint16x4_t __ret; \
22625  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
22626  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22627  __ret; \
22628})
22629#define __noswap_vset_lane_u16(__p0, __p1, __p2) __extension__ ({ \
22630  uint16_t __s0 = __p0; \
22631  uint16x4_t __s1 = __p1; \
22632  uint16x4_t __ret; \
22633  __ret = (uint16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22634  __ret; \
22635})
22636#endif
22637
22638#ifdef __LITTLE_ENDIAN__
22639#define vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22640  int8_t __s0 = __p0; \
22641  int8x8_t __s1 = __p1; \
22642  int8x8_t __ret; \
22643  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22644  __ret; \
22645})
22646#else
22647#define vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22648  int8_t __s0 = __p0; \
22649  int8x8_t __s1 = __p1; \
22650  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
22651  int8x8_t __ret; \
22652  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__rev1, __p2); \
22653  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
22654  __ret; \
22655})
22656#define __noswap_vset_lane_s8(__p0, __p1, __p2) __extension__ ({ \
22657  int8_t __s0 = __p0; \
22658  int8x8_t __s1 = __p1; \
22659  int8x8_t __ret; \
22660  __ret = (int8x8_t) __builtin_neon_vset_lane_i8(__s0, (int8x8_t)__s1, __p2); \
22661  __ret; \
22662})
22663#endif
22664
22665#ifdef __LITTLE_ENDIAN__
22666#define vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22667  float32_t __s0 = __p0; \
22668  float32x2_t __s1 = __p1; \
22669  float32x2_t __ret; \
22670  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__s1, __p2); \
22671  __ret; \
22672})
22673#else
22674#define vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22675  float32_t __s0 = __p0; \
22676  float32x2_t __s1 = __p1; \
22677  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22678  float32x2_t __ret; \
22679  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__rev1, __p2); \
22680  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22681  __ret; \
22682})
22683#define __noswap_vset_lane_f32(__p0, __p1, __p2) __extension__ ({ \
22684  float32_t __s0 = __p0; \
22685  float32x2_t __s1 = __p1; \
22686  float32x2_t __ret; \
22687  __ret = (float32x2_t) __builtin_neon_vset_lane_f32(__s0, (int8x8_t)__s1, __p2); \
22688  __ret; \
22689})
22690#endif
22691
22692#ifdef __LITTLE_ENDIAN__
22693#define vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22694  int32_t __s0 = __p0; \
22695  int32x2_t __s1 = __p1; \
22696  int32x2_t __ret; \
22697  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22698  __ret; \
22699})
22700#else
22701#define vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22702  int32_t __s0 = __p0; \
22703  int32x2_t __s1 = __p1; \
22704  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
22705  int32x2_t __ret; \
22706  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__rev1, __p2); \
22707  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
22708  __ret; \
22709})
22710#define __noswap_vset_lane_s32(__p0, __p1, __p2) __extension__ ({ \
22711  int32_t __s0 = __p0; \
22712  int32x2_t __s1 = __p1; \
22713  int32x2_t __ret; \
22714  __ret = (int32x2_t) __builtin_neon_vset_lane_i32(__s0, (int8x8_t)__s1, __p2); \
22715  __ret; \
22716})
22717#endif
22718
22719#ifdef __LITTLE_ENDIAN__
22720#define vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22721  int64_t __s0 = __p0; \
22722  int64x1_t __s1 = __p1; \
22723  int64x1_t __ret; \
22724  __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22725  __ret; \
22726})
22727#else
22728#define vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22729  int64_t __s0 = __p0; \
22730  int64x1_t __s1 = __p1; \
22731  int64x1_t __ret; \
22732  __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22733  __ret; \
22734})
22735#define __noswap_vset_lane_s64(__p0, __p1, __p2) __extension__ ({ \
22736  int64_t __s0 = __p0; \
22737  int64x1_t __s1 = __p1; \
22738  int64x1_t __ret; \
22739  __ret = (int64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
22740  __ret; \
22741})
22742#endif
22743
22744#ifdef __LITTLE_ENDIAN__
22745#define vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22746  int16_t __s0 = __p0; \
22747  int16x4_t __s1 = __p1; \
22748  int16x4_t __ret; \
22749  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22750  __ret; \
22751})
22752#else
22753#define vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22754  int16_t __s0 = __p0; \
22755  int16x4_t __s1 = __p1; \
22756  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
22757  int16x4_t __ret; \
22758  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__rev1, __p2); \
22759  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
22760  __ret; \
22761})
22762#define __noswap_vset_lane_s16(__p0, __p1, __p2) __extension__ ({ \
22763  int16_t __s0 = __p0; \
22764  int16x4_t __s1 = __p1; \
22765  int16x4_t __ret; \
22766  __ret = (int16x4_t) __builtin_neon_vset_lane_i16(__s0, (int8x8_t)__s1, __p2); \
22767  __ret; \
22768})
22769#endif
22770
22771#ifdef __LITTLE_ENDIAN__
22772__ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
22773  uint8x16_t __ret;
22774  __ret = (uint8x16_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
22775  return __ret;
22776}
22777#else
22778__ai uint8x16_t vshlq_u8(uint8x16_t __p0, int8x16_t __p1) {
22779  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22780  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22781  uint8x16_t __ret;
22782  __ret = (uint8x16_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
22783  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22784  return __ret;
22785}
22786#endif
22787
22788#ifdef __LITTLE_ENDIAN__
22789__ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
22790  uint32x4_t __ret;
22791  __ret = (uint32x4_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
22792  return __ret;
22793}
22794#else
22795__ai uint32x4_t vshlq_u32(uint32x4_t __p0, int32x4_t __p1) {
22796  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22797  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22798  uint32x4_t __ret;
22799  __ret = (uint32x4_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
22800  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22801  return __ret;
22802}
22803#endif
22804
22805#ifdef __LITTLE_ENDIAN__
22806__ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
22807  uint64x2_t __ret;
22808  __ret = (uint64x2_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
22809  return __ret;
22810}
22811#else
22812__ai uint64x2_t vshlq_u64(uint64x2_t __p0, int64x2_t __p1) {
22813  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22814  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22815  uint64x2_t __ret;
22816  __ret = (uint64x2_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
22817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22818  return __ret;
22819}
22820#endif
22821
22822#ifdef __LITTLE_ENDIAN__
22823__ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
22824  uint16x8_t __ret;
22825  __ret = (uint16x8_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
22826  return __ret;
22827}
22828#else
22829__ai uint16x8_t vshlq_u16(uint16x8_t __p0, int16x8_t __p1) {
22830  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22831  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22832  uint16x8_t __ret;
22833  __ret = (uint16x8_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
22834  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22835  return __ret;
22836}
22837#endif
22838
22839#ifdef __LITTLE_ENDIAN__
22840__ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1) {
22841  int8x16_t __ret;
22842  __ret = (int8x16_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
22843  return __ret;
22844}
22845#else
22846__ai int8x16_t vshlq_s8(int8x16_t __p0, int8x16_t __p1) {
22847  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22848  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22849  int8x16_t __ret;
22850  __ret = (int8x16_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
22851  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
22852  return __ret;
22853}
22854#endif
22855
22856#ifdef __LITTLE_ENDIAN__
22857__ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1) {
22858  int32x4_t __ret;
22859  __ret = (int32x4_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
22860  return __ret;
22861}
22862#else
22863__ai int32x4_t vshlq_s32(int32x4_t __p0, int32x4_t __p1) {
22864  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22865  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22866  int32x4_t __ret;
22867  __ret = (int32x4_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
22868  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22869  return __ret;
22870}
22871#endif
22872
22873#ifdef __LITTLE_ENDIAN__
22874__ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1) {
22875  int64x2_t __ret;
22876  __ret = (int64x2_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
22877  return __ret;
22878}
22879#else
22880__ai int64x2_t vshlq_s64(int64x2_t __p0, int64x2_t __p1) {
22881  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22882  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22883  int64x2_t __ret;
22884  __ret = (int64x2_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
22885  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22886  return __ret;
22887}
22888#endif
22889
22890#ifdef __LITTLE_ENDIAN__
22891__ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1) {
22892  int16x8_t __ret;
22893  __ret = (int16x8_t) __builtin_neon_vshlq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
22894  return __ret;
22895}
22896#else
22897__ai int16x8_t vshlq_s16(int16x8_t __p0, int16x8_t __p1) {
22898  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22899  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22900  int16x8_t __ret;
22901  __ret = (int16x8_t) __builtin_neon_vshlq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
22902  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22903  return __ret;
22904}
22905#endif
22906
22907#ifdef __LITTLE_ENDIAN__
22908__ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1) {
22909  uint8x8_t __ret;
22910  __ret = (uint8x8_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
22911  return __ret;
22912}
22913#else
22914__ai uint8x8_t vshl_u8(uint8x8_t __p0, int8x8_t __p1) {
22915  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22916  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22917  uint8x8_t __ret;
22918  __ret = (uint8x8_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
22919  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22920  return __ret;
22921}
22922#endif
22923
22924#ifdef __LITTLE_ENDIAN__
22925__ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1) {
22926  uint32x2_t __ret;
22927  __ret = (uint32x2_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
22928  return __ret;
22929}
22930#else
22931__ai uint32x2_t vshl_u32(uint32x2_t __p0, int32x2_t __p1) {
22932  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22933  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22934  uint32x2_t __ret;
22935  __ret = (uint32x2_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
22936  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
22937  return __ret;
22938}
22939#endif
22940
22941#ifdef __LITTLE_ENDIAN__
22942__ai uint64x1_t vshl_u64(uint64x1_t __p0, int64x1_t __p1) {
22943  uint64x1_t __ret;
22944  __ret = (uint64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
22945  return __ret;
22946}
22947#else
22948__ai uint64x1_t vshl_u64(uint64x1_t __p0, int64x1_t __p1) {
22949  uint64x1_t __ret;
22950  __ret = (uint64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
22951  return __ret;
22952}
22953#endif
22954
22955#ifdef __LITTLE_ENDIAN__
22956__ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1) {
22957  uint16x4_t __ret;
22958  __ret = (uint16x4_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
22959  return __ret;
22960}
22961#else
22962__ai uint16x4_t vshl_u16(uint16x4_t __p0, int16x4_t __p1) {
22963  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
22964  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
22965  uint16x4_t __ret;
22966  __ret = (uint16x4_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
22967  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
22968  return __ret;
22969}
22970#endif
22971
22972#ifdef __LITTLE_ENDIAN__
22973__ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1) {
22974  int8x8_t __ret;
22975  __ret = (int8x8_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
22976  return __ret;
22977}
22978#else
22979__ai int8x8_t vshl_s8(int8x8_t __p0, int8x8_t __p1) {
22980  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
22981  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
22982  int8x8_t __ret;
22983  __ret = (int8x8_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
22984  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
22985  return __ret;
22986}
22987#endif
22988
22989#ifdef __LITTLE_ENDIAN__
22990__ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1) {
22991  int32x2_t __ret;
22992  __ret = (int32x2_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
22993  return __ret;
22994}
22995#else
22996__ai int32x2_t vshl_s32(int32x2_t __p0, int32x2_t __p1) {
22997  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
22998  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
22999  int32x2_t __ret;
23000  __ret = (int32x2_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
23001  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
23002  return __ret;
23003}
23004#endif
23005
23006#ifdef __LITTLE_ENDIAN__
23007__ai int64x1_t vshl_s64(int64x1_t __p0, int64x1_t __p1) {
23008  int64x1_t __ret;
23009  __ret = (int64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
23010  return __ret;
23011}
23012#else
23013__ai int64x1_t vshl_s64(int64x1_t __p0, int64x1_t __p1) {
23014  int64x1_t __ret;
23015  __ret = (int64x1_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
23016  return __ret;
23017}
23018#endif
23019
23020#ifdef __LITTLE_ENDIAN__
23021__ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1) {
23022  int16x4_t __ret;
23023  __ret = (int16x4_t) __builtin_neon_vshl_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
23024  return __ret;
23025}
23026#else
23027__ai int16x4_t vshl_s16(int16x4_t __p0, int16x4_t __p1) {
23028  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
23029  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
23030  int16x4_t __ret;
23031  __ret = (int16x4_t) __builtin_neon_vshl_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
23032  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
23033  return __ret;
23034}
23035#endif
23036
23037#ifdef __LITTLE_ENDIAN__
23038#define vshlq_n_u8(__p0, __p1) __extension__ ({ \
23039  uint8x16_t __s0 = __p0; \
23040  uint8x16_t __ret; \
23041  __ret = (uint8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 48); \
23042  __ret; \
23043})
23044#else
23045#define vshlq_n_u8(__p0, __p1) __extension__ ({ \
23046  uint8x16_t __s0 = __p0; \
23047  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23048  uint8x16_t __ret; \
23049  __ret = (uint8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 48); \
23050  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23051  __ret; \
23052})
23053#endif
23054
23055#ifdef __LITTLE_ENDIAN__
23056#define vshlq_n_u32(__p0, __p1) __extension__ ({ \
23057  uint32x4_t __s0 = __p0; \
23058  uint32x4_t __ret; \
23059  __ret = (uint32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 50); \
23060  __ret; \
23061})
23062#else
23063#define vshlq_n_u32(__p0, __p1) __extension__ ({ \
23064  uint32x4_t __s0 = __p0; \
23065  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23066  uint32x4_t __ret; \
23067  __ret = (uint32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 50); \
23068  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23069  __ret; \
23070})
23071#endif
23072
23073#ifdef __LITTLE_ENDIAN__
23074#define vshlq_n_u64(__p0, __p1) __extension__ ({ \
23075  uint64x2_t __s0 = __p0; \
23076  uint64x2_t __ret; \
23077  __ret = (uint64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 51); \
23078  __ret; \
23079})
23080#else
23081#define vshlq_n_u64(__p0, __p1) __extension__ ({ \
23082  uint64x2_t __s0 = __p0; \
23083  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23084  uint64x2_t __ret; \
23085  __ret = (uint64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 51); \
23086  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23087  __ret; \
23088})
23089#endif
23090
23091#ifdef __LITTLE_ENDIAN__
23092#define vshlq_n_u16(__p0, __p1) __extension__ ({ \
23093  uint16x8_t __s0 = __p0; \
23094  uint16x8_t __ret; \
23095  __ret = (uint16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 49); \
23096  __ret; \
23097})
23098#else
23099#define vshlq_n_u16(__p0, __p1) __extension__ ({ \
23100  uint16x8_t __s0 = __p0; \
23101  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23102  uint16x8_t __ret; \
23103  __ret = (uint16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 49); \
23104  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23105  __ret; \
23106})
23107#endif
23108
23109#ifdef __LITTLE_ENDIAN__
23110#define vshlq_n_s8(__p0, __p1) __extension__ ({ \
23111  int8x16_t __s0 = __p0; \
23112  int8x16_t __ret; \
23113  __ret = (int8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 32); \
23114  __ret; \
23115})
23116#else
23117#define vshlq_n_s8(__p0, __p1) __extension__ ({ \
23118  int8x16_t __s0 = __p0; \
23119  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23120  int8x16_t __ret; \
23121  __ret = (int8x16_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 32); \
23122  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23123  __ret; \
23124})
23125#endif
23126
23127#ifdef __LITTLE_ENDIAN__
23128#define vshlq_n_s32(__p0, __p1) __extension__ ({ \
23129  int32x4_t __s0 = __p0; \
23130  int32x4_t __ret; \
23131  __ret = (int32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 34); \
23132  __ret; \
23133})
23134#else
23135#define vshlq_n_s32(__p0, __p1) __extension__ ({ \
23136  int32x4_t __s0 = __p0; \
23137  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23138  int32x4_t __ret; \
23139  __ret = (int32x4_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 34); \
23140  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23141  __ret; \
23142})
23143#endif
23144
23145#ifdef __LITTLE_ENDIAN__
23146#define vshlq_n_s64(__p0, __p1) __extension__ ({ \
23147  int64x2_t __s0 = __p0; \
23148  int64x2_t __ret; \
23149  __ret = (int64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 35); \
23150  __ret; \
23151})
23152#else
23153#define vshlq_n_s64(__p0, __p1) __extension__ ({ \
23154  int64x2_t __s0 = __p0; \
23155  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23156  int64x2_t __ret; \
23157  __ret = (int64x2_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 35); \
23158  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23159  __ret; \
23160})
23161#endif
23162
23163#ifdef __LITTLE_ENDIAN__
23164#define vshlq_n_s16(__p0, __p1) __extension__ ({ \
23165  int16x8_t __s0 = __p0; \
23166  int16x8_t __ret; \
23167  __ret = (int16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__s0, __p1, 33); \
23168  __ret; \
23169})
23170#else
23171#define vshlq_n_s16(__p0, __p1) __extension__ ({ \
23172  int16x8_t __s0 = __p0; \
23173  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23174  int16x8_t __ret; \
23175  __ret = (int16x8_t) __builtin_neon_vshlq_n_v((int8x16_t)__rev0, __p1, 33); \
23176  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23177  __ret; \
23178})
23179#endif
23180
23181#ifdef __LITTLE_ENDIAN__
23182#define vshl_n_u8(__p0, __p1) __extension__ ({ \
23183  uint8x8_t __s0 = __p0; \
23184  uint8x8_t __ret; \
23185  __ret = (uint8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 16); \
23186  __ret; \
23187})
23188#else
23189#define vshl_n_u8(__p0, __p1) __extension__ ({ \
23190  uint8x8_t __s0 = __p0; \
23191  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23192  uint8x8_t __ret; \
23193  __ret = (uint8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 16); \
23194  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23195  __ret; \
23196})
23197#endif
23198
23199#ifdef __LITTLE_ENDIAN__
23200#define vshl_n_u32(__p0, __p1) __extension__ ({ \
23201  uint32x2_t __s0 = __p0; \
23202  uint32x2_t __ret; \
23203  __ret = (uint32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 18); \
23204  __ret; \
23205})
23206#else
23207#define vshl_n_u32(__p0, __p1) __extension__ ({ \
23208  uint32x2_t __s0 = __p0; \
23209  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23210  uint32x2_t __ret; \
23211  __ret = (uint32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 18); \
23212  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23213  __ret; \
23214})
23215#endif
23216
23217#ifdef __LITTLE_ENDIAN__
23218#define vshl_n_u64(__p0, __p1) __extension__ ({ \
23219  uint64x1_t __s0 = __p0; \
23220  uint64x1_t __ret; \
23221  __ret = (uint64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 19); \
23222  __ret; \
23223})
23224#else
23225#define vshl_n_u64(__p0, __p1) __extension__ ({ \
23226  uint64x1_t __s0 = __p0; \
23227  uint64x1_t __ret; \
23228  __ret = (uint64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 19); \
23229  __ret; \
23230})
23231#endif
23232
23233#ifdef __LITTLE_ENDIAN__
23234#define vshl_n_u16(__p0, __p1) __extension__ ({ \
23235  uint16x4_t __s0 = __p0; \
23236  uint16x4_t __ret; \
23237  __ret = (uint16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 17); \
23238  __ret; \
23239})
23240#else
23241#define vshl_n_u16(__p0, __p1) __extension__ ({ \
23242  uint16x4_t __s0 = __p0; \
23243  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23244  uint16x4_t __ret; \
23245  __ret = (uint16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 17); \
23246  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23247  __ret; \
23248})
23249#endif
23250
23251#ifdef __LITTLE_ENDIAN__
23252#define vshl_n_s8(__p0, __p1) __extension__ ({ \
23253  int8x8_t __s0 = __p0; \
23254  int8x8_t __ret; \
23255  __ret = (int8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 0); \
23256  __ret; \
23257})
23258#else
23259#define vshl_n_s8(__p0, __p1) __extension__ ({ \
23260  int8x8_t __s0 = __p0; \
23261  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23262  int8x8_t __ret; \
23263  __ret = (int8x8_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 0); \
23264  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23265  __ret; \
23266})
23267#endif
23268
23269#ifdef __LITTLE_ENDIAN__
23270#define vshl_n_s32(__p0, __p1) __extension__ ({ \
23271  int32x2_t __s0 = __p0; \
23272  int32x2_t __ret; \
23273  __ret = (int32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 2); \
23274  __ret; \
23275})
23276#else
23277#define vshl_n_s32(__p0, __p1) __extension__ ({ \
23278  int32x2_t __s0 = __p0; \
23279  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23280  int32x2_t __ret; \
23281  __ret = (int32x2_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 2); \
23282  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23283  __ret; \
23284})
23285#endif
23286
23287#ifdef __LITTLE_ENDIAN__
23288#define vshl_n_s64(__p0, __p1) __extension__ ({ \
23289  int64x1_t __s0 = __p0; \
23290  int64x1_t __ret; \
23291  __ret = (int64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 3); \
23292  __ret; \
23293})
23294#else
23295#define vshl_n_s64(__p0, __p1) __extension__ ({ \
23296  int64x1_t __s0 = __p0; \
23297  int64x1_t __ret; \
23298  __ret = (int64x1_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 3); \
23299  __ret; \
23300})
23301#endif
23302
23303#ifdef __LITTLE_ENDIAN__
23304#define vshl_n_s16(__p0, __p1) __extension__ ({ \
23305  int16x4_t __s0 = __p0; \
23306  int16x4_t __ret; \
23307  __ret = (int16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__s0, __p1, 1); \
23308  __ret; \
23309})
23310#else
23311#define vshl_n_s16(__p0, __p1) __extension__ ({ \
23312  int16x4_t __s0 = __p0; \
23313  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23314  int16x4_t __ret; \
23315  __ret = (int16x4_t) __builtin_neon_vshl_n_v((int8x8_t)__rev0, __p1, 1); \
23316  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23317  __ret; \
23318})
23319#endif
23320
23321#ifdef __LITTLE_ENDIAN__
23322#define vshll_n_u8(__p0, __p1) __extension__ ({ \
23323  uint8x8_t __s0 = __p0; \
23324  uint16x8_t __ret; \
23325  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 49); \
23326  __ret; \
23327})
23328#else
23329#define vshll_n_u8(__p0, __p1) __extension__ ({ \
23330  uint8x8_t __s0 = __p0; \
23331  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23332  uint16x8_t __ret; \
23333  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 49); \
23334  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23335  __ret; \
23336})
23337#define __noswap_vshll_n_u8(__p0, __p1) __extension__ ({ \
23338  uint8x8_t __s0 = __p0; \
23339  uint16x8_t __ret; \
23340  __ret = (uint16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 49); \
23341  __ret; \
23342})
23343#endif
23344
23345#ifdef __LITTLE_ENDIAN__
23346#define vshll_n_u32(__p0, __p1) __extension__ ({ \
23347  uint32x2_t __s0 = __p0; \
23348  uint64x2_t __ret; \
23349  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 51); \
23350  __ret; \
23351})
23352#else
23353#define vshll_n_u32(__p0, __p1) __extension__ ({ \
23354  uint32x2_t __s0 = __p0; \
23355  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23356  uint64x2_t __ret; \
23357  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 51); \
23358  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23359  __ret; \
23360})
23361#define __noswap_vshll_n_u32(__p0, __p1) __extension__ ({ \
23362  uint32x2_t __s0 = __p0; \
23363  uint64x2_t __ret; \
23364  __ret = (uint64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 51); \
23365  __ret; \
23366})
23367#endif
23368
23369#ifdef __LITTLE_ENDIAN__
23370#define vshll_n_u16(__p0, __p1) __extension__ ({ \
23371  uint16x4_t __s0 = __p0; \
23372  uint32x4_t __ret; \
23373  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 50); \
23374  __ret; \
23375})
23376#else
23377#define vshll_n_u16(__p0, __p1) __extension__ ({ \
23378  uint16x4_t __s0 = __p0; \
23379  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23380  uint32x4_t __ret; \
23381  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 50); \
23382  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23383  __ret; \
23384})
23385#define __noswap_vshll_n_u16(__p0, __p1) __extension__ ({ \
23386  uint16x4_t __s0 = __p0; \
23387  uint32x4_t __ret; \
23388  __ret = (uint32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 50); \
23389  __ret; \
23390})
23391#endif
23392
23393#ifdef __LITTLE_ENDIAN__
23394#define vshll_n_s8(__p0, __p1) __extension__ ({ \
23395  int8x8_t __s0 = __p0; \
23396  int16x8_t __ret; \
23397  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 33); \
23398  __ret; \
23399})
23400#else
23401#define vshll_n_s8(__p0, __p1) __extension__ ({ \
23402  int8x8_t __s0 = __p0; \
23403  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23404  int16x8_t __ret; \
23405  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 33); \
23406  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23407  __ret; \
23408})
23409#define __noswap_vshll_n_s8(__p0, __p1) __extension__ ({ \
23410  int8x8_t __s0 = __p0; \
23411  int16x8_t __ret; \
23412  __ret = (int16x8_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 33); \
23413  __ret; \
23414})
23415#endif
23416
23417#ifdef __LITTLE_ENDIAN__
23418#define vshll_n_s32(__p0, __p1) __extension__ ({ \
23419  int32x2_t __s0 = __p0; \
23420  int64x2_t __ret; \
23421  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 35); \
23422  __ret; \
23423})
23424#else
23425#define vshll_n_s32(__p0, __p1) __extension__ ({ \
23426  int32x2_t __s0 = __p0; \
23427  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23428  int64x2_t __ret; \
23429  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 35); \
23430  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23431  __ret; \
23432})
23433#define __noswap_vshll_n_s32(__p0, __p1) __extension__ ({ \
23434  int32x2_t __s0 = __p0; \
23435  int64x2_t __ret; \
23436  __ret = (int64x2_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 35); \
23437  __ret; \
23438})
23439#endif
23440
23441#ifdef __LITTLE_ENDIAN__
23442#define vshll_n_s16(__p0, __p1) __extension__ ({ \
23443  int16x4_t __s0 = __p0; \
23444  int32x4_t __ret; \
23445  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 34); \
23446  __ret; \
23447})
23448#else
23449#define vshll_n_s16(__p0, __p1) __extension__ ({ \
23450  int16x4_t __s0 = __p0; \
23451  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23452  int32x4_t __ret; \
23453  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__rev0, __p1, 34); \
23454  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23455  __ret; \
23456})
23457#define __noswap_vshll_n_s16(__p0, __p1) __extension__ ({ \
23458  int16x4_t __s0 = __p0; \
23459  int32x4_t __ret; \
23460  __ret = (int32x4_t) __builtin_neon_vshll_n_v((int8x8_t)__s0, __p1, 34); \
23461  __ret; \
23462})
23463#endif
23464
23465#ifdef __LITTLE_ENDIAN__
23466#define vshrq_n_u8(__p0, __p1) __extension__ ({ \
23467  uint8x16_t __s0 = __p0; \
23468  uint8x16_t __ret; \
23469  __ret = (uint8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 48); \
23470  __ret; \
23471})
23472#else
23473#define vshrq_n_u8(__p0, __p1) __extension__ ({ \
23474  uint8x16_t __s0 = __p0; \
23475  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23476  uint8x16_t __ret; \
23477  __ret = (uint8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 48); \
23478  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23479  __ret; \
23480})
23481#endif
23482
23483#ifdef __LITTLE_ENDIAN__
23484#define vshrq_n_u32(__p0, __p1) __extension__ ({ \
23485  uint32x4_t __s0 = __p0; \
23486  uint32x4_t __ret; \
23487  __ret = (uint32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 50); \
23488  __ret; \
23489})
23490#else
23491#define vshrq_n_u32(__p0, __p1) __extension__ ({ \
23492  uint32x4_t __s0 = __p0; \
23493  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23494  uint32x4_t __ret; \
23495  __ret = (uint32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 50); \
23496  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23497  __ret; \
23498})
23499#endif
23500
23501#ifdef __LITTLE_ENDIAN__
23502#define vshrq_n_u64(__p0, __p1) __extension__ ({ \
23503  uint64x2_t __s0 = __p0; \
23504  uint64x2_t __ret; \
23505  __ret = (uint64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 51); \
23506  __ret; \
23507})
23508#else
23509#define vshrq_n_u64(__p0, __p1) __extension__ ({ \
23510  uint64x2_t __s0 = __p0; \
23511  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23512  uint64x2_t __ret; \
23513  __ret = (uint64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 51); \
23514  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23515  __ret; \
23516})
23517#endif
23518
23519#ifdef __LITTLE_ENDIAN__
23520#define vshrq_n_u16(__p0, __p1) __extension__ ({ \
23521  uint16x8_t __s0 = __p0; \
23522  uint16x8_t __ret; \
23523  __ret = (uint16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 49); \
23524  __ret; \
23525})
23526#else
23527#define vshrq_n_u16(__p0, __p1) __extension__ ({ \
23528  uint16x8_t __s0 = __p0; \
23529  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23530  uint16x8_t __ret; \
23531  __ret = (uint16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 49); \
23532  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23533  __ret; \
23534})
23535#endif
23536
23537#ifdef __LITTLE_ENDIAN__
23538#define vshrq_n_s8(__p0, __p1) __extension__ ({ \
23539  int8x16_t __s0 = __p0; \
23540  int8x16_t __ret; \
23541  __ret = (int8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 32); \
23542  __ret; \
23543})
23544#else
23545#define vshrq_n_s8(__p0, __p1) __extension__ ({ \
23546  int8x16_t __s0 = __p0; \
23547  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23548  int8x16_t __ret; \
23549  __ret = (int8x16_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 32); \
23550  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23551  __ret; \
23552})
23553#endif
23554
23555#ifdef __LITTLE_ENDIAN__
23556#define vshrq_n_s32(__p0, __p1) __extension__ ({ \
23557  int32x4_t __s0 = __p0; \
23558  int32x4_t __ret; \
23559  __ret = (int32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 34); \
23560  __ret; \
23561})
23562#else
23563#define vshrq_n_s32(__p0, __p1) __extension__ ({ \
23564  int32x4_t __s0 = __p0; \
23565  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23566  int32x4_t __ret; \
23567  __ret = (int32x4_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 34); \
23568  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23569  __ret; \
23570})
23571#endif
23572
23573#ifdef __LITTLE_ENDIAN__
23574#define vshrq_n_s64(__p0, __p1) __extension__ ({ \
23575  int64x2_t __s0 = __p0; \
23576  int64x2_t __ret; \
23577  __ret = (int64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 35); \
23578  __ret; \
23579})
23580#else
23581#define vshrq_n_s64(__p0, __p1) __extension__ ({ \
23582  int64x2_t __s0 = __p0; \
23583  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23584  int64x2_t __ret; \
23585  __ret = (int64x2_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 35); \
23586  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23587  __ret; \
23588})
23589#endif
23590
23591#ifdef __LITTLE_ENDIAN__
23592#define vshrq_n_s16(__p0, __p1) __extension__ ({ \
23593  int16x8_t __s0 = __p0; \
23594  int16x8_t __ret; \
23595  __ret = (int16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__s0, __p1, 33); \
23596  __ret; \
23597})
23598#else
23599#define vshrq_n_s16(__p0, __p1) __extension__ ({ \
23600  int16x8_t __s0 = __p0; \
23601  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23602  int16x8_t __ret; \
23603  __ret = (int16x8_t) __builtin_neon_vshrq_n_v((int8x16_t)__rev0, __p1, 33); \
23604  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23605  __ret; \
23606})
23607#endif
23608
23609#ifdef __LITTLE_ENDIAN__
23610#define vshr_n_u8(__p0, __p1) __extension__ ({ \
23611  uint8x8_t __s0 = __p0; \
23612  uint8x8_t __ret; \
23613  __ret = (uint8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 16); \
23614  __ret; \
23615})
23616#else
23617#define vshr_n_u8(__p0, __p1) __extension__ ({ \
23618  uint8x8_t __s0 = __p0; \
23619  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23620  uint8x8_t __ret; \
23621  __ret = (uint8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 16); \
23622  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23623  __ret; \
23624})
23625#endif
23626
23627#ifdef __LITTLE_ENDIAN__
23628#define vshr_n_u32(__p0, __p1) __extension__ ({ \
23629  uint32x2_t __s0 = __p0; \
23630  uint32x2_t __ret; \
23631  __ret = (uint32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 18); \
23632  __ret; \
23633})
23634#else
23635#define vshr_n_u32(__p0, __p1) __extension__ ({ \
23636  uint32x2_t __s0 = __p0; \
23637  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23638  uint32x2_t __ret; \
23639  __ret = (uint32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 18); \
23640  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23641  __ret; \
23642})
23643#endif
23644
23645#ifdef __LITTLE_ENDIAN__
23646#define vshr_n_u64(__p0, __p1) __extension__ ({ \
23647  uint64x1_t __s0 = __p0; \
23648  uint64x1_t __ret; \
23649  __ret = (uint64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 19); \
23650  __ret; \
23651})
23652#else
23653#define vshr_n_u64(__p0, __p1) __extension__ ({ \
23654  uint64x1_t __s0 = __p0; \
23655  uint64x1_t __ret; \
23656  __ret = (uint64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 19); \
23657  __ret; \
23658})
23659#endif
23660
23661#ifdef __LITTLE_ENDIAN__
23662#define vshr_n_u16(__p0, __p1) __extension__ ({ \
23663  uint16x4_t __s0 = __p0; \
23664  uint16x4_t __ret; \
23665  __ret = (uint16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 17); \
23666  __ret; \
23667})
23668#else
23669#define vshr_n_u16(__p0, __p1) __extension__ ({ \
23670  uint16x4_t __s0 = __p0; \
23671  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23672  uint16x4_t __ret; \
23673  __ret = (uint16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 17); \
23674  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23675  __ret; \
23676})
23677#endif
23678
23679#ifdef __LITTLE_ENDIAN__
23680#define vshr_n_s8(__p0, __p1) __extension__ ({ \
23681  int8x8_t __s0 = __p0; \
23682  int8x8_t __ret; \
23683  __ret = (int8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 0); \
23684  __ret; \
23685})
23686#else
23687#define vshr_n_s8(__p0, __p1) __extension__ ({ \
23688  int8x8_t __s0 = __p0; \
23689  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23690  int8x8_t __ret; \
23691  __ret = (int8x8_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 0); \
23692  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23693  __ret; \
23694})
23695#endif
23696
23697#ifdef __LITTLE_ENDIAN__
23698#define vshr_n_s32(__p0, __p1) __extension__ ({ \
23699  int32x2_t __s0 = __p0; \
23700  int32x2_t __ret; \
23701  __ret = (int32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 2); \
23702  __ret; \
23703})
23704#else
23705#define vshr_n_s32(__p0, __p1) __extension__ ({ \
23706  int32x2_t __s0 = __p0; \
23707  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23708  int32x2_t __ret; \
23709  __ret = (int32x2_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 2); \
23710  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23711  __ret; \
23712})
23713#endif
23714
23715#ifdef __LITTLE_ENDIAN__
23716#define vshr_n_s64(__p0, __p1) __extension__ ({ \
23717  int64x1_t __s0 = __p0; \
23718  int64x1_t __ret; \
23719  __ret = (int64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 3); \
23720  __ret; \
23721})
23722#else
23723#define vshr_n_s64(__p0, __p1) __extension__ ({ \
23724  int64x1_t __s0 = __p0; \
23725  int64x1_t __ret; \
23726  __ret = (int64x1_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 3); \
23727  __ret; \
23728})
23729#endif
23730
23731#ifdef __LITTLE_ENDIAN__
23732#define vshr_n_s16(__p0, __p1) __extension__ ({ \
23733  int16x4_t __s0 = __p0; \
23734  int16x4_t __ret; \
23735  __ret = (int16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__s0, __p1, 1); \
23736  __ret; \
23737})
23738#else
23739#define vshr_n_s16(__p0, __p1) __extension__ ({ \
23740  int16x4_t __s0 = __p0; \
23741  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23742  int16x4_t __ret; \
23743  __ret = (int16x4_t) __builtin_neon_vshr_n_v((int8x8_t)__rev0, __p1, 1); \
23744  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23745  __ret; \
23746})
23747#endif
23748
23749#ifdef __LITTLE_ENDIAN__
23750#define vshrn_n_u32(__p0, __p1) __extension__ ({ \
23751  uint32x4_t __s0 = __p0; \
23752  uint16x4_t __ret; \
23753  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 17); \
23754  __ret; \
23755})
23756#else
23757#define vshrn_n_u32(__p0, __p1) __extension__ ({ \
23758  uint32x4_t __s0 = __p0; \
23759  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23760  uint16x4_t __ret; \
23761  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 17); \
23762  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23763  __ret; \
23764})
23765#define __noswap_vshrn_n_u32(__p0, __p1) __extension__ ({ \
23766  uint32x4_t __s0 = __p0; \
23767  uint16x4_t __ret; \
23768  __ret = (uint16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 17); \
23769  __ret; \
23770})
23771#endif
23772
23773#ifdef __LITTLE_ENDIAN__
23774#define vshrn_n_u64(__p0, __p1) __extension__ ({ \
23775  uint64x2_t __s0 = __p0; \
23776  uint32x2_t __ret; \
23777  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 18); \
23778  __ret; \
23779})
23780#else
23781#define vshrn_n_u64(__p0, __p1) __extension__ ({ \
23782  uint64x2_t __s0 = __p0; \
23783  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23784  uint32x2_t __ret; \
23785  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 18); \
23786  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23787  __ret; \
23788})
23789#define __noswap_vshrn_n_u64(__p0, __p1) __extension__ ({ \
23790  uint64x2_t __s0 = __p0; \
23791  uint32x2_t __ret; \
23792  __ret = (uint32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 18); \
23793  __ret; \
23794})
23795#endif
23796
23797#ifdef __LITTLE_ENDIAN__
23798#define vshrn_n_u16(__p0, __p1) __extension__ ({ \
23799  uint16x8_t __s0 = __p0; \
23800  uint8x8_t __ret; \
23801  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 16); \
23802  __ret; \
23803})
23804#else
23805#define vshrn_n_u16(__p0, __p1) __extension__ ({ \
23806  uint16x8_t __s0 = __p0; \
23807  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23808  uint8x8_t __ret; \
23809  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 16); \
23810  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23811  __ret; \
23812})
23813#define __noswap_vshrn_n_u16(__p0, __p1) __extension__ ({ \
23814  uint16x8_t __s0 = __p0; \
23815  uint8x8_t __ret; \
23816  __ret = (uint8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 16); \
23817  __ret; \
23818})
23819#endif
23820
23821#ifdef __LITTLE_ENDIAN__
23822#define vshrn_n_s32(__p0, __p1) __extension__ ({ \
23823  int32x4_t __s0 = __p0; \
23824  int16x4_t __ret; \
23825  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 1); \
23826  __ret; \
23827})
23828#else
23829#define vshrn_n_s32(__p0, __p1) __extension__ ({ \
23830  int32x4_t __s0 = __p0; \
23831  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23832  int16x4_t __ret; \
23833  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 1); \
23834  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23835  __ret; \
23836})
23837#define __noswap_vshrn_n_s32(__p0, __p1) __extension__ ({ \
23838  int32x4_t __s0 = __p0; \
23839  int16x4_t __ret; \
23840  __ret = (int16x4_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 1); \
23841  __ret; \
23842})
23843#endif
23844
23845#ifdef __LITTLE_ENDIAN__
23846#define vshrn_n_s64(__p0, __p1) __extension__ ({ \
23847  int64x2_t __s0 = __p0; \
23848  int32x2_t __ret; \
23849  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 2); \
23850  __ret; \
23851})
23852#else
23853#define vshrn_n_s64(__p0, __p1) __extension__ ({ \
23854  int64x2_t __s0 = __p0; \
23855  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
23856  int32x2_t __ret; \
23857  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 2); \
23858  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
23859  __ret; \
23860})
23861#define __noswap_vshrn_n_s64(__p0, __p1) __extension__ ({ \
23862  int64x2_t __s0 = __p0; \
23863  int32x2_t __ret; \
23864  __ret = (int32x2_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 2); \
23865  __ret; \
23866})
23867#endif
23868
23869#ifdef __LITTLE_ENDIAN__
23870#define vshrn_n_s16(__p0, __p1) __extension__ ({ \
23871  int16x8_t __s0 = __p0; \
23872  int8x8_t __ret; \
23873  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 0); \
23874  __ret; \
23875})
23876#else
23877#define vshrn_n_s16(__p0, __p1) __extension__ ({ \
23878  int16x8_t __s0 = __p0; \
23879  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23880  int8x8_t __ret; \
23881  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__rev0, __p1, 0); \
23882  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23883  __ret; \
23884})
23885#define __noswap_vshrn_n_s16(__p0, __p1) __extension__ ({ \
23886  int16x8_t __s0 = __p0; \
23887  int8x8_t __ret; \
23888  __ret = (int8x8_t) __builtin_neon_vshrn_n_v((int8x16_t)__s0, __p1, 0); \
23889  __ret; \
23890})
23891#endif
23892
23893#ifdef __LITTLE_ENDIAN__
23894#define vsli_n_p8(__p0, __p1, __p2) __extension__ ({ \
23895  poly8x8_t __s0 = __p0; \
23896  poly8x8_t __s1 = __p1; \
23897  poly8x8_t __ret; \
23898  __ret = (poly8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
23899  __ret; \
23900})
23901#else
23902#define vsli_n_p8(__p0, __p1, __p2) __extension__ ({ \
23903  poly8x8_t __s0 = __p0; \
23904  poly8x8_t __s1 = __p1; \
23905  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23906  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23907  poly8x8_t __ret; \
23908  __ret = (poly8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
23909  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23910  __ret; \
23911})
23912#endif
23913
23914#ifdef __LITTLE_ENDIAN__
23915#define vsli_n_p16(__p0, __p1, __p2) __extension__ ({ \
23916  poly16x4_t __s0 = __p0; \
23917  poly16x4_t __s1 = __p1; \
23918  poly16x4_t __ret; \
23919  __ret = (poly16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
23920  __ret; \
23921})
23922#else
23923#define vsli_n_p16(__p0, __p1, __p2) __extension__ ({ \
23924  poly16x4_t __s0 = __p0; \
23925  poly16x4_t __s1 = __p1; \
23926  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
23927  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
23928  poly16x4_t __ret; \
23929  __ret = (poly16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
23930  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
23931  __ret; \
23932})
23933#endif
23934
23935#ifdef __LITTLE_ENDIAN__
23936#define vsliq_n_p8(__p0, __p1, __p2) __extension__ ({ \
23937  poly8x16_t __s0 = __p0; \
23938  poly8x16_t __s1 = __p1; \
23939  poly8x16_t __ret; \
23940  __ret = (poly8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
23941  __ret; \
23942})
23943#else
23944#define vsliq_n_p8(__p0, __p1, __p2) __extension__ ({ \
23945  poly8x16_t __s0 = __p0; \
23946  poly8x16_t __s1 = __p1; \
23947  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23948  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23949  poly8x16_t __ret; \
23950  __ret = (poly8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
23951  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23952  __ret; \
23953})
23954#endif
23955
23956#ifdef __LITTLE_ENDIAN__
23957#define vsliq_n_p16(__p0, __p1, __p2) __extension__ ({ \
23958  poly16x8_t __s0 = __p0; \
23959  poly16x8_t __s1 = __p1; \
23960  poly16x8_t __ret; \
23961  __ret = (poly16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
23962  __ret; \
23963})
23964#else
23965#define vsliq_n_p16(__p0, __p1, __p2) __extension__ ({ \
23966  poly16x8_t __s0 = __p0; \
23967  poly16x8_t __s1 = __p1; \
23968  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
23969  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
23970  poly16x8_t __ret; \
23971  __ret = (poly16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
23972  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
23973  __ret; \
23974})
23975#endif
23976
23977#ifdef __LITTLE_ENDIAN__
23978#define vsliq_n_u8(__p0, __p1, __p2) __extension__ ({ \
23979  uint8x16_t __s0 = __p0; \
23980  uint8x16_t __s1 = __p1; \
23981  uint8x16_t __ret; \
23982  __ret = (uint8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
23983  __ret; \
23984})
23985#else
23986#define vsliq_n_u8(__p0, __p1, __p2) __extension__ ({ \
23987  uint8x16_t __s0 = __p0; \
23988  uint8x16_t __s1 = __p1; \
23989  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23990  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23991  uint8x16_t __ret; \
23992  __ret = (uint8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
23993  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
23994  __ret; \
23995})
23996#endif
23997
23998#ifdef __LITTLE_ENDIAN__
23999#define vsliq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24000  uint32x4_t __s0 = __p0; \
24001  uint32x4_t __s1 = __p1; \
24002  uint32x4_t __ret; \
24003  __ret = (uint32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
24004  __ret; \
24005})
24006#else
24007#define vsliq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24008  uint32x4_t __s0 = __p0; \
24009  uint32x4_t __s1 = __p1; \
24010  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24011  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24012  uint32x4_t __ret; \
24013  __ret = (uint32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
24014  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24015  __ret; \
24016})
24017#endif
24018
24019#ifdef __LITTLE_ENDIAN__
24020#define vsliq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24021  uint64x2_t __s0 = __p0; \
24022  uint64x2_t __s1 = __p1; \
24023  uint64x2_t __ret; \
24024  __ret = (uint64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
24025  __ret; \
24026})
24027#else
24028#define vsliq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24029  uint64x2_t __s0 = __p0; \
24030  uint64x2_t __s1 = __p1; \
24031  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24032  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24033  uint64x2_t __ret; \
24034  __ret = (uint64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
24035  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24036  __ret; \
24037})
24038#endif
24039
24040#ifdef __LITTLE_ENDIAN__
24041#define vsliq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24042  uint16x8_t __s0 = __p0; \
24043  uint16x8_t __s1 = __p1; \
24044  uint16x8_t __ret; \
24045  __ret = (uint16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
24046  __ret; \
24047})
24048#else
24049#define vsliq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24050  uint16x8_t __s0 = __p0; \
24051  uint16x8_t __s1 = __p1; \
24052  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24053  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24054  uint16x8_t __ret; \
24055  __ret = (uint16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
24056  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24057  __ret; \
24058})
24059#endif
24060
24061#ifdef __LITTLE_ENDIAN__
24062#define vsliq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24063  int8x16_t __s0 = __p0; \
24064  int8x16_t __s1 = __p1; \
24065  int8x16_t __ret; \
24066  __ret = (int8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
24067  __ret; \
24068})
24069#else
24070#define vsliq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24071  int8x16_t __s0 = __p0; \
24072  int8x16_t __s1 = __p1; \
24073  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24074  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24075  int8x16_t __ret; \
24076  __ret = (int8x16_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
24077  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24078  __ret; \
24079})
24080#endif
24081
24082#ifdef __LITTLE_ENDIAN__
24083#define vsliq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24084  int32x4_t __s0 = __p0; \
24085  int32x4_t __s1 = __p1; \
24086  int32x4_t __ret; \
24087  __ret = (int32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
24088  __ret; \
24089})
24090#else
24091#define vsliq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24092  int32x4_t __s0 = __p0; \
24093  int32x4_t __s1 = __p1; \
24094  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24095  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24096  int32x4_t __ret; \
24097  __ret = (int32x4_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
24098  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24099  __ret; \
24100})
24101#endif
24102
24103#ifdef __LITTLE_ENDIAN__
24104#define vsliq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24105  int64x2_t __s0 = __p0; \
24106  int64x2_t __s1 = __p1; \
24107  int64x2_t __ret; \
24108  __ret = (int64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
24109  __ret; \
24110})
24111#else
24112#define vsliq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24113  int64x2_t __s0 = __p0; \
24114  int64x2_t __s1 = __p1; \
24115  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24116  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24117  int64x2_t __ret; \
24118  __ret = (int64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
24119  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24120  __ret; \
24121})
24122#endif
24123
24124#ifdef __LITTLE_ENDIAN__
24125#define vsliq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24126  int16x8_t __s0 = __p0; \
24127  int16x8_t __s1 = __p1; \
24128  int16x8_t __ret; \
24129  __ret = (int16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
24130  __ret; \
24131})
24132#else
24133#define vsliq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24134  int16x8_t __s0 = __p0; \
24135  int16x8_t __s1 = __p1; \
24136  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24137  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24138  int16x8_t __ret; \
24139  __ret = (int16x8_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
24140  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24141  __ret; \
24142})
24143#endif
24144
24145#ifdef __LITTLE_ENDIAN__
24146#define vsli_n_u8(__p0, __p1, __p2) __extension__ ({ \
24147  uint8x8_t __s0 = __p0; \
24148  uint8x8_t __s1 = __p1; \
24149  uint8x8_t __ret; \
24150  __ret = (uint8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
24151  __ret; \
24152})
24153#else
24154#define vsli_n_u8(__p0, __p1, __p2) __extension__ ({ \
24155  uint8x8_t __s0 = __p0; \
24156  uint8x8_t __s1 = __p1; \
24157  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24158  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24159  uint8x8_t __ret; \
24160  __ret = (uint8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
24161  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24162  __ret; \
24163})
24164#endif
24165
24166#ifdef __LITTLE_ENDIAN__
24167#define vsli_n_u32(__p0, __p1, __p2) __extension__ ({ \
24168  uint32x2_t __s0 = __p0; \
24169  uint32x2_t __s1 = __p1; \
24170  uint32x2_t __ret; \
24171  __ret = (uint32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
24172  __ret; \
24173})
24174#else
24175#define vsli_n_u32(__p0, __p1, __p2) __extension__ ({ \
24176  uint32x2_t __s0 = __p0; \
24177  uint32x2_t __s1 = __p1; \
24178  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24179  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24180  uint32x2_t __ret; \
24181  __ret = (uint32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
24182  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24183  __ret; \
24184})
24185#endif
24186
24187#ifdef __LITTLE_ENDIAN__
24188#define vsli_n_u64(__p0, __p1, __p2) __extension__ ({ \
24189  uint64x1_t __s0 = __p0; \
24190  uint64x1_t __s1 = __p1; \
24191  uint64x1_t __ret; \
24192  __ret = (uint64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24193  __ret; \
24194})
24195#else
24196#define vsli_n_u64(__p0, __p1, __p2) __extension__ ({ \
24197  uint64x1_t __s0 = __p0; \
24198  uint64x1_t __s1 = __p1; \
24199  uint64x1_t __ret; \
24200  __ret = (uint64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24201  __ret; \
24202})
24203#endif
24204
24205#ifdef __LITTLE_ENDIAN__
24206#define vsli_n_u16(__p0, __p1, __p2) __extension__ ({ \
24207  uint16x4_t __s0 = __p0; \
24208  uint16x4_t __s1 = __p1; \
24209  uint16x4_t __ret; \
24210  __ret = (uint16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
24211  __ret; \
24212})
24213#else
24214#define vsli_n_u16(__p0, __p1, __p2) __extension__ ({ \
24215  uint16x4_t __s0 = __p0; \
24216  uint16x4_t __s1 = __p1; \
24217  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24218  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24219  uint16x4_t __ret; \
24220  __ret = (uint16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
24221  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24222  __ret; \
24223})
24224#endif
24225
24226#ifdef __LITTLE_ENDIAN__
24227#define vsli_n_s8(__p0, __p1, __p2) __extension__ ({ \
24228  int8x8_t __s0 = __p0; \
24229  int8x8_t __s1 = __p1; \
24230  int8x8_t __ret; \
24231  __ret = (int8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
24232  __ret; \
24233})
24234#else
24235#define vsli_n_s8(__p0, __p1, __p2) __extension__ ({ \
24236  int8x8_t __s0 = __p0; \
24237  int8x8_t __s1 = __p1; \
24238  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24239  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24240  int8x8_t __ret; \
24241  __ret = (int8x8_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
24242  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24243  __ret; \
24244})
24245#endif
24246
24247#ifdef __LITTLE_ENDIAN__
24248#define vsli_n_s32(__p0, __p1, __p2) __extension__ ({ \
24249  int32x2_t __s0 = __p0; \
24250  int32x2_t __s1 = __p1; \
24251  int32x2_t __ret; \
24252  __ret = (int32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
24253  __ret; \
24254})
24255#else
24256#define vsli_n_s32(__p0, __p1, __p2) __extension__ ({ \
24257  int32x2_t __s0 = __p0; \
24258  int32x2_t __s1 = __p1; \
24259  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24260  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24261  int32x2_t __ret; \
24262  __ret = (int32x2_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
24263  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24264  __ret; \
24265})
24266#endif
24267
24268#ifdef __LITTLE_ENDIAN__
24269#define vsli_n_s64(__p0, __p1, __p2) __extension__ ({ \
24270  int64x1_t __s0 = __p0; \
24271  int64x1_t __s1 = __p1; \
24272  int64x1_t __ret; \
24273  __ret = (int64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24274  __ret; \
24275})
24276#else
24277#define vsli_n_s64(__p0, __p1, __p2) __extension__ ({ \
24278  int64x1_t __s0 = __p0; \
24279  int64x1_t __s1 = __p1; \
24280  int64x1_t __ret; \
24281  __ret = (int64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24282  __ret; \
24283})
24284#endif
24285
24286#ifdef __LITTLE_ENDIAN__
24287#define vsli_n_s16(__p0, __p1, __p2) __extension__ ({ \
24288  int16x4_t __s0 = __p0; \
24289  int16x4_t __s1 = __p1; \
24290  int16x4_t __ret; \
24291  __ret = (int16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
24292  __ret; \
24293})
24294#else
24295#define vsli_n_s16(__p0, __p1, __p2) __extension__ ({ \
24296  int16x4_t __s0 = __p0; \
24297  int16x4_t __s1 = __p1; \
24298  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24299  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24300  int16x4_t __ret; \
24301  __ret = (int16x4_t) __builtin_neon_vsli_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
24302  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24303  __ret; \
24304})
24305#endif
24306
24307#ifdef __LITTLE_ENDIAN__
24308#define vsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24309  uint8x16_t __s0 = __p0; \
24310  uint8x16_t __s1 = __p1; \
24311  uint8x16_t __ret; \
24312  __ret = (uint8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
24313  __ret; \
24314})
24315#else
24316#define vsraq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24317  uint8x16_t __s0 = __p0; \
24318  uint8x16_t __s1 = __p1; \
24319  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24320  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24321  uint8x16_t __ret; \
24322  __ret = (uint8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
24323  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24324  __ret; \
24325})
24326#endif
24327
24328#ifdef __LITTLE_ENDIAN__
24329#define vsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24330  uint32x4_t __s0 = __p0; \
24331  uint32x4_t __s1 = __p1; \
24332  uint32x4_t __ret; \
24333  __ret = (uint32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
24334  __ret; \
24335})
24336#else
24337#define vsraq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24338  uint32x4_t __s0 = __p0; \
24339  uint32x4_t __s1 = __p1; \
24340  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24341  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24342  uint32x4_t __ret; \
24343  __ret = (uint32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
24344  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24345  __ret; \
24346})
24347#endif
24348
24349#ifdef __LITTLE_ENDIAN__
24350#define vsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24351  uint64x2_t __s0 = __p0; \
24352  uint64x2_t __s1 = __p1; \
24353  uint64x2_t __ret; \
24354  __ret = (uint64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
24355  __ret; \
24356})
24357#else
24358#define vsraq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24359  uint64x2_t __s0 = __p0; \
24360  uint64x2_t __s1 = __p1; \
24361  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24362  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24363  uint64x2_t __ret; \
24364  __ret = (uint64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
24365  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24366  __ret; \
24367})
24368#endif
24369
24370#ifdef __LITTLE_ENDIAN__
24371#define vsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24372  uint16x8_t __s0 = __p0; \
24373  uint16x8_t __s1 = __p1; \
24374  uint16x8_t __ret; \
24375  __ret = (uint16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
24376  __ret; \
24377})
24378#else
24379#define vsraq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24380  uint16x8_t __s0 = __p0; \
24381  uint16x8_t __s1 = __p1; \
24382  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24383  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24384  uint16x8_t __ret; \
24385  __ret = (uint16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
24386  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24387  __ret; \
24388})
24389#endif
24390
24391#ifdef __LITTLE_ENDIAN__
24392#define vsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24393  int8x16_t __s0 = __p0; \
24394  int8x16_t __s1 = __p1; \
24395  int8x16_t __ret; \
24396  __ret = (int8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
24397  __ret; \
24398})
24399#else
24400#define vsraq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24401  int8x16_t __s0 = __p0; \
24402  int8x16_t __s1 = __p1; \
24403  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24404  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24405  int8x16_t __ret; \
24406  __ret = (int8x16_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
24407  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24408  __ret; \
24409})
24410#endif
24411
24412#ifdef __LITTLE_ENDIAN__
24413#define vsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24414  int32x4_t __s0 = __p0; \
24415  int32x4_t __s1 = __p1; \
24416  int32x4_t __ret; \
24417  __ret = (int32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
24418  __ret; \
24419})
24420#else
24421#define vsraq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24422  int32x4_t __s0 = __p0; \
24423  int32x4_t __s1 = __p1; \
24424  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24425  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24426  int32x4_t __ret; \
24427  __ret = (int32x4_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
24428  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24429  __ret; \
24430})
24431#endif
24432
24433#ifdef __LITTLE_ENDIAN__
24434#define vsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24435  int64x2_t __s0 = __p0; \
24436  int64x2_t __s1 = __p1; \
24437  int64x2_t __ret; \
24438  __ret = (int64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
24439  __ret; \
24440})
24441#else
24442#define vsraq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24443  int64x2_t __s0 = __p0; \
24444  int64x2_t __s1 = __p1; \
24445  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24446  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24447  int64x2_t __ret; \
24448  __ret = (int64x2_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
24449  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24450  __ret; \
24451})
24452#endif
24453
24454#ifdef __LITTLE_ENDIAN__
24455#define vsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24456  int16x8_t __s0 = __p0; \
24457  int16x8_t __s1 = __p1; \
24458  int16x8_t __ret; \
24459  __ret = (int16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
24460  __ret; \
24461})
24462#else
24463#define vsraq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24464  int16x8_t __s0 = __p0; \
24465  int16x8_t __s1 = __p1; \
24466  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24467  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24468  int16x8_t __ret; \
24469  __ret = (int16x8_t) __builtin_neon_vsraq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
24470  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24471  __ret; \
24472})
24473#endif
24474
24475#ifdef __LITTLE_ENDIAN__
24476#define vsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
24477  uint8x8_t __s0 = __p0; \
24478  uint8x8_t __s1 = __p1; \
24479  uint8x8_t __ret; \
24480  __ret = (uint8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
24481  __ret; \
24482})
24483#else
24484#define vsra_n_u8(__p0, __p1, __p2) __extension__ ({ \
24485  uint8x8_t __s0 = __p0; \
24486  uint8x8_t __s1 = __p1; \
24487  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24488  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24489  uint8x8_t __ret; \
24490  __ret = (uint8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
24491  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24492  __ret; \
24493})
24494#endif
24495
24496#ifdef __LITTLE_ENDIAN__
24497#define vsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
24498  uint32x2_t __s0 = __p0; \
24499  uint32x2_t __s1 = __p1; \
24500  uint32x2_t __ret; \
24501  __ret = (uint32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
24502  __ret; \
24503})
24504#else
24505#define vsra_n_u32(__p0, __p1, __p2) __extension__ ({ \
24506  uint32x2_t __s0 = __p0; \
24507  uint32x2_t __s1 = __p1; \
24508  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24509  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24510  uint32x2_t __ret; \
24511  __ret = (uint32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
24512  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24513  __ret; \
24514})
24515#endif
24516
24517#ifdef __LITTLE_ENDIAN__
24518#define vsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
24519  uint64x1_t __s0 = __p0; \
24520  uint64x1_t __s1 = __p1; \
24521  uint64x1_t __ret; \
24522  __ret = (uint64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24523  __ret; \
24524})
24525#else
24526#define vsra_n_u64(__p0, __p1, __p2) __extension__ ({ \
24527  uint64x1_t __s0 = __p0; \
24528  uint64x1_t __s1 = __p1; \
24529  uint64x1_t __ret; \
24530  __ret = (uint64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24531  __ret; \
24532})
24533#endif
24534
24535#ifdef __LITTLE_ENDIAN__
24536#define vsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
24537  uint16x4_t __s0 = __p0; \
24538  uint16x4_t __s1 = __p1; \
24539  uint16x4_t __ret; \
24540  __ret = (uint16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
24541  __ret; \
24542})
24543#else
24544#define vsra_n_u16(__p0, __p1, __p2) __extension__ ({ \
24545  uint16x4_t __s0 = __p0; \
24546  uint16x4_t __s1 = __p1; \
24547  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24548  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24549  uint16x4_t __ret; \
24550  __ret = (uint16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
24551  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24552  __ret; \
24553})
24554#endif
24555
24556#ifdef __LITTLE_ENDIAN__
24557#define vsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
24558  int8x8_t __s0 = __p0; \
24559  int8x8_t __s1 = __p1; \
24560  int8x8_t __ret; \
24561  __ret = (int8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
24562  __ret; \
24563})
24564#else
24565#define vsra_n_s8(__p0, __p1, __p2) __extension__ ({ \
24566  int8x8_t __s0 = __p0; \
24567  int8x8_t __s1 = __p1; \
24568  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24569  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24570  int8x8_t __ret; \
24571  __ret = (int8x8_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
24572  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24573  __ret; \
24574})
24575#endif
24576
24577#ifdef __LITTLE_ENDIAN__
24578#define vsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
24579  int32x2_t __s0 = __p0; \
24580  int32x2_t __s1 = __p1; \
24581  int32x2_t __ret; \
24582  __ret = (int32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
24583  __ret; \
24584})
24585#else
24586#define vsra_n_s32(__p0, __p1, __p2) __extension__ ({ \
24587  int32x2_t __s0 = __p0; \
24588  int32x2_t __s1 = __p1; \
24589  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24590  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24591  int32x2_t __ret; \
24592  __ret = (int32x2_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
24593  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24594  __ret; \
24595})
24596#endif
24597
24598#ifdef __LITTLE_ENDIAN__
24599#define vsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
24600  int64x1_t __s0 = __p0; \
24601  int64x1_t __s1 = __p1; \
24602  int64x1_t __ret; \
24603  __ret = (int64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24604  __ret; \
24605})
24606#else
24607#define vsra_n_s64(__p0, __p1, __p2) __extension__ ({ \
24608  int64x1_t __s0 = __p0; \
24609  int64x1_t __s1 = __p1; \
24610  int64x1_t __ret; \
24611  __ret = (int64x1_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
24612  __ret; \
24613})
24614#endif
24615
24616#ifdef __LITTLE_ENDIAN__
24617#define vsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
24618  int16x4_t __s0 = __p0; \
24619  int16x4_t __s1 = __p1; \
24620  int16x4_t __ret; \
24621  __ret = (int16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
24622  __ret; \
24623})
24624#else
24625#define vsra_n_s16(__p0, __p1, __p2) __extension__ ({ \
24626  int16x4_t __s0 = __p0; \
24627  int16x4_t __s1 = __p1; \
24628  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24629  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24630  int16x4_t __ret; \
24631  __ret = (int16x4_t) __builtin_neon_vsra_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
24632  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24633  __ret; \
24634})
24635#endif
24636
24637#ifdef __LITTLE_ENDIAN__
24638#define vsri_n_p8(__p0, __p1, __p2) __extension__ ({ \
24639  poly8x8_t __s0 = __p0; \
24640  poly8x8_t __s1 = __p1; \
24641  poly8x8_t __ret; \
24642  __ret = (poly8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 4); \
24643  __ret; \
24644})
24645#else
24646#define vsri_n_p8(__p0, __p1, __p2) __extension__ ({ \
24647  poly8x8_t __s0 = __p0; \
24648  poly8x8_t __s1 = __p1; \
24649  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24650  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24651  poly8x8_t __ret; \
24652  __ret = (poly8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 4); \
24653  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24654  __ret; \
24655})
24656#endif
24657
24658#ifdef __LITTLE_ENDIAN__
24659#define vsri_n_p16(__p0, __p1, __p2) __extension__ ({ \
24660  poly16x4_t __s0 = __p0; \
24661  poly16x4_t __s1 = __p1; \
24662  poly16x4_t __ret; \
24663  __ret = (poly16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 5); \
24664  __ret; \
24665})
24666#else
24667#define vsri_n_p16(__p0, __p1, __p2) __extension__ ({ \
24668  poly16x4_t __s0 = __p0; \
24669  poly16x4_t __s1 = __p1; \
24670  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24671  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24672  poly16x4_t __ret; \
24673  __ret = (poly16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 5); \
24674  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24675  __ret; \
24676})
24677#endif
24678
24679#ifdef __LITTLE_ENDIAN__
24680#define vsriq_n_p8(__p0, __p1, __p2) __extension__ ({ \
24681  poly8x16_t __s0 = __p0; \
24682  poly8x16_t __s1 = __p1; \
24683  poly8x16_t __ret; \
24684  __ret = (poly8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 36); \
24685  __ret; \
24686})
24687#else
24688#define vsriq_n_p8(__p0, __p1, __p2) __extension__ ({ \
24689  poly8x16_t __s0 = __p0; \
24690  poly8x16_t __s1 = __p1; \
24691  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24692  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24693  poly8x16_t __ret; \
24694  __ret = (poly8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 36); \
24695  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24696  __ret; \
24697})
24698#endif
24699
24700#ifdef __LITTLE_ENDIAN__
24701#define vsriq_n_p16(__p0, __p1, __p2) __extension__ ({ \
24702  poly16x8_t __s0 = __p0; \
24703  poly16x8_t __s1 = __p1; \
24704  poly16x8_t __ret; \
24705  __ret = (poly16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 37); \
24706  __ret; \
24707})
24708#else
24709#define vsriq_n_p16(__p0, __p1, __p2) __extension__ ({ \
24710  poly16x8_t __s0 = __p0; \
24711  poly16x8_t __s1 = __p1; \
24712  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24713  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24714  poly16x8_t __ret; \
24715  __ret = (poly16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 37); \
24716  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24717  __ret; \
24718})
24719#endif
24720
24721#ifdef __LITTLE_ENDIAN__
24722#define vsriq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24723  uint8x16_t __s0 = __p0; \
24724  uint8x16_t __s1 = __p1; \
24725  uint8x16_t __ret; \
24726  __ret = (uint8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 48); \
24727  __ret; \
24728})
24729#else
24730#define vsriq_n_u8(__p0, __p1, __p2) __extension__ ({ \
24731  uint8x16_t __s0 = __p0; \
24732  uint8x16_t __s1 = __p1; \
24733  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24734  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24735  uint8x16_t __ret; \
24736  __ret = (uint8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 48); \
24737  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24738  __ret; \
24739})
24740#endif
24741
24742#ifdef __LITTLE_ENDIAN__
24743#define vsriq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24744  uint32x4_t __s0 = __p0; \
24745  uint32x4_t __s1 = __p1; \
24746  uint32x4_t __ret; \
24747  __ret = (uint32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 50); \
24748  __ret; \
24749})
24750#else
24751#define vsriq_n_u32(__p0, __p1, __p2) __extension__ ({ \
24752  uint32x4_t __s0 = __p0; \
24753  uint32x4_t __s1 = __p1; \
24754  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24755  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24756  uint32x4_t __ret; \
24757  __ret = (uint32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 50); \
24758  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24759  __ret; \
24760})
24761#endif
24762
24763#ifdef __LITTLE_ENDIAN__
24764#define vsriq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24765  uint64x2_t __s0 = __p0; \
24766  uint64x2_t __s1 = __p1; \
24767  uint64x2_t __ret; \
24768  __ret = (uint64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \
24769  __ret; \
24770})
24771#else
24772#define vsriq_n_u64(__p0, __p1, __p2) __extension__ ({ \
24773  uint64x2_t __s0 = __p0; \
24774  uint64x2_t __s1 = __p1; \
24775  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24776  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24777  uint64x2_t __ret; \
24778  __ret = (uint64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \
24779  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24780  __ret; \
24781})
24782#endif
24783
24784#ifdef __LITTLE_ENDIAN__
24785#define vsriq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24786  uint16x8_t __s0 = __p0; \
24787  uint16x8_t __s1 = __p1; \
24788  uint16x8_t __ret; \
24789  __ret = (uint16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 49); \
24790  __ret; \
24791})
24792#else
24793#define vsriq_n_u16(__p0, __p1, __p2) __extension__ ({ \
24794  uint16x8_t __s0 = __p0; \
24795  uint16x8_t __s1 = __p1; \
24796  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24797  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24798  uint16x8_t __ret; \
24799  __ret = (uint16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 49); \
24800  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24801  __ret; \
24802})
24803#endif
24804
24805#ifdef __LITTLE_ENDIAN__
24806#define vsriq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24807  int8x16_t __s0 = __p0; \
24808  int8x16_t __s1 = __p1; \
24809  int8x16_t __ret; \
24810  __ret = (int8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 32); \
24811  __ret; \
24812})
24813#else
24814#define vsriq_n_s8(__p0, __p1, __p2) __extension__ ({ \
24815  int8x16_t __s0 = __p0; \
24816  int8x16_t __s1 = __p1; \
24817  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24818  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24819  int8x16_t __ret; \
24820  __ret = (int8x16_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 32); \
24821  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
24822  __ret; \
24823})
24824#endif
24825
24826#ifdef __LITTLE_ENDIAN__
24827#define vsriq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24828  int32x4_t __s0 = __p0; \
24829  int32x4_t __s1 = __p1; \
24830  int32x4_t __ret; \
24831  __ret = (int32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 34); \
24832  __ret; \
24833})
24834#else
24835#define vsriq_n_s32(__p0, __p1, __p2) __extension__ ({ \
24836  int32x4_t __s0 = __p0; \
24837  int32x4_t __s1 = __p1; \
24838  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24839  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24840  int32x4_t __ret; \
24841  __ret = (int32x4_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 34); \
24842  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24843  __ret; \
24844})
24845#endif
24846
24847#ifdef __LITTLE_ENDIAN__
24848#define vsriq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24849  int64x2_t __s0 = __p0; \
24850  int64x2_t __s1 = __p1; \
24851  int64x2_t __ret; \
24852  __ret = (int64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 35); \
24853  __ret; \
24854})
24855#else
24856#define vsriq_n_s64(__p0, __p1, __p2) __extension__ ({ \
24857  int64x2_t __s0 = __p0; \
24858  int64x2_t __s1 = __p1; \
24859  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24860  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24861  int64x2_t __ret; \
24862  __ret = (int64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 35); \
24863  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24864  __ret; \
24865})
24866#endif
24867
24868#ifdef __LITTLE_ENDIAN__
24869#define vsriq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24870  int16x8_t __s0 = __p0; \
24871  int16x8_t __s1 = __p1; \
24872  int16x8_t __ret; \
24873  __ret = (int16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 33); \
24874  __ret; \
24875})
24876#else
24877#define vsriq_n_s16(__p0, __p1, __p2) __extension__ ({ \
24878  int16x8_t __s0 = __p0; \
24879  int16x8_t __s1 = __p1; \
24880  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24881  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24882  int16x8_t __ret; \
24883  __ret = (int16x8_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 33); \
24884  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24885  __ret; \
24886})
24887#endif
24888
24889#ifdef __LITTLE_ENDIAN__
24890#define vsri_n_u8(__p0, __p1, __p2) __extension__ ({ \
24891  uint8x8_t __s0 = __p0; \
24892  uint8x8_t __s1 = __p1; \
24893  uint8x8_t __ret; \
24894  __ret = (uint8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 16); \
24895  __ret; \
24896})
24897#else
24898#define vsri_n_u8(__p0, __p1, __p2) __extension__ ({ \
24899  uint8x8_t __s0 = __p0; \
24900  uint8x8_t __s1 = __p1; \
24901  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24902  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24903  uint8x8_t __ret; \
24904  __ret = (uint8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 16); \
24905  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24906  __ret; \
24907})
24908#endif
24909
24910#ifdef __LITTLE_ENDIAN__
24911#define vsri_n_u32(__p0, __p1, __p2) __extension__ ({ \
24912  uint32x2_t __s0 = __p0; \
24913  uint32x2_t __s1 = __p1; \
24914  uint32x2_t __ret; \
24915  __ret = (uint32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 18); \
24916  __ret; \
24917})
24918#else
24919#define vsri_n_u32(__p0, __p1, __p2) __extension__ ({ \
24920  uint32x2_t __s0 = __p0; \
24921  uint32x2_t __s1 = __p1; \
24922  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
24923  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
24924  uint32x2_t __ret; \
24925  __ret = (uint32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 18); \
24926  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
24927  __ret; \
24928})
24929#endif
24930
24931#ifdef __LITTLE_ENDIAN__
24932#define vsri_n_u64(__p0, __p1, __p2) __extension__ ({ \
24933  uint64x1_t __s0 = __p0; \
24934  uint64x1_t __s1 = __p1; \
24935  uint64x1_t __ret; \
24936  __ret = (uint64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24937  __ret; \
24938})
24939#else
24940#define vsri_n_u64(__p0, __p1, __p2) __extension__ ({ \
24941  uint64x1_t __s0 = __p0; \
24942  uint64x1_t __s1 = __p1; \
24943  uint64x1_t __ret; \
24944  __ret = (uint64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 19); \
24945  __ret; \
24946})
24947#endif
24948
24949#ifdef __LITTLE_ENDIAN__
24950#define vsri_n_u16(__p0, __p1, __p2) __extension__ ({ \
24951  uint16x4_t __s0 = __p0; \
24952  uint16x4_t __s1 = __p1; \
24953  uint16x4_t __ret; \
24954  __ret = (uint16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 17); \
24955  __ret; \
24956})
24957#else
24958#define vsri_n_u16(__p0, __p1, __p2) __extension__ ({ \
24959  uint16x4_t __s0 = __p0; \
24960  uint16x4_t __s1 = __p1; \
24961  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
24962  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
24963  uint16x4_t __ret; \
24964  __ret = (uint16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 17); \
24965  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
24966  __ret; \
24967})
24968#endif
24969
24970#ifdef __LITTLE_ENDIAN__
24971#define vsri_n_s8(__p0, __p1, __p2) __extension__ ({ \
24972  int8x8_t __s0 = __p0; \
24973  int8x8_t __s1 = __p1; \
24974  int8x8_t __ret; \
24975  __ret = (int8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 0); \
24976  __ret; \
24977})
24978#else
24979#define vsri_n_s8(__p0, __p1, __p2) __extension__ ({ \
24980  int8x8_t __s0 = __p0; \
24981  int8x8_t __s1 = __p1; \
24982  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
24983  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
24984  int8x8_t __ret; \
24985  __ret = (int8x8_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 0); \
24986  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
24987  __ret; \
24988})
24989#endif
24990
24991#ifdef __LITTLE_ENDIAN__
24992#define vsri_n_s32(__p0, __p1, __p2) __extension__ ({ \
24993  int32x2_t __s0 = __p0; \
24994  int32x2_t __s1 = __p1; \
24995  int32x2_t __ret; \
24996  __ret = (int32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 2); \
24997  __ret; \
24998})
24999#else
25000#define vsri_n_s32(__p0, __p1, __p2) __extension__ ({ \
25001  int32x2_t __s0 = __p0; \
25002  int32x2_t __s1 = __p1; \
25003  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
25004  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25005  int32x2_t __ret; \
25006  __ret = (int32x2_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 2); \
25007  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
25008  __ret; \
25009})
25010#endif
25011
25012#ifdef __LITTLE_ENDIAN__
25013#define vsri_n_s64(__p0, __p1, __p2) __extension__ ({ \
25014  int64x1_t __s0 = __p0; \
25015  int64x1_t __s1 = __p1; \
25016  int64x1_t __ret; \
25017  __ret = (int64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
25018  __ret; \
25019})
25020#else
25021#define vsri_n_s64(__p0, __p1, __p2) __extension__ ({ \
25022  int64x1_t __s0 = __p0; \
25023  int64x1_t __s1 = __p1; \
25024  int64x1_t __ret; \
25025  __ret = (int64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 3); \
25026  __ret; \
25027})
25028#endif
25029
25030#ifdef __LITTLE_ENDIAN__
25031#define vsri_n_s16(__p0, __p1, __p2) __extension__ ({ \
25032  int16x4_t __s0 = __p0; \
25033  int16x4_t __s1 = __p1; \
25034  int16x4_t __ret; \
25035  __ret = (int16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 1); \
25036  __ret; \
25037})
25038#else
25039#define vsri_n_s16(__p0, __p1, __p2) __extension__ ({ \
25040  int16x4_t __s0 = __p0; \
25041  int16x4_t __s1 = __p1; \
25042  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
25043  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25044  int16x4_t __ret; \
25045  __ret = (int16x4_t) __builtin_neon_vsri_n_v((int8x8_t)__rev0, (int8x8_t)__rev1, __p2, 1); \
25046  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
25047  __ret; \
25048})
25049#endif
25050
25051#ifdef __LITTLE_ENDIAN__
25052#define vst1_p8(__p0, __p1) __extension__ ({ \
25053  poly8x8_t __s1 = __p1; \
25054  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 4); \
25055})
25056#else
25057#define vst1_p8(__p0, __p1) __extension__ ({ \
25058  poly8x8_t __s1 = __p1; \
25059  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25060  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 4); \
25061})
25062#endif
25063
25064#ifdef __LITTLE_ENDIAN__
25065#define vst1_p16(__p0, __p1) __extension__ ({ \
25066  poly16x4_t __s1 = __p1; \
25067  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 5); \
25068})
25069#else
25070#define vst1_p16(__p0, __p1) __extension__ ({ \
25071  poly16x4_t __s1 = __p1; \
25072  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25073  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 5); \
25074})
25075#endif
25076
25077#ifdef __LITTLE_ENDIAN__
25078#define vst1q_p8(__p0, __p1) __extension__ ({ \
25079  poly8x16_t __s1 = __p1; \
25080  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 36); \
25081})
25082#else
25083#define vst1q_p8(__p0, __p1) __extension__ ({ \
25084  poly8x16_t __s1 = __p1; \
25085  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25086  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 36); \
25087})
25088#endif
25089
25090#ifdef __LITTLE_ENDIAN__
25091#define vst1q_p16(__p0, __p1) __extension__ ({ \
25092  poly16x8_t __s1 = __p1; \
25093  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 37); \
25094})
25095#else
25096#define vst1q_p16(__p0, __p1) __extension__ ({ \
25097  poly16x8_t __s1 = __p1; \
25098  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25099  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 37); \
25100})
25101#endif
25102
25103#ifdef __LITTLE_ENDIAN__
25104#define vst1q_u8(__p0, __p1) __extension__ ({ \
25105  uint8x16_t __s1 = __p1; \
25106  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 48); \
25107})
25108#else
25109#define vst1q_u8(__p0, __p1) __extension__ ({ \
25110  uint8x16_t __s1 = __p1; \
25111  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25112  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 48); \
25113})
25114#endif
25115
25116#ifdef __LITTLE_ENDIAN__
25117#define vst1q_u32(__p0, __p1) __extension__ ({ \
25118  uint32x4_t __s1 = __p1; \
25119  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 50); \
25120})
25121#else
25122#define vst1q_u32(__p0, __p1) __extension__ ({ \
25123  uint32x4_t __s1 = __p1; \
25124  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25125  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 50); \
25126})
25127#endif
25128
25129#ifdef __LITTLE_ENDIAN__
25130#define vst1q_u64(__p0, __p1) __extension__ ({ \
25131  uint64x2_t __s1 = __p1; \
25132  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 51); \
25133})
25134#else
25135#define vst1q_u64(__p0, __p1) __extension__ ({ \
25136  uint64x2_t __s1 = __p1; \
25137  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25138  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 51); \
25139})
25140#endif
25141
25142#ifdef __LITTLE_ENDIAN__
25143#define vst1q_u16(__p0, __p1) __extension__ ({ \
25144  uint16x8_t __s1 = __p1; \
25145  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 49); \
25146})
25147#else
25148#define vst1q_u16(__p0, __p1) __extension__ ({ \
25149  uint16x8_t __s1 = __p1; \
25150  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25151  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 49); \
25152})
25153#endif
25154
25155#ifdef __LITTLE_ENDIAN__
25156#define vst1q_s8(__p0, __p1) __extension__ ({ \
25157  int8x16_t __s1 = __p1; \
25158  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 32); \
25159})
25160#else
25161#define vst1q_s8(__p0, __p1) __extension__ ({ \
25162  int8x16_t __s1 = __p1; \
25163  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25164  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 32); \
25165})
25166#endif
25167
25168#ifdef __LITTLE_ENDIAN__
25169#define vst1q_f32(__p0, __p1) __extension__ ({ \
25170  float32x4_t __s1 = __p1; \
25171  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 41); \
25172})
25173#else
25174#define vst1q_f32(__p0, __p1) __extension__ ({ \
25175  float32x4_t __s1 = __p1; \
25176  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25177  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 41); \
25178})
25179#endif
25180
25181#ifdef __LITTLE_ENDIAN__
25182#define vst1q_f16(__p0, __p1) __extension__ ({ \
25183  float16x8_t __s1 = __p1; \
25184  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 40); \
25185})
25186#else
25187#define vst1q_f16(__p0, __p1) __extension__ ({ \
25188  float16x8_t __s1 = __p1; \
25189  float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25190  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 40); \
25191})
25192#endif
25193
25194#ifdef __LITTLE_ENDIAN__
25195#define vst1q_s32(__p0, __p1) __extension__ ({ \
25196  int32x4_t __s1 = __p1; \
25197  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 34); \
25198})
25199#else
25200#define vst1q_s32(__p0, __p1) __extension__ ({ \
25201  int32x4_t __s1 = __p1; \
25202  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25203  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 34); \
25204})
25205#endif
25206
25207#ifdef __LITTLE_ENDIAN__
25208#define vst1q_s64(__p0, __p1) __extension__ ({ \
25209  int64x2_t __s1 = __p1; \
25210  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 35); \
25211})
25212#else
25213#define vst1q_s64(__p0, __p1) __extension__ ({ \
25214  int64x2_t __s1 = __p1; \
25215  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25216  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 35); \
25217})
25218#endif
25219
25220#ifdef __LITTLE_ENDIAN__
25221#define vst1q_s16(__p0, __p1) __extension__ ({ \
25222  int16x8_t __s1 = __p1; \
25223  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 33); \
25224})
25225#else
25226#define vst1q_s16(__p0, __p1) __extension__ ({ \
25227  int16x8_t __s1 = __p1; \
25228  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25229  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 33); \
25230})
25231#endif
25232
25233#ifdef __LITTLE_ENDIAN__
25234#define vst1_u8(__p0, __p1) __extension__ ({ \
25235  uint8x8_t __s1 = __p1; \
25236  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 16); \
25237})
25238#else
25239#define vst1_u8(__p0, __p1) __extension__ ({ \
25240  uint8x8_t __s1 = __p1; \
25241  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25242  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 16); \
25243})
25244#endif
25245
25246#ifdef __LITTLE_ENDIAN__
25247#define vst1_u32(__p0, __p1) __extension__ ({ \
25248  uint32x2_t __s1 = __p1; \
25249  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 18); \
25250})
25251#else
25252#define vst1_u32(__p0, __p1) __extension__ ({ \
25253  uint32x2_t __s1 = __p1; \
25254  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25255  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 18); \
25256})
25257#endif
25258
25259#ifdef __LITTLE_ENDIAN__
25260#define vst1_u64(__p0, __p1) __extension__ ({ \
25261  uint64x1_t __s1 = __p1; \
25262  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 19); \
25263})
25264#else
25265#define vst1_u64(__p0, __p1) __extension__ ({ \
25266  uint64x1_t __s1 = __p1; \
25267  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 19); \
25268})
25269#endif
25270
25271#ifdef __LITTLE_ENDIAN__
25272#define vst1_u16(__p0, __p1) __extension__ ({ \
25273  uint16x4_t __s1 = __p1; \
25274  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 17); \
25275})
25276#else
25277#define vst1_u16(__p0, __p1) __extension__ ({ \
25278  uint16x4_t __s1 = __p1; \
25279  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25280  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 17); \
25281})
25282#endif
25283
25284#ifdef __LITTLE_ENDIAN__
25285#define vst1_s8(__p0, __p1) __extension__ ({ \
25286  int8x8_t __s1 = __p1; \
25287  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 0); \
25288})
25289#else
25290#define vst1_s8(__p0, __p1) __extension__ ({ \
25291  int8x8_t __s1 = __p1; \
25292  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25293  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 0); \
25294})
25295#endif
25296
25297#ifdef __LITTLE_ENDIAN__
25298#define vst1_f32(__p0, __p1) __extension__ ({ \
25299  float32x2_t __s1 = __p1; \
25300  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 9); \
25301})
25302#else
25303#define vst1_f32(__p0, __p1) __extension__ ({ \
25304  float32x2_t __s1 = __p1; \
25305  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25306  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 9); \
25307})
25308#endif
25309
25310#ifdef __LITTLE_ENDIAN__
25311#define vst1_f16(__p0, __p1) __extension__ ({ \
25312  float16x4_t __s1 = __p1; \
25313  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 8); \
25314})
25315#else
25316#define vst1_f16(__p0, __p1) __extension__ ({ \
25317  float16x4_t __s1 = __p1; \
25318  float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25319  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 8); \
25320})
25321#endif
25322
25323#ifdef __LITTLE_ENDIAN__
25324#define vst1_s32(__p0, __p1) __extension__ ({ \
25325  int32x2_t __s1 = __p1; \
25326  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 2); \
25327})
25328#else
25329#define vst1_s32(__p0, __p1) __extension__ ({ \
25330  int32x2_t __s1 = __p1; \
25331  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25332  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 2); \
25333})
25334#endif
25335
25336#ifdef __LITTLE_ENDIAN__
25337#define vst1_s64(__p0, __p1) __extension__ ({ \
25338  int64x1_t __s1 = __p1; \
25339  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 3); \
25340})
25341#else
25342#define vst1_s64(__p0, __p1) __extension__ ({ \
25343  int64x1_t __s1 = __p1; \
25344  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 3); \
25345})
25346#endif
25347
25348#ifdef __LITTLE_ENDIAN__
25349#define vst1_s16(__p0, __p1) __extension__ ({ \
25350  int16x4_t __s1 = __p1; \
25351  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 1); \
25352})
25353#else
25354#define vst1_s16(__p0, __p1) __extension__ ({ \
25355  int16x4_t __s1 = __p1; \
25356  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25357  __builtin_neon_vst1_v(__p0, (int8x8_t)__rev1, 1); \
25358})
25359#endif
25360
25361#ifdef __LITTLE_ENDIAN__
25362#define vst1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25363  poly8x8_t __s1 = __p1; \
25364  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 4); \
25365})
25366#else
25367#define vst1_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25368  poly8x8_t __s1 = __p1; \
25369  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25370  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 4); \
25371})
25372#endif
25373
25374#ifdef __LITTLE_ENDIAN__
25375#define vst1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25376  poly16x4_t __s1 = __p1; \
25377  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 5); \
25378})
25379#else
25380#define vst1_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25381  poly16x4_t __s1 = __p1; \
25382  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25383  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 5); \
25384})
25385#endif
25386
25387#ifdef __LITTLE_ENDIAN__
25388#define vst1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25389  poly8x16_t __s1 = __p1; \
25390  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 36); \
25391})
25392#else
25393#define vst1q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25394  poly8x16_t __s1 = __p1; \
25395  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25396  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 36); \
25397})
25398#endif
25399
25400#ifdef __LITTLE_ENDIAN__
25401#define vst1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25402  poly16x8_t __s1 = __p1; \
25403  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 37); \
25404})
25405#else
25406#define vst1q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
25407  poly16x8_t __s1 = __p1; \
25408  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25409  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 37); \
25410})
25411#endif
25412
25413#ifdef __LITTLE_ENDIAN__
25414#define vst1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25415  uint8x16_t __s1 = __p1; \
25416  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 48); \
25417})
25418#else
25419#define vst1q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25420  uint8x16_t __s1 = __p1; \
25421  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25422  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 48); \
25423})
25424#endif
25425
25426#ifdef __LITTLE_ENDIAN__
25427#define vst1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25428  uint32x4_t __s1 = __p1; \
25429  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 50); \
25430})
25431#else
25432#define vst1q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25433  uint32x4_t __s1 = __p1; \
25434  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25435  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 50); \
25436})
25437#endif
25438
25439#ifdef __LITTLE_ENDIAN__
25440#define vst1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25441  uint64x2_t __s1 = __p1; \
25442  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 51); \
25443})
25444#else
25445#define vst1q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25446  uint64x2_t __s1 = __p1; \
25447  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25448  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 51); \
25449})
25450#endif
25451
25452#ifdef __LITTLE_ENDIAN__
25453#define vst1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25454  uint16x8_t __s1 = __p1; \
25455  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 49); \
25456})
25457#else
25458#define vst1q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25459  uint16x8_t __s1 = __p1; \
25460  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25461  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 49); \
25462})
25463#endif
25464
25465#ifdef __LITTLE_ENDIAN__
25466#define vst1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25467  int8x16_t __s1 = __p1; \
25468  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 32); \
25469})
25470#else
25471#define vst1q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25472  int8x16_t __s1 = __p1; \
25473  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25474  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 32); \
25475})
25476#endif
25477
25478#ifdef __LITTLE_ENDIAN__
25479#define vst1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25480  float32x4_t __s1 = __p1; \
25481  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 41); \
25482})
25483#else
25484#define vst1q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25485  float32x4_t __s1 = __p1; \
25486  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25487  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 41); \
25488})
25489#endif
25490
25491#ifdef __LITTLE_ENDIAN__
25492#define vst1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25493  float16x8_t __s1 = __p1; \
25494  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 40); \
25495})
25496#else
25497#define vst1q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25498  float16x8_t __s1 = __p1; \
25499  float16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25500  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 40); \
25501})
25502#endif
25503
25504#ifdef __LITTLE_ENDIAN__
25505#define vst1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25506  int32x4_t __s1 = __p1; \
25507  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 34); \
25508})
25509#else
25510#define vst1q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25511  int32x4_t __s1 = __p1; \
25512  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25513  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 34); \
25514})
25515#endif
25516
25517#ifdef __LITTLE_ENDIAN__
25518#define vst1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25519  int64x2_t __s1 = __p1; \
25520  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 35); \
25521})
25522#else
25523#define vst1q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25524  int64x2_t __s1 = __p1; \
25525  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25526  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 35); \
25527})
25528#endif
25529
25530#ifdef __LITTLE_ENDIAN__
25531#define vst1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25532  int16x8_t __s1 = __p1; \
25533  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 33); \
25534})
25535#else
25536#define vst1q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25537  int16x8_t __s1 = __p1; \
25538  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25539  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 33); \
25540})
25541#endif
25542
25543#ifdef __LITTLE_ENDIAN__
25544#define vst1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25545  uint8x8_t __s1 = __p1; \
25546  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 16); \
25547})
25548#else
25549#define vst1_lane_u8(__p0, __p1, __p2) __extension__ ({ \
25550  uint8x8_t __s1 = __p1; \
25551  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25552  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 16); \
25553})
25554#endif
25555
25556#ifdef __LITTLE_ENDIAN__
25557#define vst1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25558  uint32x2_t __s1 = __p1; \
25559  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 18); \
25560})
25561#else
25562#define vst1_lane_u32(__p0, __p1, __p2) __extension__ ({ \
25563  uint32x2_t __s1 = __p1; \
25564  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25565  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 18); \
25566})
25567#endif
25568
25569#ifdef __LITTLE_ENDIAN__
25570#define vst1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25571  uint64x1_t __s1 = __p1; \
25572  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
25573})
25574#else
25575#define vst1_lane_u64(__p0, __p1, __p2) __extension__ ({ \
25576  uint64x1_t __s1 = __p1; \
25577  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 19); \
25578})
25579#endif
25580
25581#ifdef __LITTLE_ENDIAN__
25582#define vst1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25583  uint16x4_t __s1 = __p1; \
25584  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 17); \
25585})
25586#else
25587#define vst1_lane_u16(__p0, __p1, __p2) __extension__ ({ \
25588  uint16x4_t __s1 = __p1; \
25589  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25590  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 17); \
25591})
25592#endif
25593
25594#ifdef __LITTLE_ENDIAN__
25595#define vst1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25596  int8x8_t __s1 = __p1; \
25597  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 0); \
25598})
25599#else
25600#define vst1_lane_s8(__p0, __p1, __p2) __extension__ ({ \
25601  int8x8_t __s1 = __p1; \
25602  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
25603  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 0); \
25604})
25605#endif
25606
25607#ifdef __LITTLE_ENDIAN__
25608#define vst1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25609  float32x2_t __s1 = __p1; \
25610  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 9); \
25611})
25612#else
25613#define vst1_lane_f32(__p0, __p1, __p2) __extension__ ({ \
25614  float32x2_t __s1 = __p1; \
25615  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25616  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 9); \
25617})
25618#endif
25619
25620#ifdef __LITTLE_ENDIAN__
25621#define vst1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25622  float16x4_t __s1 = __p1; \
25623  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 8); \
25624})
25625#else
25626#define vst1_lane_f16(__p0, __p1, __p2) __extension__ ({ \
25627  float16x4_t __s1 = __p1; \
25628  float16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25629  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 8); \
25630})
25631#endif
25632
25633#ifdef __LITTLE_ENDIAN__
25634#define vst1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25635  int32x2_t __s1 = __p1; \
25636  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 2); \
25637})
25638#else
25639#define vst1_lane_s32(__p0, __p1, __p2) __extension__ ({ \
25640  int32x2_t __s1 = __p1; \
25641  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
25642  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 2); \
25643})
25644#endif
25645
25646#ifdef __LITTLE_ENDIAN__
25647#define vst1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25648  int64x1_t __s1 = __p1; \
25649  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
25650})
25651#else
25652#define vst1_lane_s64(__p0, __p1, __p2) __extension__ ({ \
25653  int64x1_t __s1 = __p1; \
25654  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 3); \
25655})
25656#endif
25657
25658#ifdef __LITTLE_ENDIAN__
25659#define vst1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25660  int16x4_t __s1 = __p1; \
25661  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 1); \
25662})
25663#else
25664#define vst1_lane_s16(__p0, __p1, __p2) __extension__ ({ \
25665  int16x4_t __s1 = __p1; \
25666  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
25667  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__rev1, __p2, 1); \
25668})
25669#endif
25670
25671#ifdef __LITTLE_ENDIAN__
25672#define vst2_p8(__p0, __p1) __extension__ ({ \
25673  poly8x8x2_t __s1 = __p1; \
25674  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 4); \
25675})
25676#else
25677#define vst2_p8(__p0, __p1) __extension__ ({ \
25678  poly8x8x2_t __s1 = __p1; \
25679  poly8x8x2_t __rev1; \
25680  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25681  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25682  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 4); \
25683})
25684#endif
25685
25686#ifdef __LITTLE_ENDIAN__
25687#define vst2_p16(__p0, __p1) __extension__ ({ \
25688  poly16x4x2_t __s1 = __p1; \
25689  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 5); \
25690})
25691#else
25692#define vst2_p16(__p0, __p1) __extension__ ({ \
25693  poly16x4x2_t __s1 = __p1; \
25694  poly16x4x2_t __rev1; \
25695  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25696  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25697  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 5); \
25698})
25699#endif
25700
25701#ifdef __LITTLE_ENDIAN__
25702#define vst2q_p8(__p0, __p1) __extension__ ({ \
25703  poly8x16x2_t __s1 = __p1; \
25704  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 36); \
25705})
25706#else
25707#define vst2q_p8(__p0, __p1) __extension__ ({ \
25708  poly8x16x2_t __s1 = __p1; \
25709  poly8x16x2_t __rev1; \
25710  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25711  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25712  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 36); \
25713})
25714#endif
25715
25716#ifdef __LITTLE_ENDIAN__
25717#define vst2q_p16(__p0, __p1) __extension__ ({ \
25718  poly16x8x2_t __s1 = __p1; \
25719  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 37); \
25720})
25721#else
25722#define vst2q_p16(__p0, __p1) __extension__ ({ \
25723  poly16x8x2_t __s1 = __p1; \
25724  poly16x8x2_t __rev1; \
25725  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25726  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25727  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 37); \
25728})
25729#endif
25730
25731#ifdef __LITTLE_ENDIAN__
25732#define vst2q_u8(__p0, __p1) __extension__ ({ \
25733  uint8x16x2_t __s1 = __p1; \
25734  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 48); \
25735})
25736#else
25737#define vst2q_u8(__p0, __p1) __extension__ ({ \
25738  uint8x16x2_t __s1 = __p1; \
25739  uint8x16x2_t __rev1; \
25740  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25741  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25742  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 48); \
25743})
25744#endif
25745
25746#ifdef __LITTLE_ENDIAN__
25747#define vst2q_u32(__p0, __p1) __extension__ ({ \
25748  uint32x4x2_t __s1 = __p1; \
25749  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 50); \
25750})
25751#else
25752#define vst2q_u32(__p0, __p1) __extension__ ({ \
25753  uint32x4x2_t __s1 = __p1; \
25754  uint32x4x2_t __rev1; \
25755  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25756  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25757  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 50); \
25758})
25759#endif
25760
25761#ifdef __LITTLE_ENDIAN__
25762#define vst2q_u16(__p0, __p1) __extension__ ({ \
25763  uint16x8x2_t __s1 = __p1; \
25764  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 49); \
25765})
25766#else
25767#define vst2q_u16(__p0, __p1) __extension__ ({ \
25768  uint16x8x2_t __s1 = __p1; \
25769  uint16x8x2_t __rev1; \
25770  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25771  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25772  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 49); \
25773})
25774#endif
25775
25776#ifdef __LITTLE_ENDIAN__
25777#define vst2q_s8(__p0, __p1) __extension__ ({ \
25778  int8x16x2_t __s1 = __p1; \
25779  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 32); \
25780})
25781#else
25782#define vst2q_s8(__p0, __p1) __extension__ ({ \
25783  int8x16x2_t __s1 = __p1; \
25784  int8x16x2_t __rev1; \
25785  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25786  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
25787  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 32); \
25788})
25789#endif
25790
25791#ifdef __LITTLE_ENDIAN__
25792#define vst2q_f32(__p0, __p1) __extension__ ({ \
25793  float32x4x2_t __s1 = __p1; \
25794  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 41); \
25795})
25796#else
25797#define vst2q_f32(__p0, __p1) __extension__ ({ \
25798  float32x4x2_t __s1 = __p1; \
25799  float32x4x2_t __rev1; \
25800  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25801  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25802  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 41); \
25803})
25804#endif
25805
25806#ifdef __LITTLE_ENDIAN__
25807#define vst2q_f16(__p0, __p1) __extension__ ({ \
25808  float16x8x2_t __s1 = __p1; \
25809  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 40); \
25810})
25811#else
25812#define vst2q_f16(__p0, __p1) __extension__ ({ \
25813  float16x8x2_t __s1 = __p1; \
25814  float16x8x2_t __rev1; \
25815  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25816  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25817  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 40); \
25818})
25819#endif
25820
25821#ifdef __LITTLE_ENDIAN__
25822#define vst2q_s32(__p0, __p1) __extension__ ({ \
25823  int32x4x2_t __s1 = __p1; \
25824  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 34); \
25825})
25826#else
25827#define vst2q_s32(__p0, __p1) __extension__ ({ \
25828  int32x4x2_t __s1 = __p1; \
25829  int32x4x2_t __rev1; \
25830  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25831  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25832  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 34); \
25833})
25834#endif
25835
25836#ifdef __LITTLE_ENDIAN__
25837#define vst2q_s16(__p0, __p1) __extension__ ({ \
25838  int16x8x2_t __s1 = __p1; \
25839  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 33); \
25840})
25841#else
25842#define vst2q_s16(__p0, __p1) __extension__ ({ \
25843  int16x8x2_t __s1 = __p1; \
25844  int16x8x2_t __rev1; \
25845  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25846  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25847  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 33); \
25848})
25849#endif
25850
25851#ifdef __LITTLE_ENDIAN__
25852#define vst2_u8(__p0, __p1) __extension__ ({ \
25853  uint8x8x2_t __s1 = __p1; \
25854  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 16); \
25855})
25856#else
25857#define vst2_u8(__p0, __p1) __extension__ ({ \
25858  uint8x8x2_t __s1 = __p1; \
25859  uint8x8x2_t __rev1; \
25860  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25861  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25862  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 16); \
25863})
25864#endif
25865
25866#ifdef __LITTLE_ENDIAN__
25867#define vst2_u32(__p0, __p1) __extension__ ({ \
25868  uint32x2x2_t __s1 = __p1; \
25869  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 18); \
25870})
25871#else
25872#define vst2_u32(__p0, __p1) __extension__ ({ \
25873  uint32x2x2_t __s1 = __p1; \
25874  uint32x2x2_t __rev1; \
25875  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
25876  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
25877  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 18); \
25878})
25879#endif
25880
25881#ifdef __LITTLE_ENDIAN__
25882#define vst2_u64(__p0, __p1) __extension__ ({ \
25883  uint64x1x2_t __s1 = __p1; \
25884  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
25885})
25886#else
25887#define vst2_u64(__p0, __p1) __extension__ ({ \
25888  uint64x1x2_t __s1 = __p1; \
25889  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
25890})
25891#endif
25892
25893#ifdef __LITTLE_ENDIAN__
25894#define vst2_u16(__p0, __p1) __extension__ ({ \
25895  uint16x4x2_t __s1 = __p1; \
25896  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 17); \
25897})
25898#else
25899#define vst2_u16(__p0, __p1) __extension__ ({ \
25900  uint16x4x2_t __s1 = __p1; \
25901  uint16x4x2_t __rev1; \
25902  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25903  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25904  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 17); \
25905})
25906#endif
25907
25908#ifdef __LITTLE_ENDIAN__
25909#define vst2_s8(__p0, __p1) __extension__ ({ \
25910  int8x8x2_t __s1 = __p1; \
25911  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 0); \
25912})
25913#else
25914#define vst2_s8(__p0, __p1) __extension__ ({ \
25915  int8x8x2_t __s1 = __p1; \
25916  int8x8x2_t __rev1; \
25917  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
25918  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
25919  __builtin_neon_vst2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 0); \
25920})
25921#endif
25922
25923#ifdef __LITTLE_ENDIAN__
25924#define vst2_f32(__p0, __p1) __extension__ ({ \
25925  float32x2x2_t __s1 = __p1; \
25926  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 9); \
25927})
25928#else
25929#define vst2_f32(__p0, __p1) __extension__ ({ \
25930  float32x2x2_t __s1 = __p1; \
25931  float32x2x2_t __rev1; \
25932  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
25933  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
25934  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 9); \
25935})
25936#endif
25937
25938#ifdef __LITTLE_ENDIAN__
25939#define vst2_f16(__p0, __p1) __extension__ ({ \
25940  float16x4x2_t __s1 = __p1; \
25941  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 8); \
25942})
25943#else
25944#define vst2_f16(__p0, __p1) __extension__ ({ \
25945  float16x4x2_t __s1 = __p1; \
25946  float16x4x2_t __rev1; \
25947  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25948  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25949  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 8); \
25950})
25951#endif
25952
25953#ifdef __LITTLE_ENDIAN__
25954#define vst2_s32(__p0, __p1) __extension__ ({ \
25955  int32x2x2_t __s1 = __p1; \
25956  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 2); \
25957})
25958#else
25959#define vst2_s32(__p0, __p1) __extension__ ({ \
25960  int32x2x2_t __s1 = __p1; \
25961  int32x2x2_t __rev1; \
25962  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
25963  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
25964  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 2); \
25965})
25966#endif
25967
25968#ifdef __LITTLE_ENDIAN__
25969#define vst2_s64(__p0, __p1) __extension__ ({ \
25970  int64x1x2_t __s1 = __p1; \
25971  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 3); \
25972})
25973#else
25974#define vst2_s64(__p0, __p1) __extension__ ({ \
25975  int64x1x2_t __s1 = __p1; \
25976  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 3); \
25977})
25978#endif
25979
25980#ifdef __LITTLE_ENDIAN__
25981#define vst2_s16(__p0, __p1) __extension__ ({ \
25982  int16x4x2_t __s1 = __p1; \
25983  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 1); \
25984})
25985#else
25986#define vst2_s16(__p0, __p1) __extension__ ({ \
25987  int16x4x2_t __s1 = __p1; \
25988  int16x4x2_t __rev1; \
25989  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
25990  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
25991  __builtin_neon_vst2_v(__p0, __rev1.val[0], __rev1.val[1], 1); \
25992})
25993#endif
25994
25995#ifdef __LITTLE_ENDIAN__
25996#define vst2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
25997  poly8x8x2_t __s1 = __p1; \
25998  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 4); \
25999})
26000#else
26001#define vst2_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26002  poly8x8x2_t __s1 = __p1; \
26003  poly8x8x2_t __rev1; \
26004  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26005  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26006  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 4); \
26007})
26008#endif
26009
26010#ifdef __LITTLE_ENDIAN__
26011#define vst2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26012  poly16x4x2_t __s1 = __p1; \
26013  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 5); \
26014})
26015#else
26016#define vst2_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26017  poly16x4x2_t __s1 = __p1; \
26018  poly16x4x2_t __rev1; \
26019  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26020  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26021  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 5); \
26022})
26023#endif
26024
26025#ifdef __LITTLE_ENDIAN__
26026#define vst2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26027  poly16x8x2_t __s1 = __p1; \
26028  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 37); \
26029})
26030#else
26031#define vst2q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26032  poly16x8x2_t __s1 = __p1; \
26033  poly16x8x2_t __rev1; \
26034  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26035  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26036  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 37); \
26037})
26038#endif
26039
26040#ifdef __LITTLE_ENDIAN__
26041#define vst2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26042  uint32x4x2_t __s1 = __p1; \
26043  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 50); \
26044})
26045#else
26046#define vst2q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26047  uint32x4x2_t __s1 = __p1; \
26048  uint32x4x2_t __rev1; \
26049  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26050  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26051  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 50); \
26052})
26053#endif
26054
26055#ifdef __LITTLE_ENDIAN__
26056#define vst2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26057  uint16x8x2_t __s1 = __p1; \
26058  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 49); \
26059})
26060#else
26061#define vst2q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26062  uint16x8x2_t __s1 = __p1; \
26063  uint16x8x2_t __rev1; \
26064  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26065  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26066  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 49); \
26067})
26068#endif
26069
26070#ifdef __LITTLE_ENDIAN__
26071#define vst2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26072  float32x4x2_t __s1 = __p1; \
26073  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 41); \
26074})
26075#else
26076#define vst2q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26077  float32x4x2_t __s1 = __p1; \
26078  float32x4x2_t __rev1; \
26079  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26080  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26081  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 41); \
26082})
26083#endif
26084
26085#ifdef __LITTLE_ENDIAN__
26086#define vst2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26087  float16x8x2_t __s1 = __p1; \
26088  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 40); \
26089})
26090#else
26091#define vst2q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26092  float16x8x2_t __s1 = __p1; \
26093  float16x8x2_t __rev1; \
26094  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26095  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26096  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 40); \
26097})
26098#endif
26099
26100#ifdef __LITTLE_ENDIAN__
26101#define vst2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26102  int32x4x2_t __s1 = __p1; \
26103  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 34); \
26104})
26105#else
26106#define vst2q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26107  int32x4x2_t __s1 = __p1; \
26108  int32x4x2_t __rev1; \
26109  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26110  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26111  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 34); \
26112})
26113#endif
26114
26115#ifdef __LITTLE_ENDIAN__
26116#define vst2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26117  int16x8x2_t __s1 = __p1; \
26118  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 33); \
26119})
26120#else
26121#define vst2q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26122  int16x8x2_t __s1 = __p1; \
26123  int16x8x2_t __rev1; \
26124  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26125  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26126  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 33); \
26127})
26128#endif
26129
26130#ifdef __LITTLE_ENDIAN__
26131#define vst2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26132  uint8x8x2_t __s1 = __p1; \
26133  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 16); \
26134})
26135#else
26136#define vst2_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26137  uint8x8x2_t __s1 = __p1; \
26138  uint8x8x2_t __rev1; \
26139  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26140  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26141  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 16); \
26142})
26143#endif
26144
26145#ifdef __LITTLE_ENDIAN__
26146#define vst2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26147  uint32x2x2_t __s1 = __p1; \
26148  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 18); \
26149})
26150#else
26151#define vst2_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26152  uint32x2x2_t __s1 = __p1; \
26153  uint32x2x2_t __rev1; \
26154  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26155  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26156  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 18); \
26157})
26158#endif
26159
26160#ifdef __LITTLE_ENDIAN__
26161#define vst2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26162  uint16x4x2_t __s1 = __p1; \
26163  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 17); \
26164})
26165#else
26166#define vst2_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26167  uint16x4x2_t __s1 = __p1; \
26168  uint16x4x2_t __rev1; \
26169  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26170  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26171  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 17); \
26172})
26173#endif
26174
26175#ifdef __LITTLE_ENDIAN__
26176#define vst2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26177  int8x8x2_t __s1 = __p1; \
26178  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 0); \
26179})
26180#else
26181#define vst2_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26182  int8x8x2_t __s1 = __p1; \
26183  int8x8x2_t __rev1; \
26184  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26185  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26186  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], __p2, 0); \
26187})
26188#endif
26189
26190#ifdef __LITTLE_ENDIAN__
26191#define vst2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26192  float32x2x2_t __s1 = __p1; \
26193  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 9); \
26194})
26195#else
26196#define vst2_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26197  float32x2x2_t __s1 = __p1; \
26198  float32x2x2_t __rev1; \
26199  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26200  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26201  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 9); \
26202})
26203#endif
26204
26205#ifdef __LITTLE_ENDIAN__
26206#define vst2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26207  float16x4x2_t __s1 = __p1; \
26208  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 8); \
26209})
26210#else
26211#define vst2_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26212  float16x4x2_t __s1 = __p1; \
26213  float16x4x2_t __rev1; \
26214  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26215  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26216  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 8); \
26217})
26218#endif
26219
26220#ifdef __LITTLE_ENDIAN__
26221#define vst2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26222  int32x2x2_t __s1 = __p1; \
26223  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 2); \
26224})
26225#else
26226#define vst2_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26227  int32x2x2_t __s1 = __p1; \
26228  int32x2x2_t __rev1; \
26229  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26230  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26231  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 2); \
26232})
26233#endif
26234
26235#ifdef __LITTLE_ENDIAN__
26236#define vst2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26237  int16x4x2_t __s1 = __p1; \
26238  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 1); \
26239})
26240#else
26241#define vst2_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26242  int16x4x2_t __s1 = __p1; \
26243  int16x4x2_t __rev1; \
26244  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26245  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26246  __builtin_neon_vst2_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 1); \
26247})
26248#endif
26249
26250#ifdef __LITTLE_ENDIAN__
26251#define vst3_p8(__p0, __p1) __extension__ ({ \
26252  poly8x8x3_t __s1 = __p1; \
26253  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 4); \
26254})
26255#else
26256#define vst3_p8(__p0, __p1) __extension__ ({ \
26257  poly8x8x3_t __s1 = __p1; \
26258  poly8x8x3_t __rev1; \
26259  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26260  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26261  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26262  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 4); \
26263})
26264#endif
26265
26266#ifdef __LITTLE_ENDIAN__
26267#define vst3_p16(__p0, __p1) __extension__ ({ \
26268  poly16x4x3_t __s1 = __p1; \
26269  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 5); \
26270})
26271#else
26272#define vst3_p16(__p0, __p1) __extension__ ({ \
26273  poly16x4x3_t __s1 = __p1; \
26274  poly16x4x3_t __rev1; \
26275  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26276  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26277  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26278  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 5); \
26279})
26280#endif
26281
26282#ifdef __LITTLE_ENDIAN__
26283#define vst3q_p8(__p0, __p1) __extension__ ({ \
26284  poly8x16x3_t __s1 = __p1; \
26285  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 36); \
26286})
26287#else
26288#define vst3q_p8(__p0, __p1) __extension__ ({ \
26289  poly8x16x3_t __s1 = __p1; \
26290  poly8x16x3_t __rev1; \
26291  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26292  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26293  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26294  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 36); \
26295})
26296#endif
26297
26298#ifdef __LITTLE_ENDIAN__
26299#define vst3q_p16(__p0, __p1) __extension__ ({ \
26300  poly16x8x3_t __s1 = __p1; \
26301  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 37); \
26302})
26303#else
26304#define vst3q_p16(__p0, __p1) __extension__ ({ \
26305  poly16x8x3_t __s1 = __p1; \
26306  poly16x8x3_t __rev1; \
26307  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26308  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26309  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26310  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 37); \
26311})
26312#endif
26313
26314#ifdef __LITTLE_ENDIAN__
26315#define vst3q_u8(__p0, __p1) __extension__ ({ \
26316  uint8x16x3_t __s1 = __p1; \
26317  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 48); \
26318})
26319#else
26320#define vst3q_u8(__p0, __p1) __extension__ ({ \
26321  uint8x16x3_t __s1 = __p1; \
26322  uint8x16x3_t __rev1; \
26323  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26324  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26325  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26326  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 48); \
26327})
26328#endif
26329
26330#ifdef __LITTLE_ENDIAN__
26331#define vst3q_u32(__p0, __p1) __extension__ ({ \
26332  uint32x4x3_t __s1 = __p1; \
26333  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 50); \
26334})
26335#else
26336#define vst3q_u32(__p0, __p1) __extension__ ({ \
26337  uint32x4x3_t __s1 = __p1; \
26338  uint32x4x3_t __rev1; \
26339  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26340  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26341  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26342  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 50); \
26343})
26344#endif
26345
26346#ifdef __LITTLE_ENDIAN__
26347#define vst3q_u16(__p0, __p1) __extension__ ({ \
26348  uint16x8x3_t __s1 = __p1; \
26349  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 49); \
26350})
26351#else
26352#define vst3q_u16(__p0, __p1) __extension__ ({ \
26353  uint16x8x3_t __s1 = __p1; \
26354  uint16x8x3_t __rev1; \
26355  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26356  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26357  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26358  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 49); \
26359})
26360#endif
26361
26362#ifdef __LITTLE_ENDIAN__
26363#define vst3q_s8(__p0, __p1) __extension__ ({ \
26364  int8x16x3_t __s1 = __p1; \
26365  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 32); \
26366})
26367#else
26368#define vst3q_s8(__p0, __p1) __extension__ ({ \
26369  int8x16x3_t __s1 = __p1; \
26370  int8x16x3_t __rev1; \
26371  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26372  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26373  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26374  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 32); \
26375})
26376#endif
26377
26378#ifdef __LITTLE_ENDIAN__
26379#define vst3q_f32(__p0, __p1) __extension__ ({ \
26380  float32x4x3_t __s1 = __p1; \
26381  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 41); \
26382})
26383#else
26384#define vst3q_f32(__p0, __p1) __extension__ ({ \
26385  float32x4x3_t __s1 = __p1; \
26386  float32x4x3_t __rev1; \
26387  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26388  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26389  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26390  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 41); \
26391})
26392#endif
26393
26394#ifdef __LITTLE_ENDIAN__
26395#define vst3q_f16(__p0, __p1) __extension__ ({ \
26396  float16x8x3_t __s1 = __p1; \
26397  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 40); \
26398})
26399#else
26400#define vst3q_f16(__p0, __p1) __extension__ ({ \
26401  float16x8x3_t __s1 = __p1; \
26402  float16x8x3_t __rev1; \
26403  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26404  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26405  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26406  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 40); \
26407})
26408#endif
26409
26410#ifdef __LITTLE_ENDIAN__
26411#define vst3q_s32(__p0, __p1) __extension__ ({ \
26412  int32x4x3_t __s1 = __p1; \
26413  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 34); \
26414})
26415#else
26416#define vst3q_s32(__p0, __p1) __extension__ ({ \
26417  int32x4x3_t __s1 = __p1; \
26418  int32x4x3_t __rev1; \
26419  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26420  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26421  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26422  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 34); \
26423})
26424#endif
26425
26426#ifdef __LITTLE_ENDIAN__
26427#define vst3q_s16(__p0, __p1) __extension__ ({ \
26428  int16x8x3_t __s1 = __p1; \
26429  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 33); \
26430})
26431#else
26432#define vst3q_s16(__p0, __p1) __extension__ ({ \
26433  int16x8x3_t __s1 = __p1; \
26434  int16x8x3_t __rev1; \
26435  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26436  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26437  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26438  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 33); \
26439})
26440#endif
26441
26442#ifdef __LITTLE_ENDIAN__
26443#define vst3_u8(__p0, __p1) __extension__ ({ \
26444  uint8x8x3_t __s1 = __p1; \
26445  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 16); \
26446})
26447#else
26448#define vst3_u8(__p0, __p1) __extension__ ({ \
26449  uint8x8x3_t __s1 = __p1; \
26450  uint8x8x3_t __rev1; \
26451  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26452  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26453  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26454  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 16); \
26455})
26456#endif
26457
26458#ifdef __LITTLE_ENDIAN__
26459#define vst3_u32(__p0, __p1) __extension__ ({ \
26460  uint32x2x3_t __s1 = __p1; \
26461  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 18); \
26462})
26463#else
26464#define vst3_u32(__p0, __p1) __extension__ ({ \
26465  uint32x2x3_t __s1 = __p1; \
26466  uint32x2x3_t __rev1; \
26467  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26468  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26469  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26470  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 18); \
26471})
26472#endif
26473
26474#ifdef __LITTLE_ENDIAN__
26475#define vst3_u64(__p0, __p1) __extension__ ({ \
26476  uint64x1x3_t __s1 = __p1; \
26477  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
26478})
26479#else
26480#define vst3_u64(__p0, __p1) __extension__ ({ \
26481  uint64x1x3_t __s1 = __p1; \
26482  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
26483})
26484#endif
26485
26486#ifdef __LITTLE_ENDIAN__
26487#define vst3_u16(__p0, __p1) __extension__ ({ \
26488  uint16x4x3_t __s1 = __p1; \
26489  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 17); \
26490})
26491#else
26492#define vst3_u16(__p0, __p1) __extension__ ({ \
26493  uint16x4x3_t __s1 = __p1; \
26494  uint16x4x3_t __rev1; \
26495  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26496  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26497  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26498  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 17); \
26499})
26500#endif
26501
26502#ifdef __LITTLE_ENDIAN__
26503#define vst3_s8(__p0, __p1) __extension__ ({ \
26504  int8x8x3_t __s1 = __p1; \
26505  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 0); \
26506})
26507#else
26508#define vst3_s8(__p0, __p1) __extension__ ({ \
26509  int8x8x3_t __s1 = __p1; \
26510  int8x8x3_t __rev1; \
26511  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26512  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26513  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26514  __builtin_neon_vst3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 0); \
26515})
26516#endif
26517
26518#ifdef __LITTLE_ENDIAN__
26519#define vst3_f32(__p0, __p1) __extension__ ({ \
26520  float32x2x3_t __s1 = __p1; \
26521  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 9); \
26522})
26523#else
26524#define vst3_f32(__p0, __p1) __extension__ ({ \
26525  float32x2x3_t __s1 = __p1; \
26526  float32x2x3_t __rev1; \
26527  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26528  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26529  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26530  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 9); \
26531})
26532#endif
26533
26534#ifdef __LITTLE_ENDIAN__
26535#define vst3_f16(__p0, __p1) __extension__ ({ \
26536  float16x4x3_t __s1 = __p1; \
26537  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 8); \
26538})
26539#else
26540#define vst3_f16(__p0, __p1) __extension__ ({ \
26541  float16x4x3_t __s1 = __p1; \
26542  float16x4x3_t __rev1; \
26543  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26544  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26545  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26546  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 8); \
26547})
26548#endif
26549
26550#ifdef __LITTLE_ENDIAN__
26551#define vst3_s32(__p0, __p1) __extension__ ({ \
26552  int32x2x3_t __s1 = __p1; \
26553  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 2); \
26554})
26555#else
26556#define vst3_s32(__p0, __p1) __extension__ ({ \
26557  int32x2x3_t __s1 = __p1; \
26558  int32x2x3_t __rev1; \
26559  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26560  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26561  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26562  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 2); \
26563})
26564#endif
26565
26566#ifdef __LITTLE_ENDIAN__
26567#define vst3_s64(__p0, __p1) __extension__ ({ \
26568  int64x1x3_t __s1 = __p1; \
26569  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
26570})
26571#else
26572#define vst3_s64(__p0, __p1) __extension__ ({ \
26573  int64x1x3_t __s1 = __p1; \
26574  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
26575})
26576#endif
26577
26578#ifdef __LITTLE_ENDIAN__
26579#define vst3_s16(__p0, __p1) __extension__ ({ \
26580  int16x4x3_t __s1 = __p1; \
26581  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 1); \
26582})
26583#else
26584#define vst3_s16(__p0, __p1) __extension__ ({ \
26585  int16x4x3_t __s1 = __p1; \
26586  int16x4x3_t __rev1; \
26587  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26588  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26589  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26590  __builtin_neon_vst3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 1); \
26591})
26592#endif
26593
26594#ifdef __LITTLE_ENDIAN__
26595#define vst3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26596  poly8x8x3_t __s1 = __p1; \
26597  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 4); \
26598})
26599#else
26600#define vst3_lane_p8(__p0, __p1, __p2) __extension__ ({ \
26601  poly8x8x3_t __s1 = __p1; \
26602  poly8x8x3_t __rev1; \
26603  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26604  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26605  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26606  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 4); \
26607})
26608#endif
26609
26610#ifdef __LITTLE_ENDIAN__
26611#define vst3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26612  poly16x4x3_t __s1 = __p1; \
26613  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 5); \
26614})
26615#else
26616#define vst3_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26617  poly16x4x3_t __s1 = __p1; \
26618  poly16x4x3_t __rev1; \
26619  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26620  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26621  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26622  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 5); \
26623})
26624#endif
26625
26626#ifdef __LITTLE_ENDIAN__
26627#define vst3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26628  poly16x8x3_t __s1 = __p1; \
26629  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 37); \
26630})
26631#else
26632#define vst3q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
26633  poly16x8x3_t __s1 = __p1; \
26634  poly16x8x3_t __rev1; \
26635  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26636  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26637  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26638  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 37); \
26639})
26640#endif
26641
26642#ifdef __LITTLE_ENDIAN__
26643#define vst3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26644  uint32x4x3_t __s1 = __p1; \
26645  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 50); \
26646})
26647#else
26648#define vst3q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26649  uint32x4x3_t __s1 = __p1; \
26650  uint32x4x3_t __rev1; \
26651  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26652  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26653  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26654  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 50); \
26655})
26656#endif
26657
26658#ifdef __LITTLE_ENDIAN__
26659#define vst3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26660  uint16x8x3_t __s1 = __p1; \
26661  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 49); \
26662})
26663#else
26664#define vst3q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26665  uint16x8x3_t __s1 = __p1; \
26666  uint16x8x3_t __rev1; \
26667  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26668  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26669  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26670  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 49); \
26671})
26672#endif
26673
26674#ifdef __LITTLE_ENDIAN__
26675#define vst3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26676  float32x4x3_t __s1 = __p1; \
26677  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 41); \
26678})
26679#else
26680#define vst3q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26681  float32x4x3_t __s1 = __p1; \
26682  float32x4x3_t __rev1; \
26683  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26684  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26685  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26686  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 41); \
26687})
26688#endif
26689
26690#ifdef __LITTLE_ENDIAN__
26691#define vst3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26692  float16x8x3_t __s1 = __p1; \
26693  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 40); \
26694})
26695#else
26696#define vst3q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26697  float16x8x3_t __s1 = __p1; \
26698  float16x8x3_t __rev1; \
26699  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26700  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26701  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26702  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 40); \
26703})
26704#endif
26705
26706#ifdef __LITTLE_ENDIAN__
26707#define vst3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26708  int32x4x3_t __s1 = __p1; \
26709  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 34); \
26710})
26711#else
26712#define vst3q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26713  int32x4x3_t __s1 = __p1; \
26714  int32x4x3_t __rev1; \
26715  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26716  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26717  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26718  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 34); \
26719})
26720#endif
26721
26722#ifdef __LITTLE_ENDIAN__
26723#define vst3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26724  int16x8x3_t __s1 = __p1; \
26725  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 33); \
26726})
26727#else
26728#define vst3q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26729  int16x8x3_t __s1 = __p1; \
26730  int16x8x3_t __rev1; \
26731  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26732  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26733  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26734  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 33); \
26735})
26736#endif
26737
26738#ifdef __LITTLE_ENDIAN__
26739#define vst3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26740  uint8x8x3_t __s1 = __p1; \
26741  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 16); \
26742})
26743#else
26744#define vst3_lane_u8(__p0, __p1, __p2) __extension__ ({ \
26745  uint8x8x3_t __s1 = __p1; \
26746  uint8x8x3_t __rev1; \
26747  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26748  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26749  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26750  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 16); \
26751})
26752#endif
26753
26754#ifdef __LITTLE_ENDIAN__
26755#define vst3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26756  uint32x2x3_t __s1 = __p1; \
26757  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 18); \
26758})
26759#else
26760#define vst3_lane_u32(__p0, __p1, __p2) __extension__ ({ \
26761  uint32x2x3_t __s1 = __p1; \
26762  uint32x2x3_t __rev1; \
26763  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26764  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26765  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26766  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 18); \
26767})
26768#endif
26769
26770#ifdef __LITTLE_ENDIAN__
26771#define vst3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26772  uint16x4x3_t __s1 = __p1; \
26773  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 17); \
26774})
26775#else
26776#define vst3_lane_u16(__p0, __p1, __p2) __extension__ ({ \
26777  uint16x4x3_t __s1 = __p1; \
26778  uint16x4x3_t __rev1; \
26779  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26780  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26781  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26782  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 17); \
26783})
26784#endif
26785
26786#ifdef __LITTLE_ENDIAN__
26787#define vst3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26788  int8x8x3_t __s1 = __p1; \
26789  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 0); \
26790})
26791#else
26792#define vst3_lane_s8(__p0, __p1, __p2) __extension__ ({ \
26793  int8x8x3_t __s1 = __p1; \
26794  int8x8x3_t __rev1; \
26795  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26796  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26797  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26798  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], __p2, 0); \
26799})
26800#endif
26801
26802#ifdef __LITTLE_ENDIAN__
26803#define vst3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26804  float32x2x3_t __s1 = __p1; \
26805  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 9); \
26806})
26807#else
26808#define vst3_lane_f32(__p0, __p1, __p2) __extension__ ({ \
26809  float32x2x3_t __s1 = __p1; \
26810  float32x2x3_t __rev1; \
26811  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26812  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26813  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26814  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 9); \
26815})
26816#endif
26817
26818#ifdef __LITTLE_ENDIAN__
26819#define vst3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26820  float16x4x3_t __s1 = __p1; \
26821  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 8); \
26822})
26823#else
26824#define vst3_lane_f16(__p0, __p1, __p2) __extension__ ({ \
26825  float16x4x3_t __s1 = __p1; \
26826  float16x4x3_t __rev1; \
26827  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26828  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26829  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26830  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 8); \
26831})
26832#endif
26833
26834#ifdef __LITTLE_ENDIAN__
26835#define vst3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26836  int32x2x3_t __s1 = __p1; \
26837  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 2); \
26838})
26839#else
26840#define vst3_lane_s32(__p0, __p1, __p2) __extension__ ({ \
26841  int32x2x3_t __s1 = __p1; \
26842  int32x2x3_t __rev1; \
26843  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
26844  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
26845  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
26846  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 2); \
26847})
26848#endif
26849
26850#ifdef __LITTLE_ENDIAN__
26851#define vst3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26852  int16x4x3_t __s1 = __p1; \
26853  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 1); \
26854})
26855#else
26856#define vst3_lane_s16(__p0, __p1, __p2) __extension__ ({ \
26857  int16x4x3_t __s1 = __p1; \
26858  int16x4x3_t __rev1; \
26859  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26860  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26861  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26862  __builtin_neon_vst3_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 1); \
26863})
26864#endif
26865
26866#ifdef __LITTLE_ENDIAN__
26867#define vst4_p8(__p0, __p1) __extension__ ({ \
26868  poly8x8x4_t __s1 = __p1; \
26869  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 4); \
26870})
26871#else
26872#define vst4_p8(__p0, __p1) __extension__ ({ \
26873  poly8x8x4_t __s1 = __p1; \
26874  poly8x8x4_t __rev1; \
26875  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26876  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26877  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26878  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26879  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 4); \
26880})
26881#endif
26882
26883#ifdef __LITTLE_ENDIAN__
26884#define vst4_p16(__p0, __p1) __extension__ ({ \
26885  poly16x4x4_t __s1 = __p1; \
26886  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 5); \
26887})
26888#else
26889#define vst4_p16(__p0, __p1) __extension__ ({ \
26890  poly16x4x4_t __s1 = __p1; \
26891  poly16x4x4_t __rev1; \
26892  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26893  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26894  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26895  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
26896  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 5); \
26897})
26898#endif
26899
26900#ifdef __LITTLE_ENDIAN__
26901#define vst4q_p8(__p0, __p1) __extension__ ({ \
26902  poly8x16x4_t __s1 = __p1; \
26903  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 36); \
26904})
26905#else
26906#define vst4q_p8(__p0, __p1) __extension__ ({ \
26907  poly8x16x4_t __s1 = __p1; \
26908  poly8x16x4_t __rev1; \
26909  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26910  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26911  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26912  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26913  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 36); \
26914})
26915#endif
26916
26917#ifdef __LITTLE_ENDIAN__
26918#define vst4q_p16(__p0, __p1) __extension__ ({ \
26919  poly16x8x4_t __s1 = __p1; \
26920  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 37); \
26921})
26922#else
26923#define vst4q_p16(__p0, __p1) __extension__ ({ \
26924  poly16x8x4_t __s1 = __p1; \
26925  poly16x8x4_t __rev1; \
26926  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26927  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26928  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26929  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26930  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 37); \
26931})
26932#endif
26933
26934#ifdef __LITTLE_ENDIAN__
26935#define vst4q_u8(__p0, __p1) __extension__ ({ \
26936  uint8x16x4_t __s1 = __p1; \
26937  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 48); \
26938})
26939#else
26940#define vst4q_u8(__p0, __p1) __extension__ ({ \
26941  uint8x16x4_t __s1 = __p1; \
26942  uint8x16x4_t __rev1; \
26943  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26944  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26945  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26946  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26947  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 48); \
26948})
26949#endif
26950
26951#ifdef __LITTLE_ENDIAN__
26952#define vst4q_u32(__p0, __p1) __extension__ ({ \
26953  uint32x4x4_t __s1 = __p1; \
26954  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 50); \
26955})
26956#else
26957#define vst4q_u32(__p0, __p1) __extension__ ({ \
26958  uint32x4x4_t __s1 = __p1; \
26959  uint32x4x4_t __rev1; \
26960  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
26961  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
26962  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
26963  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
26964  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 50); \
26965})
26966#endif
26967
26968#ifdef __LITTLE_ENDIAN__
26969#define vst4q_u16(__p0, __p1) __extension__ ({ \
26970  uint16x8x4_t __s1 = __p1; \
26971  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 49); \
26972})
26973#else
26974#define vst4q_u16(__p0, __p1) __extension__ ({ \
26975  uint16x8x4_t __s1 = __p1; \
26976  uint16x8x4_t __rev1; \
26977  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
26978  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
26979  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
26980  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
26981  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 49); \
26982})
26983#endif
26984
26985#ifdef __LITTLE_ENDIAN__
26986#define vst4q_s8(__p0, __p1) __extension__ ({ \
26987  int8x16x4_t __s1 = __p1; \
26988  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 32); \
26989})
26990#else
26991#define vst4q_s8(__p0, __p1) __extension__ ({ \
26992  int8x16x4_t __s1 = __p1; \
26993  int8x16x4_t __rev1; \
26994  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26995  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26996  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26997  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
26998  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 32); \
26999})
27000#endif
27001
27002#ifdef __LITTLE_ENDIAN__
27003#define vst4q_f32(__p0, __p1) __extension__ ({ \
27004  float32x4x4_t __s1 = __p1; \
27005  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 41); \
27006})
27007#else
27008#define vst4q_f32(__p0, __p1) __extension__ ({ \
27009  float32x4x4_t __s1 = __p1; \
27010  float32x4x4_t __rev1; \
27011  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27012  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27013  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27014  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27015  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 41); \
27016})
27017#endif
27018
27019#ifdef __LITTLE_ENDIAN__
27020#define vst4q_f16(__p0, __p1) __extension__ ({ \
27021  float16x8x4_t __s1 = __p1; \
27022  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 40); \
27023})
27024#else
27025#define vst4q_f16(__p0, __p1) __extension__ ({ \
27026  float16x8x4_t __s1 = __p1; \
27027  float16x8x4_t __rev1; \
27028  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27029  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27030  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27031  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27032  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 40); \
27033})
27034#endif
27035
27036#ifdef __LITTLE_ENDIAN__
27037#define vst4q_s32(__p0, __p1) __extension__ ({ \
27038  int32x4x4_t __s1 = __p1; \
27039  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 34); \
27040})
27041#else
27042#define vst4q_s32(__p0, __p1) __extension__ ({ \
27043  int32x4x4_t __s1 = __p1; \
27044  int32x4x4_t __rev1; \
27045  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27046  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27047  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27048  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27049  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 34); \
27050})
27051#endif
27052
27053#ifdef __LITTLE_ENDIAN__
27054#define vst4q_s16(__p0, __p1) __extension__ ({ \
27055  int16x8x4_t __s1 = __p1; \
27056  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 33); \
27057})
27058#else
27059#define vst4q_s16(__p0, __p1) __extension__ ({ \
27060  int16x8x4_t __s1 = __p1; \
27061  int16x8x4_t __rev1; \
27062  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27063  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27064  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27065  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27066  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 33); \
27067})
27068#endif
27069
27070#ifdef __LITTLE_ENDIAN__
27071#define vst4_u8(__p0, __p1) __extension__ ({ \
27072  uint8x8x4_t __s1 = __p1; \
27073  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 16); \
27074})
27075#else
27076#define vst4_u8(__p0, __p1) __extension__ ({ \
27077  uint8x8x4_t __s1 = __p1; \
27078  uint8x8x4_t __rev1; \
27079  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27080  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27081  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27082  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27083  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 16); \
27084})
27085#endif
27086
27087#ifdef __LITTLE_ENDIAN__
27088#define vst4_u32(__p0, __p1) __extension__ ({ \
27089  uint32x2x4_t __s1 = __p1; \
27090  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 18); \
27091})
27092#else
27093#define vst4_u32(__p0, __p1) __extension__ ({ \
27094  uint32x2x4_t __s1 = __p1; \
27095  uint32x2x4_t __rev1; \
27096  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27097  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27098  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27099  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27100  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 18); \
27101})
27102#endif
27103
27104#ifdef __LITTLE_ENDIAN__
27105#define vst4_u64(__p0, __p1) __extension__ ({ \
27106  uint64x1x4_t __s1 = __p1; \
27107  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
27108})
27109#else
27110#define vst4_u64(__p0, __p1) __extension__ ({ \
27111  uint64x1x4_t __s1 = __p1; \
27112  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
27113})
27114#endif
27115
27116#ifdef __LITTLE_ENDIAN__
27117#define vst4_u16(__p0, __p1) __extension__ ({ \
27118  uint16x4x4_t __s1 = __p1; \
27119  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 17); \
27120})
27121#else
27122#define vst4_u16(__p0, __p1) __extension__ ({ \
27123  uint16x4x4_t __s1 = __p1; \
27124  uint16x4x4_t __rev1; \
27125  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27126  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27127  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27128  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27129  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 17); \
27130})
27131#endif
27132
27133#ifdef __LITTLE_ENDIAN__
27134#define vst4_s8(__p0, __p1) __extension__ ({ \
27135  int8x8x4_t __s1 = __p1; \
27136  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 0); \
27137})
27138#else
27139#define vst4_s8(__p0, __p1) __extension__ ({ \
27140  int8x8x4_t __s1 = __p1; \
27141  int8x8x4_t __rev1; \
27142  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27143  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27144  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27145  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27146  __builtin_neon_vst4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 0); \
27147})
27148#endif
27149
27150#ifdef __LITTLE_ENDIAN__
27151#define vst4_f32(__p0, __p1) __extension__ ({ \
27152  float32x2x4_t __s1 = __p1; \
27153  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 9); \
27154})
27155#else
27156#define vst4_f32(__p0, __p1) __extension__ ({ \
27157  float32x2x4_t __s1 = __p1; \
27158  float32x2x4_t __rev1; \
27159  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27160  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27161  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27162  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27163  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 9); \
27164})
27165#endif
27166
27167#ifdef __LITTLE_ENDIAN__
27168#define vst4_f16(__p0, __p1) __extension__ ({ \
27169  float16x4x4_t __s1 = __p1; \
27170  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 8); \
27171})
27172#else
27173#define vst4_f16(__p0, __p1) __extension__ ({ \
27174  float16x4x4_t __s1 = __p1; \
27175  float16x4x4_t __rev1; \
27176  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27177  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27178  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27179  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27180  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 8); \
27181})
27182#endif
27183
27184#ifdef __LITTLE_ENDIAN__
27185#define vst4_s32(__p0, __p1) __extension__ ({ \
27186  int32x2x4_t __s1 = __p1; \
27187  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 2); \
27188})
27189#else
27190#define vst4_s32(__p0, __p1) __extension__ ({ \
27191  int32x2x4_t __s1 = __p1; \
27192  int32x2x4_t __rev1; \
27193  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27194  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27195  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27196  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27197  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 2); \
27198})
27199#endif
27200
27201#ifdef __LITTLE_ENDIAN__
27202#define vst4_s64(__p0, __p1) __extension__ ({ \
27203  int64x1x4_t __s1 = __p1; \
27204  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
27205})
27206#else
27207#define vst4_s64(__p0, __p1) __extension__ ({ \
27208  int64x1x4_t __s1 = __p1; \
27209  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
27210})
27211#endif
27212
27213#ifdef __LITTLE_ENDIAN__
27214#define vst4_s16(__p0, __p1) __extension__ ({ \
27215  int16x4x4_t __s1 = __p1; \
27216  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 1); \
27217})
27218#else
27219#define vst4_s16(__p0, __p1) __extension__ ({ \
27220  int16x4x4_t __s1 = __p1; \
27221  int16x4x4_t __rev1; \
27222  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27223  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27224  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27225  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27226  __builtin_neon_vst4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 1); \
27227})
27228#endif
27229
27230#ifdef __LITTLE_ENDIAN__
27231#define vst4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
27232  poly8x8x4_t __s1 = __p1; \
27233  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 4); \
27234})
27235#else
27236#define vst4_lane_p8(__p0, __p1, __p2) __extension__ ({ \
27237  poly8x8x4_t __s1 = __p1; \
27238  poly8x8x4_t __rev1; \
27239  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27240  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27241  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27242  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27243  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 4); \
27244})
27245#endif
27246
27247#ifdef __LITTLE_ENDIAN__
27248#define vst4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27249  poly16x4x4_t __s1 = __p1; \
27250  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 5); \
27251})
27252#else
27253#define vst4_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27254  poly16x4x4_t __s1 = __p1; \
27255  poly16x4x4_t __rev1; \
27256  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27257  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27258  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27259  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27260  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 5); \
27261})
27262#endif
27263
27264#ifdef __LITTLE_ENDIAN__
27265#define vst4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27266  poly16x8x4_t __s1 = __p1; \
27267  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 37); \
27268})
27269#else
27270#define vst4q_lane_p16(__p0, __p1, __p2) __extension__ ({ \
27271  poly16x8x4_t __s1 = __p1; \
27272  poly16x8x4_t __rev1; \
27273  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27274  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27275  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27276  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27277  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 37); \
27278})
27279#endif
27280
27281#ifdef __LITTLE_ENDIAN__
27282#define vst4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27283  uint32x4x4_t __s1 = __p1; \
27284  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 50); \
27285})
27286#else
27287#define vst4q_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27288  uint32x4x4_t __s1 = __p1; \
27289  uint32x4x4_t __rev1; \
27290  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27291  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27292  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27293  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27294  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 50); \
27295})
27296#endif
27297
27298#ifdef __LITTLE_ENDIAN__
27299#define vst4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27300  uint16x8x4_t __s1 = __p1; \
27301  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 49); \
27302})
27303#else
27304#define vst4q_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27305  uint16x8x4_t __s1 = __p1; \
27306  uint16x8x4_t __rev1; \
27307  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27308  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27309  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27310  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27311  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 49); \
27312})
27313#endif
27314
27315#ifdef __LITTLE_ENDIAN__
27316#define vst4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27317  float32x4x4_t __s1 = __p1; \
27318  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 41); \
27319})
27320#else
27321#define vst4q_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27322  float32x4x4_t __s1 = __p1; \
27323  float32x4x4_t __rev1; \
27324  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27325  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27326  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27327  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27328  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 41); \
27329})
27330#endif
27331
27332#ifdef __LITTLE_ENDIAN__
27333#define vst4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27334  float16x8x4_t __s1 = __p1; \
27335  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 40); \
27336})
27337#else
27338#define vst4q_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27339  float16x8x4_t __s1 = __p1; \
27340  float16x8x4_t __rev1; \
27341  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27342  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27343  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27344  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27345  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 40); \
27346})
27347#endif
27348
27349#ifdef __LITTLE_ENDIAN__
27350#define vst4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27351  int32x4x4_t __s1 = __p1; \
27352  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 34); \
27353})
27354#else
27355#define vst4q_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27356  int32x4x4_t __s1 = __p1; \
27357  int32x4x4_t __rev1; \
27358  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27359  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27360  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27361  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27362  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 34); \
27363})
27364#endif
27365
27366#ifdef __LITTLE_ENDIAN__
27367#define vst4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27368  int16x8x4_t __s1 = __p1; \
27369  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 33); \
27370})
27371#else
27372#define vst4q_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27373  int16x8x4_t __s1 = __p1; \
27374  int16x8x4_t __rev1; \
27375  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27376  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27377  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27378  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27379  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 33); \
27380})
27381#endif
27382
27383#ifdef __LITTLE_ENDIAN__
27384#define vst4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
27385  uint8x8x4_t __s1 = __p1; \
27386  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 16); \
27387})
27388#else
27389#define vst4_lane_u8(__p0, __p1, __p2) __extension__ ({ \
27390  uint8x8x4_t __s1 = __p1; \
27391  uint8x8x4_t __rev1; \
27392  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27393  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27394  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27395  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27396  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 16); \
27397})
27398#endif
27399
27400#ifdef __LITTLE_ENDIAN__
27401#define vst4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27402  uint32x2x4_t __s1 = __p1; \
27403  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 18); \
27404})
27405#else
27406#define vst4_lane_u32(__p0, __p1, __p2) __extension__ ({ \
27407  uint32x2x4_t __s1 = __p1; \
27408  uint32x2x4_t __rev1; \
27409  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27410  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27411  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27412  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27413  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 18); \
27414})
27415#endif
27416
27417#ifdef __LITTLE_ENDIAN__
27418#define vst4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27419  uint16x4x4_t __s1 = __p1; \
27420  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 17); \
27421})
27422#else
27423#define vst4_lane_u16(__p0, __p1, __p2) __extension__ ({ \
27424  uint16x4x4_t __s1 = __p1; \
27425  uint16x4x4_t __rev1; \
27426  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27427  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27428  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27429  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27430  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 17); \
27431})
27432#endif
27433
27434#ifdef __LITTLE_ENDIAN__
27435#define vst4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
27436  int8x8x4_t __s1 = __p1; \
27437  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 0); \
27438})
27439#else
27440#define vst4_lane_s8(__p0, __p1, __p2) __extension__ ({ \
27441  int8x8x4_t __s1 = __p1; \
27442  int8x8x4_t __rev1; \
27443  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
27444  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
27445  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
27446  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
27447  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], __p2, 0); \
27448})
27449#endif
27450
27451#ifdef __LITTLE_ENDIAN__
27452#define vst4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27453  float32x2x4_t __s1 = __p1; \
27454  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 9); \
27455})
27456#else
27457#define vst4_lane_f32(__p0, __p1, __p2) __extension__ ({ \
27458  float32x2x4_t __s1 = __p1; \
27459  float32x2x4_t __rev1; \
27460  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27461  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27462  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27463  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27464  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 9); \
27465})
27466#endif
27467
27468#ifdef __LITTLE_ENDIAN__
27469#define vst4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27470  float16x4x4_t __s1 = __p1; \
27471  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 8); \
27472})
27473#else
27474#define vst4_lane_f16(__p0, __p1, __p2) __extension__ ({ \
27475  float16x4x4_t __s1 = __p1; \
27476  float16x4x4_t __rev1; \
27477  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27478  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27479  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27480  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27481  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 8); \
27482})
27483#endif
27484
27485#ifdef __LITTLE_ENDIAN__
27486#define vst4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27487  int32x2x4_t __s1 = __p1; \
27488  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 2); \
27489})
27490#else
27491#define vst4_lane_s32(__p0, __p1, __p2) __extension__ ({ \
27492  int32x2x4_t __s1 = __p1; \
27493  int32x2x4_t __rev1; \
27494  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
27495  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
27496  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
27497  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
27498  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 2); \
27499})
27500#endif
27501
27502#ifdef __LITTLE_ENDIAN__
27503#define vst4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27504  int16x4x4_t __s1 = __p1; \
27505  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 1); \
27506})
27507#else
27508#define vst4_lane_s16(__p0, __p1, __p2) __extension__ ({ \
27509  int16x4x4_t __s1 = __p1; \
27510  int16x4x4_t __rev1; \
27511  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
27512  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
27513  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
27514  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
27515  __builtin_neon_vst4_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 1); \
27516})
27517#endif
27518
27519#ifdef __LITTLE_ENDIAN__
27520__ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
27521  uint8x16_t __ret;
27522  __ret = __p0 - __p1;
27523  return __ret;
27524}
27525#else
27526__ai uint8x16_t vsubq_u8(uint8x16_t __p0, uint8x16_t __p1) {
27527  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27528  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27529  uint8x16_t __ret;
27530  __ret = __rev0 - __rev1;
27531  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27532  return __ret;
27533}
27534#endif
27535
27536#ifdef __LITTLE_ENDIAN__
27537__ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
27538  uint32x4_t __ret;
27539  __ret = __p0 - __p1;
27540  return __ret;
27541}
27542#else
27543__ai uint32x4_t vsubq_u32(uint32x4_t __p0, uint32x4_t __p1) {
27544  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27545  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27546  uint32x4_t __ret;
27547  __ret = __rev0 - __rev1;
27548  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27549  return __ret;
27550}
27551#endif
27552
27553#ifdef __LITTLE_ENDIAN__
27554__ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
27555  uint64x2_t __ret;
27556  __ret = __p0 - __p1;
27557  return __ret;
27558}
27559#else
27560__ai uint64x2_t vsubq_u64(uint64x2_t __p0, uint64x2_t __p1) {
27561  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27562  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27563  uint64x2_t __ret;
27564  __ret = __rev0 - __rev1;
27565  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27566  return __ret;
27567}
27568#endif
27569
27570#ifdef __LITTLE_ENDIAN__
27571__ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
27572  uint16x8_t __ret;
27573  __ret = __p0 - __p1;
27574  return __ret;
27575}
27576#else
27577__ai uint16x8_t vsubq_u16(uint16x8_t __p0, uint16x8_t __p1) {
27578  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27579  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27580  uint16x8_t __ret;
27581  __ret = __rev0 - __rev1;
27582  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27583  return __ret;
27584}
27585#endif
27586
27587#ifdef __LITTLE_ENDIAN__
27588__ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1) {
27589  int8x16_t __ret;
27590  __ret = __p0 - __p1;
27591  return __ret;
27592}
27593#else
27594__ai int8x16_t vsubq_s8(int8x16_t __p0, int8x16_t __p1) {
27595  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27596  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27597  int8x16_t __ret;
27598  __ret = __rev0 - __rev1;
27599  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
27600  return __ret;
27601}
27602#endif
27603
27604#ifdef __LITTLE_ENDIAN__
27605__ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1) {
27606  float32x4_t __ret;
27607  __ret = __p0 - __p1;
27608  return __ret;
27609}
27610#else
27611__ai float32x4_t vsubq_f32(float32x4_t __p0, float32x4_t __p1) {
27612  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27613  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27614  float32x4_t __ret;
27615  __ret = __rev0 - __rev1;
27616  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27617  return __ret;
27618}
27619#endif
27620
27621#ifdef __LITTLE_ENDIAN__
27622__ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1) {
27623  int32x4_t __ret;
27624  __ret = __p0 - __p1;
27625  return __ret;
27626}
27627#else
27628__ai int32x4_t vsubq_s32(int32x4_t __p0, int32x4_t __p1) {
27629  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27630  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27631  int32x4_t __ret;
27632  __ret = __rev0 - __rev1;
27633  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27634  return __ret;
27635}
27636#endif
27637
27638#ifdef __LITTLE_ENDIAN__
27639__ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1) {
27640  int64x2_t __ret;
27641  __ret = __p0 - __p1;
27642  return __ret;
27643}
27644#else
27645__ai int64x2_t vsubq_s64(int64x2_t __p0, int64x2_t __p1) {
27646  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27647  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27648  int64x2_t __ret;
27649  __ret = __rev0 - __rev1;
27650  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27651  return __ret;
27652}
27653#endif
27654
27655#ifdef __LITTLE_ENDIAN__
27656__ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1) {
27657  int16x8_t __ret;
27658  __ret = __p0 - __p1;
27659  return __ret;
27660}
27661#else
27662__ai int16x8_t vsubq_s16(int16x8_t __p0, int16x8_t __p1) {
27663  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27664  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27665  int16x8_t __ret;
27666  __ret = __rev0 - __rev1;
27667  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27668  return __ret;
27669}
27670#endif
27671
27672#ifdef __LITTLE_ENDIAN__
27673__ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
27674  uint8x8_t __ret;
27675  __ret = __p0 - __p1;
27676  return __ret;
27677}
27678#else
27679__ai uint8x8_t vsub_u8(uint8x8_t __p0, uint8x8_t __p1) {
27680  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27681  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27682  uint8x8_t __ret;
27683  __ret = __rev0 - __rev1;
27684  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27685  return __ret;
27686}
27687#endif
27688
27689#ifdef __LITTLE_ENDIAN__
27690__ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
27691  uint32x2_t __ret;
27692  __ret = __p0 - __p1;
27693  return __ret;
27694}
27695#else
27696__ai uint32x2_t vsub_u32(uint32x2_t __p0, uint32x2_t __p1) {
27697  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27698  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27699  uint32x2_t __ret;
27700  __ret = __rev0 - __rev1;
27701  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27702  return __ret;
27703}
27704#endif
27705
27706#ifdef __LITTLE_ENDIAN__
27707__ai uint64x1_t vsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
27708  uint64x1_t __ret;
27709  __ret = __p0 - __p1;
27710  return __ret;
27711}
27712#else
27713__ai uint64x1_t vsub_u64(uint64x1_t __p0, uint64x1_t __p1) {
27714  uint64x1_t __ret;
27715  __ret = __p0 - __p1;
27716  return __ret;
27717}
27718#endif
27719
27720#ifdef __LITTLE_ENDIAN__
27721__ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
27722  uint16x4_t __ret;
27723  __ret = __p0 - __p1;
27724  return __ret;
27725}
27726#else
27727__ai uint16x4_t vsub_u16(uint16x4_t __p0, uint16x4_t __p1) {
27728  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27729  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27730  uint16x4_t __ret;
27731  __ret = __rev0 - __rev1;
27732  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27733  return __ret;
27734}
27735#endif
27736
27737#ifdef __LITTLE_ENDIAN__
27738__ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1) {
27739  int8x8_t __ret;
27740  __ret = __p0 - __p1;
27741  return __ret;
27742}
27743#else
27744__ai int8x8_t vsub_s8(int8x8_t __p0, int8x8_t __p1) {
27745  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27746  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27747  int8x8_t __ret;
27748  __ret = __rev0 - __rev1;
27749  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27750  return __ret;
27751}
27752#endif
27753
27754#ifdef __LITTLE_ENDIAN__
27755__ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1) {
27756  float32x2_t __ret;
27757  __ret = __p0 - __p1;
27758  return __ret;
27759}
27760#else
27761__ai float32x2_t vsub_f32(float32x2_t __p0, float32x2_t __p1) {
27762  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27763  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27764  float32x2_t __ret;
27765  __ret = __rev0 - __rev1;
27766  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27767  return __ret;
27768}
27769#endif
27770
27771#ifdef __LITTLE_ENDIAN__
27772__ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1) {
27773  int32x2_t __ret;
27774  __ret = __p0 - __p1;
27775  return __ret;
27776}
27777#else
27778__ai int32x2_t vsub_s32(int32x2_t __p0, int32x2_t __p1) {
27779  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27780  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27781  int32x2_t __ret;
27782  __ret = __rev0 - __rev1;
27783  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27784  return __ret;
27785}
27786#endif
27787
27788#ifdef __LITTLE_ENDIAN__
27789__ai int64x1_t vsub_s64(int64x1_t __p0, int64x1_t __p1) {
27790  int64x1_t __ret;
27791  __ret = __p0 - __p1;
27792  return __ret;
27793}
27794#else
27795__ai int64x1_t vsub_s64(int64x1_t __p0, int64x1_t __p1) {
27796  int64x1_t __ret;
27797  __ret = __p0 - __p1;
27798  return __ret;
27799}
27800#endif
27801
27802#ifdef __LITTLE_ENDIAN__
27803__ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1) {
27804  int16x4_t __ret;
27805  __ret = __p0 - __p1;
27806  return __ret;
27807}
27808#else
27809__ai int16x4_t vsub_s16(int16x4_t __p0, int16x4_t __p1) {
27810  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27811  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27812  int16x4_t __ret;
27813  __ret = __rev0 - __rev1;
27814  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27815  return __ret;
27816}
27817#endif
27818
27819#ifdef __LITTLE_ENDIAN__
27820__ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
27821  uint16x4_t __ret;
27822  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
27823  return __ret;
27824}
27825#else
27826__ai uint16x4_t vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
27827  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27828  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27829  uint16x4_t __ret;
27830  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 17);
27831  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27832  return __ret;
27833}
27834__ai uint16x4_t __noswap_vsubhn_u32(uint32x4_t __p0, uint32x4_t __p1) {
27835  uint16x4_t __ret;
27836  __ret = (uint16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 17);
27837  return __ret;
27838}
27839#endif
27840
27841#ifdef __LITTLE_ENDIAN__
27842__ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
27843  uint32x2_t __ret;
27844  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
27845  return __ret;
27846}
27847#else
27848__ai uint32x2_t vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
27849  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27850  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27851  uint32x2_t __ret;
27852  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 18);
27853  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27854  return __ret;
27855}
27856__ai uint32x2_t __noswap_vsubhn_u64(uint64x2_t __p0, uint64x2_t __p1) {
27857  uint32x2_t __ret;
27858  __ret = (uint32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 18);
27859  return __ret;
27860}
27861#endif
27862
27863#ifdef __LITTLE_ENDIAN__
27864__ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
27865  uint8x8_t __ret;
27866  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
27867  return __ret;
27868}
27869#else
27870__ai uint8x8_t vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
27871  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27872  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27873  uint8x8_t __ret;
27874  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 16);
27875  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27876  return __ret;
27877}
27878__ai uint8x8_t __noswap_vsubhn_u16(uint16x8_t __p0, uint16x8_t __p1) {
27879  uint8x8_t __ret;
27880  __ret = (uint8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 16);
27881  return __ret;
27882}
27883#endif
27884
27885#ifdef __LITTLE_ENDIAN__
27886__ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
27887  int16x4_t __ret;
27888  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
27889  return __ret;
27890}
27891#else
27892__ai int16x4_t vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
27893  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27894  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27895  int16x4_t __ret;
27896  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 1);
27897  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27898  return __ret;
27899}
27900__ai int16x4_t __noswap_vsubhn_s32(int32x4_t __p0, int32x4_t __p1) {
27901  int16x4_t __ret;
27902  __ret = (int16x4_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 1);
27903  return __ret;
27904}
27905#endif
27906
27907#ifdef __LITTLE_ENDIAN__
27908__ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
27909  int32x2_t __ret;
27910  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
27911  return __ret;
27912}
27913#else
27914__ai int32x2_t vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
27915  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27916  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27917  int32x2_t __ret;
27918  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 2);
27919  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27920  return __ret;
27921}
27922__ai int32x2_t __noswap_vsubhn_s64(int64x2_t __p0, int64x2_t __p1) {
27923  int32x2_t __ret;
27924  __ret = (int32x2_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 2);
27925  return __ret;
27926}
27927#endif
27928
27929#ifdef __LITTLE_ENDIAN__
27930__ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
27931  int8x8_t __ret;
27932  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
27933  return __ret;
27934}
27935#else
27936__ai int8x8_t vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
27937  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27938  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27939  int8x8_t __ret;
27940  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__rev0, (int8x16_t)__rev1, 0);
27941  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27942  return __ret;
27943}
27944__ai int8x8_t __noswap_vsubhn_s16(int16x8_t __p0, int16x8_t __p1) {
27945  int8x8_t __ret;
27946  __ret = (int8x8_t) __builtin_neon_vsubhn_v((int8x16_t)__p0, (int8x16_t)__p1, 0);
27947  return __ret;
27948}
27949#endif
27950
27951#ifdef __LITTLE_ENDIAN__
27952__ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1) {
27953  uint16x8_t __ret;
27954  __ret = vmovl_u8(__p0) - vmovl_u8(__p1);
27955  return __ret;
27956}
27957#else
27958__ai uint16x8_t vsubl_u8(uint8x8_t __p0, uint8x8_t __p1) {
27959  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
27960  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
27961  uint16x8_t __ret;
27962  __ret = __noswap_vmovl_u8(__rev0) - __noswap_vmovl_u8(__rev1);
27963  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
27964  return __ret;
27965}
27966#endif
27967
27968#ifdef __LITTLE_ENDIAN__
27969__ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1) {
27970  uint64x2_t __ret;
27971  __ret = vmovl_u32(__p0) - vmovl_u32(__p1);
27972  return __ret;
27973}
27974#else
27975__ai uint64x2_t vsubl_u32(uint32x2_t __p0, uint32x2_t __p1) {
27976  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
27977  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
27978  uint64x2_t __ret;
27979  __ret = __noswap_vmovl_u32(__rev0) - __noswap_vmovl_u32(__rev1);
27980  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
27981  return __ret;
27982}
27983#endif
27984
27985#ifdef __LITTLE_ENDIAN__
27986__ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1) {
27987  uint32x4_t __ret;
27988  __ret = vmovl_u16(__p0) - vmovl_u16(__p1);
27989  return __ret;
27990}
27991#else
27992__ai uint32x4_t vsubl_u16(uint16x4_t __p0, uint16x4_t __p1) {
27993  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
27994  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
27995  uint32x4_t __ret;
27996  __ret = __noswap_vmovl_u16(__rev0) - __noswap_vmovl_u16(__rev1);
27997  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
27998  return __ret;
27999}
28000#endif
28001
28002#ifdef __LITTLE_ENDIAN__
28003__ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1) {
28004  int16x8_t __ret;
28005  __ret = vmovl_s8(__p0) - vmovl_s8(__p1);
28006  return __ret;
28007}
28008#else
28009__ai int16x8_t vsubl_s8(int8x8_t __p0, int8x8_t __p1) {
28010  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28011  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28012  int16x8_t __ret;
28013  __ret = __noswap_vmovl_s8(__rev0) - __noswap_vmovl_s8(__rev1);
28014  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28015  return __ret;
28016}
28017#endif
28018
28019#ifdef __LITTLE_ENDIAN__
28020__ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1) {
28021  int64x2_t __ret;
28022  __ret = vmovl_s32(__p0) - vmovl_s32(__p1);
28023  return __ret;
28024}
28025#else
28026__ai int64x2_t vsubl_s32(int32x2_t __p0, int32x2_t __p1) {
28027  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28028  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28029  int64x2_t __ret;
28030  __ret = __noswap_vmovl_s32(__rev0) - __noswap_vmovl_s32(__rev1);
28031  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
28032  return __ret;
28033}
28034#endif
28035
28036#ifdef __LITTLE_ENDIAN__
28037__ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1) {
28038  int32x4_t __ret;
28039  __ret = vmovl_s16(__p0) - vmovl_s16(__p1);
28040  return __ret;
28041}
28042#else
28043__ai int32x4_t vsubl_s16(int16x4_t __p0, int16x4_t __p1) {
28044  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28045  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28046  int32x4_t __ret;
28047  __ret = __noswap_vmovl_s16(__rev0) - __noswap_vmovl_s16(__rev1);
28048  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
28049  return __ret;
28050}
28051#endif
28052
28053#ifdef __LITTLE_ENDIAN__
28054__ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1) {
28055  uint16x8_t __ret;
28056  __ret = __p0 - vmovl_u8(__p1);
28057  return __ret;
28058}
28059#else
28060__ai uint16x8_t vsubw_u8(uint16x8_t __p0, uint8x8_t __p1) {
28061  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28062  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28063  uint16x8_t __ret;
28064  __ret = __rev0 - __noswap_vmovl_u8(__rev1);
28065  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28066  return __ret;
28067}
28068#endif
28069
28070#ifdef __LITTLE_ENDIAN__
28071__ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1) {
28072  uint64x2_t __ret;
28073  __ret = __p0 - vmovl_u32(__p1);
28074  return __ret;
28075}
28076#else
28077__ai uint64x2_t vsubw_u32(uint64x2_t __p0, uint32x2_t __p1) {
28078  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28079  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28080  uint64x2_t __ret;
28081  __ret = __rev0 - __noswap_vmovl_u32(__rev1);
28082  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
28083  return __ret;
28084}
28085#endif
28086
28087#ifdef __LITTLE_ENDIAN__
28088__ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1) {
28089  uint32x4_t __ret;
28090  __ret = __p0 - vmovl_u16(__p1);
28091  return __ret;
28092}
28093#else
28094__ai uint32x4_t vsubw_u16(uint32x4_t __p0, uint16x4_t __p1) {
28095  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28096  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28097  uint32x4_t __ret;
28098  __ret = __rev0 - __noswap_vmovl_u16(__rev1);
28099  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
28100  return __ret;
28101}
28102#endif
28103
28104#ifdef __LITTLE_ENDIAN__
28105__ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1) {
28106  int16x8_t __ret;
28107  __ret = __p0 - vmovl_s8(__p1);
28108  return __ret;
28109}
28110#else
28111__ai int16x8_t vsubw_s8(int16x8_t __p0, int8x8_t __p1) {
28112  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28113  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28114  int16x8_t __ret;
28115  __ret = __rev0 - __noswap_vmovl_s8(__rev1);
28116  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28117  return __ret;
28118}
28119#endif
28120
28121#ifdef __LITTLE_ENDIAN__
28122__ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1) {
28123  int64x2_t __ret;
28124  __ret = __p0 - vmovl_s32(__p1);
28125  return __ret;
28126}
28127#else
28128__ai int64x2_t vsubw_s32(int64x2_t __p0, int32x2_t __p1) {
28129  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28130  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28131  int64x2_t __ret;
28132  __ret = __rev0 - __noswap_vmovl_s32(__rev1);
28133  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
28134  return __ret;
28135}
28136#endif
28137
28138#ifdef __LITTLE_ENDIAN__
28139__ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1) {
28140  int32x4_t __ret;
28141  __ret = __p0 - vmovl_s16(__p1);
28142  return __ret;
28143}
28144#else
28145__ai int32x4_t vsubw_s16(int32x4_t __p0, int16x4_t __p1) {
28146  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28147  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28148  int32x4_t __ret;
28149  __ret = __rev0 - __noswap_vmovl_s16(__rev1);
28150  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
28151  return __ret;
28152}
28153#endif
28154
28155#ifdef __LITTLE_ENDIAN__
28156__ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1) {
28157  poly8x8_t __ret;
28158  __ret = (poly8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 4);
28159  return __ret;
28160}
28161#else
28162__ai poly8x8_t vtbl1_p8(poly8x8_t __p0, uint8x8_t __p1) {
28163  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28164  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28165  poly8x8_t __ret;
28166  __ret = (poly8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4);
28167  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28168  return __ret;
28169}
28170#endif
28171
28172#ifdef __LITTLE_ENDIAN__
28173__ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1) {
28174  uint8x8_t __ret;
28175  __ret = (uint8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
28176  return __ret;
28177}
28178#else
28179__ai uint8x8_t vtbl1_u8(uint8x8_t __p0, uint8x8_t __p1) {
28180  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28181  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28182  uint8x8_t __ret;
28183  __ret = (uint8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
28184  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28185  return __ret;
28186}
28187#endif
28188
28189#ifdef __LITTLE_ENDIAN__
28190__ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1) {
28191  int8x8_t __ret;
28192  __ret = (int8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
28193  return __ret;
28194}
28195#else
28196__ai int8x8_t vtbl1_s8(int8x8_t __p0, int8x8_t __p1) {
28197  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28198  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28199  int8x8_t __ret;
28200  __ret = (int8x8_t) __builtin_neon_vtbl1_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
28201  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28202  return __ret;
28203}
28204#endif
28205
28206#ifdef __LITTLE_ENDIAN__
28207__ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1) {
28208  poly8x8_t __ret;
28209  __ret = (poly8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 4);
28210  return __ret;
28211}
28212#else
28213__ai poly8x8_t vtbl2_p8(poly8x8x2_t __p0, uint8x8_t __p1) {
28214  poly8x8x2_t __rev0;
28215  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28216  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28217  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28218  poly8x8_t __ret;
28219  __ret = (poly8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 4);
28220  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28221  return __ret;
28222}
28223#endif
28224
28225#ifdef __LITTLE_ENDIAN__
28226__ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1) {
28227  uint8x8_t __ret;
28228  __ret = (uint8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 16);
28229  return __ret;
28230}
28231#else
28232__ai uint8x8_t vtbl2_u8(uint8x8x2_t __p0, uint8x8_t __p1) {
28233  uint8x8x2_t __rev0;
28234  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28235  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28236  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28237  uint8x8_t __ret;
28238  __ret = (uint8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 16);
28239  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28240  return __ret;
28241}
28242#endif
28243
28244#ifdef __LITTLE_ENDIAN__
28245__ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1) {
28246  int8x8_t __ret;
28247  __ret = (int8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p1, 0);
28248  return __ret;
28249}
28250#else
28251__ai int8x8_t vtbl2_s8(int8x8x2_t __p0, int8x8_t __p1) {
28252  int8x8x2_t __rev0;
28253  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28254  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28255  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28256  int8x8_t __ret;
28257  __ret = (int8x8_t) __builtin_neon_vtbl2_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev1, 0);
28258  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28259  return __ret;
28260}
28261#endif
28262
28263#ifdef __LITTLE_ENDIAN__
28264__ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1) {
28265  poly8x8_t __ret;
28266  __ret = (poly8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 4);
28267  return __ret;
28268}
28269#else
28270__ai poly8x8_t vtbl3_p8(poly8x8x3_t __p0, uint8x8_t __p1) {
28271  poly8x8x3_t __rev0;
28272  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28273  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28274  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28275  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28276  poly8x8_t __ret;
28277  __ret = (poly8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 4);
28278  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28279  return __ret;
28280}
28281#endif
28282
28283#ifdef __LITTLE_ENDIAN__
28284__ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1) {
28285  uint8x8_t __ret;
28286  __ret = (uint8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 16);
28287  return __ret;
28288}
28289#else
28290__ai uint8x8_t vtbl3_u8(uint8x8x3_t __p0, uint8x8_t __p1) {
28291  uint8x8x3_t __rev0;
28292  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28293  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28294  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28295  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28296  uint8x8_t __ret;
28297  __ret = (uint8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 16);
28298  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28299  return __ret;
28300}
28301#endif
28302
28303#ifdef __LITTLE_ENDIAN__
28304__ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1) {
28305  int8x8_t __ret;
28306  __ret = (int8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p1, 0);
28307  return __ret;
28308}
28309#else
28310__ai int8x8_t vtbl3_s8(int8x8x3_t __p0, int8x8_t __p1) {
28311  int8x8x3_t __rev0;
28312  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28313  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28314  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28315  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28316  int8x8_t __ret;
28317  __ret = (int8x8_t) __builtin_neon_vtbl3_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev1, 0);
28318  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28319  return __ret;
28320}
28321#endif
28322
28323#ifdef __LITTLE_ENDIAN__
28324__ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1) {
28325  poly8x8_t __ret;
28326  __ret = (poly8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 4);
28327  return __ret;
28328}
28329#else
28330__ai poly8x8_t vtbl4_p8(poly8x8x4_t __p0, uint8x8_t __p1) {
28331  poly8x8x4_t __rev0;
28332  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28333  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28334  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28335  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28336  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28337  poly8x8_t __ret;
28338  __ret = (poly8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 4);
28339  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28340  return __ret;
28341}
28342#endif
28343
28344#ifdef __LITTLE_ENDIAN__
28345__ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1) {
28346  uint8x8_t __ret;
28347  __ret = (uint8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 16);
28348  return __ret;
28349}
28350#else
28351__ai uint8x8_t vtbl4_u8(uint8x8x4_t __p0, uint8x8_t __p1) {
28352  uint8x8x4_t __rev0;
28353  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28354  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28355  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28356  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28357  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28358  uint8x8_t __ret;
28359  __ret = (uint8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 16);
28360  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28361  return __ret;
28362}
28363#endif
28364
28365#ifdef __LITTLE_ENDIAN__
28366__ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1) {
28367  int8x8_t __ret;
28368  __ret = (int8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__p0.val[0], (int8x8_t)__p0.val[1], (int8x8_t)__p0.val[2], (int8x8_t)__p0.val[3], (int8x8_t)__p1, 0);
28369  return __ret;
28370}
28371#else
28372__ai int8x8_t vtbl4_s8(int8x8x4_t __p0, int8x8_t __p1) {
28373  int8x8x4_t __rev0;
28374  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28375  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28376  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28377  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28378  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28379  int8x8_t __ret;
28380  __ret = (int8x8_t) __builtin_neon_vtbl4_v((int8x8_t)__rev0.val[0], (int8x8_t)__rev0.val[1], (int8x8_t)__rev0.val[2], (int8x8_t)__rev0.val[3], (int8x8_t)__rev1, 0);
28381  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28382  return __ret;
28383}
28384#endif
28385
28386#ifdef __LITTLE_ENDIAN__
28387__ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2) {
28388  poly8x8_t __ret;
28389  __ret = (poly8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 4);
28390  return __ret;
28391}
28392#else
28393__ai poly8x8_t vtbx1_p8(poly8x8_t __p0, poly8x8_t __p1, uint8x8_t __p2) {
28394  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28395  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28396  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28397  poly8x8_t __ret;
28398  __ret = (poly8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 4);
28399  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28400  return __ret;
28401}
28402#endif
28403
28404#ifdef __LITTLE_ENDIAN__
28405__ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
28406  uint8x8_t __ret;
28407  __ret = (uint8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 16);
28408  return __ret;
28409}
28410#else
28411__ai uint8x8_t vtbx1_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
28412  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28413  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28414  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28415  uint8x8_t __ret;
28416  __ret = (uint8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 16);
28417  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28418  return __ret;
28419}
28420#endif
28421
28422#ifdef __LITTLE_ENDIAN__
28423__ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
28424  int8x8_t __ret;
28425  __ret = (int8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 0);
28426  return __ret;
28427}
28428#else
28429__ai int8x8_t vtbx1_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
28430  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28431  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28432  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28433  int8x8_t __ret;
28434  __ret = (int8x8_t) __builtin_neon_vtbx1_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 0);
28435  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28436  return __ret;
28437}
28438#endif
28439
28440#ifdef __LITTLE_ENDIAN__
28441__ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2) {
28442  poly8x8_t __ret;
28443  __ret = (poly8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 4);
28444  return __ret;
28445}
28446#else
28447__ai poly8x8_t vtbx2_p8(poly8x8_t __p0, poly8x8x2_t __p1, uint8x8_t __p2) {
28448  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28449  poly8x8x2_t __rev1;
28450  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28451  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28452  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28453  poly8x8_t __ret;
28454  __ret = (poly8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 4);
28455  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28456  return __ret;
28457}
28458#endif
28459
28460#ifdef __LITTLE_ENDIAN__
28461__ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2) {
28462  uint8x8_t __ret;
28463  __ret = (uint8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 16);
28464  return __ret;
28465}
28466#else
28467__ai uint8x8_t vtbx2_u8(uint8x8_t __p0, uint8x8x2_t __p1, uint8x8_t __p2) {
28468  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28469  uint8x8x2_t __rev1;
28470  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28471  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28472  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28473  uint8x8_t __ret;
28474  __ret = (uint8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 16);
28475  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28476  return __ret;
28477}
28478#endif
28479
28480#ifdef __LITTLE_ENDIAN__
28481__ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2) {
28482  int8x8_t __ret;
28483  __ret = (int8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p2, 0);
28484  return __ret;
28485}
28486#else
28487__ai int8x8_t vtbx2_s8(int8x8_t __p0, int8x8x2_t __p1, int8x8_t __p2) {
28488  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28489  int8x8x2_t __rev1;
28490  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28491  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28492  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28493  int8x8_t __ret;
28494  __ret = (int8x8_t) __builtin_neon_vtbx2_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev2, 0);
28495  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28496  return __ret;
28497}
28498#endif
28499
28500#ifdef __LITTLE_ENDIAN__
28501__ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2) {
28502  poly8x8_t __ret;
28503  __ret = (poly8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 4);
28504  return __ret;
28505}
28506#else
28507__ai poly8x8_t vtbx3_p8(poly8x8_t __p0, poly8x8x3_t __p1, uint8x8_t __p2) {
28508  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28509  poly8x8x3_t __rev1;
28510  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28511  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28512  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28513  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28514  poly8x8_t __ret;
28515  __ret = (poly8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 4);
28516  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28517  return __ret;
28518}
28519#endif
28520
28521#ifdef __LITTLE_ENDIAN__
28522__ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2) {
28523  uint8x8_t __ret;
28524  __ret = (uint8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 16);
28525  return __ret;
28526}
28527#else
28528__ai uint8x8_t vtbx3_u8(uint8x8_t __p0, uint8x8x3_t __p1, uint8x8_t __p2) {
28529  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28530  uint8x8x3_t __rev1;
28531  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28532  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28533  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28534  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28535  uint8x8_t __ret;
28536  __ret = (uint8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 16);
28537  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28538  return __ret;
28539}
28540#endif
28541
28542#ifdef __LITTLE_ENDIAN__
28543__ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2) {
28544  int8x8_t __ret;
28545  __ret = (int8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p2, 0);
28546  return __ret;
28547}
28548#else
28549__ai int8x8_t vtbx3_s8(int8x8_t __p0, int8x8x3_t __p1, int8x8_t __p2) {
28550  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28551  int8x8x3_t __rev1;
28552  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28553  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28554  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28555  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28556  int8x8_t __ret;
28557  __ret = (int8x8_t) __builtin_neon_vtbx3_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev2, 0);
28558  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28559  return __ret;
28560}
28561#endif
28562
28563#ifdef __LITTLE_ENDIAN__
28564__ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2) {
28565  poly8x8_t __ret;
28566  __ret = (poly8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 4);
28567  return __ret;
28568}
28569#else
28570__ai poly8x8_t vtbx4_p8(poly8x8_t __p0, poly8x8x4_t __p1, uint8x8_t __p2) {
28571  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28572  poly8x8x4_t __rev1;
28573  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28574  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28575  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28576  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28577  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28578  poly8x8_t __ret;
28579  __ret = (poly8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 4);
28580  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28581  return __ret;
28582}
28583#endif
28584
28585#ifdef __LITTLE_ENDIAN__
28586__ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2) {
28587  uint8x8_t __ret;
28588  __ret = (uint8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 16);
28589  return __ret;
28590}
28591#else
28592__ai uint8x8_t vtbx4_u8(uint8x8_t __p0, uint8x8x4_t __p1, uint8x8_t __p2) {
28593  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28594  uint8x8x4_t __rev1;
28595  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28596  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28597  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28598  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28599  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28600  uint8x8_t __ret;
28601  __ret = (uint8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 16);
28602  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28603  return __ret;
28604}
28605#endif
28606
28607#ifdef __LITTLE_ENDIAN__
28608__ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2) {
28609  int8x8_t __ret;
28610  __ret = (int8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__p0, (int8x8_t)__p1.val[0], (int8x8_t)__p1.val[1], (int8x8_t)__p1.val[2], (int8x8_t)__p1.val[3], (int8x8_t)__p2, 0);
28611  return __ret;
28612}
28613#else
28614__ai int8x8_t vtbx4_s8(int8x8_t __p0, int8x8x4_t __p1, int8x8_t __p2) {
28615  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28616  int8x8x4_t __rev1;
28617  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28618  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28619  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 7, 6, 5, 4, 3, 2, 1, 0);
28620  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 7, 6, 5, 4, 3, 2, 1, 0);
28621  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
28622  int8x8_t __ret;
28623  __ret = (int8x8_t) __builtin_neon_vtbx4_v((int8x8_t)__rev0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], (int8x8_t)__rev2, 0);
28624  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28625  return __ret;
28626}
28627#endif
28628
28629#ifdef __LITTLE_ENDIAN__
28630__ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1) {
28631  poly8x8x2_t __ret;
28632  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
28633  return __ret;
28634}
28635#else
28636__ai poly8x8x2_t vtrn_p8(poly8x8_t __p0, poly8x8_t __p1) {
28637  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28638  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28639  poly8x8x2_t __ret;
28640  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
28641
28642  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28643  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28644  return __ret;
28645}
28646#endif
28647
28648#ifdef __LITTLE_ENDIAN__
28649__ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1) {
28650  poly16x4x2_t __ret;
28651  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
28652  return __ret;
28653}
28654#else
28655__ai poly16x4x2_t vtrn_p16(poly16x4_t __p0, poly16x4_t __p1) {
28656  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28657  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28658  poly16x4x2_t __ret;
28659  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
28660
28661  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28662  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28663  return __ret;
28664}
28665#endif
28666
28667#ifdef __LITTLE_ENDIAN__
28668__ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1) {
28669  poly8x16x2_t __ret;
28670  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
28671  return __ret;
28672}
28673#else
28674__ai poly8x16x2_t vtrnq_p8(poly8x16_t __p0, poly8x16_t __p1) {
28675  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28676  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28677  poly8x16x2_t __ret;
28678  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
28679
28680  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28681  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28682  return __ret;
28683}
28684#endif
28685
28686#ifdef __LITTLE_ENDIAN__
28687__ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1) {
28688  poly16x8x2_t __ret;
28689  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
28690  return __ret;
28691}
28692#else
28693__ai poly16x8x2_t vtrnq_p16(poly16x8_t __p0, poly16x8_t __p1) {
28694  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28695  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28696  poly16x8x2_t __ret;
28697  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
28698
28699  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28700  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28701  return __ret;
28702}
28703#endif
28704
28705#ifdef __LITTLE_ENDIAN__
28706__ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1) {
28707  uint8x16x2_t __ret;
28708  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
28709  return __ret;
28710}
28711#else
28712__ai uint8x16x2_t vtrnq_u8(uint8x16_t __p0, uint8x16_t __p1) {
28713  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28714  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28715  uint8x16x2_t __ret;
28716  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
28717
28718  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28719  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28720  return __ret;
28721}
28722#endif
28723
28724#ifdef __LITTLE_ENDIAN__
28725__ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1) {
28726  uint32x4x2_t __ret;
28727  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
28728  return __ret;
28729}
28730#else
28731__ai uint32x4x2_t vtrnq_u32(uint32x4_t __p0, uint32x4_t __p1) {
28732  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28733  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28734  uint32x4x2_t __ret;
28735  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
28736
28737  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28738  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28739  return __ret;
28740}
28741#endif
28742
28743#ifdef __LITTLE_ENDIAN__
28744__ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1) {
28745  uint16x8x2_t __ret;
28746  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
28747  return __ret;
28748}
28749#else
28750__ai uint16x8x2_t vtrnq_u16(uint16x8_t __p0, uint16x8_t __p1) {
28751  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28752  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28753  uint16x8x2_t __ret;
28754  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
28755
28756  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28757  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28758  return __ret;
28759}
28760#endif
28761
28762#ifdef __LITTLE_ENDIAN__
28763__ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1) {
28764  int8x16x2_t __ret;
28765  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
28766  return __ret;
28767}
28768#else
28769__ai int8x16x2_t vtrnq_s8(int8x16_t __p0, int8x16_t __p1) {
28770  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28771  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28772  int8x16x2_t __ret;
28773  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
28774
28775  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28776  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
28777  return __ret;
28778}
28779#endif
28780
28781#ifdef __LITTLE_ENDIAN__
28782__ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1) {
28783  float32x4x2_t __ret;
28784  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
28785  return __ret;
28786}
28787#else
28788__ai float32x4x2_t vtrnq_f32(float32x4_t __p0, float32x4_t __p1) {
28789  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28790  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28791  float32x4x2_t __ret;
28792  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
28793
28794  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28795  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28796  return __ret;
28797}
28798#endif
28799
28800#ifdef __LITTLE_ENDIAN__
28801__ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1) {
28802  int32x4x2_t __ret;
28803  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
28804  return __ret;
28805}
28806#else
28807__ai int32x4x2_t vtrnq_s32(int32x4_t __p0, int32x4_t __p1) {
28808  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28809  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28810  int32x4x2_t __ret;
28811  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
28812
28813  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28814  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28815  return __ret;
28816}
28817#endif
28818
28819#ifdef __LITTLE_ENDIAN__
28820__ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1) {
28821  int16x8x2_t __ret;
28822  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
28823  return __ret;
28824}
28825#else
28826__ai int16x8x2_t vtrnq_s16(int16x8_t __p0, int16x8_t __p1) {
28827  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28828  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28829  int16x8x2_t __ret;
28830  __builtin_neon_vtrnq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
28831
28832  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28833  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28834  return __ret;
28835}
28836#endif
28837
28838#ifdef __LITTLE_ENDIAN__
28839__ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1) {
28840  uint8x8x2_t __ret;
28841  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
28842  return __ret;
28843}
28844#else
28845__ai uint8x8x2_t vtrn_u8(uint8x8_t __p0, uint8x8_t __p1) {
28846  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28847  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28848  uint8x8x2_t __ret;
28849  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
28850
28851  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28852  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28853  return __ret;
28854}
28855#endif
28856
28857#ifdef __LITTLE_ENDIAN__
28858__ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1) {
28859  uint32x2x2_t __ret;
28860  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
28861  return __ret;
28862}
28863#else
28864__ai uint32x2x2_t vtrn_u32(uint32x2_t __p0, uint32x2_t __p1) {
28865  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28866  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28867  uint32x2x2_t __ret;
28868  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
28869
28870  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
28871  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
28872  return __ret;
28873}
28874#endif
28875
28876#ifdef __LITTLE_ENDIAN__
28877__ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1) {
28878  uint16x4x2_t __ret;
28879  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
28880  return __ret;
28881}
28882#else
28883__ai uint16x4x2_t vtrn_u16(uint16x4_t __p0, uint16x4_t __p1) {
28884  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28885  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28886  uint16x4x2_t __ret;
28887  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
28888
28889  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28890  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28891  return __ret;
28892}
28893#endif
28894
28895#ifdef __LITTLE_ENDIAN__
28896__ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1) {
28897  int8x8x2_t __ret;
28898  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
28899  return __ret;
28900}
28901#else
28902__ai int8x8x2_t vtrn_s8(int8x8_t __p0, int8x8_t __p1) {
28903  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28904  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28905  int8x8x2_t __ret;
28906  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
28907
28908  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
28909  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
28910  return __ret;
28911}
28912#endif
28913
28914#ifdef __LITTLE_ENDIAN__
28915__ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1) {
28916  float32x2x2_t __ret;
28917  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
28918  return __ret;
28919}
28920#else
28921__ai float32x2x2_t vtrn_f32(float32x2_t __p0, float32x2_t __p1) {
28922  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28923  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28924  float32x2x2_t __ret;
28925  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
28926
28927  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
28928  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
28929  return __ret;
28930}
28931#endif
28932
28933#ifdef __LITTLE_ENDIAN__
28934__ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1) {
28935  int32x2x2_t __ret;
28936  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
28937  return __ret;
28938}
28939#else
28940__ai int32x2x2_t vtrn_s32(int32x2_t __p0, int32x2_t __p1) {
28941  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
28942  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
28943  int32x2x2_t __ret;
28944  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
28945
28946  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
28947  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
28948  return __ret;
28949}
28950#endif
28951
28952#ifdef __LITTLE_ENDIAN__
28953__ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1) {
28954  int16x4x2_t __ret;
28955  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
28956  return __ret;
28957}
28958#else
28959__ai int16x4x2_t vtrn_s16(int16x4_t __p0, int16x4_t __p1) {
28960  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28961  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28962  int16x4x2_t __ret;
28963  __builtin_neon_vtrn_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
28964
28965  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
28966  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
28967  return __ret;
28968}
28969#endif
28970
28971#ifdef __LITTLE_ENDIAN__
28972__ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1) {
28973  uint8x8_t __ret;
28974  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
28975  return __ret;
28976}
28977#else
28978__ai uint8x8_t vtst_p8(poly8x8_t __p0, poly8x8_t __p1) {
28979  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
28980  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
28981  uint8x8_t __ret;
28982  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
28983  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
28984  return __ret;
28985}
28986#endif
28987
28988#ifdef __LITTLE_ENDIAN__
28989__ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1) {
28990  uint16x4_t __ret;
28991  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
28992  return __ret;
28993}
28994#else
28995__ai uint16x4_t vtst_p16(poly16x4_t __p0, poly16x4_t __p1) {
28996  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
28997  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
28998  uint16x4_t __ret;
28999  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29000  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29001  return __ret;
29002}
29003#endif
29004
29005#ifdef __LITTLE_ENDIAN__
29006__ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29007  uint8x16_t __ret;
29008  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
29009  return __ret;
29010}
29011#else
29012__ai uint8x16_t vtstq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29013  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29014  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29015  uint8x16_t __ret;
29016  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29017  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29018  return __ret;
29019}
29020#endif
29021
29022#ifdef __LITTLE_ENDIAN__
29023__ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29024  uint16x8_t __ret;
29025  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
29026  return __ret;
29027}
29028#else
29029__ai uint16x8_t vtstq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29030  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29031  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29032  uint16x8_t __ret;
29033  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29034  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29035  return __ret;
29036}
29037#endif
29038
29039#ifdef __LITTLE_ENDIAN__
29040__ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29041  uint8x16_t __ret;
29042  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
29043  return __ret;
29044}
29045#else
29046__ai uint8x16_t vtstq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29047  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29048  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29049  uint8x16_t __ret;
29050  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29051  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29052  return __ret;
29053}
29054#endif
29055
29056#ifdef __LITTLE_ENDIAN__
29057__ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29058  uint32x4_t __ret;
29059  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
29060  return __ret;
29061}
29062#else
29063__ai uint32x4_t vtstq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29064  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29065  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29066  uint32x4_t __ret;
29067  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29068  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29069  return __ret;
29070}
29071#endif
29072
29073#ifdef __LITTLE_ENDIAN__
29074__ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29075  uint16x8_t __ret;
29076  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
29077  return __ret;
29078}
29079#else
29080__ai uint16x8_t vtstq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29081  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29082  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29083  uint16x8_t __ret;
29084  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29085  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29086  return __ret;
29087}
29088#endif
29089
29090#ifdef __LITTLE_ENDIAN__
29091__ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1) {
29092  uint8x16_t __ret;
29093  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
29094  return __ret;
29095}
29096#else
29097__ai uint8x16_t vtstq_s8(int8x16_t __p0, int8x16_t __p1) {
29098  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29099  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29100  uint8x16_t __ret;
29101  __ret = (uint8x16_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29102  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29103  return __ret;
29104}
29105#endif
29106
29107#ifdef __LITTLE_ENDIAN__
29108__ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1) {
29109  uint32x4_t __ret;
29110  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
29111  return __ret;
29112}
29113#else
29114__ai uint32x4_t vtstq_s32(int32x4_t __p0, int32x4_t __p1) {
29115  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29116  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29117  uint32x4_t __ret;
29118  __ret = (uint32x4_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29119  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29120  return __ret;
29121}
29122#endif
29123
29124#ifdef __LITTLE_ENDIAN__
29125__ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1) {
29126  uint16x8_t __ret;
29127  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
29128  return __ret;
29129}
29130#else
29131__ai uint16x8_t vtstq_s16(int16x8_t __p0, int16x8_t __p1) {
29132  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29133  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29134  uint16x8_t __ret;
29135  __ret = (uint16x8_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29136  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29137  return __ret;
29138}
29139#endif
29140
29141#ifdef __LITTLE_ENDIAN__
29142__ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1) {
29143  uint8x8_t __ret;
29144  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
29145  return __ret;
29146}
29147#else
29148__ai uint8x8_t vtst_u8(uint8x8_t __p0, uint8x8_t __p1) {
29149  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29150  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29151  uint8x8_t __ret;
29152  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29153  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29154  return __ret;
29155}
29156#endif
29157
29158#ifdef __LITTLE_ENDIAN__
29159__ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1) {
29160  uint32x2_t __ret;
29161  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
29162  return __ret;
29163}
29164#else
29165__ai uint32x2_t vtst_u32(uint32x2_t __p0, uint32x2_t __p1) {
29166  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29167  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29168  uint32x2_t __ret;
29169  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29170  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29171  return __ret;
29172}
29173#endif
29174
29175#ifdef __LITTLE_ENDIAN__
29176__ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1) {
29177  uint16x4_t __ret;
29178  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
29179  return __ret;
29180}
29181#else
29182__ai uint16x4_t vtst_u16(uint16x4_t __p0, uint16x4_t __p1) {
29183  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29184  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29185  uint16x4_t __ret;
29186  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29187  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29188  return __ret;
29189}
29190#endif
29191
29192#ifdef __LITTLE_ENDIAN__
29193__ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1) {
29194  uint8x8_t __ret;
29195  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
29196  return __ret;
29197}
29198#else
29199__ai uint8x8_t vtst_s8(int8x8_t __p0, int8x8_t __p1) {
29200  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29201  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29202  uint8x8_t __ret;
29203  __ret = (uint8x8_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29204  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
29205  return __ret;
29206}
29207#endif
29208
29209#ifdef __LITTLE_ENDIAN__
29210__ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1) {
29211  uint32x2_t __ret;
29212  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
29213  return __ret;
29214}
29215#else
29216__ai uint32x2_t vtst_s32(int32x2_t __p0, int32x2_t __p1) {
29217  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29218  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29219  uint32x2_t __ret;
29220  __ret = (uint32x2_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29221  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
29222  return __ret;
29223}
29224#endif
29225
29226#ifdef __LITTLE_ENDIAN__
29227__ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1) {
29228  uint16x4_t __ret;
29229  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
29230  return __ret;
29231}
29232#else
29233__ai uint16x4_t vtst_s16(int16x4_t __p0, int16x4_t __p1) {
29234  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29235  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29236  uint16x4_t __ret;
29237  __ret = (uint16x4_t) __builtin_neon_vtst_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29238  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
29239  return __ret;
29240}
29241#endif
29242
29243#ifdef __LITTLE_ENDIAN__
29244__ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1) {
29245  poly8x8x2_t __ret;
29246  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
29247  return __ret;
29248}
29249#else
29250__ai poly8x8x2_t vuzp_p8(poly8x8_t __p0, poly8x8_t __p1) {
29251  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29252  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29253  poly8x8x2_t __ret;
29254  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
29255
29256  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29257  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29258  return __ret;
29259}
29260#endif
29261
29262#ifdef __LITTLE_ENDIAN__
29263__ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1) {
29264  poly16x4x2_t __ret;
29265  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
29266  return __ret;
29267}
29268#else
29269__ai poly16x4x2_t vuzp_p16(poly16x4_t __p0, poly16x4_t __p1) {
29270  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29271  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29272  poly16x4x2_t __ret;
29273  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
29274
29275  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29276  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29277  return __ret;
29278}
29279#endif
29280
29281#ifdef __LITTLE_ENDIAN__
29282__ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29283  poly8x16x2_t __ret;
29284  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
29285  return __ret;
29286}
29287#else
29288__ai poly8x16x2_t vuzpq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29289  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29290  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29291  poly8x16x2_t __ret;
29292  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
29293
29294  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29295  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29296  return __ret;
29297}
29298#endif
29299
29300#ifdef __LITTLE_ENDIAN__
29301__ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29302  poly16x8x2_t __ret;
29303  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
29304  return __ret;
29305}
29306#else
29307__ai poly16x8x2_t vuzpq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29308  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29309  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29310  poly16x8x2_t __ret;
29311  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
29312
29313  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29314  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29315  return __ret;
29316}
29317#endif
29318
29319#ifdef __LITTLE_ENDIAN__
29320__ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29321  uint8x16x2_t __ret;
29322  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
29323  return __ret;
29324}
29325#else
29326__ai uint8x16x2_t vuzpq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29327  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29328  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29329  uint8x16x2_t __ret;
29330  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29331
29332  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29333  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29334  return __ret;
29335}
29336#endif
29337
29338#ifdef __LITTLE_ENDIAN__
29339__ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29340  uint32x4x2_t __ret;
29341  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
29342  return __ret;
29343}
29344#else
29345__ai uint32x4x2_t vuzpq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29346  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29347  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29348  uint32x4x2_t __ret;
29349  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29350
29351  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29352  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29353  return __ret;
29354}
29355#endif
29356
29357#ifdef __LITTLE_ENDIAN__
29358__ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29359  uint16x8x2_t __ret;
29360  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
29361  return __ret;
29362}
29363#else
29364__ai uint16x8x2_t vuzpq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29365  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29366  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29367  uint16x8x2_t __ret;
29368  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29369
29370  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29371  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29372  return __ret;
29373}
29374#endif
29375
29376#ifdef __LITTLE_ENDIAN__
29377__ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1) {
29378  int8x16x2_t __ret;
29379  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
29380  return __ret;
29381}
29382#else
29383__ai int8x16x2_t vuzpq_s8(int8x16_t __p0, int8x16_t __p1) {
29384  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29385  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29386  int8x16x2_t __ret;
29387  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
29388
29389  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29390  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29391  return __ret;
29392}
29393#endif
29394
29395#ifdef __LITTLE_ENDIAN__
29396__ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1) {
29397  float32x4x2_t __ret;
29398  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
29399  return __ret;
29400}
29401#else
29402__ai float32x4x2_t vuzpq_f32(float32x4_t __p0, float32x4_t __p1) {
29403  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29404  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29405  float32x4x2_t __ret;
29406  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
29407
29408  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29409  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29410  return __ret;
29411}
29412#endif
29413
29414#ifdef __LITTLE_ENDIAN__
29415__ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1) {
29416  int32x4x2_t __ret;
29417  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
29418  return __ret;
29419}
29420#else
29421__ai int32x4x2_t vuzpq_s32(int32x4_t __p0, int32x4_t __p1) {
29422  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29423  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29424  int32x4x2_t __ret;
29425  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
29426
29427  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29428  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29429  return __ret;
29430}
29431#endif
29432
29433#ifdef __LITTLE_ENDIAN__
29434__ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1) {
29435  int16x8x2_t __ret;
29436  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
29437  return __ret;
29438}
29439#else
29440__ai int16x8x2_t vuzpq_s16(int16x8_t __p0, int16x8_t __p1) {
29441  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29442  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29443  int16x8x2_t __ret;
29444  __builtin_neon_vuzpq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
29445
29446  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29447  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29448  return __ret;
29449}
29450#endif
29451
29452#ifdef __LITTLE_ENDIAN__
29453__ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1) {
29454  uint8x8x2_t __ret;
29455  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
29456  return __ret;
29457}
29458#else
29459__ai uint8x8x2_t vuzp_u8(uint8x8_t __p0, uint8x8_t __p1) {
29460  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29461  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29462  uint8x8x2_t __ret;
29463  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29464
29465  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29466  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29467  return __ret;
29468}
29469#endif
29470
29471#ifdef __LITTLE_ENDIAN__
29472__ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1) {
29473  uint32x2x2_t __ret;
29474  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
29475  return __ret;
29476}
29477#else
29478__ai uint32x2x2_t vuzp_u32(uint32x2_t __p0, uint32x2_t __p1) {
29479  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29480  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29481  uint32x2x2_t __ret;
29482  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29483
29484  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29485  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29486  return __ret;
29487}
29488#endif
29489
29490#ifdef __LITTLE_ENDIAN__
29491__ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1) {
29492  uint16x4x2_t __ret;
29493  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
29494  return __ret;
29495}
29496#else
29497__ai uint16x4x2_t vuzp_u16(uint16x4_t __p0, uint16x4_t __p1) {
29498  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29499  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29500  uint16x4x2_t __ret;
29501  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29502
29503  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29504  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29505  return __ret;
29506}
29507#endif
29508
29509#ifdef __LITTLE_ENDIAN__
29510__ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1) {
29511  int8x8x2_t __ret;
29512  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
29513  return __ret;
29514}
29515#else
29516__ai int8x8x2_t vuzp_s8(int8x8_t __p0, int8x8_t __p1) {
29517  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29518  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29519  int8x8x2_t __ret;
29520  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
29521
29522  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29523  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29524  return __ret;
29525}
29526#endif
29527
29528#ifdef __LITTLE_ENDIAN__
29529__ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1) {
29530  float32x2x2_t __ret;
29531  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
29532  return __ret;
29533}
29534#else
29535__ai float32x2x2_t vuzp_f32(float32x2_t __p0, float32x2_t __p1) {
29536  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29537  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29538  float32x2x2_t __ret;
29539  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
29540
29541  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29542  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29543  return __ret;
29544}
29545#endif
29546
29547#ifdef __LITTLE_ENDIAN__
29548__ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1) {
29549  int32x2x2_t __ret;
29550  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
29551  return __ret;
29552}
29553#else
29554__ai int32x2x2_t vuzp_s32(int32x2_t __p0, int32x2_t __p1) {
29555  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29556  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29557  int32x2x2_t __ret;
29558  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
29559
29560  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29561  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29562  return __ret;
29563}
29564#endif
29565
29566#ifdef __LITTLE_ENDIAN__
29567__ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1) {
29568  int16x4x2_t __ret;
29569  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
29570  return __ret;
29571}
29572#else
29573__ai int16x4x2_t vuzp_s16(int16x4_t __p0, int16x4_t __p1) {
29574  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29575  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29576  int16x4x2_t __ret;
29577  __builtin_neon_vuzp_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
29578
29579  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29580  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29581  return __ret;
29582}
29583#endif
29584
29585#ifdef __LITTLE_ENDIAN__
29586__ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1) {
29587  poly8x8x2_t __ret;
29588  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 4);
29589  return __ret;
29590}
29591#else
29592__ai poly8x8x2_t vzip_p8(poly8x8_t __p0, poly8x8_t __p1) {
29593  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29594  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29595  poly8x8x2_t __ret;
29596  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 4);
29597
29598  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29599  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29600  return __ret;
29601}
29602#endif
29603
29604#ifdef __LITTLE_ENDIAN__
29605__ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1) {
29606  poly16x4x2_t __ret;
29607  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 5);
29608  return __ret;
29609}
29610#else
29611__ai poly16x4x2_t vzip_p16(poly16x4_t __p0, poly16x4_t __p1) {
29612  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29613  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29614  poly16x4x2_t __ret;
29615  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 5);
29616
29617  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29618  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29619  return __ret;
29620}
29621#endif
29622
29623#ifdef __LITTLE_ENDIAN__
29624__ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29625  poly8x16x2_t __ret;
29626  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 36);
29627  return __ret;
29628}
29629#else
29630__ai poly8x16x2_t vzipq_p8(poly8x16_t __p0, poly8x16_t __p1) {
29631  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29632  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29633  poly8x16x2_t __ret;
29634  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 36);
29635
29636  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29637  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29638  return __ret;
29639}
29640#endif
29641
29642#ifdef __LITTLE_ENDIAN__
29643__ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29644  poly16x8x2_t __ret;
29645  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 37);
29646  return __ret;
29647}
29648#else
29649__ai poly16x8x2_t vzipq_p16(poly16x8_t __p0, poly16x8_t __p1) {
29650  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29651  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29652  poly16x8x2_t __ret;
29653  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 37);
29654
29655  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29656  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29657  return __ret;
29658}
29659#endif
29660
29661#ifdef __LITTLE_ENDIAN__
29662__ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29663  uint8x16x2_t __ret;
29664  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 48);
29665  return __ret;
29666}
29667#else
29668__ai uint8x16x2_t vzipq_u8(uint8x16_t __p0, uint8x16_t __p1) {
29669  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29670  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29671  uint8x16x2_t __ret;
29672  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 48);
29673
29674  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29675  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29676  return __ret;
29677}
29678#endif
29679
29680#ifdef __LITTLE_ENDIAN__
29681__ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29682  uint32x4x2_t __ret;
29683  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 50);
29684  return __ret;
29685}
29686#else
29687__ai uint32x4x2_t vzipq_u32(uint32x4_t __p0, uint32x4_t __p1) {
29688  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29689  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29690  uint32x4x2_t __ret;
29691  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 50);
29692
29693  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29694  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29695  return __ret;
29696}
29697#endif
29698
29699#ifdef __LITTLE_ENDIAN__
29700__ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29701  uint16x8x2_t __ret;
29702  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 49);
29703  return __ret;
29704}
29705#else
29706__ai uint16x8x2_t vzipq_u16(uint16x8_t __p0, uint16x8_t __p1) {
29707  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29708  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29709  uint16x8x2_t __ret;
29710  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 49);
29711
29712  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29713  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29714  return __ret;
29715}
29716#endif
29717
29718#ifdef __LITTLE_ENDIAN__
29719__ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1) {
29720  int8x16x2_t __ret;
29721  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 32);
29722  return __ret;
29723}
29724#else
29725__ai int8x16x2_t vzipq_s8(int8x16_t __p0, int8x16_t __p1) {
29726  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29727  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29728  int8x16x2_t __ret;
29729  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 32);
29730
29731  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29732  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
29733  return __ret;
29734}
29735#endif
29736
29737#ifdef __LITTLE_ENDIAN__
29738__ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1) {
29739  float32x4x2_t __ret;
29740  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 41);
29741  return __ret;
29742}
29743#else
29744__ai float32x4x2_t vzipq_f32(float32x4_t __p0, float32x4_t __p1) {
29745  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29746  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29747  float32x4x2_t __ret;
29748  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 41);
29749
29750  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29751  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29752  return __ret;
29753}
29754#endif
29755
29756#ifdef __LITTLE_ENDIAN__
29757__ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1) {
29758  int32x4x2_t __ret;
29759  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 34);
29760  return __ret;
29761}
29762#else
29763__ai int32x4x2_t vzipq_s32(int32x4_t __p0, int32x4_t __p1) {
29764  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29765  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29766  int32x4x2_t __ret;
29767  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 34);
29768
29769  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29770  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29771  return __ret;
29772}
29773#endif
29774
29775#ifdef __LITTLE_ENDIAN__
29776__ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1) {
29777  int16x8x2_t __ret;
29778  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__p0, (int8x16_t)__p1, 33);
29779  return __ret;
29780}
29781#else
29782__ai int16x8x2_t vzipq_s16(int16x8_t __p0, int16x8_t __p1) {
29783  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29784  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29785  int16x8x2_t __ret;
29786  __builtin_neon_vzipq_v(&__ret, (int8x16_t)__rev0, (int8x16_t)__rev1, 33);
29787
29788  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29789  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29790  return __ret;
29791}
29792#endif
29793
29794#ifdef __LITTLE_ENDIAN__
29795__ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1) {
29796  uint8x8x2_t __ret;
29797  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 16);
29798  return __ret;
29799}
29800#else
29801__ai uint8x8x2_t vzip_u8(uint8x8_t __p0, uint8x8_t __p1) {
29802  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29803  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29804  uint8x8x2_t __ret;
29805  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 16);
29806
29807  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29808  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29809  return __ret;
29810}
29811#endif
29812
29813#ifdef __LITTLE_ENDIAN__
29814__ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1) {
29815  uint32x2x2_t __ret;
29816  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 18);
29817  return __ret;
29818}
29819#else
29820__ai uint32x2x2_t vzip_u32(uint32x2_t __p0, uint32x2_t __p1) {
29821  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29822  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29823  uint32x2x2_t __ret;
29824  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 18);
29825
29826  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29827  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29828  return __ret;
29829}
29830#endif
29831
29832#ifdef __LITTLE_ENDIAN__
29833__ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1) {
29834  uint16x4x2_t __ret;
29835  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 17);
29836  return __ret;
29837}
29838#else
29839__ai uint16x4x2_t vzip_u16(uint16x4_t __p0, uint16x4_t __p1) {
29840  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29841  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29842  uint16x4x2_t __ret;
29843  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 17);
29844
29845  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29846  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29847  return __ret;
29848}
29849#endif
29850
29851#ifdef __LITTLE_ENDIAN__
29852__ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1) {
29853  int8x8x2_t __ret;
29854  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 0);
29855  return __ret;
29856}
29857#else
29858__ai int8x8x2_t vzip_s8(int8x8_t __p0, int8x8_t __p1) {
29859  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
29860  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
29861  int8x8x2_t __ret;
29862  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 0);
29863
29864  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0);
29865  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0);
29866  return __ret;
29867}
29868#endif
29869
29870#ifdef __LITTLE_ENDIAN__
29871__ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1) {
29872  float32x2x2_t __ret;
29873  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 9);
29874  return __ret;
29875}
29876#else
29877__ai float32x2x2_t vzip_f32(float32x2_t __p0, float32x2_t __p1) {
29878  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29879  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29880  float32x2x2_t __ret;
29881  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 9);
29882
29883  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29884  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29885  return __ret;
29886}
29887#endif
29888
29889#ifdef __LITTLE_ENDIAN__
29890__ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1) {
29891  int32x2x2_t __ret;
29892  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 2);
29893  return __ret;
29894}
29895#else
29896__ai int32x2x2_t vzip_s32(int32x2_t __p0, int32x2_t __p1) {
29897  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
29898  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
29899  int32x2x2_t __ret;
29900  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 2);
29901
29902  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0);
29903  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0);
29904  return __ret;
29905}
29906#endif
29907
29908#ifdef __LITTLE_ENDIAN__
29909__ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1) {
29910  int16x4x2_t __ret;
29911  __builtin_neon_vzip_v(&__ret, (int8x8_t)__p0, (int8x8_t)__p1, 1);
29912  return __ret;
29913}
29914#else
29915__ai int16x4x2_t vzip_s16(int16x4_t __p0, int16x4_t __p1) {
29916  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
29917  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
29918  int16x4x2_t __ret;
29919  __builtin_neon_vzip_v(&__ret, (int8x8_t)__rev0, (int8x8_t)__rev1, 1);
29920
29921  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0);
29922  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0);
29923  return __ret;
29924}
29925#endif
29926
29927#if !defined(__aarch64__)
29928#ifdef __LITTLE_ENDIAN__
29929__ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
29930  poly8x8_t __ret;
29931  __ret = (poly8x8_t)(__p0);
29932  return __ret;
29933}
29934#else
29935__ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
29936  poly8x8_t __ret;
29937  __ret = (poly8x8_t)(__p0);
29938  return __ret;
29939}
29940#endif
29941
29942#ifdef __LITTLE_ENDIAN__
29943__ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
29944  poly8x8_t __ret;
29945  __ret = (poly8x8_t)(__p0);
29946  return __ret;
29947}
29948#else
29949__ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
29950  poly8x8_t __ret;
29951  __ret = (poly8x8_t)(__p0);
29952  return __ret;
29953}
29954#endif
29955
29956#ifdef __LITTLE_ENDIAN__
29957__ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
29958  poly8x8_t __ret;
29959  __ret = (poly8x8_t)(__p0);
29960  return __ret;
29961}
29962#else
29963__ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
29964  poly8x8_t __ret;
29965  __ret = (poly8x8_t)(__p0);
29966  return __ret;
29967}
29968#endif
29969
29970#ifdef __LITTLE_ENDIAN__
29971__ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
29972  poly8x8_t __ret;
29973  __ret = (poly8x8_t)(__p0);
29974  return __ret;
29975}
29976#else
29977__ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
29978  poly8x8_t __ret;
29979  __ret = (poly8x8_t)(__p0);
29980  return __ret;
29981}
29982#endif
29983
29984#ifdef __LITTLE_ENDIAN__
29985__ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
29986  poly8x8_t __ret;
29987  __ret = (poly8x8_t)(__p0);
29988  return __ret;
29989}
29990#else
29991__ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
29992  poly8x8_t __ret;
29993  __ret = (poly8x8_t)(__p0);
29994  return __ret;
29995}
29996#endif
29997
29998#ifdef __LITTLE_ENDIAN__
29999__ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
30000  poly8x8_t __ret;
30001  __ret = (poly8x8_t)(__p0);
30002  return __ret;
30003}
30004#else
30005__ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
30006  poly8x8_t __ret;
30007  __ret = (poly8x8_t)(__p0);
30008  return __ret;
30009}
30010#endif
30011
30012#ifdef __LITTLE_ENDIAN__
30013__ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
30014  poly8x8_t __ret;
30015  __ret = (poly8x8_t)(__p0);
30016  return __ret;
30017}
30018#else
30019__ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
30020  poly8x8_t __ret;
30021  __ret = (poly8x8_t)(__p0);
30022  return __ret;
30023}
30024#endif
30025
30026#ifdef __LITTLE_ENDIAN__
30027__ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
30028  poly8x8_t __ret;
30029  __ret = (poly8x8_t)(__p0);
30030  return __ret;
30031}
30032#else
30033__ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
30034  poly8x8_t __ret;
30035  __ret = (poly8x8_t)(__p0);
30036  return __ret;
30037}
30038#endif
30039
30040#ifdef __LITTLE_ENDIAN__
30041__ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
30042  poly8x8_t __ret;
30043  __ret = (poly8x8_t)(__p0);
30044  return __ret;
30045}
30046#else
30047__ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
30048  poly8x8_t __ret;
30049  __ret = (poly8x8_t)(__p0);
30050  return __ret;
30051}
30052#endif
30053
30054#ifdef __LITTLE_ENDIAN__
30055__ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
30056  poly8x8_t __ret;
30057  __ret = (poly8x8_t)(__p0);
30058  return __ret;
30059}
30060#else
30061__ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
30062  poly8x8_t __ret;
30063  __ret = (poly8x8_t)(__p0);
30064  return __ret;
30065}
30066#endif
30067
30068#ifdef __LITTLE_ENDIAN__
30069__ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
30070  poly8x8_t __ret;
30071  __ret = (poly8x8_t)(__p0);
30072  return __ret;
30073}
30074#else
30075__ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
30076  poly8x8_t __ret;
30077  __ret = (poly8x8_t)(__p0);
30078  return __ret;
30079}
30080#endif
30081
30082#ifdef __LITTLE_ENDIAN__
30083__ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
30084  poly16x4_t __ret;
30085  __ret = (poly16x4_t)(__p0);
30086  return __ret;
30087}
30088#else
30089__ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
30090  poly16x4_t __ret;
30091  __ret = (poly16x4_t)(__p0);
30092  return __ret;
30093}
30094#endif
30095
30096#ifdef __LITTLE_ENDIAN__
30097__ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
30098  poly16x4_t __ret;
30099  __ret = (poly16x4_t)(__p0);
30100  return __ret;
30101}
30102#else
30103__ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
30104  poly16x4_t __ret;
30105  __ret = (poly16x4_t)(__p0);
30106  return __ret;
30107}
30108#endif
30109
30110#ifdef __LITTLE_ENDIAN__
30111__ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
30112  poly16x4_t __ret;
30113  __ret = (poly16x4_t)(__p0);
30114  return __ret;
30115}
30116#else
30117__ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
30118  poly16x4_t __ret;
30119  __ret = (poly16x4_t)(__p0);
30120  return __ret;
30121}
30122#endif
30123
30124#ifdef __LITTLE_ENDIAN__
30125__ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
30126  poly16x4_t __ret;
30127  __ret = (poly16x4_t)(__p0);
30128  return __ret;
30129}
30130#else
30131__ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
30132  poly16x4_t __ret;
30133  __ret = (poly16x4_t)(__p0);
30134  return __ret;
30135}
30136#endif
30137
30138#ifdef __LITTLE_ENDIAN__
30139__ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
30140  poly16x4_t __ret;
30141  __ret = (poly16x4_t)(__p0);
30142  return __ret;
30143}
30144#else
30145__ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
30146  poly16x4_t __ret;
30147  __ret = (poly16x4_t)(__p0);
30148  return __ret;
30149}
30150#endif
30151
30152#ifdef __LITTLE_ENDIAN__
30153__ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
30154  poly16x4_t __ret;
30155  __ret = (poly16x4_t)(__p0);
30156  return __ret;
30157}
30158#else
30159__ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
30160  poly16x4_t __ret;
30161  __ret = (poly16x4_t)(__p0);
30162  return __ret;
30163}
30164#endif
30165
30166#ifdef __LITTLE_ENDIAN__
30167__ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
30168  poly16x4_t __ret;
30169  __ret = (poly16x4_t)(__p0);
30170  return __ret;
30171}
30172#else
30173__ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
30174  poly16x4_t __ret;
30175  __ret = (poly16x4_t)(__p0);
30176  return __ret;
30177}
30178#endif
30179
30180#ifdef __LITTLE_ENDIAN__
30181__ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
30182  poly16x4_t __ret;
30183  __ret = (poly16x4_t)(__p0);
30184  return __ret;
30185}
30186#else
30187__ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
30188  poly16x4_t __ret;
30189  __ret = (poly16x4_t)(__p0);
30190  return __ret;
30191}
30192#endif
30193
30194#ifdef __LITTLE_ENDIAN__
30195__ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
30196  poly16x4_t __ret;
30197  __ret = (poly16x4_t)(__p0);
30198  return __ret;
30199}
30200#else
30201__ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
30202  poly16x4_t __ret;
30203  __ret = (poly16x4_t)(__p0);
30204  return __ret;
30205}
30206#endif
30207
30208#ifdef __LITTLE_ENDIAN__
30209__ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
30210  poly16x4_t __ret;
30211  __ret = (poly16x4_t)(__p0);
30212  return __ret;
30213}
30214#else
30215__ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
30216  poly16x4_t __ret;
30217  __ret = (poly16x4_t)(__p0);
30218  return __ret;
30219}
30220#endif
30221
30222#ifdef __LITTLE_ENDIAN__
30223__ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
30224  poly16x4_t __ret;
30225  __ret = (poly16x4_t)(__p0);
30226  return __ret;
30227}
30228#else
30229__ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
30230  poly16x4_t __ret;
30231  __ret = (poly16x4_t)(__p0);
30232  return __ret;
30233}
30234#endif
30235
30236#ifdef __LITTLE_ENDIAN__
30237__ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
30238  poly8x16_t __ret;
30239  __ret = (poly8x16_t)(__p0);
30240  return __ret;
30241}
30242#else
30243__ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
30244  poly8x16_t __ret;
30245  __ret = (poly8x16_t)(__p0);
30246  return __ret;
30247}
30248#endif
30249
30250#ifdef __LITTLE_ENDIAN__
30251__ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
30252  poly8x16_t __ret;
30253  __ret = (poly8x16_t)(__p0);
30254  return __ret;
30255}
30256#else
30257__ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
30258  poly8x16_t __ret;
30259  __ret = (poly8x16_t)(__p0);
30260  return __ret;
30261}
30262#endif
30263
30264#ifdef __LITTLE_ENDIAN__
30265__ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
30266  poly8x16_t __ret;
30267  __ret = (poly8x16_t)(__p0);
30268  return __ret;
30269}
30270#else
30271__ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
30272  poly8x16_t __ret;
30273  __ret = (poly8x16_t)(__p0);
30274  return __ret;
30275}
30276#endif
30277
30278#ifdef __LITTLE_ENDIAN__
30279__ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
30280  poly8x16_t __ret;
30281  __ret = (poly8x16_t)(__p0);
30282  return __ret;
30283}
30284#else
30285__ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
30286  poly8x16_t __ret;
30287  __ret = (poly8x16_t)(__p0);
30288  return __ret;
30289}
30290#endif
30291
30292#ifdef __LITTLE_ENDIAN__
30293__ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
30294  poly8x16_t __ret;
30295  __ret = (poly8x16_t)(__p0);
30296  return __ret;
30297}
30298#else
30299__ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
30300  poly8x16_t __ret;
30301  __ret = (poly8x16_t)(__p0);
30302  return __ret;
30303}
30304#endif
30305
30306#ifdef __LITTLE_ENDIAN__
30307__ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
30308  poly8x16_t __ret;
30309  __ret = (poly8x16_t)(__p0);
30310  return __ret;
30311}
30312#else
30313__ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
30314  poly8x16_t __ret;
30315  __ret = (poly8x16_t)(__p0);
30316  return __ret;
30317}
30318#endif
30319
30320#ifdef __LITTLE_ENDIAN__
30321__ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
30322  poly8x16_t __ret;
30323  __ret = (poly8x16_t)(__p0);
30324  return __ret;
30325}
30326#else
30327__ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
30328  poly8x16_t __ret;
30329  __ret = (poly8x16_t)(__p0);
30330  return __ret;
30331}
30332#endif
30333
30334#ifdef __LITTLE_ENDIAN__
30335__ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
30336  poly8x16_t __ret;
30337  __ret = (poly8x16_t)(__p0);
30338  return __ret;
30339}
30340#else
30341__ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
30342  poly8x16_t __ret;
30343  __ret = (poly8x16_t)(__p0);
30344  return __ret;
30345}
30346#endif
30347
30348#ifdef __LITTLE_ENDIAN__
30349__ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
30350  poly8x16_t __ret;
30351  __ret = (poly8x16_t)(__p0);
30352  return __ret;
30353}
30354#else
30355__ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
30356  poly8x16_t __ret;
30357  __ret = (poly8x16_t)(__p0);
30358  return __ret;
30359}
30360#endif
30361
30362#ifdef __LITTLE_ENDIAN__
30363__ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
30364  poly8x16_t __ret;
30365  __ret = (poly8x16_t)(__p0);
30366  return __ret;
30367}
30368#else
30369__ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
30370  poly8x16_t __ret;
30371  __ret = (poly8x16_t)(__p0);
30372  return __ret;
30373}
30374#endif
30375
30376#ifdef __LITTLE_ENDIAN__
30377__ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
30378  poly8x16_t __ret;
30379  __ret = (poly8x16_t)(__p0);
30380  return __ret;
30381}
30382#else
30383__ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
30384  poly8x16_t __ret;
30385  __ret = (poly8x16_t)(__p0);
30386  return __ret;
30387}
30388#endif
30389
30390#ifdef __LITTLE_ENDIAN__
30391__ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
30392  poly16x8_t __ret;
30393  __ret = (poly16x8_t)(__p0);
30394  return __ret;
30395}
30396#else
30397__ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
30398  poly16x8_t __ret;
30399  __ret = (poly16x8_t)(__p0);
30400  return __ret;
30401}
30402#endif
30403
30404#ifdef __LITTLE_ENDIAN__
30405__ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
30406  poly16x8_t __ret;
30407  __ret = (poly16x8_t)(__p0);
30408  return __ret;
30409}
30410#else
30411__ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
30412  poly16x8_t __ret;
30413  __ret = (poly16x8_t)(__p0);
30414  return __ret;
30415}
30416#endif
30417
30418#ifdef __LITTLE_ENDIAN__
30419__ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
30420  poly16x8_t __ret;
30421  __ret = (poly16x8_t)(__p0);
30422  return __ret;
30423}
30424#else
30425__ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
30426  poly16x8_t __ret;
30427  __ret = (poly16x8_t)(__p0);
30428  return __ret;
30429}
30430#endif
30431
30432#ifdef __LITTLE_ENDIAN__
30433__ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
30434  poly16x8_t __ret;
30435  __ret = (poly16x8_t)(__p0);
30436  return __ret;
30437}
30438#else
30439__ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
30440  poly16x8_t __ret;
30441  __ret = (poly16x8_t)(__p0);
30442  return __ret;
30443}
30444#endif
30445
30446#ifdef __LITTLE_ENDIAN__
30447__ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
30448  poly16x8_t __ret;
30449  __ret = (poly16x8_t)(__p0);
30450  return __ret;
30451}
30452#else
30453__ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
30454  poly16x8_t __ret;
30455  __ret = (poly16x8_t)(__p0);
30456  return __ret;
30457}
30458#endif
30459
30460#ifdef __LITTLE_ENDIAN__
30461__ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
30462  poly16x8_t __ret;
30463  __ret = (poly16x8_t)(__p0);
30464  return __ret;
30465}
30466#else
30467__ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
30468  poly16x8_t __ret;
30469  __ret = (poly16x8_t)(__p0);
30470  return __ret;
30471}
30472#endif
30473
30474#ifdef __LITTLE_ENDIAN__
30475__ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
30476  poly16x8_t __ret;
30477  __ret = (poly16x8_t)(__p0);
30478  return __ret;
30479}
30480#else
30481__ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
30482  poly16x8_t __ret;
30483  __ret = (poly16x8_t)(__p0);
30484  return __ret;
30485}
30486#endif
30487
30488#ifdef __LITTLE_ENDIAN__
30489__ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
30490  poly16x8_t __ret;
30491  __ret = (poly16x8_t)(__p0);
30492  return __ret;
30493}
30494#else
30495__ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
30496  poly16x8_t __ret;
30497  __ret = (poly16x8_t)(__p0);
30498  return __ret;
30499}
30500#endif
30501
30502#ifdef __LITTLE_ENDIAN__
30503__ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
30504  poly16x8_t __ret;
30505  __ret = (poly16x8_t)(__p0);
30506  return __ret;
30507}
30508#else
30509__ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
30510  poly16x8_t __ret;
30511  __ret = (poly16x8_t)(__p0);
30512  return __ret;
30513}
30514#endif
30515
30516#ifdef __LITTLE_ENDIAN__
30517__ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
30518  poly16x8_t __ret;
30519  __ret = (poly16x8_t)(__p0);
30520  return __ret;
30521}
30522#else
30523__ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
30524  poly16x8_t __ret;
30525  __ret = (poly16x8_t)(__p0);
30526  return __ret;
30527}
30528#endif
30529
30530#ifdef __LITTLE_ENDIAN__
30531__ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
30532  poly16x8_t __ret;
30533  __ret = (poly16x8_t)(__p0);
30534  return __ret;
30535}
30536#else
30537__ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
30538  poly16x8_t __ret;
30539  __ret = (poly16x8_t)(__p0);
30540  return __ret;
30541}
30542#endif
30543
30544#ifdef __LITTLE_ENDIAN__
30545__ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
30546  uint8x16_t __ret;
30547  __ret = (uint8x16_t)(__p0);
30548  return __ret;
30549}
30550#else
30551__ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
30552  uint8x16_t __ret;
30553  __ret = (uint8x16_t)(__p0);
30554  return __ret;
30555}
30556#endif
30557
30558#ifdef __LITTLE_ENDIAN__
30559__ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
30560  uint8x16_t __ret;
30561  __ret = (uint8x16_t)(__p0);
30562  return __ret;
30563}
30564#else
30565__ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
30566  uint8x16_t __ret;
30567  __ret = (uint8x16_t)(__p0);
30568  return __ret;
30569}
30570#endif
30571
30572#ifdef __LITTLE_ENDIAN__
30573__ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
30574  uint8x16_t __ret;
30575  __ret = (uint8x16_t)(__p0);
30576  return __ret;
30577}
30578#else
30579__ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
30580  uint8x16_t __ret;
30581  __ret = (uint8x16_t)(__p0);
30582  return __ret;
30583}
30584#endif
30585
30586#ifdef __LITTLE_ENDIAN__
30587__ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
30588  uint8x16_t __ret;
30589  __ret = (uint8x16_t)(__p0);
30590  return __ret;
30591}
30592#else
30593__ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
30594  uint8x16_t __ret;
30595  __ret = (uint8x16_t)(__p0);
30596  return __ret;
30597}
30598#endif
30599
30600#ifdef __LITTLE_ENDIAN__
30601__ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
30602  uint8x16_t __ret;
30603  __ret = (uint8x16_t)(__p0);
30604  return __ret;
30605}
30606#else
30607__ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
30608  uint8x16_t __ret;
30609  __ret = (uint8x16_t)(__p0);
30610  return __ret;
30611}
30612#endif
30613
30614#ifdef __LITTLE_ENDIAN__
30615__ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
30616  uint8x16_t __ret;
30617  __ret = (uint8x16_t)(__p0);
30618  return __ret;
30619}
30620#else
30621__ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
30622  uint8x16_t __ret;
30623  __ret = (uint8x16_t)(__p0);
30624  return __ret;
30625}
30626#endif
30627
30628#ifdef __LITTLE_ENDIAN__
30629__ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
30630  uint8x16_t __ret;
30631  __ret = (uint8x16_t)(__p0);
30632  return __ret;
30633}
30634#else
30635__ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
30636  uint8x16_t __ret;
30637  __ret = (uint8x16_t)(__p0);
30638  return __ret;
30639}
30640#endif
30641
30642#ifdef __LITTLE_ENDIAN__
30643__ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
30644  uint8x16_t __ret;
30645  __ret = (uint8x16_t)(__p0);
30646  return __ret;
30647}
30648#else
30649__ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
30650  uint8x16_t __ret;
30651  __ret = (uint8x16_t)(__p0);
30652  return __ret;
30653}
30654#endif
30655
30656#ifdef __LITTLE_ENDIAN__
30657__ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
30658  uint8x16_t __ret;
30659  __ret = (uint8x16_t)(__p0);
30660  return __ret;
30661}
30662#else
30663__ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
30664  uint8x16_t __ret;
30665  __ret = (uint8x16_t)(__p0);
30666  return __ret;
30667}
30668#endif
30669
30670#ifdef __LITTLE_ENDIAN__
30671__ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
30672  uint8x16_t __ret;
30673  __ret = (uint8x16_t)(__p0);
30674  return __ret;
30675}
30676#else
30677__ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
30678  uint8x16_t __ret;
30679  __ret = (uint8x16_t)(__p0);
30680  return __ret;
30681}
30682#endif
30683
30684#ifdef __LITTLE_ENDIAN__
30685__ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
30686  uint8x16_t __ret;
30687  __ret = (uint8x16_t)(__p0);
30688  return __ret;
30689}
30690#else
30691__ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
30692  uint8x16_t __ret;
30693  __ret = (uint8x16_t)(__p0);
30694  return __ret;
30695}
30696#endif
30697
30698#ifdef __LITTLE_ENDIAN__
30699__ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
30700  uint32x4_t __ret;
30701  __ret = (uint32x4_t)(__p0);
30702  return __ret;
30703}
30704#else
30705__ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
30706  uint32x4_t __ret;
30707  __ret = (uint32x4_t)(__p0);
30708  return __ret;
30709}
30710#endif
30711
30712#ifdef __LITTLE_ENDIAN__
30713__ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
30714  uint32x4_t __ret;
30715  __ret = (uint32x4_t)(__p0);
30716  return __ret;
30717}
30718#else
30719__ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
30720  uint32x4_t __ret;
30721  __ret = (uint32x4_t)(__p0);
30722  return __ret;
30723}
30724#endif
30725
30726#ifdef __LITTLE_ENDIAN__
30727__ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
30728  uint32x4_t __ret;
30729  __ret = (uint32x4_t)(__p0);
30730  return __ret;
30731}
30732#else
30733__ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
30734  uint32x4_t __ret;
30735  __ret = (uint32x4_t)(__p0);
30736  return __ret;
30737}
30738#endif
30739
30740#ifdef __LITTLE_ENDIAN__
30741__ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
30742  uint32x4_t __ret;
30743  __ret = (uint32x4_t)(__p0);
30744  return __ret;
30745}
30746#else
30747__ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
30748  uint32x4_t __ret;
30749  __ret = (uint32x4_t)(__p0);
30750  return __ret;
30751}
30752#endif
30753
30754#ifdef __LITTLE_ENDIAN__
30755__ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
30756  uint32x4_t __ret;
30757  __ret = (uint32x4_t)(__p0);
30758  return __ret;
30759}
30760#else
30761__ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
30762  uint32x4_t __ret;
30763  __ret = (uint32x4_t)(__p0);
30764  return __ret;
30765}
30766#endif
30767
30768#ifdef __LITTLE_ENDIAN__
30769__ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
30770  uint32x4_t __ret;
30771  __ret = (uint32x4_t)(__p0);
30772  return __ret;
30773}
30774#else
30775__ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
30776  uint32x4_t __ret;
30777  __ret = (uint32x4_t)(__p0);
30778  return __ret;
30779}
30780#endif
30781
30782#ifdef __LITTLE_ENDIAN__
30783__ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
30784  uint32x4_t __ret;
30785  __ret = (uint32x4_t)(__p0);
30786  return __ret;
30787}
30788#else
30789__ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
30790  uint32x4_t __ret;
30791  __ret = (uint32x4_t)(__p0);
30792  return __ret;
30793}
30794#endif
30795
30796#ifdef __LITTLE_ENDIAN__
30797__ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
30798  uint32x4_t __ret;
30799  __ret = (uint32x4_t)(__p0);
30800  return __ret;
30801}
30802#else
30803__ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
30804  uint32x4_t __ret;
30805  __ret = (uint32x4_t)(__p0);
30806  return __ret;
30807}
30808#endif
30809
30810#ifdef __LITTLE_ENDIAN__
30811__ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
30812  uint32x4_t __ret;
30813  __ret = (uint32x4_t)(__p0);
30814  return __ret;
30815}
30816#else
30817__ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
30818  uint32x4_t __ret;
30819  __ret = (uint32x4_t)(__p0);
30820  return __ret;
30821}
30822#endif
30823
30824#ifdef __LITTLE_ENDIAN__
30825__ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
30826  uint32x4_t __ret;
30827  __ret = (uint32x4_t)(__p0);
30828  return __ret;
30829}
30830#else
30831__ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
30832  uint32x4_t __ret;
30833  __ret = (uint32x4_t)(__p0);
30834  return __ret;
30835}
30836#endif
30837
30838#ifdef __LITTLE_ENDIAN__
30839__ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
30840  uint32x4_t __ret;
30841  __ret = (uint32x4_t)(__p0);
30842  return __ret;
30843}
30844#else
30845__ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
30846  uint32x4_t __ret;
30847  __ret = (uint32x4_t)(__p0);
30848  return __ret;
30849}
30850#endif
30851
30852#ifdef __LITTLE_ENDIAN__
30853__ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
30854  uint64x2_t __ret;
30855  __ret = (uint64x2_t)(__p0);
30856  return __ret;
30857}
30858#else
30859__ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
30860  uint64x2_t __ret;
30861  __ret = (uint64x2_t)(__p0);
30862  return __ret;
30863}
30864#endif
30865
30866#ifdef __LITTLE_ENDIAN__
30867__ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
30868  uint64x2_t __ret;
30869  __ret = (uint64x2_t)(__p0);
30870  return __ret;
30871}
30872#else
30873__ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
30874  uint64x2_t __ret;
30875  __ret = (uint64x2_t)(__p0);
30876  return __ret;
30877}
30878#endif
30879
30880#ifdef __LITTLE_ENDIAN__
30881__ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
30882  uint64x2_t __ret;
30883  __ret = (uint64x2_t)(__p0);
30884  return __ret;
30885}
30886#else
30887__ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
30888  uint64x2_t __ret;
30889  __ret = (uint64x2_t)(__p0);
30890  return __ret;
30891}
30892#endif
30893
30894#ifdef __LITTLE_ENDIAN__
30895__ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
30896  uint64x2_t __ret;
30897  __ret = (uint64x2_t)(__p0);
30898  return __ret;
30899}
30900#else
30901__ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
30902  uint64x2_t __ret;
30903  __ret = (uint64x2_t)(__p0);
30904  return __ret;
30905}
30906#endif
30907
30908#ifdef __LITTLE_ENDIAN__
30909__ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
30910  uint64x2_t __ret;
30911  __ret = (uint64x2_t)(__p0);
30912  return __ret;
30913}
30914#else
30915__ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
30916  uint64x2_t __ret;
30917  __ret = (uint64x2_t)(__p0);
30918  return __ret;
30919}
30920#endif
30921
30922#ifdef __LITTLE_ENDIAN__
30923__ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
30924  uint64x2_t __ret;
30925  __ret = (uint64x2_t)(__p0);
30926  return __ret;
30927}
30928#else
30929__ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
30930  uint64x2_t __ret;
30931  __ret = (uint64x2_t)(__p0);
30932  return __ret;
30933}
30934#endif
30935
30936#ifdef __LITTLE_ENDIAN__
30937__ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
30938  uint64x2_t __ret;
30939  __ret = (uint64x2_t)(__p0);
30940  return __ret;
30941}
30942#else
30943__ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
30944  uint64x2_t __ret;
30945  __ret = (uint64x2_t)(__p0);
30946  return __ret;
30947}
30948#endif
30949
30950#ifdef __LITTLE_ENDIAN__
30951__ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
30952  uint64x2_t __ret;
30953  __ret = (uint64x2_t)(__p0);
30954  return __ret;
30955}
30956#else
30957__ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
30958  uint64x2_t __ret;
30959  __ret = (uint64x2_t)(__p0);
30960  return __ret;
30961}
30962#endif
30963
30964#ifdef __LITTLE_ENDIAN__
30965__ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
30966  uint64x2_t __ret;
30967  __ret = (uint64x2_t)(__p0);
30968  return __ret;
30969}
30970#else
30971__ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
30972  uint64x2_t __ret;
30973  __ret = (uint64x2_t)(__p0);
30974  return __ret;
30975}
30976#endif
30977
30978#ifdef __LITTLE_ENDIAN__
30979__ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
30980  uint64x2_t __ret;
30981  __ret = (uint64x2_t)(__p0);
30982  return __ret;
30983}
30984#else
30985__ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
30986  uint64x2_t __ret;
30987  __ret = (uint64x2_t)(__p0);
30988  return __ret;
30989}
30990#endif
30991
30992#ifdef __LITTLE_ENDIAN__
30993__ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
30994  uint64x2_t __ret;
30995  __ret = (uint64x2_t)(__p0);
30996  return __ret;
30997}
30998#else
30999__ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
31000  uint64x2_t __ret;
31001  __ret = (uint64x2_t)(__p0);
31002  return __ret;
31003}
31004#endif
31005
31006#ifdef __LITTLE_ENDIAN__
31007__ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
31008  uint16x8_t __ret;
31009  __ret = (uint16x8_t)(__p0);
31010  return __ret;
31011}
31012#else
31013__ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
31014  uint16x8_t __ret;
31015  __ret = (uint16x8_t)(__p0);
31016  return __ret;
31017}
31018#endif
31019
31020#ifdef __LITTLE_ENDIAN__
31021__ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
31022  uint16x8_t __ret;
31023  __ret = (uint16x8_t)(__p0);
31024  return __ret;
31025}
31026#else
31027__ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
31028  uint16x8_t __ret;
31029  __ret = (uint16x8_t)(__p0);
31030  return __ret;
31031}
31032#endif
31033
31034#ifdef __LITTLE_ENDIAN__
31035__ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
31036  uint16x8_t __ret;
31037  __ret = (uint16x8_t)(__p0);
31038  return __ret;
31039}
31040#else
31041__ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
31042  uint16x8_t __ret;
31043  __ret = (uint16x8_t)(__p0);
31044  return __ret;
31045}
31046#endif
31047
31048#ifdef __LITTLE_ENDIAN__
31049__ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
31050  uint16x8_t __ret;
31051  __ret = (uint16x8_t)(__p0);
31052  return __ret;
31053}
31054#else
31055__ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
31056  uint16x8_t __ret;
31057  __ret = (uint16x8_t)(__p0);
31058  return __ret;
31059}
31060#endif
31061
31062#ifdef __LITTLE_ENDIAN__
31063__ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
31064  uint16x8_t __ret;
31065  __ret = (uint16x8_t)(__p0);
31066  return __ret;
31067}
31068#else
31069__ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
31070  uint16x8_t __ret;
31071  __ret = (uint16x8_t)(__p0);
31072  return __ret;
31073}
31074#endif
31075
31076#ifdef __LITTLE_ENDIAN__
31077__ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
31078  uint16x8_t __ret;
31079  __ret = (uint16x8_t)(__p0);
31080  return __ret;
31081}
31082#else
31083__ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
31084  uint16x8_t __ret;
31085  __ret = (uint16x8_t)(__p0);
31086  return __ret;
31087}
31088#endif
31089
31090#ifdef __LITTLE_ENDIAN__
31091__ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
31092  uint16x8_t __ret;
31093  __ret = (uint16x8_t)(__p0);
31094  return __ret;
31095}
31096#else
31097__ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
31098  uint16x8_t __ret;
31099  __ret = (uint16x8_t)(__p0);
31100  return __ret;
31101}
31102#endif
31103
31104#ifdef __LITTLE_ENDIAN__
31105__ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
31106  uint16x8_t __ret;
31107  __ret = (uint16x8_t)(__p0);
31108  return __ret;
31109}
31110#else
31111__ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
31112  uint16x8_t __ret;
31113  __ret = (uint16x8_t)(__p0);
31114  return __ret;
31115}
31116#endif
31117
31118#ifdef __LITTLE_ENDIAN__
31119__ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
31120  uint16x8_t __ret;
31121  __ret = (uint16x8_t)(__p0);
31122  return __ret;
31123}
31124#else
31125__ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
31126  uint16x8_t __ret;
31127  __ret = (uint16x8_t)(__p0);
31128  return __ret;
31129}
31130#endif
31131
31132#ifdef __LITTLE_ENDIAN__
31133__ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
31134  uint16x8_t __ret;
31135  __ret = (uint16x8_t)(__p0);
31136  return __ret;
31137}
31138#else
31139__ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
31140  uint16x8_t __ret;
31141  __ret = (uint16x8_t)(__p0);
31142  return __ret;
31143}
31144#endif
31145
31146#ifdef __LITTLE_ENDIAN__
31147__ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
31148  uint16x8_t __ret;
31149  __ret = (uint16x8_t)(__p0);
31150  return __ret;
31151}
31152#else
31153__ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
31154  uint16x8_t __ret;
31155  __ret = (uint16x8_t)(__p0);
31156  return __ret;
31157}
31158#endif
31159
31160#ifdef __LITTLE_ENDIAN__
31161__ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
31162  int8x16_t __ret;
31163  __ret = (int8x16_t)(__p0);
31164  return __ret;
31165}
31166#else
31167__ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
31168  int8x16_t __ret;
31169  __ret = (int8x16_t)(__p0);
31170  return __ret;
31171}
31172#endif
31173
31174#ifdef __LITTLE_ENDIAN__
31175__ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
31176  int8x16_t __ret;
31177  __ret = (int8x16_t)(__p0);
31178  return __ret;
31179}
31180#else
31181__ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
31182  int8x16_t __ret;
31183  __ret = (int8x16_t)(__p0);
31184  return __ret;
31185}
31186#endif
31187
31188#ifdef __LITTLE_ENDIAN__
31189__ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
31190  int8x16_t __ret;
31191  __ret = (int8x16_t)(__p0);
31192  return __ret;
31193}
31194#else
31195__ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
31196  int8x16_t __ret;
31197  __ret = (int8x16_t)(__p0);
31198  return __ret;
31199}
31200#endif
31201
31202#ifdef __LITTLE_ENDIAN__
31203__ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
31204  int8x16_t __ret;
31205  __ret = (int8x16_t)(__p0);
31206  return __ret;
31207}
31208#else
31209__ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
31210  int8x16_t __ret;
31211  __ret = (int8x16_t)(__p0);
31212  return __ret;
31213}
31214#endif
31215
31216#ifdef __LITTLE_ENDIAN__
31217__ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
31218  int8x16_t __ret;
31219  __ret = (int8x16_t)(__p0);
31220  return __ret;
31221}
31222#else
31223__ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
31224  int8x16_t __ret;
31225  __ret = (int8x16_t)(__p0);
31226  return __ret;
31227}
31228#endif
31229
31230#ifdef __LITTLE_ENDIAN__
31231__ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
31232  int8x16_t __ret;
31233  __ret = (int8x16_t)(__p0);
31234  return __ret;
31235}
31236#else
31237__ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
31238  int8x16_t __ret;
31239  __ret = (int8x16_t)(__p0);
31240  return __ret;
31241}
31242#endif
31243
31244#ifdef __LITTLE_ENDIAN__
31245__ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
31246  int8x16_t __ret;
31247  __ret = (int8x16_t)(__p0);
31248  return __ret;
31249}
31250#else
31251__ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
31252  int8x16_t __ret;
31253  __ret = (int8x16_t)(__p0);
31254  return __ret;
31255}
31256#endif
31257
31258#ifdef __LITTLE_ENDIAN__
31259__ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
31260  int8x16_t __ret;
31261  __ret = (int8x16_t)(__p0);
31262  return __ret;
31263}
31264#else
31265__ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
31266  int8x16_t __ret;
31267  __ret = (int8x16_t)(__p0);
31268  return __ret;
31269}
31270#endif
31271
31272#ifdef __LITTLE_ENDIAN__
31273__ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
31274  int8x16_t __ret;
31275  __ret = (int8x16_t)(__p0);
31276  return __ret;
31277}
31278#else
31279__ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
31280  int8x16_t __ret;
31281  __ret = (int8x16_t)(__p0);
31282  return __ret;
31283}
31284#endif
31285
31286#ifdef __LITTLE_ENDIAN__
31287__ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
31288  int8x16_t __ret;
31289  __ret = (int8x16_t)(__p0);
31290  return __ret;
31291}
31292#else
31293__ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
31294  int8x16_t __ret;
31295  __ret = (int8x16_t)(__p0);
31296  return __ret;
31297}
31298#endif
31299
31300#ifdef __LITTLE_ENDIAN__
31301__ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
31302  int8x16_t __ret;
31303  __ret = (int8x16_t)(__p0);
31304  return __ret;
31305}
31306#else
31307__ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
31308  int8x16_t __ret;
31309  __ret = (int8x16_t)(__p0);
31310  return __ret;
31311}
31312#endif
31313
31314#ifdef __LITTLE_ENDIAN__
31315__ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
31316  float32x4_t __ret;
31317  __ret = (float32x4_t)(__p0);
31318  return __ret;
31319}
31320#else
31321__ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
31322  float32x4_t __ret;
31323  __ret = (float32x4_t)(__p0);
31324  return __ret;
31325}
31326#endif
31327
31328#ifdef __LITTLE_ENDIAN__
31329__ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
31330  float32x4_t __ret;
31331  __ret = (float32x4_t)(__p0);
31332  return __ret;
31333}
31334#else
31335__ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
31336  float32x4_t __ret;
31337  __ret = (float32x4_t)(__p0);
31338  return __ret;
31339}
31340#endif
31341
31342#ifdef __LITTLE_ENDIAN__
31343__ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
31344  float32x4_t __ret;
31345  __ret = (float32x4_t)(__p0);
31346  return __ret;
31347}
31348#else
31349__ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
31350  float32x4_t __ret;
31351  __ret = (float32x4_t)(__p0);
31352  return __ret;
31353}
31354#endif
31355
31356#ifdef __LITTLE_ENDIAN__
31357__ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
31358  float32x4_t __ret;
31359  __ret = (float32x4_t)(__p0);
31360  return __ret;
31361}
31362#else
31363__ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
31364  float32x4_t __ret;
31365  __ret = (float32x4_t)(__p0);
31366  return __ret;
31367}
31368#endif
31369
31370#ifdef __LITTLE_ENDIAN__
31371__ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
31372  float32x4_t __ret;
31373  __ret = (float32x4_t)(__p0);
31374  return __ret;
31375}
31376#else
31377__ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
31378  float32x4_t __ret;
31379  __ret = (float32x4_t)(__p0);
31380  return __ret;
31381}
31382#endif
31383
31384#ifdef __LITTLE_ENDIAN__
31385__ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
31386  float32x4_t __ret;
31387  __ret = (float32x4_t)(__p0);
31388  return __ret;
31389}
31390#else
31391__ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
31392  float32x4_t __ret;
31393  __ret = (float32x4_t)(__p0);
31394  return __ret;
31395}
31396#endif
31397
31398#ifdef __LITTLE_ENDIAN__
31399__ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
31400  float32x4_t __ret;
31401  __ret = (float32x4_t)(__p0);
31402  return __ret;
31403}
31404#else
31405__ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
31406  float32x4_t __ret;
31407  __ret = (float32x4_t)(__p0);
31408  return __ret;
31409}
31410#endif
31411
31412#ifdef __LITTLE_ENDIAN__
31413__ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
31414  float32x4_t __ret;
31415  __ret = (float32x4_t)(__p0);
31416  return __ret;
31417}
31418#else
31419__ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
31420  float32x4_t __ret;
31421  __ret = (float32x4_t)(__p0);
31422  return __ret;
31423}
31424#endif
31425
31426#ifdef __LITTLE_ENDIAN__
31427__ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
31428  float32x4_t __ret;
31429  __ret = (float32x4_t)(__p0);
31430  return __ret;
31431}
31432#else
31433__ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
31434  float32x4_t __ret;
31435  __ret = (float32x4_t)(__p0);
31436  return __ret;
31437}
31438#endif
31439
31440#ifdef __LITTLE_ENDIAN__
31441__ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
31442  float32x4_t __ret;
31443  __ret = (float32x4_t)(__p0);
31444  return __ret;
31445}
31446#else
31447__ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
31448  float32x4_t __ret;
31449  __ret = (float32x4_t)(__p0);
31450  return __ret;
31451}
31452#endif
31453
31454#ifdef __LITTLE_ENDIAN__
31455__ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
31456  float32x4_t __ret;
31457  __ret = (float32x4_t)(__p0);
31458  return __ret;
31459}
31460#else
31461__ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
31462  float32x4_t __ret;
31463  __ret = (float32x4_t)(__p0);
31464  return __ret;
31465}
31466#endif
31467
31468#ifdef __LITTLE_ENDIAN__
31469__ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
31470  float16x8_t __ret;
31471  __ret = (float16x8_t)(__p0);
31472  return __ret;
31473}
31474#else
31475__ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
31476  float16x8_t __ret;
31477  __ret = (float16x8_t)(__p0);
31478  return __ret;
31479}
31480#endif
31481
31482#ifdef __LITTLE_ENDIAN__
31483__ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
31484  float16x8_t __ret;
31485  __ret = (float16x8_t)(__p0);
31486  return __ret;
31487}
31488#else
31489__ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
31490  float16x8_t __ret;
31491  __ret = (float16x8_t)(__p0);
31492  return __ret;
31493}
31494#endif
31495
31496#ifdef __LITTLE_ENDIAN__
31497__ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
31498  float16x8_t __ret;
31499  __ret = (float16x8_t)(__p0);
31500  return __ret;
31501}
31502#else
31503__ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
31504  float16x8_t __ret;
31505  __ret = (float16x8_t)(__p0);
31506  return __ret;
31507}
31508#endif
31509
31510#ifdef __LITTLE_ENDIAN__
31511__ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
31512  float16x8_t __ret;
31513  __ret = (float16x8_t)(__p0);
31514  return __ret;
31515}
31516#else
31517__ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
31518  float16x8_t __ret;
31519  __ret = (float16x8_t)(__p0);
31520  return __ret;
31521}
31522#endif
31523
31524#ifdef __LITTLE_ENDIAN__
31525__ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
31526  float16x8_t __ret;
31527  __ret = (float16x8_t)(__p0);
31528  return __ret;
31529}
31530#else
31531__ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
31532  float16x8_t __ret;
31533  __ret = (float16x8_t)(__p0);
31534  return __ret;
31535}
31536#endif
31537
31538#ifdef __LITTLE_ENDIAN__
31539__ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
31540  float16x8_t __ret;
31541  __ret = (float16x8_t)(__p0);
31542  return __ret;
31543}
31544#else
31545__ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
31546  float16x8_t __ret;
31547  __ret = (float16x8_t)(__p0);
31548  return __ret;
31549}
31550#endif
31551
31552#ifdef __LITTLE_ENDIAN__
31553__ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
31554  float16x8_t __ret;
31555  __ret = (float16x8_t)(__p0);
31556  return __ret;
31557}
31558#else
31559__ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
31560  float16x8_t __ret;
31561  __ret = (float16x8_t)(__p0);
31562  return __ret;
31563}
31564#endif
31565
31566#ifdef __LITTLE_ENDIAN__
31567__ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
31568  float16x8_t __ret;
31569  __ret = (float16x8_t)(__p0);
31570  return __ret;
31571}
31572#else
31573__ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
31574  float16x8_t __ret;
31575  __ret = (float16x8_t)(__p0);
31576  return __ret;
31577}
31578#endif
31579
31580#ifdef __LITTLE_ENDIAN__
31581__ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
31582  float16x8_t __ret;
31583  __ret = (float16x8_t)(__p0);
31584  return __ret;
31585}
31586#else
31587__ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
31588  float16x8_t __ret;
31589  __ret = (float16x8_t)(__p0);
31590  return __ret;
31591}
31592#endif
31593
31594#ifdef __LITTLE_ENDIAN__
31595__ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
31596  float16x8_t __ret;
31597  __ret = (float16x8_t)(__p0);
31598  return __ret;
31599}
31600#else
31601__ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
31602  float16x8_t __ret;
31603  __ret = (float16x8_t)(__p0);
31604  return __ret;
31605}
31606#endif
31607
31608#ifdef __LITTLE_ENDIAN__
31609__ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
31610  float16x8_t __ret;
31611  __ret = (float16x8_t)(__p0);
31612  return __ret;
31613}
31614#else
31615__ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
31616  float16x8_t __ret;
31617  __ret = (float16x8_t)(__p0);
31618  return __ret;
31619}
31620#endif
31621
31622#ifdef __LITTLE_ENDIAN__
31623__ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
31624  int32x4_t __ret;
31625  __ret = (int32x4_t)(__p0);
31626  return __ret;
31627}
31628#else
31629__ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
31630  int32x4_t __ret;
31631  __ret = (int32x4_t)(__p0);
31632  return __ret;
31633}
31634#endif
31635
31636#ifdef __LITTLE_ENDIAN__
31637__ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
31638  int32x4_t __ret;
31639  __ret = (int32x4_t)(__p0);
31640  return __ret;
31641}
31642#else
31643__ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
31644  int32x4_t __ret;
31645  __ret = (int32x4_t)(__p0);
31646  return __ret;
31647}
31648#endif
31649
31650#ifdef __LITTLE_ENDIAN__
31651__ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
31652  int32x4_t __ret;
31653  __ret = (int32x4_t)(__p0);
31654  return __ret;
31655}
31656#else
31657__ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
31658  int32x4_t __ret;
31659  __ret = (int32x4_t)(__p0);
31660  return __ret;
31661}
31662#endif
31663
31664#ifdef __LITTLE_ENDIAN__
31665__ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
31666  int32x4_t __ret;
31667  __ret = (int32x4_t)(__p0);
31668  return __ret;
31669}
31670#else
31671__ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
31672  int32x4_t __ret;
31673  __ret = (int32x4_t)(__p0);
31674  return __ret;
31675}
31676#endif
31677
31678#ifdef __LITTLE_ENDIAN__
31679__ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
31680  int32x4_t __ret;
31681  __ret = (int32x4_t)(__p0);
31682  return __ret;
31683}
31684#else
31685__ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
31686  int32x4_t __ret;
31687  __ret = (int32x4_t)(__p0);
31688  return __ret;
31689}
31690#endif
31691
31692#ifdef __LITTLE_ENDIAN__
31693__ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
31694  int32x4_t __ret;
31695  __ret = (int32x4_t)(__p0);
31696  return __ret;
31697}
31698#else
31699__ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
31700  int32x4_t __ret;
31701  __ret = (int32x4_t)(__p0);
31702  return __ret;
31703}
31704#endif
31705
31706#ifdef __LITTLE_ENDIAN__
31707__ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
31708  int32x4_t __ret;
31709  __ret = (int32x4_t)(__p0);
31710  return __ret;
31711}
31712#else
31713__ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
31714  int32x4_t __ret;
31715  __ret = (int32x4_t)(__p0);
31716  return __ret;
31717}
31718#endif
31719
31720#ifdef __LITTLE_ENDIAN__
31721__ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
31722  int32x4_t __ret;
31723  __ret = (int32x4_t)(__p0);
31724  return __ret;
31725}
31726#else
31727__ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
31728  int32x4_t __ret;
31729  __ret = (int32x4_t)(__p0);
31730  return __ret;
31731}
31732#endif
31733
31734#ifdef __LITTLE_ENDIAN__
31735__ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
31736  int32x4_t __ret;
31737  __ret = (int32x4_t)(__p0);
31738  return __ret;
31739}
31740#else
31741__ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
31742  int32x4_t __ret;
31743  __ret = (int32x4_t)(__p0);
31744  return __ret;
31745}
31746#endif
31747
31748#ifdef __LITTLE_ENDIAN__
31749__ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
31750  int32x4_t __ret;
31751  __ret = (int32x4_t)(__p0);
31752  return __ret;
31753}
31754#else
31755__ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
31756  int32x4_t __ret;
31757  __ret = (int32x4_t)(__p0);
31758  return __ret;
31759}
31760#endif
31761
31762#ifdef __LITTLE_ENDIAN__
31763__ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
31764  int32x4_t __ret;
31765  __ret = (int32x4_t)(__p0);
31766  return __ret;
31767}
31768#else
31769__ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
31770  int32x4_t __ret;
31771  __ret = (int32x4_t)(__p0);
31772  return __ret;
31773}
31774#endif
31775
31776#ifdef __LITTLE_ENDIAN__
31777__ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
31778  int64x2_t __ret;
31779  __ret = (int64x2_t)(__p0);
31780  return __ret;
31781}
31782#else
31783__ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
31784  int64x2_t __ret;
31785  __ret = (int64x2_t)(__p0);
31786  return __ret;
31787}
31788#endif
31789
31790#ifdef __LITTLE_ENDIAN__
31791__ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
31792  int64x2_t __ret;
31793  __ret = (int64x2_t)(__p0);
31794  return __ret;
31795}
31796#else
31797__ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
31798  int64x2_t __ret;
31799  __ret = (int64x2_t)(__p0);
31800  return __ret;
31801}
31802#endif
31803
31804#ifdef __LITTLE_ENDIAN__
31805__ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
31806  int64x2_t __ret;
31807  __ret = (int64x2_t)(__p0);
31808  return __ret;
31809}
31810#else
31811__ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
31812  int64x2_t __ret;
31813  __ret = (int64x2_t)(__p0);
31814  return __ret;
31815}
31816#endif
31817
31818#ifdef __LITTLE_ENDIAN__
31819__ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
31820  int64x2_t __ret;
31821  __ret = (int64x2_t)(__p0);
31822  return __ret;
31823}
31824#else
31825__ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
31826  int64x2_t __ret;
31827  __ret = (int64x2_t)(__p0);
31828  return __ret;
31829}
31830#endif
31831
31832#ifdef __LITTLE_ENDIAN__
31833__ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
31834  int64x2_t __ret;
31835  __ret = (int64x2_t)(__p0);
31836  return __ret;
31837}
31838#else
31839__ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
31840  int64x2_t __ret;
31841  __ret = (int64x2_t)(__p0);
31842  return __ret;
31843}
31844#endif
31845
31846#ifdef __LITTLE_ENDIAN__
31847__ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
31848  int64x2_t __ret;
31849  __ret = (int64x2_t)(__p0);
31850  return __ret;
31851}
31852#else
31853__ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
31854  int64x2_t __ret;
31855  __ret = (int64x2_t)(__p0);
31856  return __ret;
31857}
31858#endif
31859
31860#ifdef __LITTLE_ENDIAN__
31861__ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
31862  int64x2_t __ret;
31863  __ret = (int64x2_t)(__p0);
31864  return __ret;
31865}
31866#else
31867__ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
31868  int64x2_t __ret;
31869  __ret = (int64x2_t)(__p0);
31870  return __ret;
31871}
31872#endif
31873
31874#ifdef __LITTLE_ENDIAN__
31875__ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
31876  int64x2_t __ret;
31877  __ret = (int64x2_t)(__p0);
31878  return __ret;
31879}
31880#else
31881__ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
31882  int64x2_t __ret;
31883  __ret = (int64x2_t)(__p0);
31884  return __ret;
31885}
31886#endif
31887
31888#ifdef __LITTLE_ENDIAN__
31889__ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
31890  int64x2_t __ret;
31891  __ret = (int64x2_t)(__p0);
31892  return __ret;
31893}
31894#else
31895__ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
31896  int64x2_t __ret;
31897  __ret = (int64x2_t)(__p0);
31898  return __ret;
31899}
31900#endif
31901
31902#ifdef __LITTLE_ENDIAN__
31903__ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
31904  int64x2_t __ret;
31905  __ret = (int64x2_t)(__p0);
31906  return __ret;
31907}
31908#else
31909__ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
31910  int64x2_t __ret;
31911  __ret = (int64x2_t)(__p0);
31912  return __ret;
31913}
31914#endif
31915
31916#ifdef __LITTLE_ENDIAN__
31917__ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
31918  int64x2_t __ret;
31919  __ret = (int64x2_t)(__p0);
31920  return __ret;
31921}
31922#else
31923__ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
31924  int64x2_t __ret;
31925  __ret = (int64x2_t)(__p0);
31926  return __ret;
31927}
31928#endif
31929
31930#ifdef __LITTLE_ENDIAN__
31931__ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
31932  int16x8_t __ret;
31933  __ret = (int16x8_t)(__p0);
31934  return __ret;
31935}
31936#else
31937__ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
31938  int16x8_t __ret;
31939  __ret = (int16x8_t)(__p0);
31940  return __ret;
31941}
31942#endif
31943
31944#ifdef __LITTLE_ENDIAN__
31945__ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
31946  int16x8_t __ret;
31947  __ret = (int16x8_t)(__p0);
31948  return __ret;
31949}
31950#else
31951__ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
31952  int16x8_t __ret;
31953  __ret = (int16x8_t)(__p0);
31954  return __ret;
31955}
31956#endif
31957
31958#ifdef __LITTLE_ENDIAN__
31959__ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
31960  int16x8_t __ret;
31961  __ret = (int16x8_t)(__p0);
31962  return __ret;
31963}
31964#else
31965__ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
31966  int16x8_t __ret;
31967  __ret = (int16x8_t)(__p0);
31968  return __ret;
31969}
31970#endif
31971
31972#ifdef __LITTLE_ENDIAN__
31973__ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
31974  int16x8_t __ret;
31975  __ret = (int16x8_t)(__p0);
31976  return __ret;
31977}
31978#else
31979__ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
31980  int16x8_t __ret;
31981  __ret = (int16x8_t)(__p0);
31982  return __ret;
31983}
31984#endif
31985
31986#ifdef __LITTLE_ENDIAN__
31987__ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
31988  int16x8_t __ret;
31989  __ret = (int16x8_t)(__p0);
31990  return __ret;
31991}
31992#else
31993__ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
31994  int16x8_t __ret;
31995  __ret = (int16x8_t)(__p0);
31996  return __ret;
31997}
31998#endif
31999
32000#ifdef __LITTLE_ENDIAN__
32001__ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
32002  int16x8_t __ret;
32003  __ret = (int16x8_t)(__p0);
32004  return __ret;
32005}
32006#else
32007__ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
32008  int16x8_t __ret;
32009  __ret = (int16x8_t)(__p0);
32010  return __ret;
32011}
32012#endif
32013
32014#ifdef __LITTLE_ENDIAN__
32015__ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
32016  int16x8_t __ret;
32017  __ret = (int16x8_t)(__p0);
32018  return __ret;
32019}
32020#else
32021__ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
32022  int16x8_t __ret;
32023  __ret = (int16x8_t)(__p0);
32024  return __ret;
32025}
32026#endif
32027
32028#ifdef __LITTLE_ENDIAN__
32029__ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
32030  int16x8_t __ret;
32031  __ret = (int16x8_t)(__p0);
32032  return __ret;
32033}
32034#else
32035__ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
32036  int16x8_t __ret;
32037  __ret = (int16x8_t)(__p0);
32038  return __ret;
32039}
32040#endif
32041
32042#ifdef __LITTLE_ENDIAN__
32043__ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
32044  int16x8_t __ret;
32045  __ret = (int16x8_t)(__p0);
32046  return __ret;
32047}
32048#else
32049__ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
32050  int16x8_t __ret;
32051  __ret = (int16x8_t)(__p0);
32052  return __ret;
32053}
32054#endif
32055
32056#ifdef __LITTLE_ENDIAN__
32057__ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
32058  int16x8_t __ret;
32059  __ret = (int16x8_t)(__p0);
32060  return __ret;
32061}
32062#else
32063__ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
32064  int16x8_t __ret;
32065  __ret = (int16x8_t)(__p0);
32066  return __ret;
32067}
32068#endif
32069
32070#ifdef __LITTLE_ENDIAN__
32071__ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
32072  int16x8_t __ret;
32073  __ret = (int16x8_t)(__p0);
32074  return __ret;
32075}
32076#else
32077__ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
32078  int16x8_t __ret;
32079  __ret = (int16x8_t)(__p0);
32080  return __ret;
32081}
32082#endif
32083
32084#ifdef __LITTLE_ENDIAN__
32085__ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
32086  uint8x8_t __ret;
32087  __ret = (uint8x8_t)(__p0);
32088  return __ret;
32089}
32090#else
32091__ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
32092  uint8x8_t __ret;
32093  __ret = (uint8x8_t)(__p0);
32094  return __ret;
32095}
32096#endif
32097
32098#ifdef __LITTLE_ENDIAN__
32099__ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
32100  uint8x8_t __ret;
32101  __ret = (uint8x8_t)(__p0);
32102  return __ret;
32103}
32104#else
32105__ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
32106  uint8x8_t __ret;
32107  __ret = (uint8x8_t)(__p0);
32108  return __ret;
32109}
32110#endif
32111
32112#ifdef __LITTLE_ENDIAN__
32113__ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
32114  uint8x8_t __ret;
32115  __ret = (uint8x8_t)(__p0);
32116  return __ret;
32117}
32118#else
32119__ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
32120  uint8x8_t __ret;
32121  __ret = (uint8x8_t)(__p0);
32122  return __ret;
32123}
32124#endif
32125
32126#ifdef __LITTLE_ENDIAN__
32127__ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
32128  uint8x8_t __ret;
32129  __ret = (uint8x8_t)(__p0);
32130  return __ret;
32131}
32132#else
32133__ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
32134  uint8x8_t __ret;
32135  __ret = (uint8x8_t)(__p0);
32136  return __ret;
32137}
32138#endif
32139
32140#ifdef __LITTLE_ENDIAN__
32141__ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
32142  uint8x8_t __ret;
32143  __ret = (uint8x8_t)(__p0);
32144  return __ret;
32145}
32146#else
32147__ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
32148  uint8x8_t __ret;
32149  __ret = (uint8x8_t)(__p0);
32150  return __ret;
32151}
32152#endif
32153
32154#ifdef __LITTLE_ENDIAN__
32155__ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
32156  uint8x8_t __ret;
32157  __ret = (uint8x8_t)(__p0);
32158  return __ret;
32159}
32160#else
32161__ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
32162  uint8x8_t __ret;
32163  __ret = (uint8x8_t)(__p0);
32164  return __ret;
32165}
32166#endif
32167
32168#ifdef __LITTLE_ENDIAN__
32169__ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
32170  uint8x8_t __ret;
32171  __ret = (uint8x8_t)(__p0);
32172  return __ret;
32173}
32174#else
32175__ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
32176  uint8x8_t __ret;
32177  __ret = (uint8x8_t)(__p0);
32178  return __ret;
32179}
32180#endif
32181
32182#ifdef __LITTLE_ENDIAN__
32183__ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
32184  uint8x8_t __ret;
32185  __ret = (uint8x8_t)(__p0);
32186  return __ret;
32187}
32188#else
32189__ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
32190  uint8x8_t __ret;
32191  __ret = (uint8x8_t)(__p0);
32192  return __ret;
32193}
32194#endif
32195
32196#ifdef __LITTLE_ENDIAN__
32197__ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
32198  uint8x8_t __ret;
32199  __ret = (uint8x8_t)(__p0);
32200  return __ret;
32201}
32202#else
32203__ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
32204  uint8x8_t __ret;
32205  __ret = (uint8x8_t)(__p0);
32206  return __ret;
32207}
32208#endif
32209
32210#ifdef __LITTLE_ENDIAN__
32211__ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
32212  uint8x8_t __ret;
32213  __ret = (uint8x8_t)(__p0);
32214  return __ret;
32215}
32216#else
32217__ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
32218  uint8x8_t __ret;
32219  __ret = (uint8x8_t)(__p0);
32220  return __ret;
32221}
32222#endif
32223
32224#ifdef __LITTLE_ENDIAN__
32225__ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
32226  uint8x8_t __ret;
32227  __ret = (uint8x8_t)(__p0);
32228  return __ret;
32229}
32230#else
32231__ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
32232  uint8x8_t __ret;
32233  __ret = (uint8x8_t)(__p0);
32234  return __ret;
32235}
32236#endif
32237
32238#ifdef __LITTLE_ENDIAN__
32239__ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
32240  uint32x2_t __ret;
32241  __ret = (uint32x2_t)(__p0);
32242  return __ret;
32243}
32244#else
32245__ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
32246  uint32x2_t __ret;
32247  __ret = (uint32x2_t)(__p0);
32248  return __ret;
32249}
32250#endif
32251
32252#ifdef __LITTLE_ENDIAN__
32253__ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
32254  uint32x2_t __ret;
32255  __ret = (uint32x2_t)(__p0);
32256  return __ret;
32257}
32258#else
32259__ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
32260  uint32x2_t __ret;
32261  __ret = (uint32x2_t)(__p0);
32262  return __ret;
32263}
32264#endif
32265
32266#ifdef __LITTLE_ENDIAN__
32267__ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
32268  uint32x2_t __ret;
32269  __ret = (uint32x2_t)(__p0);
32270  return __ret;
32271}
32272#else
32273__ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
32274  uint32x2_t __ret;
32275  __ret = (uint32x2_t)(__p0);
32276  return __ret;
32277}
32278#endif
32279
32280#ifdef __LITTLE_ENDIAN__
32281__ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
32282  uint32x2_t __ret;
32283  __ret = (uint32x2_t)(__p0);
32284  return __ret;
32285}
32286#else
32287__ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
32288  uint32x2_t __ret;
32289  __ret = (uint32x2_t)(__p0);
32290  return __ret;
32291}
32292#endif
32293
32294#ifdef __LITTLE_ENDIAN__
32295__ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
32296  uint32x2_t __ret;
32297  __ret = (uint32x2_t)(__p0);
32298  return __ret;
32299}
32300#else
32301__ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
32302  uint32x2_t __ret;
32303  __ret = (uint32x2_t)(__p0);
32304  return __ret;
32305}
32306#endif
32307
32308#ifdef __LITTLE_ENDIAN__
32309__ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
32310  uint32x2_t __ret;
32311  __ret = (uint32x2_t)(__p0);
32312  return __ret;
32313}
32314#else
32315__ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
32316  uint32x2_t __ret;
32317  __ret = (uint32x2_t)(__p0);
32318  return __ret;
32319}
32320#endif
32321
32322#ifdef __LITTLE_ENDIAN__
32323__ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
32324  uint32x2_t __ret;
32325  __ret = (uint32x2_t)(__p0);
32326  return __ret;
32327}
32328#else
32329__ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
32330  uint32x2_t __ret;
32331  __ret = (uint32x2_t)(__p0);
32332  return __ret;
32333}
32334#endif
32335
32336#ifdef __LITTLE_ENDIAN__
32337__ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
32338  uint32x2_t __ret;
32339  __ret = (uint32x2_t)(__p0);
32340  return __ret;
32341}
32342#else
32343__ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
32344  uint32x2_t __ret;
32345  __ret = (uint32x2_t)(__p0);
32346  return __ret;
32347}
32348#endif
32349
32350#ifdef __LITTLE_ENDIAN__
32351__ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
32352  uint32x2_t __ret;
32353  __ret = (uint32x2_t)(__p0);
32354  return __ret;
32355}
32356#else
32357__ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
32358  uint32x2_t __ret;
32359  __ret = (uint32x2_t)(__p0);
32360  return __ret;
32361}
32362#endif
32363
32364#ifdef __LITTLE_ENDIAN__
32365__ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
32366  uint32x2_t __ret;
32367  __ret = (uint32x2_t)(__p0);
32368  return __ret;
32369}
32370#else
32371__ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
32372  uint32x2_t __ret;
32373  __ret = (uint32x2_t)(__p0);
32374  return __ret;
32375}
32376#endif
32377
32378#ifdef __LITTLE_ENDIAN__
32379__ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
32380  uint32x2_t __ret;
32381  __ret = (uint32x2_t)(__p0);
32382  return __ret;
32383}
32384#else
32385__ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
32386  uint32x2_t __ret;
32387  __ret = (uint32x2_t)(__p0);
32388  return __ret;
32389}
32390#endif
32391
32392#ifdef __LITTLE_ENDIAN__
32393__ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
32394  uint64x1_t __ret;
32395  __ret = (uint64x1_t)(__p0);
32396  return __ret;
32397}
32398#else
32399__ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
32400  uint64x1_t __ret;
32401  __ret = (uint64x1_t)(__p0);
32402  return __ret;
32403}
32404#endif
32405
32406#ifdef __LITTLE_ENDIAN__
32407__ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
32408  uint64x1_t __ret;
32409  __ret = (uint64x1_t)(__p0);
32410  return __ret;
32411}
32412#else
32413__ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
32414  uint64x1_t __ret;
32415  __ret = (uint64x1_t)(__p0);
32416  return __ret;
32417}
32418#endif
32419
32420#ifdef __LITTLE_ENDIAN__
32421__ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
32422  uint64x1_t __ret;
32423  __ret = (uint64x1_t)(__p0);
32424  return __ret;
32425}
32426#else
32427__ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
32428  uint64x1_t __ret;
32429  __ret = (uint64x1_t)(__p0);
32430  return __ret;
32431}
32432#endif
32433
32434#ifdef __LITTLE_ENDIAN__
32435__ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
32436  uint64x1_t __ret;
32437  __ret = (uint64x1_t)(__p0);
32438  return __ret;
32439}
32440#else
32441__ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
32442  uint64x1_t __ret;
32443  __ret = (uint64x1_t)(__p0);
32444  return __ret;
32445}
32446#endif
32447
32448#ifdef __LITTLE_ENDIAN__
32449__ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
32450  uint64x1_t __ret;
32451  __ret = (uint64x1_t)(__p0);
32452  return __ret;
32453}
32454#else
32455__ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
32456  uint64x1_t __ret;
32457  __ret = (uint64x1_t)(__p0);
32458  return __ret;
32459}
32460#endif
32461
32462#ifdef __LITTLE_ENDIAN__
32463__ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
32464  uint64x1_t __ret;
32465  __ret = (uint64x1_t)(__p0);
32466  return __ret;
32467}
32468#else
32469__ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
32470  uint64x1_t __ret;
32471  __ret = (uint64x1_t)(__p0);
32472  return __ret;
32473}
32474#endif
32475
32476#ifdef __LITTLE_ENDIAN__
32477__ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
32478  uint64x1_t __ret;
32479  __ret = (uint64x1_t)(__p0);
32480  return __ret;
32481}
32482#else
32483__ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
32484  uint64x1_t __ret;
32485  __ret = (uint64x1_t)(__p0);
32486  return __ret;
32487}
32488#endif
32489
32490#ifdef __LITTLE_ENDIAN__
32491__ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
32492  uint64x1_t __ret;
32493  __ret = (uint64x1_t)(__p0);
32494  return __ret;
32495}
32496#else
32497__ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
32498  uint64x1_t __ret;
32499  __ret = (uint64x1_t)(__p0);
32500  return __ret;
32501}
32502#endif
32503
32504#ifdef __LITTLE_ENDIAN__
32505__ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
32506  uint64x1_t __ret;
32507  __ret = (uint64x1_t)(__p0);
32508  return __ret;
32509}
32510#else
32511__ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
32512  uint64x1_t __ret;
32513  __ret = (uint64x1_t)(__p0);
32514  return __ret;
32515}
32516#endif
32517
32518#ifdef __LITTLE_ENDIAN__
32519__ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
32520  uint64x1_t __ret;
32521  __ret = (uint64x1_t)(__p0);
32522  return __ret;
32523}
32524#else
32525__ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
32526  uint64x1_t __ret;
32527  __ret = (uint64x1_t)(__p0);
32528  return __ret;
32529}
32530#endif
32531
32532#ifdef __LITTLE_ENDIAN__
32533__ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
32534  uint64x1_t __ret;
32535  __ret = (uint64x1_t)(__p0);
32536  return __ret;
32537}
32538#else
32539__ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
32540  uint64x1_t __ret;
32541  __ret = (uint64x1_t)(__p0);
32542  return __ret;
32543}
32544#endif
32545
32546#ifdef __LITTLE_ENDIAN__
32547__ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
32548  uint16x4_t __ret;
32549  __ret = (uint16x4_t)(__p0);
32550  return __ret;
32551}
32552#else
32553__ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
32554  uint16x4_t __ret;
32555  __ret = (uint16x4_t)(__p0);
32556  return __ret;
32557}
32558#endif
32559
32560#ifdef __LITTLE_ENDIAN__
32561__ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
32562  uint16x4_t __ret;
32563  __ret = (uint16x4_t)(__p0);
32564  return __ret;
32565}
32566#else
32567__ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
32568  uint16x4_t __ret;
32569  __ret = (uint16x4_t)(__p0);
32570  return __ret;
32571}
32572#endif
32573
32574#ifdef __LITTLE_ENDIAN__
32575__ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
32576  uint16x4_t __ret;
32577  __ret = (uint16x4_t)(__p0);
32578  return __ret;
32579}
32580#else
32581__ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
32582  uint16x4_t __ret;
32583  __ret = (uint16x4_t)(__p0);
32584  return __ret;
32585}
32586#endif
32587
32588#ifdef __LITTLE_ENDIAN__
32589__ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
32590  uint16x4_t __ret;
32591  __ret = (uint16x4_t)(__p0);
32592  return __ret;
32593}
32594#else
32595__ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
32596  uint16x4_t __ret;
32597  __ret = (uint16x4_t)(__p0);
32598  return __ret;
32599}
32600#endif
32601
32602#ifdef __LITTLE_ENDIAN__
32603__ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
32604  uint16x4_t __ret;
32605  __ret = (uint16x4_t)(__p0);
32606  return __ret;
32607}
32608#else
32609__ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
32610  uint16x4_t __ret;
32611  __ret = (uint16x4_t)(__p0);
32612  return __ret;
32613}
32614#endif
32615
32616#ifdef __LITTLE_ENDIAN__
32617__ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
32618  uint16x4_t __ret;
32619  __ret = (uint16x4_t)(__p0);
32620  return __ret;
32621}
32622#else
32623__ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
32624  uint16x4_t __ret;
32625  __ret = (uint16x4_t)(__p0);
32626  return __ret;
32627}
32628#endif
32629
32630#ifdef __LITTLE_ENDIAN__
32631__ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
32632  uint16x4_t __ret;
32633  __ret = (uint16x4_t)(__p0);
32634  return __ret;
32635}
32636#else
32637__ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
32638  uint16x4_t __ret;
32639  __ret = (uint16x4_t)(__p0);
32640  return __ret;
32641}
32642#endif
32643
32644#ifdef __LITTLE_ENDIAN__
32645__ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
32646  uint16x4_t __ret;
32647  __ret = (uint16x4_t)(__p0);
32648  return __ret;
32649}
32650#else
32651__ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
32652  uint16x4_t __ret;
32653  __ret = (uint16x4_t)(__p0);
32654  return __ret;
32655}
32656#endif
32657
32658#ifdef __LITTLE_ENDIAN__
32659__ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
32660  uint16x4_t __ret;
32661  __ret = (uint16x4_t)(__p0);
32662  return __ret;
32663}
32664#else
32665__ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
32666  uint16x4_t __ret;
32667  __ret = (uint16x4_t)(__p0);
32668  return __ret;
32669}
32670#endif
32671
32672#ifdef __LITTLE_ENDIAN__
32673__ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
32674  uint16x4_t __ret;
32675  __ret = (uint16x4_t)(__p0);
32676  return __ret;
32677}
32678#else
32679__ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
32680  uint16x4_t __ret;
32681  __ret = (uint16x4_t)(__p0);
32682  return __ret;
32683}
32684#endif
32685
32686#ifdef __LITTLE_ENDIAN__
32687__ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
32688  uint16x4_t __ret;
32689  __ret = (uint16x4_t)(__p0);
32690  return __ret;
32691}
32692#else
32693__ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
32694  uint16x4_t __ret;
32695  __ret = (uint16x4_t)(__p0);
32696  return __ret;
32697}
32698#endif
32699
32700#ifdef __LITTLE_ENDIAN__
32701__ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
32702  int8x8_t __ret;
32703  __ret = (int8x8_t)(__p0);
32704  return __ret;
32705}
32706#else
32707__ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
32708  int8x8_t __ret;
32709  __ret = (int8x8_t)(__p0);
32710  return __ret;
32711}
32712#endif
32713
32714#ifdef __LITTLE_ENDIAN__
32715__ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
32716  int8x8_t __ret;
32717  __ret = (int8x8_t)(__p0);
32718  return __ret;
32719}
32720#else
32721__ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
32722  int8x8_t __ret;
32723  __ret = (int8x8_t)(__p0);
32724  return __ret;
32725}
32726#endif
32727
32728#ifdef __LITTLE_ENDIAN__
32729__ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
32730  int8x8_t __ret;
32731  __ret = (int8x8_t)(__p0);
32732  return __ret;
32733}
32734#else
32735__ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
32736  int8x8_t __ret;
32737  __ret = (int8x8_t)(__p0);
32738  return __ret;
32739}
32740#endif
32741
32742#ifdef __LITTLE_ENDIAN__
32743__ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
32744  int8x8_t __ret;
32745  __ret = (int8x8_t)(__p0);
32746  return __ret;
32747}
32748#else
32749__ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
32750  int8x8_t __ret;
32751  __ret = (int8x8_t)(__p0);
32752  return __ret;
32753}
32754#endif
32755
32756#ifdef __LITTLE_ENDIAN__
32757__ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
32758  int8x8_t __ret;
32759  __ret = (int8x8_t)(__p0);
32760  return __ret;
32761}
32762#else
32763__ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
32764  int8x8_t __ret;
32765  __ret = (int8x8_t)(__p0);
32766  return __ret;
32767}
32768#endif
32769
32770#ifdef __LITTLE_ENDIAN__
32771__ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
32772  int8x8_t __ret;
32773  __ret = (int8x8_t)(__p0);
32774  return __ret;
32775}
32776#else
32777__ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
32778  int8x8_t __ret;
32779  __ret = (int8x8_t)(__p0);
32780  return __ret;
32781}
32782#endif
32783
32784#ifdef __LITTLE_ENDIAN__
32785__ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
32786  int8x8_t __ret;
32787  __ret = (int8x8_t)(__p0);
32788  return __ret;
32789}
32790#else
32791__ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
32792  int8x8_t __ret;
32793  __ret = (int8x8_t)(__p0);
32794  return __ret;
32795}
32796#endif
32797
32798#ifdef __LITTLE_ENDIAN__
32799__ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
32800  int8x8_t __ret;
32801  __ret = (int8x8_t)(__p0);
32802  return __ret;
32803}
32804#else
32805__ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
32806  int8x8_t __ret;
32807  __ret = (int8x8_t)(__p0);
32808  return __ret;
32809}
32810#endif
32811
32812#ifdef __LITTLE_ENDIAN__
32813__ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
32814  int8x8_t __ret;
32815  __ret = (int8x8_t)(__p0);
32816  return __ret;
32817}
32818#else
32819__ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
32820  int8x8_t __ret;
32821  __ret = (int8x8_t)(__p0);
32822  return __ret;
32823}
32824#endif
32825
32826#ifdef __LITTLE_ENDIAN__
32827__ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
32828  int8x8_t __ret;
32829  __ret = (int8x8_t)(__p0);
32830  return __ret;
32831}
32832#else
32833__ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
32834  int8x8_t __ret;
32835  __ret = (int8x8_t)(__p0);
32836  return __ret;
32837}
32838#endif
32839
32840#ifdef __LITTLE_ENDIAN__
32841__ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
32842  int8x8_t __ret;
32843  __ret = (int8x8_t)(__p0);
32844  return __ret;
32845}
32846#else
32847__ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
32848  int8x8_t __ret;
32849  __ret = (int8x8_t)(__p0);
32850  return __ret;
32851}
32852#endif
32853
32854#ifdef __LITTLE_ENDIAN__
32855__ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
32856  float32x2_t __ret;
32857  __ret = (float32x2_t)(__p0);
32858  return __ret;
32859}
32860#else
32861__ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
32862  float32x2_t __ret;
32863  __ret = (float32x2_t)(__p0);
32864  return __ret;
32865}
32866#endif
32867
32868#ifdef __LITTLE_ENDIAN__
32869__ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
32870  float32x2_t __ret;
32871  __ret = (float32x2_t)(__p0);
32872  return __ret;
32873}
32874#else
32875__ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
32876  float32x2_t __ret;
32877  __ret = (float32x2_t)(__p0);
32878  return __ret;
32879}
32880#endif
32881
32882#ifdef __LITTLE_ENDIAN__
32883__ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
32884  float32x2_t __ret;
32885  __ret = (float32x2_t)(__p0);
32886  return __ret;
32887}
32888#else
32889__ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
32890  float32x2_t __ret;
32891  __ret = (float32x2_t)(__p0);
32892  return __ret;
32893}
32894#endif
32895
32896#ifdef __LITTLE_ENDIAN__
32897__ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
32898  float32x2_t __ret;
32899  __ret = (float32x2_t)(__p0);
32900  return __ret;
32901}
32902#else
32903__ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
32904  float32x2_t __ret;
32905  __ret = (float32x2_t)(__p0);
32906  return __ret;
32907}
32908#endif
32909
32910#ifdef __LITTLE_ENDIAN__
32911__ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
32912  float32x2_t __ret;
32913  __ret = (float32x2_t)(__p0);
32914  return __ret;
32915}
32916#else
32917__ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
32918  float32x2_t __ret;
32919  __ret = (float32x2_t)(__p0);
32920  return __ret;
32921}
32922#endif
32923
32924#ifdef __LITTLE_ENDIAN__
32925__ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
32926  float32x2_t __ret;
32927  __ret = (float32x2_t)(__p0);
32928  return __ret;
32929}
32930#else
32931__ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
32932  float32x2_t __ret;
32933  __ret = (float32x2_t)(__p0);
32934  return __ret;
32935}
32936#endif
32937
32938#ifdef __LITTLE_ENDIAN__
32939__ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
32940  float32x2_t __ret;
32941  __ret = (float32x2_t)(__p0);
32942  return __ret;
32943}
32944#else
32945__ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
32946  float32x2_t __ret;
32947  __ret = (float32x2_t)(__p0);
32948  return __ret;
32949}
32950#endif
32951
32952#ifdef __LITTLE_ENDIAN__
32953__ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
32954  float32x2_t __ret;
32955  __ret = (float32x2_t)(__p0);
32956  return __ret;
32957}
32958#else
32959__ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
32960  float32x2_t __ret;
32961  __ret = (float32x2_t)(__p0);
32962  return __ret;
32963}
32964#endif
32965
32966#ifdef __LITTLE_ENDIAN__
32967__ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
32968  float32x2_t __ret;
32969  __ret = (float32x2_t)(__p0);
32970  return __ret;
32971}
32972#else
32973__ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
32974  float32x2_t __ret;
32975  __ret = (float32x2_t)(__p0);
32976  return __ret;
32977}
32978#endif
32979
32980#ifdef __LITTLE_ENDIAN__
32981__ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
32982  float32x2_t __ret;
32983  __ret = (float32x2_t)(__p0);
32984  return __ret;
32985}
32986#else
32987__ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
32988  float32x2_t __ret;
32989  __ret = (float32x2_t)(__p0);
32990  return __ret;
32991}
32992#endif
32993
32994#ifdef __LITTLE_ENDIAN__
32995__ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
32996  float32x2_t __ret;
32997  __ret = (float32x2_t)(__p0);
32998  return __ret;
32999}
33000#else
33001__ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
33002  float32x2_t __ret;
33003  __ret = (float32x2_t)(__p0);
33004  return __ret;
33005}
33006#endif
33007
33008#ifdef __LITTLE_ENDIAN__
33009__ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
33010  float16x4_t __ret;
33011  __ret = (float16x4_t)(__p0);
33012  return __ret;
33013}
33014#else
33015__ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
33016  float16x4_t __ret;
33017  __ret = (float16x4_t)(__p0);
33018  return __ret;
33019}
33020#endif
33021
33022#ifdef __LITTLE_ENDIAN__
33023__ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
33024  float16x4_t __ret;
33025  __ret = (float16x4_t)(__p0);
33026  return __ret;
33027}
33028#else
33029__ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
33030  float16x4_t __ret;
33031  __ret = (float16x4_t)(__p0);
33032  return __ret;
33033}
33034#endif
33035
33036#ifdef __LITTLE_ENDIAN__
33037__ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
33038  float16x4_t __ret;
33039  __ret = (float16x4_t)(__p0);
33040  return __ret;
33041}
33042#else
33043__ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
33044  float16x4_t __ret;
33045  __ret = (float16x4_t)(__p0);
33046  return __ret;
33047}
33048#endif
33049
33050#ifdef __LITTLE_ENDIAN__
33051__ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
33052  float16x4_t __ret;
33053  __ret = (float16x4_t)(__p0);
33054  return __ret;
33055}
33056#else
33057__ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
33058  float16x4_t __ret;
33059  __ret = (float16x4_t)(__p0);
33060  return __ret;
33061}
33062#endif
33063
33064#ifdef __LITTLE_ENDIAN__
33065__ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
33066  float16x4_t __ret;
33067  __ret = (float16x4_t)(__p0);
33068  return __ret;
33069}
33070#else
33071__ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
33072  float16x4_t __ret;
33073  __ret = (float16x4_t)(__p0);
33074  return __ret;
33075}
33076#endif
33077
33078#ifdef __LITTLE_ENDIAN__
33079__ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
33080  float16x4_t __ret;
33081  __ret = (float16x4_t)(__p0);
33082  return __ret;
33083}
33084#else
33085__ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
33086  float16x4_t __ret;
33087  __ret = (float16x4_t)(__p0);
33088  return __ret;
33089}
33090#endif
33091
33092#ifdef __LITTLE_ENDIAN__
33093__ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
33094  float16x4_t __ret;
33095  __ret = (float16x4_t)(__p0);
33096  return __ret;
33097}
33098#else
33099__ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
33100  float16x4_t __ret;
33101  __ret = (float16x4_t)(__p0);
33102  return __ret;
33103}
33104#endif
33105
33106#ifdef __LITTLE_ENDIAN__
33107__ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
33108  float16x4_t __ret;
33109  __ret = (float16x4_t)(__p0);
33110  return __ret;
33111}
33112#else
33113__ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
33114  float16x4_t __ret;
33115  __ret = (float16x4_t)(__p0);
33116  return __ret;
33117}
33118#endif
33119
33120#ifdef __LITTLE_ENDIAN__
33121__ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
33122  float16x4_t __ret;
33123  __ret = (float16x4_t)(__p0);
33124  return __ret;
33125}
33126#else
33127__ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
33128  float16x4_t __ret;
33129  __ret = (float16x4_t)(__p0);
33130  return __ret;
33131}
33132#endif
33133
33134#ifdef __LITTLE_ENDIAN__
33135__ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
33136  float16x4_t __ret;
33137  __ret = (float16x4_t)(__p0);
33138  return __ret;
33139}
33140#else
33141__ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
33142  float16x4_t __ret;
33143  __ret = (float16x4_t)(__p0);
33144  return __ret;
33145}
33146#endif
33147
33148#ifdef __LITTLE_ENDIAN__
33149__ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
33150  float16x4_t __ret;
33151  __ret = (float16x4_t)(__p0);
33152  return __ret;
33153}
33154#else
33155__ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
33156  float16x4_t __ret;
33157  __ret = (float16x4_t)(__p0);
33158  return __ret;
33159}
33160#endif
33161
33162#ifdef __LITTLE_ENDIAN__
33163__ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
33164  int32x2_t __ret;
33165  __ret = (int32x2_t)(__p0);
33166  return __ret;
33167}
33168#else
33169__ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
33170  int32x2_t __ret;
33171  __ret = (int32x2_t)(__p0);
33172  return __ret;
33173}
33174#endif
33175
33176#ifdef __LITTLE_ENDIAN__
33177__ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
33178  int32x2_t __ret;
33179  __ret = (int32x2_t)(__p0);
33180  return __ret;
33181}
33182#else
33183__ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
33184  int32x2_t __ret;
33185  __ret = (int32x2_t)(__p0);
33186  return __ret;
33187}
33188#endif
33189
33190#ifdef __LITTLE_ENDIAN__
33191__ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
33192  int32x2_t __ret;
33193  __ret = (int32x2_t)(__p0);
33194  return __ret;
33195}
33196#else
33197__ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
33198  int32x2_t __ret;
33199  __ret = (int32x2_t)(__p0);
33200  return __ret;
33201}
33202#endif
33203
33204#ifdef __LITTLE_ENDIAN__
33205__ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
33206  int32x2_t __ret;
33207  __ret = (int32x2_t)(__p0);
33208  return __ret;
33209}
33210#else
33211__ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
33212  int32x2_t __ret;
33213  __ret = (int32x2_t)(__p0);
33214  return __ret;
33215}
33216#endif
33217
33218#ifdef __LITTLE_ENDIAN__
33219__ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
33220  int32x2_t __ret;
33221  __ret = (int32x2_t)(__p0);
33222  return __ret;
33223}
33224#else
33225__ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
33226  int32x2_t __ret;
33227  __ret = (int32x2_t)(__p0);
33228  return __ret;
33229}
33230#endif
33231
33232#ifdef __LITTLE_ENDIAN__
33233__ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
33234  int32x2_t __ret;
33235  __ret = (int32x2_t)(__p0);
33236  return __ret;
33237}
33238#else
33239__ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
33240  int32x2_t __ret;
33241  __ret = (int32x2_t)(__p0);
33242  return __ret;
33243}
33244#endif
33245
33246#ifdef __LITTLE_ENDIAN__
33247__ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
33248  int32x2_t __ret;
33249  __ret = (int32x2_t)(__p0);
33250  return __ret;
33251}
33252#else
33253__ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
33254  int32x2_t __ret;
33255  __ret = (int32x2_t)(__p0);
33256  return __ret;
33257}
33258#endif
33259
33260#ifdef __LITTLE_ENDIAN__
33261__ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
33262  int32x2_t __ret;
33263  __ret = (int32x2_t)(__p0);
33264  return __ret;
33265}
33266#else
33267__ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
33268  int32x2_t __ret;
33269  __ret = (int32x2_t)(__p0);
33270  return __ret;
33271}
33272#endif
33273
33274#ifdef __LITTLE_ENDIAN__
33275__ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
33276  int32x2_t __ret;
33277  __ret = (int32x2_t)(__p0);
33278  return __ret;
33279}
33280#else
33281__ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
33282  int32x2_t __ret;
33283  __ret = (int32x2_t)(__p0);
33284  return __ret;
33285}
33286#endif
33287
33288#ifdef __LITTLE_ENDIAN__
33289__ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
33290  int32x2_t __ret;
33291  __ret = (int32x2_t)(__p0);
33292  return __ret;
33293}
33294#else
33295__ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
33296  int32x2_t __ret;
33297  __ret = (int32x2_t)(__p0);
33298  return __ret;
33299}
33300#endif
33301
33302#ifdef __LITTLE_ENDIAN__
33303__ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
33304  int32x2_t __ret;
33305  __ret = (int32x2_t)(__p0);
33306  return __ret;
33307}
33308#else
33309__ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
33310  int32x2_t __ret;
33311  __ret = (int32x2_t)(__p0);
33312  return __ret;
33313}
33314#endif
33315
33316#ifdef __LITTLE_ENDIAN__
33317__ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
33318  int64x1_t __ret;
33319  __ret = (int64x1_t)(__p0);
33320  return __ret;
33321}
33322#else
33323__ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
33324  int64x1_t __ret;
33325  __ret = (int64x1_t)(__p0);
33326  return __ret;
33327}
33328#endif
33329
33330#ifdef __LITTLE_ENDIAN__
33331__ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
33332  int64x1_t __ret;
33333  __ret = (int64x1_t)(__p0);
33334  return __ret;
33335}
33336#else
33337__ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
33338  int64x1_t __ret;
33339  __ret = (int64x1_t)(__p0);
33340  return __ret;
33341}
33342#endif
33343
33344#ifdef __LITTLE_ENDIAN__
33345__ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
33346  int64x1_t __ret;
33347  __ret = (int64x1_t)(__p0);
33348  return __ret;
33349}
33350#else
33351__ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
33352  int64x1_t __ret;
33353  __ret = (int64x1_t)(__p0);
33354  return __ret;
33355}
33356#endif
33357
33358#ifdef __LITTLE_ENDIAN__
33359__ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
33360  int64x1_t __ret;
33361  __ret = (int64x1_t)(__p0);
33362  return __ret;
33363}
33364#else
33365__ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
33366  int64x1_t __ret;
33367  __ret = (int64x1_t)(__p0);
33368  return __ret;
33369}
33370#endif
33371
33372#ifdef __LITTLE_ENDIAN__
33373__ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
33374  int64x1_t __ret;
33375  __ret = (int64x1_t)(__p0);
33376  return __ret;
33377}
33378#else
33379__ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
33380  int64x1_t __ret;
33381  __ret = (int64x1_t)(__p0);
33382  return __ret;
33383}
33384#endif
33385
33386#ifdef __LITTLE_ENDIAN__
33387__ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
33388  int64x1_t __ret;
33389  __ret = (int64x1_t)(__p0);
33390  return __ret;
33391}
33392#else
33393__ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
33394  int64x1_t __ret;
33395  __ret = (int64x1_t)(__p0);
33396  return __ret;
33397}
33398#endif
33399
33400#ifdef __LITTLE_ENDIAN__
33401__ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
33402  int64x1_t __ret;
33403  __ret = (int64x1_t)(__p0);
33404  return __ret;
33405}
33406#else
33407__ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
33408  int64x1_t __ret;
33409  __ret = (int64x1_t)(__p0);
33410  return __ret;
33411}
33412#endif
33413
33414#ifdef __LITTLE_ENDIAN__
33415__ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
33416  int64x1_t __ret;
33417  __ret = (int64x1_t)(__p0);
33418  return __ret;
33419}
33420#else
33421__ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
33422  int64x1_t __ret;
33423  __ret = (int64x1_t)(__p0);
33424  return __ret;
33425}
33426#endif
33427
33428#ifdef __LITTLE_ENDIAN__
33429__ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
33430  int64x1_t __ret;
33431  __ret = (int64x1_t)(__p0);
33432  return __ret;
33433}
33434#else
33435__ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
33436  int64x1_t __ret;
33437  __ret = (int64x1_t)(__p0);
33438  return __ret;
33439}
33440#endif
33441
33442#ifdef __LITTLE_ENDIAN__
33443__ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
33444  int64x1_t __ret;
33445  __ret = (int64x1_t)(__p0);
33446  return __ret;
33447}
33448#else
33449__ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
33450  int64x1_t __ret;
33451  __ret = (int64x1_t)(__p0);
33452  return __ret;
33453}
33454#endif
33455
33456#ifdef __LITTLE_ENDIAN__
33457__ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
33458  int64x1_t __ret;
33459  __ret = (int64x1_t)(__p0);
33460  return __ret;
33461}
33462#else
33463__ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
33464  int64x1_t __ret;
33465  __ret = (int64x1_t)(__p0);
33466  return __ret;
33467}
33468#endif
33469
33470#ifdef __LITTLE_ENDIAN__
33471__ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
33472  int16x4_t __ret;
33473  __ret = (int16x4_t)(__p0);
33474  return __ret;
33475}
33476#else
33477__ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
33478  int16x4_t __ret;
33479  __ret = (int16x4_t)(__p0);
33480  return __ret;
33481}
33482#endif
33483
33484#ifdef __LITTLE_ENDIAN__
33485__ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
33486  int16x4_t __ret;
33487  __ret = (int16x4_t)(__p0);
33488  return __ret;
33489}
33490#else
33491__ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
33492  int16x4_t __ret;
33493  __ret = (int16x4_t)(__p0);
33494  return __ret;
33495}
33496#endif
33497
33498#ifdef __LITTLE_ENDIAN__
33499__ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
33500  int16x4_t __ret;
33501  __ret = (int16x4_t)(__p0);
33502  return __ret;
33503}
33504#else
33505__ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
33506  int16x4_t __ret;
33507  __ret = (int16x4_t)(__p0);
33508  return __ret;
33509}
33510#endif
33511
33512#ifdef __LITTLE_ENDIAN__
33513__ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
33514  int16x4_t __ret;
33515  __ret = (int16x4_t)(__p0);
33516  return __ret;
33517}
33518#else
33519__ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
33520  int16x4_t __ret;
33521  __ret = (int16x4_t)(__p0);
33522  return __ret;
33523}
33524#endif
33525
33526#ifdef __LITTLE_ENDIAN__
33527__ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
33528  int16x4_t __ret;
33529  __ret = (int16x4_t)(__p0);
33530  return __ret;
33531}
33532#else
33533__ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
33534  int16x4_t __ret;
33535  __ret = (int16x4_t)(__p0);
33536  return __ret;
33537}
33538#endif
33539
33540#ifdef __LITTLE_ENDIAN__
33541__ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
33542  int16x4_t __ret;
33543  __ret = (int16x4_t)(__p0);
33544  return __ret;
33545}
33546#else
33547__ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
33548  int16x4_t __ret;
33549  __ret = (int16x4_t)(__p0);
33550  return __ret;
33551}
33552#endif
33553
33554#ifdef __LITTLE_ENDIAN__
33555__ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
33556  int16x4_t __ret;
33557  __ret = (int16x4_t)(__p0);
33558  return __ret;
33559}
33560#else
33561__ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
33562  int16x4_t __ret;
33563  __ret = (int16x4_t)(__p0);
33564  return __ret;
33565}
33566#endif
33567
33568#ifdef __LITTLE_ENDIAN__
33569__ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
33570  int16x4_t __ret;
33571  __ret = (int16x4_t)(__p0);
33572  return __ret;
33573}
33574#else
33575__ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
33576  int16x4_t __ret;
33577  __ret = (int16x4_t)(__p0);
33578  return __ret;
33579}
33580#endif
33581
33582#ifdef __LITTLE_ENDIAN__
33583__ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
33584  int16x4_t __ret;
33585  __ret = (int16x4_t)(__p0);
33586  return __ret;
33587}
33588#else
33589__ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
33590  int16x4_t __ret;
33591  __ret = (int16x4_t)(__p0);
33592  return __ret;
33593}
33594#endif
33595
33596#ifdef __LITTLE_ENDIAN__
33597__ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
33598  int16x4_t __ret;
33599  __ret = (int16x4_t)(__p0);
33600  return __ret;
33601}
33602#else
33603__ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
33604  int16x4_t __ret;
33605  __ret = (int16x4_t)(__p0);
33606  return __ret;
33607}
33608#endif
33609
33610#ifdef __LITTLE_ENDIAN__
33611__ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
33612  int16x4_t __ret;
33613  __ret = (int16x4_t)(__p0);
33614  return __ret;
33615}
33616#else
33617__ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
33618  int16x4_t __ret;
33619  __ret = (int16x4_t)(__p0);
33620  return __ret;
33621}
33622#endif
33623
33624#endif
33625#if __ARM_ARCH >= 8
33626#ifdef __LITTLE_ENDIAN__
33627__ai int32x4_t vcvtaq_s32_f32(float32x4_t __p0) {
33628  int32x4_t __ret;
33629  __ret = (int32x4_t) __builtin_neon_vcvtaq_s32_v((int8x16_t)__p0, 34);
33630  return __ret;
33631}
33632#else
33633__ai int32x4_t vcvtaq_s32_f32(float32x4_t __p0) {
33634  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33635  int32x4_t __ret;
33636  __ret = (int32x4_t) __builtin_neon_vcvtaq_s32_v((int8x16_t)__rev0, 34);
33637  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33638  return __ret;
33639}
33640#endif
33641
33642#ifdef __LITTLE_ENDIAN__
33643__ai int32x2_t vcvta_s32_f32(float32x2_t __p0) {
33644  int32x2_t __ret;
33645  __ret = (int32x2_t) __builtin_neon_vcvta_s32_v((int8x8_t)__p0, 2);
33646  return __ret;
33647}
33648#else
33649__ai int32x2_t vcvta_s32_f32(float32x2_t __p0) {
33650  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33651  int32x2_t __ret;
33652  __ret = (int32x2_t) __builtin_neon_vcvta_s32_v((int8x8_t)__rev0, 2);
33653  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33654  return __ret;
33655}
33656#endif
33657
33658#ifdef __LITTLE_ENDIAN__
33659__ai uint32x4_t vcvtaq_u32_f32(float32x4_t __p0) {
33660  uint32x4_t __ret;
33661  __ret = (uint32x4_t) __builtin_neon_vcvtaq_u32_v((int8x16_t)__p0, 50);
33662  return __ret;
33663}
33664#else
33665__ai uint32x4_t vcvtaq_u32_f32(float32x4_t __p0) {
33666  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33667  uint32x4_t __ret;
33668  __ret = (uint32x4_t) __builtin_neon_vcvtaq_u32_v((int8x16_t)__rev0, 50);
33669  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33670  return __ret;
33671}
33672#endif
33673
33674#ifdef __LITTLE_ENDIAN__
33675__ai uint32x2_t vcvta_u32_f32(float32x2_t __p0) {
33676  uint32x2_t __ret;
33677  __ret = (uint32x2_t) __builtin_neon_vcvta_u32_v((int8x8_t)__p0, 18);
33678  return __ret;
33679}
33680#else
33681__ai uint32x2_t vcvta_u32_f32(float32x2_t __p0) {
33682  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33683  uint32x2_t __ret;
33684  __ret = (uint32x2_t) __builtin_neon_vcvta_u32_v((int8x8_t)__rev0, 18);
33685  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33686  return __ret;
33687}
33688#endif
33689
33690#ifdef __LITTLE_ENDIAN__
33691__ai int32x4_t vcvtmq_s32_f32(float32x4_t __p0) {
33692  int32x4_t __ret;
33693  __ret = (int32x4_t) __builtin_neon_vcvtmq_s32_v((int8x16_t)__p0, 34);
33694  return __ret;
33695}
33696#else
33697__ai int32x4_t vcvtmq_s32_f32(float32x4_t __p0) {
33698  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33699  int32x4_t __ret;
33700  __ret = (int32x4_t) __builtin_neon_vcvtmq_s32_v((int8x16_t)__rev0, 34);
33701  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33702  return __ret;
33703}
33704#endif
33705
33706#ifdef __LITTLE_ENDIAN__
33707__ai int32x2_t vcvtm_s32_f32(float32x2_t __p0) {
33708  int32x2_t __ret;
33709  __ret = (int32x2_t) __builtin_neon_vcvtm_s32_v((int8x8_t)__p0, 2);
33710  return __ret;
33711}
33712#else
33713__ai int32x2_t vcvtm_s32_f32(float32x2_t __p0) {
33714  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33715  int32x2_t __ret;
33716  __ret = (int32x2_t) __builtin_neon_vcvtm_s32_v((int8x8_t)__rev0, 2);
33717  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33718  return __ret;
33719}
33720#endif
33721
33722#ifdef __LITTLE_ENDIAN__
33723__ai uint32x4_t vcvtmq_u32_f32(float32x4_t __p0) {
33724  uint32x4_t __ret;
33725  __ret = (uint32x4_t) __builtin_neon_vcvtmq_u32_v((int8x16_t)__p0, 50);
33726  return __ret;
33727}
33728#else
33729__ai uint32x4_t vcvtmq_u32_f32(float32x4_t __p0) {
33730  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33731  uint32x4_t __ret;
33732  __ret = (uint32x4_t) __builtin_neon_vcvtmq_u32_v((int8x16_t)__rev0, 50);
33733  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33734  return __ret;
33735}
33736#endif
33737
33738#ifdef __LITTLE_ENDIAN__
33739__ai uint32x2_t vcvtm_u32_f32(float32x2_t __p0) {
33740  uint32x2_t __ret;
33741  __ret = (uint32x2_t) __builtin_neon_vcvtm_u32_v((int8x8_t)__p0, 18);
33742  return __ret;
33743}
33744#else
33745__ai uint32x2_t vcvtm_u32_f32(float32x2_t __p0) {
33746  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33747  uint32x2_t __ret;
33748  __ret = (uint32x2_t) __builtin_neon_vcvtm_u32_v((int8x8_t)__rev0, 18);
33749  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33750  return __ret;
33751}
33752#endif
33753
33754#ifdef __LITTLE_ENDIAN__
33755__ai int32x4_t vcvtnq_s32_f32(float32x4_t __p0) {
33756  int32x4_t __ret;
33757  __ret = (int32x4_t) __builtin_neon_vcvtnq_s32_v((int8x16_t)__p0, 34);
33758  return __ret;
33759}
33760#else
33761__ai int32x4_t vcvtnq_s32_f32(float32x4_t __p0) {
33762  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33763  int32x4_t __ret;
33764  __ret = (int32x4_t) __builtin_neon_vcvtnq_s32_v((int8x16_t)__rev0, 34);
33765  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33766  return __ret;
33767}
33768#endif
33769
33770#ifdef __LITTLE_ENDIAN__
33771__ai int32x2_t vcvtn_s32_f32(float32x2_t __p0) {
33772  int32x2_t __ret;
33773  __ret = (int32x2_t) __builtin_neon_vcvtn_s32_v((int8x8_t)__p0, 2);
33774  return __ret;
33775}
33776#else
33777__ai int32x2_t vcvtn_s32_f32(float32x2_t __p0) {
33778  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33779  int32x2_t __ret;
33780  __ret = (int32x2_t) __builtin_neon_vcvtn_s32_v((int8x8_t)__rev0, 2);
33781  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33782  return __ret;
33783}
33784#endif
33785
33786#ifdef __LITTLE_ENDIAN__
33787__ai uint32x4_t vcvtnq_u32_f32(float32x4_t __p0) {
33788  uint32x4_t __ret;
33789  __ret = (uint32x4_t) __builtin_neon_vcvtnq_u32_v((int8x16_t)__p0, 50);
33790  return __ret;
33791}
33792#else
33793__ai uint32x4_t vcvtnq_u32_f32(float32x4_t __p0) {
33794  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33795  uint32x4_t __ret;
33796  __ret = (uint32x4_t) __builtin_neon_vcvtnq_u32_v((int8x16_t)__rev0, 50);
33797  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33798  return __ret;
33799}
33800#endif
33801
33802#ifdef __LITTLE_ENDIAN__
33803__ai uint32x2_t vcvtn_u32_f32(float32x2_t __p0) {
33804  uint32x2_t __ret;
33805  __ret = (uint32x2_t) __builtin_neon_vcvtn_u32_v((int8x8_t)__p0, 18);
33806  return __ret;
33807}
33808#else
33809__ai uint32x2_t vcvtn_u32_f32(float32x2_t __p0) {
33810  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33811  uint32x2_t __ret;
33812  __ret = (uint32x2_t) __builtin_neon_vcvtn_u32_v((int8x8_t)__rev0, 18);
33813  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33814  return __ret;
33815}
33816#endif
33817
33818#ifdef __LITTLE_ENDIAN__
33819__ai int32x4_t vcvtpq_s32_f32(float32x4_t __p0) {
33820  int32x4_t __ret;
33821  __ret = (int32x4_t) __builtin_neon_vcvtpq_s32_v((int8x16_t)__p0, 34);
33822  return __ret;
33823}
33824#else
33825__ai int32x4_t vcvtpq_s32_f32(float32x4_t __p0) {
33826  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33827  int32x4_t __ret;
33828  __ret = (int32x4_t) __builtin_neon_vcvtpq_s32_v((int8x16_t)__rev0, 34);
33829  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33830  return __ret;
33831}
33832#endif
33833
33834#ifdef __LITTLE_ENDIAN__
33835__ai int32x2_t vcvtp_s32_f32(float32x2_t __p0) {
33836  int32x2_t __ret;
33837  __ret = (int32x2_t) __builtin_neon_vcvtp_s32_v((int8x8_t)__p0, 2);
33838  return __ret;
33839}
33840#else
33841__ai int32x2_t vcvtp_s32_f32(float32x2_t __p0) {
33842  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33843  int32x2_t __ret;
33844  __ret = (int32x2_t) __builtin_neon_vcvtp_s32_v((int8x8_t)__rev0, 2);
33845  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33846  return __ret;
33847}
33848#endif
33849
33850#ifdef __LITTLE_ENDIAN__
33851__ai uint32x4_t vcvtpq_u32_f32(float32x4_t __p0) {
33852  uint32x4_t __ret;
33853  __ret = (uint32x4_t) __builtin_neon_vcvtpq_u32_v((int8x16_t)__p0, 50);
33854  return __ret;
33855}
33856#else
33857__ai uint32x4_t vcvtpq_u32_f32(float32x4_t __p0) {
33858  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33859  uint32x4_t __ret;
33860  __ret = (uint32x4_t) __builtin_neon_vcvtpq_u32_v((int8x16_t)__rev0, 50);
33861  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33862  return __ret;
33863}
33864#endif
33865
33866#ifdef __LITTLE_ENDIAN__
33867__ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) {
33868  uint32x2_t __ret;
33869  __ret = (uint32x2_t) __builtin_neon_vcvtp_u32_v((int8x8_t)__p0, 18);
33870  return __ret;
33871}
33872#else
33873__ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) {
33874  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33875  uint32x2_t __ret;
33876  __ret = (uint32x2_t) __builtin_neon_vcvtp_u32_v((int8x8_t)__rev0, 18);
33877  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33878  return __ret;
33879}
33880#endif
33881
33882#endif
33883#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_DIRECTED_ROUNDING)
33884#ifdef __LITTLE_ENDIAN__
33885__ai float32x4_t vrndq_f32(float32x4_t __p0) {
33886  float32x4_t __ret;
33887  __ret = (float32x4_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 41);
33888  return __ret;
33889}
33890#else
33891__ai float32x4_t vrndq_f32(float32x4_t __p0) {
33892  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33893  float32x4_t __ret;
33894  __ret = (float32x4_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 41);
33895  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33896  return __ret;
33897}
33898#endif
33899
33900#ifdef __LITTLE_ENDIAN__
33901__ai float32x2_t vrnd_f32(float32x2_t __p0) {
33902  float32x2_t __ret;
33903  __ret = (float32x2_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 9);
33904  return __ret;
33905}
33906#else
33907__ai float32x2_t vrnd_f32(float32x2_t __p0) {
33908  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33909  float32x2_t __ret;
33910  __ret = (float32x2_t) __builtin_neon_vrnd_v((int8x8_t)__rev0, 9);
33911  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33912  return __ret;
33913}
33914#endif
33915
33916#ifdef __LITTLE_ENDIAN__
33917__ai float32x4_t vrndaq_f32(float32x4_t __p0) {
33918  float32x4_t __ret;
33919  __ret = (float32x4_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 41);
33920  return __ret;
33921}
33922#else
33923__ai float32x4_t vrndaq_f32(float32x4_t __p0) {
33924  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33925  float32x4_t __ret;
33926  __ret = (float32x4_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 41);
33927  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33928  return __ret;
33929}
33930#endif
33931
33932#ifdef __LITTLE_ENDIAN__
33933__ai float32x2_t vrnda_f32(float32x2_t __p0) {
33934  float32x2_t __ret;
33935  __ret = (float32x2_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 9);
33936  return __ret;
33937}
33938#else
33939__ai float32x2_t vrnda_f32(float32x2_t __p0) {
33940  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33941  float32x2_t __ret;
33942  __ret = (float32x2_t) __builtin_neon_vrnda_v((int8x8_t)__rev0, 9);
33943  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33944  return __ret;
33945}
33946#endif
33947
33948#ifdef __LITTLE_ENDIAN__
33949__ai float32x4_t vrndmq_f32(float32x4_t __p0) {
33950  float32x4_t __ret;
33951  __ret = (float32x4_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 41);
33952  return __ret;
33953}
33954#else
33955__ai float32x4_t vrndmq_f32(float32x4_t __p0) {
33956  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33957  float32x4_t __ret;
33958  __ret = (float32x4_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 41);
33959  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33960  return __ret;
33961}
33962#endif
33963
33964#ifdef __LITTLE_ENDIAN__
33965__ai float32x2_t vrndm_f32(float32x2_t __p0) {
33966  float32x2_t __ret;
33967  __ret = (float32x2_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 9);
33968  return __ret;
33969}
33970#else
33971__ai float32x2_t vrndm_f32(float32x2_t __p0) {
33972  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
33973  float32x2_t __ret;
33974  __ret = (float32x2_t) __builtin_neon_vrndm_v((int8x8_t)__rev0, 9);
33975  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
33976  return __ret;
33977}
33978#endif
33979
33980#ifdef __LITTLE_ENDIAN__
33981__ai float32x4_t vrndnq_f32(float32x4_t __p0) {
33982  float32x4_t __ret;
33983  __ret = (float32x4_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 41);
33984  return __ret;
33985}
33986#else
33987__ai float32x4_t vrndnq_f32(float32x4_t __p0) {
33988  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
33989  float32x4_t __ret;
33990  __ret = (float32x4_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 41);
33991  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
33992  return __ret;
33993}
33994#endif
33995
33996#ifdef __LITTLE_ENDIAN__
33997__ai float32x2_t vrndn_f32(float32x2_t __p0) {
33998  float32x2_t __ret;
33999  __ret = (float32x2_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 9);
34000  return __ret;
34001}
34002#else
34003__ai float32x2_t vrndn_f32(float32x2_t __p0) {
34004  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34005  float32x2_t __ret;
34006  __ret = (float32x2_t) __builtin_neon_vrndn_v((int8x8_t)__rev0, 9);
34007  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34008  return __ret;
34009}
34010#endif
34011
34012#ifdef __LITTLE_ENDIAN__
34013__ai float32x4_t vrndpq_f32(float32x4_t __p0) {
34014  float32x4_t __ret;
34015  __ret = (float32x4_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 41);
34016  return __ret;
34017}
34018#else
34019__ai float32x4_t vrndpq_f32(float32x4_t __p0) {
34020  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34021  float32x4_t __ret;
34022  __ret = (float32x4_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 41);
34023  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34024  return __ret;
34025}
34026#endif
34027
34028#ifdef __LITTLE_ENDIAN__
34029__ai float32x2_t vrndp_f32(float32x2_t __p0) {
34030  float32x2_t __ret;
34031  __ret = (float32x2_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 9);
34032  return __ret;
34033}
34034#else
34035__ai float32x2_t vrndp_f32(float32x2_t __p0) {
34036  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34037  float32x2_t __ret;
34038  __ret = (float32x2_t) __builtin_neon_vrndp_v((int8x8_t)__rev0, 9);
34039  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34040  return __ret;
34041}
34042#endif
34043
34044#ifdef __LITTLE_ENDIAN__
34045__ai float32x4_t vrndxq_f32(float32x4_t __p0) {
34046  float32x4_t __ret;
34047  __ret = (float32x4_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 41);
34048  return __ret;
34049}
34050#else
34051__ai float32x4_t vrndxq_f32(float32x4_t __p0) {
34052  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34053  float32x4_t __ret;
34054  __ret = (float32x4_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 41);
34055  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34056  return __ret;
34057}
34058#endif
34059
34060#ifdef __LITTLE_ENDIAN__
34061__ai float32x2_t vrndx_f32(float32x2_t __p0) {
34062  float32x2_t __ret;
34063  __ret = (float32x2_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 9);
34064  return __ret;
34065}
34066#else
34067__ai float32x2_t vrndx_f32(float32x2_t __p0) {
34068  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34069  float32x2_t __ret;
34070  __ret = (float32x2_t) __builtin_neon_vrndx_v((int8x8_t)__rev0, 9);
34071  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34072  return __ret;
34073}
34074#endif
34075
34076#endif
34077#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_NUMERIC_MAXMIN)
34078#ifdef __LITTLE_ENDIAN__
34079__ai float32x4_t vmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
34080  float32x4_t __ret;
34081  __ret = (float32x4_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
34082  return __ret;
34083}
34084#else
34085__ai float32x4_t vmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
34086  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34087  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
34088  float32x4_t __ret;
34089  __ret = (float32x4_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
34090  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34091  return __ret;
34092}
34093#endif
34094
34095#ifdef __LITTLE_ENDIAN__
34096__ai float32x2_t vmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
34097  float32x2_t __ret;
34098  __ret = (float32x2_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
34099  return __ret;
34100}
34101#else
34102__ai float32x2_t vmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
34103  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34104  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
34105  float32x2_t __ret;
34106  __ret = (float32x2_t) __builtin_neon_vmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
34107  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34108  return __ret;
34109}
34110#endif
34111
34112#ifdef __LITTLE_ENDIAN__
34113__ai float32x4_t vminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
34114  float32x4_t __ret;
34115  __ret = (float32x4_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
34116  return __ret;
34117}
34118#else
34119__ai float32x4_t vminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
34120  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
34121  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
34122  float32x4_t __ret;
34123  __ret = (float32x4_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
34124  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
34125  return __ret;
34126}
34127#endif
34128
34129#ifdef __LITTLE_ENDIAN__
34130__ai float32x2_t vminnm_f32(float32x2_t __p0, float32x2_t __p1) {
34131  float32x2_t __ret;
34132  __ret = (float32x2_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
34133  return __ret;
34134}
34135#else
34136__ai float32x2_t vminnm_f32(float32x2_t __p0, float32x2_t __p1) {
34137  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34138  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
34139  float32x2_t __ret;
34140  __ret = (float32x2_t) __builtin_neon_vminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
34141  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34142  return __ret;
34143}
34144#endif
34145
34146#endif
34147#if __ARM_ARCH >= 8 && defined(__aarch64__)
34148#ifdef __LITTLE_ENDIAN__
34149__ai int64x2_t vcvtaq_s64_f64(float64x2_t __p0) {
34150  int64x2_t __ret;
34151  __ret = (int64x2_t) __builtin_neon_vcvtaq_s64_v((int8x16_t)__p0, 35);
34152  return __ret;
34153}
34154#else
34155__ai int64x2_t vcvtaq_s64_f64(float64x2_t __p0) {
34156  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34157  int64x2_t __ret;
34158  __ret = (int64x2_t) __builtin_neon_vcvtaq_s64_v((int8x16_t)__rev0, 35);
34159  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34160  return __ret;
34161}
34162#endif
34163
34164#ifdef __LITTLE_ENDIAN__
34165__ai int64x1_t vcvta_s64_f64(float64x1_t __p0) {
34166  int64x1_t __ret;
34167  __ret = (int64x1_t) __builtin_neon_vcvta_s64_v((int8x8_t)__p0, 3);
34168  return __ret;
34169}
34170#else
34171__ai int64x1_t vcvta_s64_f64(float64x1_t __p0) {
34172  int64x1_t __ret;
34173  __ret = (int64x1_t) __builtin_neon_vcvta_s64_v((int8x8_t)__p0, 3);
34174  return __ret;
34175}
34176#endif
34177
34178#ifdef __LITTLE_ENDIAN__
34179__ai uint64x2_t vcvtaq_u64_f64(float64x2_t __p0) {
34180  uint64x2_t __ret;
34181  __ret = (uint64x2_t) __builtin_neon_vcvtaq_u64_v((int8x16_t)__p0, 51);
34182  return __ret;
34183}
34184#else
34185__ai uint64x2_t vcvtaq_u64_f64(float64x2_t __p0) {
34186  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34187  uint64x2_t __ret;
34188  __ret = (uint64x2_t) __builtin_neon_vcvtaq_u64_v((int8x16_t)__rev0, 51);
34189  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34190  return __ret;
34191}
34192#endif
34193
34194#ifdef __LITTLE_ENDIAN__
34195__ai uint64x1_t vcvta_u64_f64(float64x1_t __p0) {
34196  uint64x1_t __ret;
34197  __ret = (uint64x1_t) __builtin_neon_vcvta_u64_v((int8x8_t)__p0, 19);
34198  return __ret;
34199}
34200#else
34201__ai uint64x1_t vcvta_u64_f64(float64x1_t __p0) {
34202  uint64x1_t __ret;
34203  __ret = (uint64x1_t) __builtin_neon_vcvta_u64_v((int8x8_t)__p0, 19);
34204  return __ret;
34205}
34206#endif
34207
34208#ifdef __LITTLE_ENDIAN__
34209__ai int64x2_t vcvtmq_s64_f64(float64x2_t __p0) {
34210  int64x2_t __ret;
34211  __ret = (int64x2_t) __builtin_neon_vcvtmq_s64_v((int8x16_t)__p0, 35);
34212  return __ret;
34213}
34214#else
34215__ai int64x2_t vcvtmq_s64_f64(float64x2_t __p0) {
34216  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34217  int64x2_t __ret;
34218  __ret = (int64x2_t) __builtin_neon_vcvtmq_s64_v((int8x16_t)__rev0, 35);
34219  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34220  return __ret;
34221}
34222#endif
34223
34224#ifdef __LITTLE_ENDIAN__
34225__ai int64x1_t vcvtm_s64_f64(float64x1_t __p0) {
34226  int64x1_t __ret;
34227  __ret = (int64x1_t) __builtin_neon_vcvtm_s64_v((int8x8_t)__p0, 3);
34228  return __ret;
34229}
34230#else
34231__ai int64x1_t vcvtm_s64_f64(float64x1_t __p0) {
34232  int64x1_t __ret;
34233  __ret = (int64x1_t) __builtin_neon_vcvtm_s64_v((int8x8_t)__p0, 3);
34234  return __ret;
34235}
34236#endif
34237
34238#ifdef __LITTLE_ENDIAN__
34239__ai uint64x2_t vcvtmq_u64_f64(float64x2_t __p0) {
34240  uint64x2_t __ret;
34241  __ret = (uint64x2_t) __builtin_neon_vcvtmq_u64_v((int8x16_t)__p0, 51);
34242  return __ret;
34243}
34244#else
34245__ai uint64x2_t vcvtmq_u64_f64(float64x2_t __p0) {
34246  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34247  uint64x2_t __ret;
34248  __ret = (uint64x2_t) __builtin_neon_vcvtmq_u64_v((int8x16_t)__rev0, 51);
34249  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34250  return __ret;
34251}
34252#endif
34253
34254#ifdef __LITTLE_ENDIAN__
34255__ai uint64x1_t vcvtm_u64_f64(float64x1_t __p0) {
34256  uint64x1_t __ret;
34257  __ret = (uint64x1_t) __builtin_neon_vcvtm_u64_v((int8x8_t)__p0, 19);
34258  return __ret;
34259}
34260#else
34261__ai uint64x1_t vcvtm_u64_f64(float64x1_t __p0) {
34262  uint64x1_t __ret;
34263  __ret = (uint64x1_t) __builtin_neon_vcvtm_u64_v((int8x8_t)__p0, 19);
34264  return __ret;
34265}
34266#endif
34267
34268#ifdef __LITTLE_ENDIAN__
34269__ai int64x2_t vcvtnq_s64_f64(float64x2_t __p0) {
34270  int64x2_t __ret;
34271  __ret = (int64x2_t) __builtin_neon_vcvtnq_s64_v((int8x16_t)__p0, 35);
34272  return __ret;
34273}
34274#else
34275__ai int64x2_t vcvtnq_s64_f64(float64x2_t __p0) {
34276  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34277  int64x2_t __ret;
34278  __ret = (int64x2_t) __builtin_neon_vcvtnq_s64_v((int8x16_t)__rev0, 35);
34279  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34280  return __ret;
34281}
34282#endif
34283
34284#ifdef __LITTLE_ENDIAN__
34285__ai int64x1_t vcvtn_s64_f64(float64x1_t __p0) {
34286  int64x1_t __ret;
34287  __ret = (int64x1_t) __builtin_neon_vcvtn_s64_v((int8x8_t)__p0, 3);
34288  return __ret;
34289}
34290#else
34291__ai int64x1_t vcvtn_s64_f64(float64x1_t __p0) {
34292  int64x1_t __ret;
34293  __ret = (int64x1_t) __builtin_neon_vcvtn_s64_v((int8x8_t)__p0, 3);
34294  return __ret;
34295}
34296#endif
34297
34298#ifdef __LITTLE_ENDIAN__
34299__ai uint64x2_t vcvtnq_u64_f64(float64x2_t __p0) {
34300  uint64x2_t __ret;
34301  __ret = (uint64x2_t) __builtin_neon_vcvtnq_u64_v((int8x16_t)__p0, 51);
34302  return __ret;
34303}
34304#else
34305__ai uint64x2_t vcvtnq_u64_f64(float64x2_t __p0) {
34306  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34307  uint64x2_t __ret;
34308  __ret = (uint64x2_t) __builtin_neon_vcvtnq_u64_v((int8x16_t)__rev0, 51);
34309  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34310  return __ret;
34311}
34312#endif
34313
34314#ifdef __LITTLE_ENDIAN__
34315__ai uint64x1_t vcvtn_u64_f64(float64x1_t __p0) {
34316  uint64x1_t __ret;
34317  __ret = (uint64x1_t) __builtin_neon_vcvtn_u64_v((int8x8_t)__p0, 19);
34318  return __ret;
34319}
34320#else
34321__ai uint64x1_t vcvtn_u64_f64(float64x1_t __p0) {
34322  uint64x1_t __ret;
34323  __ret = (uint64x1_t) __builtin_neon_vcvtn_u64_v((int8x8_t)__p0, 19);
34324  return __ret;
34325}
34326#endif
34327
34328#ifdef __LITTLE_ENDIAN__
34329__ai int64x2_t vcvtpq_s64_f64(float64x2_t __p0) {
34330  int64x2_t __ret;
34331  __ret = (int64x2_t) __builtin_neon_vcvtpq_s64_v((int8x16_t)__p0, 35);
34332  return __ret;
34333}
34334#else
34335__ai int64x2_t vcvtpq_s64_f64(float64x2_t __p0) {
34336  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34337  int64x2_t __ret;
34338  __ret = (int64x2_t) __builtin_neon_vcvtpq_s64_v((int8x16_t)__rev0, 35);
34339  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34340  return __ret;
34341}
34342#endif
34343
34344#ifdef __LITTLE_ENDIAN__
34345__ai int64x1_t vcvtp_s64_f64(float64x1_t __p0) {
34346  int64x1_t __ret;
34347  __ret = (int64x1_t) __builtin_neon_vcvtp_s64_v((int8x8_t)__p0, 3);
34348  return __ret;
34349}
34350#else
34351__ai int64x1_t vcvtp_s64_f64(float64x1_t __p0) {
34352  int64x1_t __ret;
34353  __ret = (int64x1_t) __builtin_neon_vcvtp_s64_v((int8x8_t)__p0, 3);
34354  return __ret;
34355}
34356#endif
34357
34358#ifdef __LITTLE_ENDIAN__
34359__ai uint64x2_t vcvtpq_u64_f64(float64x2_t __p0) {
34360  uint64x2_t __ret;
34361  __ret = (uint64x2_t) __builtin_neon_vcvtpq_u64_v((int8x16_t)__p0, 51);
34362  return __ret;
34363}
34364#else
34365__ai uint64x2_t vcvtpq_u64_f64(float64x2_t __p0) {
34366  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
34367  uint64x2_t __ret;
34368  __ret = (uint64x2_t) __builtin_neon_vcvtpq_u64_v((int8x16_t)__rev0, 51);
34369  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
34370  return __ret;
34371}
34372#endif
34373
34374#ifdef __LITTLE_ENDIAN__
34375__ai uint64x1_t vcvtp_u64_f64(float64x1_t __p0) {
34376  uint64x1_t __ret;
34377  __ret = (uint64x1_t) __builtin_neon_vcvtp_u64_v((int8x8_t)__p0, 19);
34378  return __ret;
34379}
34380#else
34381__ai uint64x1_t vcvtp_u64_f64(float64x1_t __p0) {
34382  uint64x1_t __ret;
34383  __ret = (uint64x1_t) __builtin_neon_vcvtp_u64_v((int8x8_t)__p0, 19);
34384  return __ret;
34385}
34386#endif
34387
34388#ifdef __LITTLE_ENDIAN__
34389__ai poly8x8_t vreinterpret_p8_p64(poly64x1_t __p0) {
34390  poly8x8_t __ret;
34391  __ret = (poly8x8_t)(__p0);
34392  return __ret;
34393}
34394#else
34395__ai poly8x8_t vreinterpret_p8_p64(poly64x1_t __p0) {
34396  poly8x8_t __ret;
34397  __ret = (poly8x8_t)(__p0);
34398  return __ret;
34399}
34400#endif
34401
34402#ifdef __LITTLE_ENDIAN__
34403__ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
34404  poly8x8_t __ret;
34405  __ret = (poly8x8_t)(__p0);
34406  return __ret;
34407}
34408#else
34409__ai poly8x8_t vreinterpret_p8_p16(poly16x4_t __p0) {
34410  poly8x8_t __ret;
34411  __ret = (poly8x8_t)(__p0);
34412  return __ret;
34413}
34414#endif
34415
34416#ifdef __LITTLE_ENDIAN__
34417__ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
34418  poly8x8_t __ret;
34419  __ret = (poly8x8_t)(__p0);
34420  return __ret;
34421}
34422#else
34423__ai poly8x8_t vreinterpret_p8_u8(uint8x8_t __p0) {
34424  poly8x8_t __ret;
34425  __ret = (poly8x8_t)(__p0);
34426  return __ret;
34427}
34428#endif
34429
34430#ifdef __LITTLE_ENDIAN__
34431__ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
34432  poly8x8_t __ret;
34433  __ret = (poly8x8_t)(__p0);
34434  return __ret;
34435}
34436#else
34437__ai poly8x8_t vreinterpret_p8_u32(uint32x2_t __p0) {
34438  poly8x8_t __ret;
34439  __ret = (poly8x8_t)(__p0);
34440  return __ret;
34441}
34442#endif
34443
34444#ifdef __LITTLE_ENDIAN__
34445__ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
34446  poly8x8_t __ret;
34447  __ret = (poly8x8_t)(__p0);
34448  return __ret;
34449}
34450#else
34451__ai poly8x8_t vreinterpret_p8_u64(uint64x1_t __p0) {
34452  poly8x8_t __ret;
34453  __ret = (poly8x8_t)(__p0);
34454  return __ret;
34455}
34456#endif
34457
34458#ifdef __LITTLE_ENDIAN__
34459__ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
34460  poly8x8_t __ret;
34461  __ret = (poly8x8_t)(__p0);
34462  return __ret;
34463}
34464#else
34465__ai poly8x8_t vreinterpret_p8_u16(uint16x4_t __p0) {
34466  poly8x8_t __ret;
34467  __ret = (poly8x8_t)(__p0);
34468  return __ret;
34469}
34470#endif
34471
34472#ifdef __LITTLE_ENDIAN__
34473__ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
34474  poly8x8_t __ret;
34475  __ret = (poly8x8_t)(__p0);
34476  return __ret;
34477}
34478#else
34479__ai poly8x8_t vreinterpret_p8_s8(int8x8_t __p0) {
34480  poly8x8_t __ret;
34481  __ret = (poly8x8_t)(__p0);
34482  return __ret;
34483}
34484#endif
34485
34486#ifdef __LITTLE_ENDIAN__
34487__ai poly8x8_t vreinterpret_p8_f64(float64x1_t __p0) {
34488  poly8x8_t __ret;
34489  __ret = (poly8x8_t)(__p0);
34490  return __ret;
34491}
34492#else
34493__ai poly8x8_t vreinterpret_p8_f64(float64x1_t __p0) {
34494  poly8x8_t __ret;
34495  __ret = (poly8x8_t)(__p0);
34496  return __ret;
34497}
34498#endif
34499
34500#ifdef __LITTLE_ENDIAN__
34501__ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
34502  poly8x8_t __ret;
34503  __ret = (poly8x8_t)(__p0);
34504  return __ret;
34505}
34506#else
34507__ai poly8x8_t vreinterpret_p8_f32(float32x2_t __p0) {
34508  poly8x8_t __ret;
34509  __ret = (poly8x8_t)(__p0);
34510  return __ret;
34511}
34512#endif
34513
34514#ifdef __LITTLE_ENDIAN__
34515__ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
34516  poly8x8_t __ret;
34517  __ret = (poly8x8_t)(__p0);
34518  return __ret;
34519}
34520#else
34521__ai poly8x8_t vreinterpret_p8_f16(float16x4_t __p0) {
34522  poly8x8_t __ret;
34523  __ret = (poly8x8_t)(__p0);
34524  return __ret;
34525}
34526#endif
34527
34528#ifdef __LITTLE_ENDIAN__
34529__ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
34530  poly8x8_t __ret;
34531  __ret = (poly8x8_t)(__p0);
34532  return __ret;
34533}
34534#else
34535__ai poly8x8_t vreinterpret_p8_s32(int32x2_t __p0) {
34536  poly8x8_t __ret;
34537  __ret = (poly8x8_t)(__p0);
34538  return __ret;
34539}
34540#endif
34541
34542#ifdef __LITTLE_ENDIAN__
34543__ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
34544  poly8x8_t __ret;
34545  __ret = (poly8x8_t)(__p0);
34546  return __ret;
34547}
34548#else
34549__ai poly8x8_t vreinterpret_p8_s64(int64x1_t __p0) {
34550  poly8x8_t __ret;
34551  __ret = (poly8x8_t)(__p0);
34552  return __ret;
34553}
34554#endif
34555
34556#ifdef __LITTLE_ENDIAN__
34557__ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
34558  poly8x8_t __ret;
34559  __ret = (poly8x8_t)(__p0);
34560  return __ret;
34561}
34562#else
34563__ai poly8x8_t vreinterpret_p8_s16(int16x4_t __p0) {
34564  poly8x8_t __ret;
34565  __ret = (poly8x8_t)(__p0);
34566  return __ret;
34567}
34568#endif
34569
34570#ifdef __LITTLE_ENDIAN__
34571__ai poly64x1_t vreinterpret_p64_p8(poly8x8_t __p0) {
34572  poly64x1_t __ret;
34573  __ret = (poly64x1_t)(__p0);
34574  return __ret;
34575}
34576#else
34577__ai poly64x1_t vreinterpret_p64_p8(poly8x8_t __p0) {
34578  poly64x1_t __ret;
34579  __ret = (poly64x1_t)(__p0);
34580  return __ret;
34581}
34582#endif
34583
34584#ifdef __LITTLE_ENDIAN__
34585__ai poly64x1_t vreinterpret_p64_p16(poly16x4_t __p0) {
34586  poly64x1_t __ret;
34587  __ret = (poly64x1_t)(__p0);
34588  return __ret;
34589}
34590#else
34591__ai poly64x1_t vreinterpret_p64_p16(poly16x4_t __p0) {
34592  poly64x1_t __ret;
34593  __ret = (poly64x1_t)(__p0);
34594  return __ret;
34595}
34596#endif
34597
34598#ifdef __LITTLE_ENDIAN__
34599__ai poly64x1_t vreinterpret_p64_u8(uint8x8_t __p0) {
34600  poly64x1_t __ret;
34601  __ret = (poly64x1_t)(__p0);
34602  return __ret;
34603}
34604#else
34605__ai poly64x1_t vreinterpret_p64_u8(uint8x8_t __p0) {
34606  poly64x1_t __ret;
34607  __ret = (poly64x1_t)(__p0);
34608  return __ret;
34609}
34610#endif
34611
34612#ifdef __LITTLE_ENDIAN__
34613__ai poly64x1_t vreinterpret_p64_u32(uint32x2_t __p0) {
34614  poly64x1_t __ret;
34615  __ret = (poly64x1_t)(__p0);
34616  return __ret;
34617}
34618#else
34619__ai poly64x1_t vreinterpret_p64_u32(uint32x2_t __p0) {
34620  poly64x1_t __ret;
34621  __ret = (poly64x1_t)(__p0);
34622  return __ret;
34623}
34624#endif
34625
34626#ifdef __LITTLE_ENDIAN__
34627__ai poly64x1_t vreinterpret_p64_u64(uint64x1_t __p0) {
34628  poly64x1_t __ret;
34629  __ret = (poly64x1_t)(__p0);
34630  return __ret;
34631}
34632#else
34633__ai poly64x1_t vreinterpret_p64_u64(uint64x1_t __p0) {
34634  poly64x1_t __ret;
34635  __ret = (poly64x1_t)(__p0);
34636  return __ret;
34637}
34638#endif
34639
34640#ifdef __LITTLE_ENDIAN__
34641__ai poly64x1_t vreinterpret_p64_u16(uint16x4_t __p0) {
34642  poly64x1_t __ret;
34643  __ret = (poly64x1_t)(__p0);
34644  return __ret;
34645}
34646#else
34647__ai poly64x1_t vreinterpret_p64_u16(uint16x4_t __p0) {
34648  poly64x1_t __ret;
34649  __ret = (poly64x1_t)(__p0);
34650  return __ret;
34651}
34652#endif
34653
34654#ifdef __LITTLE_ENDIAN__
34655__ai poly64x1_t vreinterpret_p64_s8(int8x8_t __p0) {
34656  poly64x1_t __ret;
34657  __ret = (poly64x1_t)(__p0);
34658  return __ret;
34659}
34660#else
34661__ai poly64x1_t vreinterpret_p64_s8(int8x8_t __p0) {
34662  poly64x1_t __ret;
34663  __ret = (poly64x1_t)(__p0);
34664  return __ret;
34665}
34666#endif
34667
34668#ifdef __LITTLE_ENDIAN__
34669__ai poly64x1_t vreinterpret_p64_f64(float64x1_t __p0) {
34670  poly64x1_t __ret;
34671  __ret = (poly64x1_t)(__p0);
34672  return __ret;
34673}
34674#else
34675__ai poly64x1_t vreinterpret_p64_f64(float64x1_t __p0) {
34676  poly64x1_t __ret;
34677  __ret = (poly64x1_t)(__p0);
34678  return __ret;
34679}
34680#endif
34681
34682#ifdef __LITTLE_ENDIAN__
34683__ai poly64x1_t vreinterpret_p64_f32(float32x2_t __p0) {
34684  poly64x1_t __ret;
34685  __ret = (poly64x1_t)(__p0);
34686  return __ret;
34687}
34688#else
34689__ai poly64x1_t vreinterpret_p64_f32(float32x2_t __p0) {
34690  poly64x1_t __ret;
34691  __ret = (poly64x1_t)(__p0);
34692  return __ret;
34693}
34694#endif
34695
34696#ifdef __LITTLE_ENDIAN__
34697__ai poly64x1_t vreinterpret_p64_f16(float16x4_t __p0) {
34698  poly64x1_t __ret;
34699  __ret = (poly64x1_t)(__p0);
34700  return __ret;
34701}
34702#else
34703__ai poly64x1_t vreinterpret_p64_f16(float16x4_t __p0) {
34704  poly64x1_t __ret;
34705  __ret = (poly64x1_t)(__p0);
34706  return __ret;
34707}
34708#endif
34709
34710#ifdef __LITTLE_ENDIAN__
34711__ai poly64x1_t vreinterpret_p64_s32(int32x2_t __p0) {
34712  poly64x1_t __ret;
34713  __ret = (poly64x1_t)(__p0);
34714  return __ret;
34715}
34716#else
34717__ai poly64x1_t vreinterpret_p64_s32(int32x2_t __p0) {
34718  poly64x1_t __ret;
34719  __ret = (poly64x1_t)(__p0);
34720  return __ret;
34721}
34722#endif
34723
34724#ifdef __LITTLE_ENDIAN__
34725__ai poly64x1_t vreinterpret_p64_s64(int64x1_t __p0) {
34726  poly64x1_t __ret;
34727  __ret = (poly64x1_t)(__p0);
34728  return __ret;
34729}
34730#else
34731__ai poly64x1_t vreinterpret_p64_s64(int64x1_t __p0) {
34732  poly64x1_t __ret;
34733  __ret = (poly64x1_t)(__p0);
34734  return __ret;
34735}
34736#endif
34737
34738#ifdef __LITTLE_ENDIAN__
34739__ai poly64x1_t vreinterpret_p64_s16(int16x4_t __p0) {
34740  poly64x1_t __ret;
34741  __ret = (poly64x1_t)(__p0);
34742  return __ret;
34743}
34744#else
34745__ai poly64x1_t vreinterpret_p64_s16(int16x4_t __p0) {
34746  poly64x1_t __ret;
34747  __ret = (poly64x1_t)(__p0);
34748  return __ret;
34749}
34750#endif
34751
34752#ifdef __LITTLE_ENDIAN__
34753__ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
34754  poly16x4_t __ret;
34755  __ret = (poly16x4_t)(__p0);
34756  return __ret;
34757}
34758#else
34759__ai poly16x4_t vreinterpret_p16_p8(poly8x8_t __p0) {
34760  poly16x4_t __ret;
34761  __ret = (poly16x4_t)(__p0);
34762  return __ret;
34763}
34764#endif
34765
34766#ifdef __LITTLE_ENDIAN__
34767__ai poly16x4_t vreinterpret_p16_p64(poly64x1_t __p0) {
34768  poly16x4_t __ret;
34769  __ret = (poly16x4_t)(__p0);
34770  return __ret;
34771}
34772#else
34773__ai poly16x4_t vreinterpret_p16_p64(poly64x1_t __p0) {
34774  poly16x4_t __ret;
34775  __ret = (poly16x4_t)(__p0);
34776  return __ret;
34777}
34778#endif
34779
34780#ifdef __LITTLE_ENDIAN__
34781__ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
34782  poly16x4_t __ret;
34783  __ret = (poly16x4_t)(__p0);
34784  return __ret;
34785}
34786#else
34787__ai poly16x4_t vreinterpret_p16_u8(uint8x8_t __p0) {
34788  poly16x4_t __ret;
34789  __ret = (poly16x4_t)(__p0);
34790  return __ret;
34791}
34792#endif
34793
34794#ifdef __LITTLE_ENDIAN__
34795__ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
34796  poly16x4_t __ret;
34797  __ret = (poly16x4_t)(__p0);
34798  return __ret;
34799}
34800#else
34801__ai poly16x4_t vreinterpret_p16_u32(uint32x2_t __p0) {
34802  poly16x4_t __ret;
34803  __ret = (poly16x4_t)(__p0);
34804  return __ret;
34805}
34806#endif
34807
34808#ifdef __LITTLE_ENDIAN__
34809__ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
34810  poly16x4_t __ret;
34811  __ret = (poly16x4_t)(__p0);
34812  return __ret;
34813}
34814#else
34815__ai poly16x4_t vreinterpret_p16_u64(uint64x1_t __p0) {
34816  poly16x4_t __ret;
34817  __ret = (poly16x4_t)(__p0);
34818  return __ret;
34819}
34820#endif
34821
34822#ifdef __LITTLE_ENDIAN__
34823__ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
34824  poly16x4_t __ret;
34825  __ret = (poly16x4_t)(__p0);
34826  return __ret;
34827}
34828#else
34829__ai poly16x4_t vreinterpret_p16_u16(uint16x4_t __p0) {
34830  poly16x4_t __ret;
34831  __ret = (poly16x4_t)(__p0);
34832  return __ret;
34833}
34834#endif
34835
34836#ifdef __LITTLE_ENDIAN__
34837__ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
34838  poly16x4_t __ret;
34839  __ret = (poly16x4_t)(__p0);
34840  return __ret;
34841}
34842#else
34843__ai poly16x4_t vreinterpret_p16_s8(int8x8_t __p0) {
34844  poly16x4_t __ret;
34845  __ret = (poly16x4_t)(__p0);
34846  return __ret;
34847}
34848#endif
34849
34850#ifdef __LITTLE_ENDIAN__
34851__ai poly16x4_t vreinterpret_p16_f64(float64x1_t __p0) {
34852  poly16x4_t __ret;
34853  __ret = (poly16x4_t)(__p0);
34854  return __ret;
34855}
34856#else
34857__ai poly16x4_t vreinterpret_p16_f64(float64x1_t __p0) {
34858  poly16x4_t __ret;
34859  __ret = (poly16x4_t)(__p0);
34860  return __ret;
34861}
34862#endif
34863
34864#ifdef __LITTLE_ENDIAN__
34865__ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
34866  poly16x4_t __ret;
34867  __ret = (poly16x4_t)(__p0);
34868  return __ret;
34869}
34870#else
34871__ai poly16x4_t vreinterpret_p16_f32(float32x2_t __p0) {
34872  poly16x4_t __ret;
34873  __ret = (poly16x4_t)(__p0);
34874  return __ret;
34875}
34876#endif
34877
34878#ifdef __LITTLE_ENDIAN__
34879__ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
34880  poly16x4_t __ret;
34881  __ret = (poly16x4_t)(__p0);
34882  return __ret;
34883}
34884#else
34885__ai poly16x4_t vreinterpret_p16_f16(float16x4_t __p0) {
34886  poly16x4_t __ret;
34887  __ret = (poly16x4_t)(__p0);
34888  return __ret;
34889}
34890#endif
34891
34892#ifdef __LITTLE_ENDIAN__
34893__ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
34894  poly16x4_t __ret;
34895  __ret = (poly16x4_t)(__p0);
34896  return __ret;
34897}
34898#else
34899__ai poly16x4_t vreinterpret_p16_s32(int32x2_t __p0) {
34900  poly16x4_t __ret;
34901  __ret = (poly16x4_t)(__p0);
34902  return __ret;
34903}
34904#endif
34905
34906#ifdef __LITTLE_ENDIAN__
34907__ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
34908  poly16x4_t __ret;
34909  __ret = (poly16x4_t)(__p0);
34910  return __ret;
34911}
34912#else
34913__ai poly16x4_t vreinterpret_p16_s64(int64x1_t __p0) {
34914  poly16x4_t __ret;
34915  __ret = (poly16x4_t)(__p0);
34916  return __ret;
34917}
34918#endif
34919
34920#ifdef __LITTLE_ENDIAN__
34921__ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
34922  poly16x4_t __ret;
34923  __ret = (poly16x4_t)(__p0);
34924  return __ret;
34925}
34926#else
34927__ai poly16x4_t vreinterpret_p16_s16(int16x4_t __p0) {
34928  poly16x4_t __ret;
34929  __ret = (poly16x4_t)(__p0);
34930  return __ret;
34931}
34932#endif
34933
34934#ifdef __LITTLE_ENDIAN__
34935__ai poly8x16_t vreinterpretq_p8_p128(poly128_t __p0) {
34936  poly8x16_t __ret;
34937  __ret = (poly8x16_t)(__p0);
34938  return __ret;
34939}
34940#else
34941__ai poly8x16_t vreinterpretq_p8_p128(poly128_t __p0) {
34942  poly8x16_t __ret;
34943  __ret = (poly8x16_t)(__p0);
34944  return __ret;
34945}
34946#endif
34947
34948#ifdef __LITTLE_ENDIAN__
34949__ai poly8x16_t vreinterpretq_p8_p64(poly64x2_t __p0) {
34950  poly8x16_t __ret;
34951  __ret = (poly8x16_t)(__p0);
34952  return __ret;
34953}
34954#else
34955__ai poly8x16_t vreinterpretq_p8_p64(poly64x2_t __p0) {
34956  poly8x16_t __ret;
34957  __ret = (poly8x16_t)(__p0);
34958  return __ret;
34959}
34960#endif
34961
34962#ifdef __LITTLE_ENDIAN__
34963__ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
34964  poly8x16_t __ret;
34965  __ret = (poly8x16_t)(__p0);
34966  return __ret;
34967}
34968#else
34969__ai poly8x16_t vreinterpretq_p8_p16(poly16x8_t __p0) {
34970  poly8x16_t __ret;
34971  __ret = (poly8x16_t)(__p0);
34972  return __ret;
34973}
34974#endif
34975
34976#ifdef __LITTLE_ENDIAN__
34977__ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
34978  poly8x16_t __ret;
34979  __ret = (poly8x16_t)(__p0);
34980  return __ret;
34981}
34982#else
34983__ai poly8x16_t vreinterpretq_p8_u8(uint8x16_t __p0) {
34984  poly8x16_t __ret;
34985  __ret = (poly8x16_t)(__p0);
34986  return __ret;
34987}
34988#endif
34989
34990#ifdef __LITTLE_ENDIAN__
34991__ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
34992  poly8x16_t __ret;
34993  __ret = (poly8x16_t)(__p0);
34994  return __ret;
34995}
34996#else
34997__ai poly8x16_t vreinterpretq_p8_u32(uint32x4_t __p0) {
34998  poly8x16_t __ret;
34999  __ret = (poly8x16_t)(__p0);
35000  return __ret;
35001}
35002#endif
35003
35004#ifdef __LITTLE_ENDIAN__
35005__ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
35006  poly8x16_t __ret;
35007  __ret = (poly8x16_t)(__p0);
35008  return __ret;
35009}
35010#else
35011__ai poly8x16_t vreinterpretq_p8_u64(uint64x2_t __p0) {
35012  poly8x16_t __ret;
35013  __ret = (poly8x16_t)(__p0);
35014  return __ret;
35015}
35016#endif
35017
35018#ifdef __LITTLE_ENDIAN__
35019__ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
35020  poly8x16_t __ret;
35021  __ret = (poly8x16_t)(__p0);
35022  return __ret;
35023}
35024#else
35025__ai poly8x16_t vreinterpretq_p8_u16(uint16x8_t __p0) {
35026  poly8x16_t __ret;
35027  __ret = (poly8x16_t)(__p0);
35028  return __ret;
35029}
35030#endif
35031
35032#ifdef __LITTLE_ENDIAN__
35033__ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
35034  poly8x16_t __ret;
35035  __ret = (poly8x16_t)(__p0);
35036  return __ret;
35037}
35038#else
35039__ai poly8x16_t vreinterpretq_p8_s8(int8x16_t __p0) {
35040  poly8x16_t __ret;
35041  __ret = (poly8x16_t)(__p0);
35042  return __ret;
35043}
35044#endif
35045
35046#ifdef __LITTLE_ENDIAN__
35047__ai poly8x16_t vreinterpretq_p8_f64(float64x2_t __p0) {
35048  poly8x16_t __ret;
35049  __ret = (poly8x16_t)(__p0);
35050  return __ret;
35051}
35052#else
35053__ai poly8x16_t vreinterpretq_p8_f64(float64x2_t __p0) {
35054  poly8x16_t __ret;
35055  __ret = (poly8x16_t)(__p0);
35056  return __ret;
35057}
35058#endif
35059
35060#ifdef __LITTLE_ENDIAN__
35061__ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
35062  poly8x16_t __ret;
35063  __ret = (poly8x16_t)(__p0);
35064  return __ret;
35065}
35066#else
35067__ai poly8x16_t vreinterpretq_p8_f32(float32x4_t __p0) {
35068  poly8x16_t __ret;
35069  __ret = (poly8x16_t)(__p0);
35070  return __ret;
35071}
35072#endif
35073
35074#ifdef __LITTLE_ENDIAN__
35075__ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
35076  poly8x16_t __ret;
35077  __ret = (poly8x16_t)(__p0);
35078  return __ret;
35079}
35080#else
35081__ai poly8x16_t vreinterpretq_p8_f16(float16x8_t __p0) {
35082  poly8x16_t __ret;
35083  __ret = (poly8x16_t)(__p0);
35084  return __ret;
35085}
35086#endif
35087
35088#ifdef __LITTLE_ENDIAN__
35089__ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
35090  poly8x16_t __ret;
35091  __ret = (poly8x16_t)(__p0);
35092  return __ret;
35093}
35094#else
35095__ai poly8x16_t vreinterpretq_p8_s32(int32x4_t __p0) {
35096  poly8x16_t __ret;
35097  __ret = (poly8x16_t)(__p0);
35098  return __ret;
35099}
35100#endif
35101
35102#ifdef __LITTLE_ENDIAN__
35103__ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
35104  poly8x16_t __ret;
35105  __ret = (poly8x16_t)(__p0);
35106  return __ret;
35107}
35108#else
35109__ai poly8x16_t vreinterpretq_p8_s64(int64x2_t __p0) {
35110  poly8x16_t __ret;
35111  __ret = (poly8x16_t)(__p0);
35112  return __ret;
35113}
35114#endif
35115
35116#ifdef __LITTLE_ENDIAN__
35117__ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
35118  poly8x16_t __ret;
35119  __ret = (poly8x16_t)(__p0);
35120  return __ret;
35121}
35122#else
35123__ai poly8x16_t vreinterpretq_p8_s16(int16x8_t __p0) {
35124  poly8x16_t __ret;
35125  __ret = (poly8x16_t)(__p0);
35126  return __ret;
35127}
35128#endif
35129
35130#ifdef __LITTLE_ENDIAN__
35131__ai poly128_t vreinterpretq_p128_p8(poly8x16_t __p0) {
35132  poly128_t __ret;
35133  __ret = (poly128_t)(__p0);
35134  return __ret;
35135}
35136#else
35137__ai poly128_t vreinterpretq_p128_p8(poly8x16_t __p0) {
35138  poly128_t __ret;
35139  __ret = (poly128_t)(__p0);
35140  return __ret;
35141}
35142#endif
35143
35144#ifdef __LITTLE_ENDIAN__
35145__ai poly128_t vreinterpretq_p128_p64(poly64x2_t __p0) {
35146  poly128_t __ret;
35147  __ret = (poly128_t)(__p0);
35148  return __ret;
35149}
35150#else
35151__ai poly128_t vreinterpretq_p128_p64(poly64x2_t __p0) {
35152  poly128_t __ret;
35153  __ret = (poly128_t)(__p0);
35154  return __ret;
35155}
35156#endif
35157
35158#ifdef __LITTLE_ENDIAN__
35159__ai poly128_t vreinterpretq_p128_p16(poly16x8_t __p0) {
35160  poly128_t __ret;
35161  __ret = (poly128_t)(__p0);
35162  return __ret;
35163}
35164#else
35165__ai poly128_t vreinterpretq_p128_p16(poly16x8_t __p0) {
35166  poly128_t __ret;
35167  __ret = (poly128_t)(__p0);
35168  return __ret;
35169}
35170#endif
35171
35172#ifdef __LITTLE_ENDIAN__
35173__ai poly128_t vreinterpretq_p128_u8(uint8x16_t __p0) {
35174  poly128_t __ret;
35175  __ret = (poly128_t)(__p0);
35176  return __ret;
35177}
35178#else
35179__ai poly128_t vreinterpretq_p128_u8(uint8x16_t __p0) {
35180  poly128_t __ret;
35181  __ret = (poly128_t)(__p0);
35182  return __ret;
35183}
35184#endif
35185
35186#ifdef __LITTLE_ENDIAN__
35187__ai poly128_t vreinterpretq_p128_u32(uint32x4_t __p0) {
35188  poly128_t __ret;
35189  __ret = (poly128_t)(__p0);
35190  return __ret;
35191}
35192#else
35193__ai poly128_t vreinterpretq_p128_u32(uint32x4_t __p0) {
35194  poly128_t __ret;
35195  __ret = (poly128_t)(__p0);
35196  return __ret;
35197}
35198#endif
35199
35200#ifdef __LITTLE_ENDIAN__
35201__ai poly128_t vreinterpretq_p128_u64(uint64x2_t __p0) {
35202  poly128_t __ret;
35203  __ret = (poly128_t)(__p0);
35204  return __ret;
35205}
35206#else
35207__ai poly128_t vreinterpretq_p128_u64(uint64x2_t __p0) {
35208  poly128_t __ret;
35209  __ret = (poly128_t)(__p0);
35210  return __ret;
35211}
35212#endif
35213
35214#ifdef __LITTLE_ENDIAN__
35215__ai poly128_t vreinterpretq_p128_u16(uint16x8_t __p0) {
35216  poly128_t __ret;
35217  __ret = (poly128_t)(__p0);
35218  return __ret;
35219}
35220#else
35221__ai poly128_t vreinterpretq_p128_u16(uint16x8_t __p0) {
35222  poly128_t __ret;
35223  __ret = (poly128_t)(__p0);
35224  return __ret;
35225}
35226#endif
35227
35228#ifdef __LITTLE_ENDIAN__
35229__ai poly128_t vreinterpretq_p128_s8(int8x16_t __p0) {
35230  poly128_t __ret;
35231  __ret = (poly128_t)(__p0);
35232  return __ret;
35233}
35234#else
35235__ai poly128_t vreinterpretq_p128_s8(int8x16_t __p0) {
35236  poly128_t __ret;
35237  __ret = (poly128_t)(__p0);
35238  return __ret;
35239}
35240#endif
35241
35242#ifdef __LITTLE_ENDIAN__
35243__ai poly128_t vreinterpretq_p128_f64(float64x2_t __p0) {
35244  poly128_t __ret;
35245  __ret = (poly128_t)(__p0);
35246  return __ret;
35247}
35248#else
35249__ai poly128_t vreinterpretq_p128_f64(float64x2_t __p0) {
35250  poly128_t __ret;
35251  __ret = (poly128_t)(__p0);
35252  return __ret;
35253}
35254#endif
35255
35256#ifdef __LITTLE_ENDIAN__
35257__ai poly128_t vreinterpretq_p128_f32(float32x4_t __p0) {
35258  poly128_t __ret;
35259  __ret = (poly128_t)(__p0);
35260  return __ret;
35261}
35262#else
35263__ai poly128_t vreinterpretq_p128_f32(float32x4_t __p0) {
35264  poly128_t __ret;
35265  __ret = (poly128_t)(__p0);
35266  return __ret;
35267}
35268#endif
35269
35270#ifdef __LITTLE_ENDIAN__
35271__ai poly128_t vreinterpretq_p128_f16(float16x8_t __p0) {
35272  poly128_t __ret;
35273  __ret = (poly128_t)(__p0);
35274  return __ret;
35275}
35276#else
35277__ai poly128_t vreinterpretq_p128_f16(float16x8_t __p0) {
35278  poly128_t __ret;
35279  __ret = (poly128_t)(__p0);
35280  return __ret;
35281}
35282#endif
35283
35284#ifdef __LITTLE_ENDIAN__
35285__ai poly128_t vreinterpretq_p128_s32(int32x4_t __p0) {
35286  poly128_t __ret;
35287  __ret = (poly128_t)(__p0);
35288  return __ret;
35289}
35290#else
35291__ai poly128_t vreinterpretq_p128_s32(int32x4_t __p0) {
35292  poly128_t __ret;
35293  __ret = (poly128_t)(__p0);
35294  return __ret;
35295}
35296#endif
35297
35298#ifdef __LITTLE_ENDIAN__
35299__ai poly128_t vreinterpretq_p128_s64(int64x2_t __p0) {
35300  poly128_t __ret;
35301  __ret = (poly128_t)(__p0);
35302  return __ret;
35303}
35304#else
35305__ai poly128_t vreinterpretq_p128_s64(int64x2_t __p0) {
35306  poly128_t __ret;
35307  __ret = (poly128_t)(__p0);
35308  return __ret;
35309}
35310#endif
35311
35312#ifdef __LITTLE_ENDIAN__
35313__ai poly128_t vreinterpretq_p128_s16(int16x8_t __p0) {
35314  poly128_t __ret;
35315  __ret = (poly128_t)(__p0);
35316  return __ret;
35317}
35318#else
35319__ai poly128_t vreinterpretq_p128_s16(int16x8_t __p0) {
35320  poly128_t __ret;
35321  __ret = (poly128_t)(__p0);
35322  return __ret;
35323}
35324#endif
35325
35326#ifdef __LITTLE_ENDIAN__
35327__ai poly64x2_t vreinterpretq_p64_p8(poly8x16_t __p0) {
35328  poly64x2_t __ret;
35329  __ret = (poly64x2_t)(__p0);
35330  return __ret;
35331}
35332#else
35333__ai poly64x2_t vreinterpretq_p64_p8(poly8x16_t __p0) {
35334  poly64x2_t __ret;
35335  __ret = (poly64x2_t)(__p0);
35336  return __ret;
35337}
35338#endif
35339
35340#ifdef __LITTLE_ENDIAN__
35341__ai poly64x2_t vreinterpretq_p64_p128(poly128_t __p0) {
35342  poly64x2_t __ret;
35343  __ret = (poly64x2_t)(__p0);
35344  return __ret;
35345}
35346#else
35347__ai poly64x2_t vreinterpretq_p64_p128(poly128_t __p0) {
35348  poly64x2_t __ret;
35349  __ret = (poly64x2_t)(__p0);
35350  return __ret;
35351}
35352#endif
35353
35354#ifdef __LITTLE_ENDIAN__
35355__ai poly64x2_t vreinterpretq_p64_p16(poly16x8_t __p0) {
35356  poly64x2_t __ret;
35357  __ret = (poly64x2_t)(__p0);
35358  return __ret;
35359}
35360#else
35361__ai poly64x2_t vreinterpretq_p64_p16(poly16x8_t __p0) {
35362  poly64x2_t __ret;
35363  __ret = (poly64x2_t)(__p0);
35364  return __ret;
35365}
35366#endif
35367
35368#ifdef __LITTLE_ENDIAN__
35369__ai poly64x2_t vreinterpretq_p64_u8(uint8x16_t __p0) {
35370  poly64x2_t __ret;
35371  __ret = (poly64x2_t)(__p0);
35372  return __ret;
35373}
35374#else
35375__ai poly64x2_t vreinterpretq_p64_u8(uint8x16_t __p0) {
35376  poly64x2_t __ret;
35377  __ret = (poly64x2_t)(__p0);
35378  return __ret;
35379}
35380#endif
35381
35382#ifdef __LITTLE_ENDIAN__
35383__ai poly64x2_t vreinterpretq_p64_u32(uint32x4_t __p0) {
35384  poly64x2_t __ret;
35385  __ret = (poly64x2_t)(__p0);
35386  return __ret;
35387}
35388#else
35389__ai poly64x2_t vreinterpretq_p64_u32(uint32x4_t __p0) {
35390  poly64x2_t __ret;
35391  __ret = (poly64x2_t)(__p0);
35392  return __ret;
35393}
35394#endif
35395
35396#ifdef __LITTLE_ENDIAN__
35397__ai poly64x2_t vreinterpretq_p64_u64(uint64x2_t __p0) {
35398  poly64x2_t __ret;
35399  __ret = (poly64x2_t)(__p0);
35400  return __ret;
35401}
35402#else
35403__ai poly64x2_t vreinterpretq_p64_u64(uint64x2_t __p0) {
35404  poly64x2_t __ret;
35405  __ret = (poly64x2_t)(__p0);
35406  return __ret;
35407}
35408#endif
35409
35410#ifdef __LITTLE_ENDIAN__
35411__ai poly64x2_t vreinterpretq_p64_u16(uint16x8_t __p0) {
35412  poly64x2_t __ret;
35413  __ret = (poly64x2_t)(__p0);
35414  return __ret;
35415}
35416#else
35417__ai poly64x2_t vreinterpretq_p64_u16(uint16x8_t __p0) {
35418  poly64x2_t __ret;
35419  __ret = (poly64x2_t)(__p0);
35420  return __ret;
35421}
35422#endif
35423
35424#ifdef __LITTLE_ENDIAN__
35425__ai poly64x2_t vreinterpretq_p64_s8(int8x16_t __p0) {
35426  poly64x2_t __ret;
35427  __ret = (poly64x2_t)(__p0);
35428  return __ret;
35429}
35430#else
35431__ai poly64x2_t vreinterpretq_p64_s8(int8x16_t __p0) {
35432  poly64x2_t __ret;
35433  __ret = (poly64x2_t)(__p0);
35434  return __ret;
35435}
35436#endif
35437
35438#ifdef __LITTLE_ENDIAN__
35439__ai poly64x2_t vreinterpretq_p64_f64(float64x2_t __p0) {
35440  poly64x2_t __ret;
35441  __ret = (poly64x2_t)(__p0);
35442  return __ret;
35443}
35444#else
35445__ai poly64x2_t vreinterpretq_p64_f64(float64x2_t __p0) {
35446  poly64x2_t __ret;
35447  __ret = (poly64x2_t)(__p0);
35448  return __ret;
35449}
35450#endif
35451
35452#ifdef __LITTLE_ENDIAN__
35453__ai poly64x2_t vreinterpretq_p64_f32(float32x4_t __p0) {
35454  poly64x2_t __ret;
35455  __ret = (poly64x2_t)(__p0);
35456  return __ret;
35457}
35458#else
35459__ai poly64x2_t vreinterpretq_p64_f32(float32x4_t __p0) {
35460  poly64x2_t __ret;
35461  __ret = (poly64x2_t)(__p0);
35462  return __ret;
35463}
35464#endif
35465
35466#ifdef __LITTLE_ENDIAN__
35467__ai poly64x2_t vreinterpretq_p64_f16(float16x8_t __p0) {
35468  poly64x2_t __ret;
35469  __ret = (poly64x2_t)(__p0);
35470  return __ret;
35471}
35472#else
35473__ai poly64x2_t vreinterpretq_p64_f16(float16x8_t __p0) {
35474  poly64x2_t __ret;
35475  __ret = (poly64x2_t)(__p0);
35476  return __ret;
35477}
35478#endif
35479
35480#ifdef __LITTLE_ENDIAN__
35481__ai poly64x2_t vreinterpretq_p64_s32(int32x4_t __p0) {
35482  poly64x2_t __ret;
35483  __ret = (poly64x2_t)(__p0);
35484  return __ret;
35485}
35486#else
35487__ai poly64x2_t vreinterpretq_p64_s32(int32x4_t __p0) {
35488  poly64x2_t __ret;
35489  __ret = (poly64x2_t)(__p0);
35490  return __ret;
35491}
35492#endif
35493
35494#ifdef __LITTLE_ENDIAN__
35495__ai poly64x2_t vreinterpretq_p64_s64(int64x2_t __p0) {
35496  poly64x2_t __ret;
35497  __ret = (poly64x2_t)(__p0);
35498  return __ret;
35499}
35500#else
35501__ai poly64x2_t vreinterpretq_p64_s64(int64x2_t __p0) {
35502  poly64x2_t __ret;
35503  __ret = (poly64x2_t)(__p0);
35504  return __ret;
35505}
35506#endif
35507
35508#ifdef __LITTLE_ENDIAN__
35509__ai poly64x2_t vreinterpretq_p64_s16(int16x8_t __p0) {
35510  poly64x2_t __ret;
35511  __ret = (poly64x2_t)(__p0);
35512  return __ret;
35513}
35514#else
35515__ai poly64x2_t vreinterpretq_p64_s16(int16x8_t __p0) {
35516  poly64x2_t __ret;
35517  __ret = (poly64x2_t)(__p0);
35518  return __ret;
35519}
35520#endif
35521
35522#ifdef __LITTLE_ENDIAN__
35523__ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
35524  poly16x8_t __ret;
35525  __ret = (poly16x8_t)(__p0);
35526  return __ret;
35527}
35528#else
35529__ai poly16x8_t vreinterpretq_p16_p8(poly8x16_t __p0) {
35530  poly16x8_t __ret;
35531  __ret = (poly16x8_t)(__p0);
35532  return __ret;
35533}
35534#endif
35535
35536#ifdef __LITTLE_ENDIAN__
35537__ai poly16x8_t vreinterpretq_p16_p128(poly128_t __p0) {
35538  poly16x8_t __ret;
35539  __ret = (poly16x8_t)(__p0);
35540  return __ret;
35541}
35542#else
35543__ai poly16x8_t vreinterpretq_p16_p128(poly128_t __p0) {
35544  poly16x8_t __ret;
35545  __ret = (poly16x8_t)(__p0);
35546  return __ret;
35547}
35548#endif
35549
35550#ifdef __LITTLE_ENDIAN__
35551__ai poly16x8_t vreinterpretq_p16_p64(poly64x2_t __p0) {
35552  poly16x8_t __ret;
35553  __ret = (poly16x8_t)(__p0);
35554  return __ret;
35555}
35556#else
35557__ai poly16x8_t vreinterpretq_p16_p64(poly64x2_t __p0) {
35558  poly16x8_t __ret;
35559  __ret = (poly16x8_t)(__p0);
35560  return __ret;
35561}
35562#endif
35563
35564#ifdef __LITTLE_ENDIAN__
35565__ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
35566  poly16x8_t __ret;
35567  __ret = (poly16x8_t)(__p0);
35568  return __ret;
35569}
35570#else
35571__ai poly16x8_t vreinterpretq_p16_u8(uint8x16_t __p0) {
35572  poly16x8_t __ret;
35573  __ret = (poly16x8_t)(__p0);
35574  return __ret;
35575}
35576#endif
35577
35578#ifdef __LITTLE_ENDIAN__
35579__ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
35580  poly16x8_t __ret;
35581  __ret = (poly16x8_t)(__p0);
35582  return __ret;
35583}
35584#else
35585__ai poly16x8_t vreinterpretq_p16_u32(uint32x4_t __p0) {
35586  poly16x8_t __ret;
35587  __ret = (poly16x8_t)(__p0);
35588  return __ret;
35589}
35590#endif
35591
35592#ifdef __LITTLE_ENDIAN__
35593__ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
35594  poly16x8_t __ret;
35595  __ret = (poly16x8_t)(__p0);
35596  return __ret;
35597}
35598#else
35599__ai poly16x8_t vreinterpretq_p16_u64(uint64x2_t __p0) {
35600  poly16x8_t __ret;
35601  __ret = (poly16x8_t)(__p0);
35602  return __ret;
35603}
35604#endif
35605
35606#ifdef __LITTLE_ENDIAN__
35607__ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
35608  poly16x8_t __ret;
35609  __ret = (poly16x8_t)(__p0);
35610  return __ret;
35611}
35612#else
35613__ai poly16x8_t vreinterpretq_p16_u16(uint16x8_t __p0) {
35614  poly16x8_t __ret;
35615  __ret = (poly16x8_t)(__p0);
35616  return __ret;
35617}
35618#endif
35619
35620#ifdef __LITTLE_ENDIAN__
35621__ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
35622  poly16x8_t __ret;
35623  __ret = (poly16x8_t)(__p0);
35624  return __ret;
35625}
35626#else
35627__ai poly16x8_t vreinterpretq_p16_s8(int8x16_t __p0) {
35628  poly16x8_t __ret;
35629  __ret = (poly16x8_t)(__p0);
35630  return __ret;
35631}
35632#endif
35633
35634#ifdef __LITTLE_ENDIAN__
35635__ai poly16x8_t vreinterpretq_p16_f64(float64x2_t __p0) {
35636  poly16x8_t __ret;
35637  __ret = (poly16x8_t)(__p0);
35638  return __ret;
35639}
35640#else
35641__ai poly16x8_t vreinterpretq_p16_f64(float64x2_t __p0) {
35642  poly16x8_t __ret;
35643  __ret = (poly16x8_t)(__p0);
35644  return __ret;
35645}
35646#endif
35647
35648#ifdef __LITTLE_ENDIAN__
35649__ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
35650  poly16x8_t __ret;
35651  __ret = (poly16x8_t)(__p0);
35652  return __ret;
35653}
35654#else
35655__ai poly16x8_t vreinterpretq_p16_f32(float32x4_t __p0) {
35656  poly16x8_t __ret;
35657  __ret = (poly16x8_t)(__p0);
35658  return __ret;
35659}
35660#endif
35661
35662#ifdef __LITTLE_ENDIAN__
35663__ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
35664  poly16x8_t __ret;
35665  __ret = (poly16x8_t)(__p0);
35666  return __ret;
35667}
35668#else
35669__ai poly16x8_t vreinterpretq_p16_f16(float16x8_t __p0) {
35670  poly16x8_t __ret;
35671  __ret = (poly16x8_t)(__p0);
35672  return __ret;
35673}
35674#endif
35675
35676#ifdef __LITTLE_ENDIAN__
35677__ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
35678  poly16x8_t __ret;
35679  __ret = (poly16x8_t)(__p0);
35680  return __ret;
35681}
35682#else
35683__ai poly16x8_t vreinterpretq_p16_s32(int32x4_t __p0) {
35684  poly16x8_t __ret;
35685  __ret = (poly16x8_t)(__p0);
35686  return __ret;
35687}
35688#endif
35689
35690#ifdef __LITTLE_ENDIAN__
35691__ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
35692  poly16x8_t __ret;
35693  __ret = (poly16x8_t)(__p0);
35694  return __ret;
35695}
35696#else
35697__ai poly16x8_t vreinterpretq_p16_s64(int64x2_t __p0) {
35698  poly16x8_t __ret;
35699  __ret = (poly16x8_t)(__p0);
35700  return __ret;
35701}
35702#endif
35703
35704#ifdef __LITTLE_ENDIAN__
35705__ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
35706  poly16x8_t __ret;
35707  __ret = (poly16x8_t)(__p0);
35708  return __ret;
35709}
35710#else
35711__ai poly16x8_t vreinterpretq_p16_s16(int16x8_t __p0) {
35712  poly16x8_t __ret;
35713  __ret = (poly16x8_t)(__p0);
35714  return __ret;
35715}
35716#endif
35717
35718#ifdef __LITTLE_ENDIAN__
35719__ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
35720  uint8x16_t __ret;
35721  __ret = (uint8x16_t)(__p0);
35722  return __ret;
35723}
35724#else
35725__ai uint8x16_t vreinterpretq_u8_p8(poly8x16_t __p0) {
35726  uint8x16_t __ret;
35727  __ret = (uint8x16_t)(__p0);
35728  return __ret;
35729}
35730#endif
35731
35732#ifdef __LITTLE_ENDIAN__
35733__ai uint8x16_t vreinterpretq_u8_p128(poly128_t __p0) {
35734  uint8x16_t __ret;
35735  __ret = (uint8x16_t)(__p0);
35736  return __ret;
35737}
35738#else
35739__ai uint8x16_t vreinterpretq_u8_p128(poly128_t __p0) {
35740  uint8x16_t __ret;
35741  __ret = (uint8x16_t)(__p0);
35742  return __ret;
35743}
35744#endif
35745
35746#ifdef __LITTLE_ENDIAN__
35747__ai uint8x16_t vreinterpretq_u8_p64(poly64x2_t __p0) {
35748  uint8x16_t __ret;
35749  __ret = (uint8x16_t)(__p0);
35750  return __ret;
35751}
35752#else
35753__ai uint8x16_t vreinterpretq_u8_p64(poly64x2_t __p0) {
35754  uint8x16_t __ret;
35755  __ret = (uint8x16_t)(__p0);
35756  return __ret;
35757}
35758#endif
35759
35760#ifdef __LITTLE_ENDIAN__
35761__ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
35762  uint8x16_t __ret;
35763  __ret = (uint8x16_t)(__p0);
35764  return __ret;
35765}
35766#else
35767__ai uint8x16_t vreinterpretq_u8_p16(poly16x8_t __p0) {
35768  uint8x16_t __ret;
35769  __ret = (uint8x16_t)(__p0);
35770  return __ret;
35771}
35772#endif
35773
35774#ifdef __LITTLE_ENDIAN__
35775__ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
35776  uint8x16_t __ret;
35777  __ret = (uint8x16_t)(__p0);
35778  return __ret;
35779}
35780#else
35781__ai uint8x16_t vreinterpretq_u8_u32(uint32x4_t __p0) {
35782  uint8x16_t __ret;
35783  __ret = (uint8x16_t)(__p0);
35784  return __ret;
35785}
35786#endif
35787
35788#ifdef __LITTLE_ENDIAN__
35789__ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
35790  uint8x16_t __ret;
35791  __ret = (uint8x16_t)(__p0);
35792  return __ret;
35793}
35794#else
35795__ai uint8x16_t vreinterpretq_u8_u64(uint64x2_t __p0) {
35796  uint8x16_t __ret;
35797  __ret = (uint8x16_t)(__p0);
35798  return __ret;
35799}
35800#endif
35801
35802#ifdef __LITTLE_ENDIAN__
35803__ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
35804  uint8x16_t __ret;
35805  __ret = (uint8x16_t)(__p0);
35806  return __ret;
35807}
35808#else
35809__ai uint8x16_t vreinterpretq_u8_u16(uint16x8_t __p0) {
35810  uint8x16_t __ret;
35811  __ret = (uint8x16_t)(__p0);
35812  return __ret;
35813}
35814#endif
35815
35816#ifdef __LITTLE_ENDIAN__
35817__ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
35818  uint8x16_t __ret;
35819  __ret = (uint8x16_t)(__p0);
35820  return __ret;
35821}
35822#else
35823__ai uint8x16_t vreinterpretq_u8_s8(int8x16_t __p0) {
35824  uint8x16_t __ret;
35825  __ret = (uint8x16_t)(__p0);
35826  return __ret;
35827}
35828#endif
35829
35830#ifdef __LITTLE_ENDIAN__
35831__ai uint8x16_t vreinterpretq_u8_f64(float64x2_t __p0) {
35832  uint8x16_t __ret;
35833  __ret = (uint8x16_t)(__p0);
35834  return __ret;
35835}
35836#else
35837__ai uint8x16_t vreinterpretq_u8_f64(float64x2_t __p0) {
35838  uint8x16_t __ret;
35839  __ret = (uint8x16_t)(__p0);
35840  return __ret;
35841}
35842#endif
35843
35844#ifdef __LITTLE_ENDIAN__
35845__ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
35846  uint8x16_t __ret;
35847  __ret = (uint8x16_t)(__p0);
35848  return __ret;
35849}
35850#else
35851__ai uint8x16_t vreinterpretq_u8_f32(float32x4_t __p0) {
35852  uint8x16_t __ret;
35853  __ret = (uint8x16_t)(__p0);
35854  return __ret;
35855}
35856#endif
35857
35858#ifdef __LITTLE_ENDIAN__
35859__ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
35860  uint8x16_t __ret;
35861  __ret = (uint8x16_t)(__p0);
35862  return __ret;
35863}
35864#else
35865__ai uint8x16_t vreinterpretq_u8_f16(float16x8_t __p0) {
35866  uint8x16_t __ret;
35867  __ret = (uint8x16_t)(__p0);
35868  return __ret;
35869}
35870#endif
35871
35872#ifdef __LITTLE_ENDIAN__
35873__ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
35874  uint8x16_t __ret;
35875  __ret = (uint8x16_t)(__p0);
35876  return __ret;
35877}
35878#else
35879__ai uint8x16_t vreinterpretq_u8_s32(int32x4_t __p0) {
35880  uint8x16_t __ret;
35881  __ret = (uint8x16_t)(__p0);
35882  return __ret;
35883}
35884#endif
35885
35886#ifdef __LITTLE_ENDIAN__
35887__ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
35888  uint8x16_t __ret;
35889  __ret = (uint8x16_t)(__p0);
35890  return __ret;
35891}
35892#else
35893__ai uint8x16_t vreinterpretq_u8_s64(int64x2_t __p0) {
35894  uint8x16_t __ret;
35895  __ret = (uint8x16_t)(__p0);
35896  return __ret;
35897}
35898#endif
35899
35900#ifdef __LITTLE_ENDIAN__
35901__ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
35902  uint8x16_t __ret;
35903  __ret = (uint8x16_t)(__p0);
35904  return __ret;
35905}
35906#else
35907__ai uint8x16_t vreinterpretq_u8_s16(int16x8_t __p0) {
35908  uint8x16_t __ret;
35909  __ret = (uint8x16_t)(__p0);
35910  return __ret;
35911}
35912#endif
35913
35914#ifdef __LITTLE_ENDIAN__
35915__ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
35916  uint32x4_t __ret;
35917  __ret = (uint32x4_t)(__p0);
35918  return __ret;
35919}
35920#else
35921__ai uint32x4_t vreinterpretq_u32_p8(poly8x16_t __p0) {
35922  uint32x4_t __ret;
35923  __ret = (uint32x4_t)(__p0);
35924  return __ret;
35925}
35926#endif
35927
35928#ifdef __LITTLE_ENDIAN__
35929__ai uint32x4_t vreinterpretq_u32_p128(poly128_t __p0) {
35930  uint32x4_t __ret;
35931  __ret = (uint32x4_t)(__p0);
35932  return __ret;
35933}
35934#else
35935__ai uint32x4_t vreinterpretq_u32_p128(poly128_t __p0) {
35936  uint32x4_t __ret;
35937  __ret = (uint32x4_t)(__p0);
35938  return __ret;
35939}
35940#endif
35941
35942#ifdef __LITTLE_ENDIAN__
35943__ai uint32x4_t vreinterpretq_u32_p64(poly64x2_t __p0) {
35944  uint32x4_t __ret;
35945  __ret = (uint32x4_t)(__p0);
35946  return __ret;
35947}
35948#else
35949__ai uint32x4_t vreinterpretq_u32_p64(poly64x2_t __p0) {
35950  uint32x4_t __ret;
35951  __ret = (uint32x4_t)(__p0);
35952  return __ret;
35953}
35954#endif
35955
35956#ifdef __LITTLE_ENDIAN__
35957__ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
35958  uint32x4_t __ret;
35959  __ret = (uint32x4_t)(__p0);
35960  return __ret;
35961}
35962#else
35963__ai uint32x4_t vreinterpretq_u32_p16(poly16x8_t __p0) {
35964  uint32x4_t __ret;
35965  __ret = (uint32x4_t)(__p0);
35966  return __ret;
35967}
35968#endif
35969
35970#ifdef __LITTLE_ENDIAN__
35971__ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
35972  uint32x4_t __ret;
35973  __ret = (uint32x4_t)(__p0);
35974  return __ret;
35975}
35976#else
35977__ai uint32x4_t vreinterpretq_u32_u8(uint8x16_t __p0) {
35978  uint32x4_t __ret;
35979  __ret = (uint32x4_t)(__p0);
35980  return __ret;
35981}
35982#endif
35983
35984#ifdef __LITTLE_ENDIAN__
35985__ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
35986  uint32x4_t __ret;
35987  __ret = (uint32x4_t)(__p0);
35988  return __ret;
35989}
35990#else
35991__ai uint32x4_t vreinterpretq_u32_u64(uint64x2_t __p0) {
35992  uint32x4_t __ret;
35993  __ret = (uint32x4_t)(__p0);
35994  return __ret;
35995}
35996#endif
35997
35998#ifdef __LITTLE_ENDIAN__
35999__ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
36000  uint32x4_t __ret;
36001  __ret = (uint32x4_t)(__p0);
36002  return __ret;
36003}
36004#else
36005__ai uint32x4_t vreinterpretq_u32_u16(uint16x8_t __p0) {
36006  uint32x4_t __ret;
36007  __ret = (uint32x4_t)(__p0);
36008  return __ret;
36009}
36010#endif
36011
36012#ifdef __LITTLE_ENDIAN__
36013__ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
36014  uint32x4_t __ret;
36015  __ret = (uint32x4_t)(__p0);
36016  return __ret;
36017}
36018#else
36019__ai uint32x4_t vreinterpretq_u32_s8(int8x16_t __p0) {
36020  uint32x4_t __ret;
36021  __ret = (uint32x4_t)(__p0);
36022  return __ret;
36023}
36024#endif
36025
36026#ifdef __LITTLE_ENDIAN__
36027__ai uint32x4_t vreinterpretq_u32_f64(float64x2_t __p0) {
36028  uint32x4_t __ret;
36029  __ret = (uint32x4_t)(__p0);
36030  return __ret;
36031}
36032#else
36033__ai uint32x4_t vreinterpretq_u32_f64(float64x2_t __p0) {
36034  uint32x4_t __ret;
36035  __ret = (uint32x4_t)(__p0);
36036  return __ret;
36037}
36038#endif
36039
36040#ifdef __LITTLE_ENDIAN__
36041__ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
36042  uint32x4_t __ret;
36043  __ret = (uint32x4_t)(__p0);
36044  return __ret;
36045}
36046#else
36047__ai uint32x4_t vreinterpretq_u32_f32(float32x4_t __p0) {
36048  uint32x4_t __ret;
36049  __ret = (uint32x4_t)(__p0);
36050  return __ret;
36051}
36052#endif
36053
36054#ifdef __LITTLE_ENDIAN__
36055__ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
36056  uint32x4_t __ret;
36057  __ret = (uint32x4_t)(__p0);
36058  return __ret;
36059}
36060#else
36061__ai uint32x4_t vreinterpretq_u32_f16(float16x8_t __p0) {
36062  uint32x4_t __ret;
36063  __ret = (uint32x4_t)(__p0);
36064  return __ret;
36065}
36066#endif
36067
36068#ifdef __LITTLE_ENDIAN__
36069__ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
36070  uint32x4_t __ret;
36071  __ret = (uint32x4_t)(__p0);
36072  return __ret;
36073}
36074#else
36075__ai uint32x4_t vreinterpretq_u32_s32(int32x4_t __p0) {
36076  uint32x4_t __ret;
36077  __ret = (uint32x4_t)(__p0);
36078  return __ret;
36079}
36080#endif
36081
36082#ifdef __LITTLE_ENDIAN__
36083__ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
36084  uint32x4_t __ret;
36085  __ret = (uint32x4_t)(__p0);
36086  return __ret;
36087}
36088#else
36089__ai uint32x4_t vreinterpretq_u32_s64(int64x2_t __p0) {
36090  uint32x4_t __ret;
36091  __ret = (uint32x4_t)(__p0);
36092  return __ret;
36093}
36094#endif
36095
36096#ifdef __LITTLE_ENDIAN__
36097__ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
36098  uint32x4_t __ret;
36099  __ret = (uint32x4_t)(__p0);
36100  return __ret;
36101}
36102#else
36103__ai uint32x4_t vreinterpretq_u32_s16(int16x8_t __p0) {
36104  uint32x4_t __ret;
36105  __ret = (uint32x4_t)(__p0);
36106  return __ret;
36107}
36108#endif
36109
36110#ifdef __LITTLE_ENDIAN__
36111__ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
36112  uint64x2_t __ret;
36113  __ret = (uint64x2_t)(__p0);
36114  return __ret;
36115}
36116#else
36117__ai uint64x2_t vreinterpretq_u64_p8(poly8x16_t __p0) {
36118  uint64x2_t __ret;
36119  __ret = (uint64x2_t)(__p0);
36120  return __ret;
36121}
36122#endif
36123
36124#ifdef __LITTLE_ENDIAN__
36125__ai uint64x2_t vreinterpretq_u64_p128(poly128_t __p0) {
36126  uint64x2_t __ret;
36127  __ret = (uint64x2_t)(__p0);
36128  return __ret;
36129}
36130#else
36131__ai uint64x2_t vreinterpretq_u64_p128(poly128_t __p0) {
36132  uint64x2_t __ret;
36133  __ret = (uint64x2_t)(__p0);
36134  return __ret;
36135}
36136#endif
36137
36138#ifdef __LITTLE_ENDIAN__
36139__ai uint64x2_t vreinterpretq_u64_p64(poly64x2_t __p0) {
36140  uint64x2_t __ret;
36141  __ret = (uint64x2_t)(__p0);
36142  return __ret;
36143}
36144#else
36145__ai uint64x2_t vreinterpretq_u64_p64(poly64x2_t __p0) {
36146  uint64x2_t __ret;
36147  __ret = (uint64x2_t)(__p0);
36148  return __ret;
36149}
36150#endif
36151
36152#ifdef __LITTLE_ENDIAN__
36153__ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
36154  uint64x2_t __ret;
36155  __ret = (uint64x2_t)(__p0);
36156  return __ret;
36157}
36158#else
36159__ai uint64x2_t vreinterpretq_u64_p16(poly16x8_t __p0) {
36160  uint64x2_t __ret;
36161  __ret = (uint64x2_t)(__p0);
36162  return __ret;
36163}
36164#endif
36165
36166#ifdef __LITTLE_ENDIAN__
36167__ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
36168  uint64x2_t __ret;
36169  __ret = (uint64x2_t)(__p0);
36170  return __ret;
36171}
36172#else
36173__ai uint64x2_t vreinterpretq_u64_u8(uint8x16_t __p0) {
36174  uint64x2_t __ret;
36175  __ret = (uint64x2_t)(__p0);
36176  return __ret;
36177}
36178#endif
36179
36180#ifdef __LITTLE_ENDIAN__
36181__ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
36182  uint64x2_t __ret;
36183  __ret = (uint64x2_t)(__p0);
36184  return __ret;
36185}
36186#else
36187__ai uint64x2_t vreinterpretq_u64_u32(uint32x4_t __p0) {
36188  uint64x2_t __ret;
36189  __ret = (uint64x2_t)(__p0);
36190  return __ret;
36191}
36192#endif
36193
36194#ifdef __LITTLE_ENDIAN__
36195__ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
36196  uint64x2_t __ret;
36197  __ret = (uint64x2_t)(__p0);
36198  return __ret;
36199}
36200#else
36201__ai uint64x2_t vreinterpretq_u64_u16(uint16x8_t __p0) {
36202  uint64x2_t __ret;
36203  __ret = (uint64x2_t)(__p0);
36204  return __ret;
36205}
36206#endif
36207
36208#ifdef __LITTLE_ENDIAN__
36209__ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
36210  uint64x2_t __ret;
36211  __ret = (uint64x2_t)(__p0);
36212  return __ret;
36213}
36214#else
36215__ai uint64x2_t vreinterpretq_u64_s8(int8x16_t __p0) {
36216  uint64x2_t __ret;
36217  __ret = (uint64x2_t)(__p0);
36218  return __ret;
36219}
36220#endif
36221
36222#ifdef __LITTLE_ENDIAN__
36223__ai uint64x2_t vreinterpretq_u64_f64(float64x2_t __p0) {
36224  uint64x2_t __ret;
36225  __ret = (uint64x2_t)(__p0);
36226  return __ret;
36227}
36228#else
36229__ai uint64x2_t vreinterpretq_u64_f64(float64x2_t __p0) {
36230  uint64x2_t __ret;
36231  __ret = (uint64x2_t)(__p0);
36232  return __ret;
36233}
36234#endif
36235
36236#ifdef __LITTLE_ENDIAN__
36237__ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
36238  uint64x2_t __ret;
36239  __ret = (uint64x2_t)(__p0);
36240  return __ret;
36241}
36242#else
36243__ai uint64x2_t vreinterpretq_u64_f32(float32x4_t __p0) {
36244  uint64x2_t __ret;
36245  __ret = (uint64x2_t)(__p0);
36246  return __ret;
36247}
36248#endif
36249
36250#ifdef __LITTLE_ENDIAN__
36251__ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
36252  uint64x2_t __ret;
36253  __ret = (uint64x2_t)(__p0);
36254  return __ret;
36255}
36256#else
36257__ai uint64x2_t vreinterpretq_u64_f16(float16x8_t __p0) {
36258  uint64x2_t __ret;
36259  __ret = (uint64x2_t)(__p0);
36260  return __ret;
36261}
36262#endif
36263
36264#ifdef __LITTLE_ENDIAN__
36265__ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
36266  uint64x2_t __ret;
36267  __ret = (uint64x2_t)(__p0);
36268  return __ret;
36269}
36270#else
36271__ai uint64x2_t vreinterpretq_u64_s32(int32x4_t __p0) {
36272  uint64x2_t __ret;
36273  __ret = (uint64x2_t)(__p0);
36274  return __ret;
36275}
36276#endif
36277
36278#ifdef __LITTLE_ENDIAN__
36279__ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
36280  uint64x2_t __ret;
36281  __ret = (uint64x2_t)(__p0);
36282  return __ret;
36283}
36284#else
36285__ai uint64x2_t vreinterpretq_u64_s64(int64x2_t __p0) {
36286  uint64x2_t __ret;
36287  __ret = (uint64x2_t)(__p0);
36288  return __ret;
36289}
36290#endif
36291
36292#ifdef __LITTLE_ENDIAN__
36293__ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
36294  uint64x2_t __ret;
36295  __ret = (uint64x2_t)(__p0);
36296  return __ret;
36297}
36298#else
36299__ai uint64x2_t vreinterpretq_u64_s16(int16x8_t __p0) {
36300  uint64x2_t __ret;
36301  __ret = (uint64x2_t)(__p0);
36302  return __ret;
36303}
36304#endif
36305
36306#ifdef __LITTLE_ENDIAN__
36307__ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
36308  uint16x8_t __ret;
36309  __ret = (uint16x8_t)(__p0);
36310  return __ret;
36311}
36312#else
36313__ai uint16x8_t vreinterpretq_u16_p8(poly8x16_t __p0) {
36314  uint16x8_t __ret;
36315  __ret = (uint16x8_t)(__p0);
36316  return __ret;
36317}
36318#endif
36319
36320#ifdef __LITTLE_ENDIAN__
36321__ai uint16x8_t vreinterpretq_u16_p128(poly128_t __p0) {
36322  uint16x8_t __ret;
36323  __ret = (uint16x8_t)(__p0);
36324  return __ret;
36325}
36326#else
36327__ai uint16x8_t vreinterpretq_u16_p128(poly128_t __p0) {
36328  uint16x8_t __ret;
36329  __ret = (uint16x8_t)(__p0);
36330  return __ret;
36331}
36332#endif
36333
36334#ifdef __LITTLE_ENDIAN__
36335__ai uint16x8_t vreinterpretq_u16_p64(poly64x2_t __p0) {
36336  uint16x8_t __ret;
36337  __ret = (uint16x8_t)(__p0);
36338  return __ret;
36339}
36340#else
36341__ai uint16x8_t vreinterpretq_u16_p64(poly64x2_t __p0) {
36342  uint16x8_t __ret;
36343  __ret = (uint16x8_t)(__p0);
36344  return __ret;
36345}
36346#endif
36347
36348#ifdef __LITTLE_ENDIAN__
36349__ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
36350  uint16x8_t __ret;
36351  __ret = (uint16x8_t)(__p0);
36352  return __ret;
36353}
36354#else
36355__ai uint16x8_t vreinterpretq_u16_p16(poly16x8_t __p0) {
36356  uint16x8_t __ret;
36357  __ret = (uint16x8_t)(__p0);
36358  return __ret;
36359}
36360#endif
36361
36362#ifdef __LITTLE_ENDIAN__
36363__ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
36364  uint16x8_t __ret;
36365  __ret = (uint16x8_t)(__p0);
36366  return __ret;
36367}
36368#else
36369__ai uint16x8_t vreinterpretq_u16_u8(uint8x16_t __p0) {
36370  uint16x8_t __ret;
36371  __ret = (uint16x8_t)(__p0);
36372  return __ret;
36373}
36374#endif
36375
36376#ifdef __LITTLE_ENDIAN__
36377__ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
36378  uint16x8_t __ret;
36379  __ret = (uint16x8_t)(__p0);
36380  return __ret;
36381}
36382#else
36383__ai uint16x8_t vreinterpretq_u16_u32(uint32x4_t __p0) {
36384  uint16x8_t __ret;
36385  __ret = (uint16x8_t)(__p0);
36386  return __ret;
36387}
36388#endif
36389
36390#ifdef __LITTLE_ENDIAN__
36391__ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
36392  uint16x8_t __ret;
36393  __ret = (uint16x8_t)(__p0);
36394  return __ret;
36395}
36396#else
36397__ai uint16x8_t vreinterpretq_u16_u64(uint64x2_t __p0) {
36398  uint16x8_t __ret;
36399  __ret = (uint16x8_t)(__p0);
36400  return __ret;
36401}
36402#endif
36403
36404#ifdef __LITTLE_ENDIAN__
36405__ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
36406  uint16x8_t __ret;
36407  __ret = (uint16x8_t)(__p0);
36408  return __ret;
36409}
36410#else
36411__ai uint16x8_t vreinterpretq_u16_s8(int8x16_t __p0) {
36412  uint16x8_t __ret;
36413  __ret = (uint16x8_t)(__p0);
36414  return __ret;
36415}
36416#endif
36417
36418#ifdef __LITTLE_ENDIAN__
36419__ai uint16x8_t vreinterpretq_u16_f64(float64x2_t __p0) {
36420  uint16x8_t __ret;
36421  __ret = (uint16x8_t)(__p0);
36422  return __ret;
36423}
36424#else
36425__ai uint16x8_t vreinterpretq_u16_f64(float64x2_t __p0) {
36426  uint16x8_t __ret;
36427  __ret = (uint16x8_t)(__p0);
36428  return __ret;
36429}
36430#endif
36431
36432#ifdef __LITTLE_ENDIAN__
36433__ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
36434  uint16x8_t __ret;
36435  __ret = (uint16x8_t)(__p0);
36436  return __ret;
36437}
36438#else
36439__ai uint16x8_t vreinterpretq_u16_f32(float32x4_t __p0) {
36440  uint16x8_t __ret;
36441  __ret = (uint16x8_t)(__p0);
36442  return __ret;
36443}
36444#endif
36445
36446#ifdef __LITTLE_ENDIAN__
36447__ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
36448  uint16x8_t __ret;
36449  __ret = (uint16x8_t)(__p0);
36450  return __ret;
36451}
36452#else
36453__ai uint16x8_t vreinterpretq_u16_f16(float16x8_t __p0) {
36454  uint16x8_t __ret;
36455  __ret = (uint16x8_t)(__p0);
36456  return __ret;
36457}
36458#endif
36459
36460#ifdef __LITTLE_ENDIAN__
36461__ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
36462  uint16x8_t __ret;
36463  __ret = (uint16x8_t)(__p0);
36464  return __ret;
36465}
36466#else
36467__ai uint16x8_t vreinterpretq_u16_s32(int32x4_t __p0) {
36468  uint16x8_t __ret;
36469  __ret = (uint16x8_t)(__p0);
36470  return __ret;
36471}
36472#endif
36473
36474#ifdef __LITTLE_ENDIAN__
36475__ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
36476  uint16x8_t __ret;
36477  __ret = (uint16x8_t)(__p0);
36478  return __ret;
36479}
36480#else
36481__ai uint16x8_t vreinterpretq_u16_s64(int64x2_t __p0) {
36482  uint16x8_t __ret;
36483  __ret = (uint16x8_t)(__p0);
36484  return __ret;
36485}
36486#endif
36487
36488#ifdef __LITTLE_ENDIAN__
36489__ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
36490  uint16x8_t __ret;
36491  __ret = (uint16x8_t)(__p0);
36492  return __ret;
36493}
36494#else
36495__ai uint16x8_t vreinterpretq_u16_s16(int16x8_t __p0) {
36496  uint16x8_t __ret;
36497  __ret = (uint16x8_t)(__p0);
36498  return __ret;
36499}
36500#endif
36501
36502#ifdef __LITTLE_ENDIAN__
36503__ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
36504  int8x16_t __ret;
36505  __ret = (int8x16_t)(__p0);
36506  return __ret;
36507}
36508#else
36509__ai int8x16_t vreinterpretq_s8_p8(poly8x16_t __p0) {
36510  int8x16_t __ret;
36511  __ret = (int8x16_t)(__p0);
36512  return __ret;
36513}
36514#endif
36515
36516#ifdef __LITTLE_ENDIAN__
36517__ai int8x16_t vreinterpretq_s8_p128(poly128_t __p0) {
36518  int8x16_t __ret;
36519  __ret = (int8x16_t)(__p0);
36520  return __ret;
36521}
36522#else
36523__ai int8x16_t vreinterpretq_s8_p128(poly128_t __p0) {
36524  int8x16_t __ret;
36525  __ret = (int8x16_t)(__p0);
36526  return __ret;
36527}
36528#endif
36529
36530#ifdef __LITTLE_ENDIAN__
36531__ai int8x16_t vreinterpretq_s8_p64(poly64x2_t __p0) {
36532  int8x16_t __ret;
36533  __ret = (int8x16_t)(__p0);
36534  return __ret;
36535}
36536#else
36537__ai int8x16_t vreinterpretq_s8_p64(poly64x2_t __p0) {
36538  int8x16_t __ret;
36539  __ret = (int8x16_t)(__p0);
36540  return __ret;
36541}
36542#endif
36543
36544#ifdef __LITTLE_ENDIAN__
36545__ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
36546  int8x16_t __ret;
36547  __ret = (int8x16_t)(__p0);
36548  return __ret;
36549}
36550#else
36551__ai int8x16_t vreinterpretq_s8_p16(poly16x8_t __p0) {
36552  int8x16_t __ret;
36553  __ret = (int8x16_t)(__p0);
36554  return __ret;
36555}
36556#endif
36557
36558#ifdef __LITTLE_ENDIAN__
36559__ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
36560  int8x16_t __ret;
36561  __ret = (int8x16_t)(__p0);
36562  return __ret;
36563}
36564#else
36565__ai int8x16_t vreinterpretq_s8_u8(uint8x16_t __p0) {
36566  int8x16_t __ret;
36567  __ret = (int8x16_t)(__p0);
36568  return __ret;
36569}
36570#endif
36571
36572#ifdef __LITTLE_ENDIAN__
36573__ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
36574  int8x16_t __ret;
36575  __ret = (int8x16_t)(__p0);
36576  return __ret;
36577}
36578#else
36579__ai int8x16_t vreinterpretq_s8_u32(uint32x4_t __p0) {
36580  int8x16_t __ret;
36581  __ret = (int8x16_t)(__p0);
36582  return __ret;
36583}
36584#endif
36585
36586#ifdef __LITTLE_ENDIAN__
36587__ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
36588  int8x16_t __ret;
36589  __ret = (int8x16_t)(__p0);
36590  return __ret;
36591}
36592#else
36593__ai int8x16_t vreinterpretq_s8_u64(uint64x2_t __p0) {
36594  int8x16_t __ret;
36595  __ret = (int8x16_t)(__p0);
36596  return __ret;
36597}
36598#endif
36599
36600#ifdef __LITTLE_ENDIAN__
36601__ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
36602  int8x16_t __ret;
36603  __ret = (int8x16_t)(__p0);
36604  return __ret;
36605}
36606#else
36607__ai int8x16_t vreinterpretq_s8_u16(uint16x8_t __p0) {
36608  int8x16_t __ret;
36609  __ret = (int8x16_t)(__p0);
36610  return __ret;
36611}
36612#endif
36613
36614#ifdef __LITTLE_ENDIAN__
36615__ai int8x16_t vreinterpretq_s8_f64(float64x2_t __p0) {
36616  int8x16_t __ret;
36617  __ret = (int8x16_t)(__p0);
36618  return __ret;
36619}
36620#else
36621__ai int8x16_t vreinterpretq_s8_f64(float64x2_t __p0) {
36622  int8x16_t __ret;
36623  __ret = (int8x16_t)(__p0);
36624  return __ret;
36625}
36626#endif
36627
36628#ifdef __LITTLE_ENDIAN__
36629__ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
36630  int8x16_t __ret;
36631  __ret = (int8x16_t)(__p0);
36632  return __ret;
36633}
36634#else
36635__ai int8x16_t vreinterpretq_s8_f32(float32x4_t __p0) {
36636  int8x16_t __ret;
36637  __ret = (int8x16_t)(__p0);
36638  return __ret;
36639}
36640#endif
36641
36642#ifdef __LITTLE_ENDIAN__
36643__ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
36644  int8x16_t __ret;
36645  __ret = (int8x16_t)(__p0);
36646  return __ret;
36647}
36648#else
36649__ai int8x16_t vreinterpretq_s8_f16(float16x8_t __p0) {
36650  int8x16_t __ret;
36651  __ret = (int8x16_t)(__p0);
36652  return __ret;
36653}
36654#endif
36655
36656#ifdef __LITTLE_ENDIAN__
36657__ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
36658  int8x16_t __ret;
36659  __ret = (int8x16_t)(__p0);
36660  return __ret;
36661}
36662#else
36663__ai int8x16_t vreinterpretq_s8_s32(int32x4_t __p0) {
36664  int8x16_t __ret;
36665  __ret = (int8x16_t)(__p0);
36666  return __ret;
36667}
36668#endif
36669
36670#ifdef __LITTLE_ENDIAN__
36671__ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
36672  int8x16_t __ret;
36673  __ret = (int8x16_t)(__p0);
36674  return __ret;
36675}
36676#else
36677__ai int8x16_t vreinterpretq_s8_s64(int64x2_t __p0) {
36678  int8x16_t __ret;
36679  __ret = (int8x16_t)(__p0);
36680  return __ret;
36681}
36682#endif
36683
36684#ifdef __LITTLE_ENDIAN__
36685__ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
36686  int8x16_t __ret;
36687  __ret = (int8x16_t)(__p0);
36688  return __ret;
36689}
36690#else
36691__ai int8x16_t vreinterpretq_s8_s16(int16x8_t __p0) {
36692  int8x16_t __ret;
36693  __ret = (int8x16_t)(__p0);
36694  return __ret;
36695}
36696#endif
36697
36698#ifdef __LITTLE_ENDIAN__
36699__ai float64x2_t vreinterpretq_f64_p8(poly8x16_t __p0) {
36700  float64x2_t __ret;
36701  __ret = (float64x2_t)(__p0);
36702  return __ret;
36703}
36704#else
36705__ai float64x2_t vreinterpretq_f64_p8(poly8x16_t __p0) {
36706  float64x2_t __ret;
36707  __ret = (float64x2_t)(__p0);
36708  return __ret;
36709}
36710#endif
36711
36712#ifdef __LITTLE_ENDIAN__
36713__ai float64x2_t vreinterpretq_f64_p128(poly128_t __p0) {
36714  float64x2_t __ret;
36715  __ret = (float64x2_t)(__p0);
36716  return __ret;
36717}
36718#else
36719__ai float64x2_t vreinterpretq_f64_p128(poly128_t __p0) {
36720  float64x2_t __ret;
36721  __ret = (float64x2_t)(__p0);
36722  return __ret;
36723}
36724#endif
36725
36726#ifdef __LITTLE_ENDIAN__
36727__ai float64x2_t vreinterpretq_f64_p64(poly64x2_t __p0) {
36728  float64x2_t __ret;
36729  __ret = (float64x2_t)(__p0);
36730  return __ret;
36731}
36732#else
36733__ai float64x2_t vreinterpretq_f64_p64(poly64x2_t __p0) {
36734  float64x2_t __ret;
36735  __ret = (float64x2_t)(__p0);
36736  return __ret;
36737}
36738#endif
36739
36740#ifdef __LITTLE_ENDIAN__
36741__ai float64x2_t vreinterpretq_f64_p16(poly16x8_t __p0) {
36742  float64x2_t __ret;
36743  __ret = (float64x2_t)(__p0);
36744  return __ret;
36745}
36746#else
36747__ai float64x2_t vreinterpretq_f64_p16(poly16x8_t __p0) {
36748  float64x2_t __ret;
36749  __ret = (float64x2_t)(__p0);
36750  return __ret;
36751}
36752#endif
36753
36754#ifdef __LITTLE_ENDIAN__
36755__ai float64x2_t vreinterpretq_f64_u8(uint8x16_t __p0) {
36756  float64x2_t __ret;
36757  __ret = (float64x2_t)(__p0);
36758  return __ret;
36759}
36760#else
36761__ai float64x2_t vreinterpretq_f64_u8(uint8x16_t __p0) {
36762  float64x2_t __ret;
36763  __ret = (float64x2_t)(__p0);
36764  return __ret;
36765}
36766#endif
36767
36768#ifdef __LITTLE_ENDIAN__
36769__ai float64x2_t vreinterpretq_f64_u32(uint32x4_t __p0) {
36770  float64x2_t __ret;
36771  __ret = (float64x2_t)(__p0);
36772  return __ret;
36773}
36774#else
36775__ai float64x2_t vreinterpretq_f64_u32(uint32x4_t __p0) {
36776  float64x2_t __ret;
36777  __ret = (float64x2_t)(__p0);
36778  return __ret;
36779}
36780#endif
36781
36782#ifdef __LITTLE_ENDIAN__
36783__ai float64x2_t vreinterpretq_f64_u64(uint64x2_t __p0) {
36784  float64x2_t __ret;
36785  __ret = (float64x2_t)(__p0);
36786  return __ret;
36787}
36788#else
36789__ai float64x2_t vreinterpretq_f64_u64(uint64x2_t __p0) {
36790  float64x2_t __ret;
36791  __ret = (float64x2_t)(__p0);
36792  return __ret;
36793}
36794#endif
36795
36796#ifdef __LITTLE_ENDIAN__
36797__ai float64x2_t vreinterpretq_f64_u16(uint16x8_t __p0) {
36798  float64x2_t __ret;
36799  __ret = (float64x2_t)(__p0);
36800  return __ret;
36801}
36802#else
36803__ai float64x2_t vreinterpretq_f64_u16(uint16x8_t __p0) {
36804  float64x2_t __ret;
36805  __ret = (float64x2_t)(__p0);
36806  return __ret;
36807}
36808#endif
36809
36810#ifdef __LITTLE_ENDIAN__
36811__ai float64x2_t vreinterpretq_f64_s8(int8x16_t __p0) {
36812  float64x2_t __ret;
36813  __ret = (float64x2_t)(__p0);
36814  return __ret;
36815}
36816#else
36817__ai float64x2_t vreinterpretq_f64_s8(int8x16_t __p0) {
36818  float64x2_t __ret;
36819  __ret = (float64x2_t)(__p0);
36820  return __ret;
36821}
36822#endif
36823
36824#ifdef __LITTLE_ENDIAN__
36825__ai float64x2_t vreinterpretq_f64_f32(float32x4_t __p0) {
36826  float64x2_t __ret;
36827  __ret = (float64x2_t)(__p0);
36828  return __ret;
36829}
36830#else
36831__ai float64x2_t vreinterpretq_f64_f32(float32x4_t __p0) {
36832  float64x2_t __ret;
36833  __ret = (float64x2_t)(__p0);
36834  return __ret;
36835}
36836#endif
36837
36838#ifdef __LITTLE_ENDIAN__
36839__ai float64x2_t vreinterpretq_f64_f16(float16x8_t __p0) {
36840  float64x2_t __ret;
36841  __ret = (float64x2_t)(__p0);
36842  return __ret;
36843}
36844#else
36845__ai float64x2_t vreinterpretq_f64_f16(float16x8_t __p0) {
36846  float64x2_t __ret;
36847  __ret = (float64x2_t)(__p0);
36848  return __ret;
36849}
36850#endif
36851
36852#ifdef __LITTLE_ENDIAN__
36853__ai float64x2_t vreinterpretq_f64_s32(int32x4_t __p0) {
36854  float64x2_t __ret;
36855  __ret = (float64x2_t)(__p0);
36856  return __ret;
36857}
36858#else
36859__ai float64x2_t vreinterpretq_f64_s32(int32x4_t __p0) {
36860  float64x2_t __ret;
36861  __ret = (float64x2_t)(__p0);
36862  return __ret;
36863}
36864#endif
36865
36866#ifdef __LITTLE_ENDIAN__
36867__ai float64x2_t vreinterpretq_f64_s64(int64x2_t __p0) {
36868  float64x2_t __ret;
36869  __ret = (float64x2_t)(__p0);
36870  return __ret;
36871}
36872#else
36873__ai float64x2_t vreinterpretq_f64_s64(int64x2_t __p0) {
36874  float64x2_t __ret;
36875  __ret = (float64x2_t)(__p0);
36876  return __ret;
36877}
36878#endif
36879
36880#ifdef __LITTLE_ENDIAN__
36881__ai float64x2_t vreinterpretq_f64_s16(int16x8_t __p0) {
36882  float64x2_t __ret;
36883  __ret = (float64x2_t)(__p0);
36884  return __ret;
36885}
36886#else
36887__ai float64x2_t vreinterpretq_f64_s16(int16x8_t __p0) {
36888  float64x2_t __ret;
36889  __ret = (float64x2_t)(__p0);
36890  return __ret;
36891}
36892#endif
36893
36894#ifdef __LITTLE_ENDIAN__
36895__ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
36896  float32x4_t __ret;
36897  __ret = (float32x4_t)(__p0);
36898  return __ret;
36899}
36900#else
36901__ai float32x4_t vreinterpretq_f32_p8(poly8x16_t __p0) {
36902  float32x4_t __ret;
36903  __ret = (float32x4_t)(__p0);
36904  return __ret;
36905}
36906#endif
36907
36908#ifdef __LITTLE_ENDIAN__
36909__ai float32x4_t vreinterpretq_f32_p128(poly128_t __p0) {
36910  float32x4_t __ret;
36911  __ret = (float32x4_t)(__p0);
36912  return __ret;
36913}
36914#else
36915__ai float32x4_t vreinterpretq_f32_p128(poly128_t __p0) {
36916  float32x4_t __ret;
36917  __ret = (float32x4_t)(__p0);
36918  return __ret;
36919}
36920#endif
36921
36922#ifdef __LITTLE_ENDIAN__
36923__ai float32x4_t vreinterpretq_f32_p64(poly64x2_t __p0) {
36924  float32x4_t __ret;
36925  __ret = (float32x4_t)(__p0);
36926  return __ret;
36927}
36928#else
36929__ai float32x4_t vreinterpretq_f32_p64(poly64x2_t __p0) {
36930  float32x4_t __ret;
36931  __ret = (float32x4_t)(__p0);
36932  return __ret;
36933}
36934#endif
36935
36936#ifdef __LITTLE_ENDIAN__
36937__ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
36938  float32x4_t __ret;
36939  __ret = (float32x4_t)(__p0);
36940  return __ret;
36941}
36942#else
36943__ai float32x4_t vreinterpretq_f32_p16(poly16x8_t __p0) {
36944  float32x4_t __ret;
36945  __ret = (float32x4_t)(__p0);
36946  return __ret;
36947}
36948#endif
36949
36950#ifdef __LITTLE_ENDIAN__
36951__ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
36952  float32x4_t __ret;
36953  __ret = (float32x4_t)(__p0);
36954  return __ret;
36955}
36956#else
36957__ai float32x4_t vreinterpretq_f32_u8(uint8x16_t __p0) {
36958  float32x4_t __ret;
36959  __ret = (float32x4_t)(__p0);
36960  return __ret;
36961}
36962#endif
36963
36964#ifdef __LITTLE_ENDIAN__
36965__ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
36966  float32x4_t __ret;
36967  __ret = (float32x4_t)(__p0);
36968  return __ret;
36969}
36970#else
36971__ai float32x4_t vreinterpretq_f32_u32(uint32x4_t __p0) {
36972  float32x4_t __ret;
36973  __ret = (float32x4_t)(__p0);
36974  return __ret;
36975}
36976#endif
36977
36978#ifdef __LITTLE_ENDIAN__
36979__ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
36980  float32x4_t __ret;
36981  __ret = (float32x4_t)(__p0);
36982  return __ret;
36983}
36984#else
36985__ai float32x4_t vreinterpretq_f32_u64(uint64x2_t __p0) {
36986  float32x4_t __ret;
36987  __ret = (float32x4_t)(__p0);
36988  return __ret;
36989}
36990#endif
36991
36992#ifdef __LITTLE_ENDIAN__
36993__ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
36994  float32x4_t __ret;
36995  __ret = (float32x4_t)(__p0);
36996  return __ret;
36997}
36998#else
36999__ai float32x4_t vreinterpretq_f32_u16(uint16x8_t __p0) {
37000  float32x4_t __ret;
37001  __ret = (float32x4_t)(__p0);
37002  return __ret;
37003}
37004#endif
37005
37006#ifdef __LITTLE_ENDIAN__
37007__ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
37008  float32x4_t __ret;
37009  __ret = (float32x4_t)(__p0);
37010  return __ret;
37011}
37012#else
37013__ai float32x4_t vreinterpretq_f32_s8(int8x16_t __p0) {
37014  float32x4_t __ret;
37015  __ret = (float32x4_t)(__p0);
37016  return __ret;
37017}
37018#endif
37019
37020#ifdef __LITTLE_ENDIAN__
37021__ai float32x4_t vreinterpretq_f32_f64(float64x2_t __p0) {
37022  float32x4_t __ret;
37023  __ret = (float32x4_t)(__p0);
37024  return __ret;
37025}
37026#else
37027__ai float32x4_t vreinterpretq_f32_f64(float64x2_t __p0) {
37028  float32x4_t __ret;
37029  __ret = (float32x4_t)(__p0);
37030  return __ret;
37031}
37032#endif
37033
37034#ifdef __LITTLE_ENDIAN__
37035__ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
37036  float32x4_t __ret;
37037  __ret = (float32x4_t)(__p0);
37038  return __ret;
37039}
37040#else
37041__ai float32x4_t vreinterpretq_f32_f16(float16x8_t __p0) {
37042  float32x4_t __ret;
37043  __ret = (float32x4_t)(__p0);
37044  return __ret;
37045}
37046#endif
37047
37048#ifdef __LITTLE_ENDIAN__
37049__ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
37050  float32x4_t __ret;
37051  __ret = (float32x4_t)(__p0);
37052  return __ret;
37053}
37054#else
37055__ai float32x4_t vreinterpretq_f32_s32(int32x4_t __p0) {
37056  float32x4_t __ret;
37057  __ret = (float32x4_t)(__p0);
37058  return __ret;
37059}
37060#endif
37061
37062#ifdef __LITTLE_ENDIAN__
37063__ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
37064  float32x4_t __ret;
37065  __ret = (float32x4_t)(__p0);
37066  return __ret;
37067}
37068#else
37069__ai float32x4_t vreinterpretq_f32_s64(int64x2_t __p0) {
37070  float32x4_t __ret;
37071  __ret = (float32x4_t)(__p0);
37072  return __ret;
37073}
37074#endif
37075
37076#ifdef __LITTLE_ENDIAN__
37077__ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
37078  float32x4_t __ret;
37079  __ret = (float32x4_t)(__p0);
37080  return __ret;
37081}
37082#else
37083__ai float32x4_t vreinterpretq_f32_s16(int16x8_t __p0) {
37084  float32x4_t __ret;
37085  __ret = (float32x4_t)(__p0);
37086  return __ret;
37087}
37088#endif
37089
37090#ifdef __LITTLE_ENDIAN__
37091__ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
37092  float16x8_t __ret;
37093  __ret = (float16x8_t)(__p0);
37094  return __ret;
37095}
37096#else
37097__ai float16x8_t vreinterpretq_f16_p8(poly8x16_t __p0) {
37098  float16x8_t __ret;
37099  __ret = (float16x8_t)(__p0);
37100  return __ret;
37101}
37102#endif
37103
37104#ifdef __LITTLE_ENDIAN__
37105__ai float16x8_t vreinterpretq_f16_p128(poly128_t __p0) {
37106  float16x8_t __ret;
37107  __ret = (float16x8_t)(__p0);
37108  return __ret;
37109}
37110#else
37111__ai float16x8_t vreinterpretq_f16_p128(poly128_t __p0) {
37112  float16x8_t __ret;
37113  __ret = (float16x8_t)(__p0);
37114  return __ret;
37115}
37116#endif
37117
37118#ifdef __LITTLE_ENDIAN__
37119__ai float16x8_t vreinterpretq_f16_p64(poly64x2_t __p0) {
37120  float16x8_t __ret;
37121  __ret = (float16x8_t)(__p0);
37122  return __ret;
37123}
37124#else
37125__ai float16x8_t vreinterpretq_f16_p64(poly64x2_t __p0) {
37126  float16x8_t __ret;
37127  __ret = (float16x8_t)(__p0);
37128  return __ret;
37129}
37130#endif
37131
37132#ifdef __LITTLE_ENDIAN__
37133__ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
37134  float16x8_t __ret;
37135  __ret = (float16x8_t)(__p0);
37136  return __ret;
37137}
37138#else
37139__ai float16x8_t vreinterpretq_f16_p16(poly16x8_t __p0) {
37140  float16x8_t __ret;
37141  __ret = (float16x8_t)(__p0);
37142  return __ret;
37143}
37144#endif
37145
37146#ifdef __LITTLE_ENDIAN__
37147__ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
37148  float16x8_t __ret;
37149  __ret = (float16x8_t)(__p0);
37150  return __ret;
37151}
37152#else
37153__ai float16x8_t vreinterpretq_f16_u8(uint8x16_t __p0) {
37154  float16x8_t __ret;
37155  __ret = (float16x8_t)(__p0);
37156  return __ret;
37157}
37158#endif
37159
37160#ifdef __LITTLE_ENDIAN__
37161__ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
37162  float16x8_t __ret;
37163  __ret = (float16x8_t)(__p0);
37164  return __ret;
37165}
37166#else
37167__ai float16x8_t vreinterpretq_f16_u32(uint32x4_t __p0) {
37168  float16x8_t __ret;
37169  __ret = (float16x8_t)(__p0);
37170  return __ret;
37171}
37172#endif
37173
37174#ifdef __LITTLE_ENDIAN__
37175__ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
37176  float16x8_t __ret;
37177  __ret = (float16x8_t)(__p0);
37178  return __ret;
37179}
37180#else
37181__ai float16x8_t vreinterpretq_f16_u64(uint64x2_t __p0) {
37182  float16x8_t __ret;
37183  __ret = (float16x8_t)(__p0);
37184  return __ret;
37185}
37186#endif
37187
37188#ifdef __LITTLE_ENDIAN__
37189__ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
37190  float16x8_t __ret;
37191  __ret = (float16x8_t)(__p0);
37192  return __ret;
37193}
37194#else
37195__ai float16x8_t vreinterpretq_f16_u16(uint16x8_t __p0) {
37196  float16x8_t __ret;
37197  __ret = (float16x8_t)(__p0);
37198  return __ret;
37199}
37200#endif
37201
37202#ifdef __LITTLE_ENDIAN__
37203__ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
37204  float16x8_t __ret;
37205  __ret = (float16x8_t)(__p0);
37206  return __ret;
37207}
37208#else
37209__ai float16x8_t vreinterpretq_f16_s8(int8x16_t __p0) {
37210  float16x8_t __ret;
37211  __ret = (float16x8_t)(__p0);
37212  return __ret;
37213}
37214#endif
37215
37216#ifdef __LITTLE_ENDIAN__
37217__ai float16x8_t vreinterpretq_f16_f64(float64x2_t __p0) {
37218  float16x8_t __ret;
37219  __ret = (float16x8_t)(__p0);
37220  return __ret;
37221}
37222#else
37223__ai float16x8_t vreinterpretq_f16_f64(float64x2_t __p0) {
37224  float16x8_t __ret;
37225  __ret = (float16x8_t)(__p0);
37226  return __ret;
37227}
37228#endif
37229
37230#ifdef __LITTLE_ENDIAN__
37231__ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
37232  float16x8_t __ret;
37233  __ret = (float16x8_t)(__p0);
37234  return __ret;
37235}
37236#else
37237__ai float16x8_t vreinterpretq_f16_f32(float32x4_t __p0) {
37238  float16x8_t __ret;
37239  __ret = (float16x8_t)(__p0);
37240  return __ret;
37241}
37242#endif
37243
37244#ifdef __LITTLE_ENDIAN__
37245__ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
37246  float16x8_t __ret;
37247  __ret = (float16x8_t)(__p0);
37248  return __ret;
37249}
37250#else
37251__ai float16x8_t vreinterpretq_f16_s32(int32x4_t __p0) {
37252  float16x8_t __ret;
37253  __ret = (float16x8_t)(__p0);
37254  return __ret;
37255}
37256#endif
37257
37258#ifdef __LITTLE_ENDIAN__
37259__ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
37260  float16x8_t __ret;
37261  __ret = (float16x8_t)(__p0);
37262  return __ret;
37263}
37264#else
37265__ai float16x8_t vreinterpretq_f16_s64(int64x2_t __p0) {
37266  float16x8_t __ret;
37267  __ret = (float16x8_t)(__p0);
37268  return __ret;
37269}
37270#endif
37271
37272#ifdef __LITTLE_ENDIAN__
37273__ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
37274  float16x8_t __ret;
37275  __ret = (float16x8_t)(__p0);
37276  return __ret;
37277}
37278#else
37279__ai float16x8_t vreinterpretq_f16_s16(int16x8_t __p0) {
37280  float16x8_t __ret;
37281  __ret = (float16x8_t)(__p0);
37282  return __ret;
37283}
37284#endif
37285
37286#ifdef __LITTLE_ENDIAN__
37287__ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
37288  int32x4_t __ret;
37289  __ret = (int32x4_t)(__p0);
37290  return __ret;
37291}
37292#else
37293__ai int32x4_t vreinterpretq_s32_p8(poly8x16_t __p0) {
37294  int32x4_t __ret;
37295  __ret = (int32x4_t)(__p0);
37296  return __ret;
37297}
37298#endif
37299
37300#ifdef __LITTLE_ENDIAN__
37301__ai int32x4_t vreinterpretq_s32_p128(poly128_t __p0) {
37302  int32x4_t __ret;
37303  __ret = (int32x4_t)(__p0);
37304  return __ret;
37305}
37306#else
37307__ai int32x4_t vreinterpretq_s32_p128(poly128_t __p0) {
37308  int32x4_t __ret;
37309  __ret = (int32x4_t)(__p0);
37310  return __ret;
37311}
37312#endif
37313
37314#ifdef __LITTLE_ENDIAN__
37315__ai int32x4_t vreinterpretq_s32_p64(poly64x2_t __p0) {
37316  int32x4_t __ret;
37317  __ret = (int32x4_t)(__p0);
37318  return __ret;
37319}
37320#else
37321__ai int32x4_t vreinterpretq_s32_p64(poly64x2_t __p0) {
37322  int32x4_t __ret;
37323  __ret = (int32x4_t)(__p0);
37324  return __ret;
37325}
37326#endif
37327
37328#ifdef __LITTLE_ENDIAN__
37329__ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
37330  int32x4_t __ret;
37331  __ret = (int32x4_t)(__p0);
37332  return __ret;
37333}
37334#else
37335__ai int32x4_t vreinterpretq_s32_p16(poly16x8_t __p0) {
37336  int32x4_t __ret;
37337  __ret = (int32x4_t)(__p0);
37338  return __ret;
37339}
37340#endif
37341
37342#ifdef __LITTLE_ENDIAN__
37343__ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
37344  int32x4_t __ret;
37345  __ret = (int32x4_t)(__p0);
37346  return __ret;
37347}
37348#else
37349__ai int32x4_t vreinterpretq_s32_u8(uint8x16_t __p0) {
37350  int32x4_t __ret;
37351  __ret = (int32x4_t)(__p0);
37352  return __ret;
37353}
37354#endif
37355
37356#ifdef __LITTLE_ENDIAN__
37357__ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
37358  int32x4_t __ret;
37359  __ret = (int32x4_t)(__p0);
37360  return __ret;
37361}
37362#else
37363__ai int32x4_t vreinterpretq_s32_u32(uint32x4_t __p0) {
37364  int32x4_t __ret;
37365  __ret = (int32x4_t)(__p0);
37366  return __ret;
37367}
37368#endif
37369
37370#ifdef __LITTLE_ENDIAN__
37371__ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
37372  int32x4_t __ret;
37373  __ret = (int32x4_t)(__p0);
37374  return __ret;
37375}
37376#else
37377__ai int32x4_t vreinterpretq_s32_u64(uint64x2_t __p0) {
37378  int32x4_t __ret;
37379  __ret = (int32x4_t)(__p0);
37380  return __ret;
37381}
37382#endif
37383
37384#ifdef __LITTLE_ENDIAN__
37385__ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
37386  int32x4_t __ret;
37387  __ret = (int32x4_t)(__p0);
37388  return __ret;
37389}
37390#else
37391__ai int32x4_t vreinterpretq_s32_u16(uint16x8_t __p0) {
37392  int32x4_t __ret;
37393  __ret = (int32x4_t)(__p0);
37394  return __ret;
37395}
37396#endif
37397
37398#ifdef __LITTLE_ENDIAN__
37399__ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
37400  int32x4_t __ret;
37401  __ret = (int32x4_t)(__p0);
37402  return __ret;
37403}
37404#else
37405__ai int32x4_t vreinterpretq_s32_s8(int8x16_t __p0) {
37406  int32x4_t __ret;
37407  __ret = (int32x4_t)(__p0);
37408  return __ret;
37409}
37410#endif
37411
37412#ifdef __LITTLE_ENDIAN__
37413__ai int32x4_t vreinterpretq_s32_f64(float64x2_t __p0) {
37414  int32x4_t __ret;
37415  __ret = (int32x4_t)(__p0);
37416  return __ret;
37417}
37418#else
37419__ai int32x4_t vreinterpretq_s32_f64(float64x2_t __p0) {
37420  int32x4_t __ret;
37421  __ret = (int32x4_t)(__p0);
37422  return __ret;
37423}
37424#endif
37425
37426#ifdef __LITTLE_ENDIAN__
37427__ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
37428  int32x4_t __ret;
37429  __ret = (int32x4_t)(__p0);
37430  return __ret;
37431}
37432#else
37433__ai int32x4_t vreinterpretq_s32_f32(float32x4_t __p0) {
37434  int32x4_t __ret;
37435  __ret = (int32x4_t)(__p0);
37436  return __ret;
37437}
37438#endif
37439
37440#ifdef __LITTLE_ENDIAN__
37441__ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
37442  int32x4_t __ret;
37443  __ret = (int32x4_t)(__p0);
37444  return __ret;
37445}
37446#else
37447__ai int32x4_t vreinterpretq_s32_f16(float16x8_t __p0) {
37448  int32x4_t __ret;
37449  __ret = (int32x4_t)(__p0);
37450  return __ret;
37451}
37452#endif
37453
37454#ifdef __LITTLE_ENDIAN__
37455__ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
37456  int32x4_t __ret;
37457  __ret = (int32x4_t)(__p0);
37458  return __ret;
37459}
37460#else
37461__ai int32x4_t vreinterpretq_s32_s64(int64x2_t __p0) {
37462  int32x4_t __ret;
37463  __ret = (int32x4_t)(__p0);
37464  return __ret;
37465}
37466#endif
37467
37468#ifdef __LITTLE_ENDIAN__
37469__ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
37470  int32x4_t __ret;
37471  __ret = (int32x4_t)(__p0);
37472  return __ret;
37473}
37474#else
37475__ai int32x4_t vreinterpretq_s32_s16(int16x8_t __p0) {
37476  int32x4_t __ret;
37477  __ret = (int32x4_t)(__p0);
37478  return __ret;
37479}
37480#endif
37481
37482#ifdef __LITTLE_ENDIAN__
37483__ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
37484  int64x2_t __ret;
37485  __ret = (int64x2_t)(__p0);
37486  return __ret;
37487}
37488#else
37489__ai int64x2_t vreinterpretq_s64_p8(poly8x16_t __p0) {
37490  int64x2_t __ret;
37491  __ret = (int64x2_t)(__p0);
37492  return __ret;
37493}
37494#endif
37495
37496#ifdef __LITTLE_ENDIAN__
37497__ai int64x2_t vreinterpretq_s64_p128(poly128_t __p0) {
37498  int64x2_t __ret;
37499  __ret = (int64x2_t)(__p0);
37500  return __ret;
37501}
37502#else
37503__ai int64x2_t vreinterpretq_s64_p128(poly128_t __p0) {
37504  int64x2_t __ret;
37505  __ret = (int64x2_t)(__p0);
37506  return __ret;
37507}
37508#endif
37509
37510#ifdef __LITTLE_ENDIAN__
37511__ai int64x2_t vreinterpretq_s64_p64(poly64x2_t __p0) {
37512  int64x2_t __ret;
37513  __ret = (int64x2_t)(__p0);
37514  return __ret;
37515}
37516#else
37517__ai int64x2_t vreinterpretq_s64_p64(poly64x2_t __p0) {
37518  int64x2_t __ret;
37519  __ret = (int64x2_t)(__p0);
37520  return __ret;
37521}
37522#endif
37523
37524#ifdef __LITTLE_ENDIAN__
37525__ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
37526  int64x2_t __ret;
37527  __ret = (int64x2_t)(__p0);
37528  return __ret;
37529}
37530#else
37531__ai int64x2_t vreinterpretq_s64_p16(poly16x8_t __p0) {
37532  int64x2_t __ret;
37533  __ret = (int64x2_t)(__p0);
37534  return __ret;
37535}
37536#endif
37537
37538#ifdef __LITTLE_ENDIAN__
37539__ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
37540  int64x2_t __ret;
37541  __ret = (int64x2_t)(__p0);
37542  return __ret;
37543}
37544#else
37545__ai int64x2_t vreinterpretq_s64_u8(uint8x16_t __p0) {
37546  int64x2_t __ret;
37547  __ret = (int64x2_t)(__p0);
37548  return __ret;
37549}
37550#endif
37551
37552#ifdef __LITTLE_ENDIAN__
37553__ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
37554  int64x2_t __ret;
37555  __ret = (int64x2_t)(__p0);
37556  return __ret;
37557}
37558#else
37559__ai int64x2_t vreinterpretq_s64_u32(uint32x4_t __p0) {
37560  int64x2_t __ret;
37561  __ret = (int64x2_t)(__p0);
37562  return __ret;
37563}
37564#endif
37565
37566#ifdef __LITTLE_ENDIAN__
37567__ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
37568  int64x2_t __ret;
37569  __ret = (int64x2_t)(__p0);
37570  return __ret;
37571}
37572#else
37573__ai int64x2_t vreinterpretq_s64_u64(uint64x2_t __p0) {
37574  int64x2_t __ret;
37575  __ret = (int64x2_t)(__p0);
37576  return __ret;
37577}
37578#endif
37579
37580#ifdef __LITTLE_ENDIAN__
37581__ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
37582  int64x2_t __ret;
37583  __ret = (int64x2_t)(__p0);
37584  return __ret;
37585}
37586#else
37587__ai int64x2_t vreinterpretq_s64_u16(uint16x8_t __p0) {
37588  int64x2_t __ret;
37589  __ret = (int64x2_t)(__p0);
37590  return __ret;
37591}
37592#endif
37593
37594#ifdef __LITTLE_ENDIAN__
37595__ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
37596  int64x2_t __ret;
37597  __ret = (int64x2_t)(__p0);
37598  return __ret;
37599}
37600#else
37601__ai int64x2_t vreinterpretq_s64_s8(int8x16_t __p0) {
37602  int64x2_t __ret;
37603  __ret = (int64x2_t)(__p0);
37604  return __ret;
37605}
37606#endif
37607
37608#ifdef __LITTLE_ENDIAN__
37609__ai int64x2_t vreinterpretq_s64_f64(float64x2_t __p0) {
37610  int64x2_t __ret;
37611  __ret = (int64x2_t)(__p0);
37612  return __ret;
37613}
37614#else
37615__ai int64x2_t vreinterpretq_s64_f64(float64x2_t __p0) {
37616  int64x2_t __ret;
37617  __ret = (int64x2_t)(__p0);
37618  return __ret;
37619}
37620#endif
37621
37622#ifdef __LITTLE_ENDIAN__
37623__ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
37624  int64x2_t __ret;
37625  __ret = (int64x2_t)(__p0);
37626  return __ret;
37627}
37628#else
37629__ai int64x2_t vreinterpretq_s64_f32(float32x4_t __p0) {
37630  int64x2_t __ret;
37631  __ret = (int64x2_t)(__p0);
37632  return __ret;
37633}
37634#endif
37635
37636#ifdef __LITTLE_ENDIAN__
37637__ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
37638  int64x2_t __ret;
37639  __ret = (int64x2_t)(__p0);
37640  return __ret;
37641}
37642#else
37643__ai int64x2_t vreinterpretq_s64_f16(float16x8_t __p0) {
37644  int64x2_t __ret;
37645  __ret = (int64x2_t)(__p0);
37646  return __ret;
37647}
37648#endif
37649
37650#ifdef __LITTLE_ENDIAN__
37651__ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
37652  int64x2_t __ret;
37653  __ret = (int64x2_t)(__p0);
37654  return __ret;
37655}
37656#else
37657__ai int64x2_t vreinterpretq_s64_s32(int32x4_t __p0) {
37658  int64x2_t __ret;
37659  __ret = (int64x2_t)(__p0);
37660  return __ret;
37661}
37662#endif
37663
37664#ifdef __LITTLE_ENDIAN__
37665__ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
37666  int64x2_t __ret;
37667  __ret = (int64x2_t)(__p0);
37668  return __ret;
37669}
37670#else
37671__ai int64x2_t vreinterpretq_s64_s16(int16x8_t __p0) {
37672  int64x2_t __ret;
37673  __ret = (int64x2_t)(__p0);
37674  return __ret;
37675}
37676#endif
37677
37678#ifdef __LITTLE_ENDIAN__
37679__ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
37680  int16x8_t __ret;
37681  __ret = (int16x8_t)(__p0);
37682  return __ret;
37683}
37684#else
37685__ai int16x8_t vreinterpretq_s16_p8(poly8x16_t __p0) {
37686  int16x8_t __ret;
37687  __ret = (int16x8_t)(__p0);
37688  return __ret;
37689}
37690#endif
37691
37692#ifdef __LITTLE_ENDIAN__
37693__ai int16x8_t vreinterpretq_s16_p128(poly128_t __p0) {
37694  int16x8_t __ret;
37695  __ret = (int16x8_t)(__p0);
37696  return __ret;
37697}
37698#else
37699__ai int16x8_t vreinterpretq_s16_p128(poly128_t __p0) {
37700  int16x8_t __ret;
37701  __ret = (int16x8_t)(__p0);
37702  return __ret;
37703}
37704#endif
37705
37706#ifdef __LITTLE_ENDIAN__
37707__ai int16x8_t vreinterpretq_s16_p64(poly64x2_t __p0) {
37708  int16x8_t __ret;
37709  __ret = (int16x8_t)(__p0);
37710  return __ret;
37711}
37712#else
37713__ai int16x8_t vreinterpretq_s16_p64(poly64x2_t __p0) {
37714  int16x8_t __ret;
37715  __ret = (int16x8_t)(__p0);
37716  return __ret;
37717}
37718#endif
37719
37720#ifdef __LITTLE_ENDIAN__
37721__ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
37722  int16x8_t __ret;
37723  __ret = (int16x8_t)(__p0);
37724  return __ret;
37725}
37726#else
37727__ai int16x8_t vreinterpretq_s16_p16(poly16x8_t __p0) {
37728  int16x8_t __ret;
37729  __ret = (int16x8_t)(__p0);
37730  return __ret;
37731}
37732#endif
37733
37734#ifdef __LITTLE_ENDIAN__
37735__ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
37736  int16x8_t __ret;
37737  __ret = (int16x8_t)(__p0);
37738  return __ret;
37739}
37740#else
37741__ai int16x8_t vreinterpretq_s16_u8(uint8x16_t __p0) {
37742  int16x8_t __ret;
37743  __ret = (int16x8_t)(__p0);
37744  return __ret;
37745}
37746#endif
37747
37748#ifdef __LITTLE_ENDIAN__
37749__ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
37750  int16x8_t __ret;
37751  __ret = (int16x8_t)(__p0);
37752  return __ret;
37753}
37754#else
37755__ai int16x8_t vreinterpretq_s16_u32(uint32x4_t __p0) {
37756  int16x8_t __ret;
37757  __ret = (int16x8_t)(__p0);
37758  return __ret;
37759}
37760#endif
37761
37762#ifdef __LITTLE_ENDIAN__
37763__ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
37764  int16x8_t __ret;
37765  __ret = (int16x8_t)(__p0);
37766  return __ret;
37767}
37768#else
37769__ai int16x8_t vreinterpretq_s16_u64(uint64x2_t __p0) {
37770  int16x8_t __ret;
37771  __ret = (int16x8_t)(__p0);
37772  return __ret;
37773}
37774#endif
37775
37776#ifdef __LITTLE_ENDIAN__
37777__ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
37778  int16x8_t __ret;
37779  __ret = (int16x8_t)(__p0);
37780  return __ret;
37781}
37782#else
37783__ai int16x8_t vreinterpretq_s16_u16(uint16x8_t __p0) {
37784  int16x8_t __ret;
37785  __ret = (int16x8_t)(__p0);
37786  return __ret;
37787}
37788#endif
37789
37790#ifdef __LITTLE_ENDIAN__
37791__ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
37792  int16x8_t __ret;
37793  __ret = (int16x8_t)(__p0);
37794  return __ret;
37795}
37796#else
37797__ai int16x8_t vreinterpretq_s16_s8(int8x16_t __p0) {
37798  int16x8_t __ret;
37799  __ret = (int16x8_t)(__p0);
37800  return __ret;
37801}
37802#endif
37803
37804#ifdef __LITTLE_ENDIAN__
37805__ai int16x8_t vreinterpretq_s16_f64(float64x2_t __p0) {
37806  int16x8_t __ret;
37807  __ret = (int16x8_t)(__p0);
37808  return __ret;
37809}
37810#else
37811__ai int16x8_t vreinterpretq_s16_f64(float64x2_t __p0) {
37812  int16x8_t __ret;
37813  __ret = (int16x8_t)(__p0);
37814  return __ret;
37815}
37816#endif
37817
37818#ifdef __LITTLE_ENDIAN__
37819__ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
37820  int16x8_t __ret;
37821  __ret = (int16x8_t)(__p0);
37822  return __ret;
37823}
37824#else
37825__ai int16x8_t vreinterpretq_s16_f32(float32x4_t __p0) {
37826  int16x8_t __ret;
37827  __ret = (int16x8_t)(__p0);
37828  return __ret;
37829}
37830#endif
37831
37832#ifdef __LITTLE_ENDIAN__
37833__ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
37834  int16x8_t __ret;
37835  __ret = (int16x8_t)(__p0);
37836  return __ret;
37837}
37838#else
37839__ai int16x8_t vreinterpretq_s16_f16(float16x8_t __p0) {
37840  int16x8_t __ret;
37841  __ret = (int16x8_t)(__p0);
37842  return __ret;
37843}
37844#endif
37845
37846#ifdef __LITTLE_ENDIAN__
37847__ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
37848  int16x8_t __ret;
37849  __ret = (int16x8_t)(__p0);
37850  return __ret;
37851}
37852#else
37853__ai int16x8_t vreinterpretq_s16_s32(int32x4_t __p0) {
37854  int16x8_t __ret;
37855  __ret = (int16x8_t)(__p0);
37856  return __ret;
37857}
37858#endif
37859
37860#ifdef __LITTLE_ENDIAN__
37861__ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
37862  int16x8_t __ret;
37863  __ret = (int16x8_t)(__p0);
37864  return __ret;
37865}
37866#else
37867__ai int16x8_t vreinterpretq_s16_s64(int64x2_t __p0) {
37868  int16x8_t __ret;
37869  __ret = (int16x8_t)(__p0);
37870  return __ret;
37871}
37872#endif
37873
37874#ifdef __LITTLE_ENDIAN__
37875__ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
37876  uint8x8_t __ret;
37877  __ret = (uint8x8_t)(__p0);
37878  return __ret;
37879}
37880#else
37881__ai uint8x8_t vreinterpret_u8_p8(poly8x8_t __p0) {
37882  uint8x8_t __ret;
37883  __ret = (uint8x8_t)(__p0);
37884  return __ret;
37885}
37886#endif
37887
37888#ifdef __LITTLE_ENDIAN__
37889__ai uint8x8_t vreinterpret_u8_p64(poly64x1_t __p0) {
37890  uint8x8_t __ret;
37891  __ret = (uint8x8_t)(__p0);
37892  return __ret;
37893}
37894#else
37895__ai uint8x8_t vreinterpret_u8_p64(poly64x1_t __p0) {
37896  uint8x8_t __ret;
37897  __ret = (uint8x8_t)(__p0);
37898  return __ret;
37899}
37900#endif
37901
37902#ifdef __LITTLE_ENDIAN__
37903__ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
37904  uint8x8_t __ret;
37905  __ret = (uint8x8_t)(__p0);
37906  return __ret;
37907}
37908#else
37909__ai uint8x8_t vreinterpret_u8_p16(poly16x4_t __p0) {
37910  uint8x8_t __ret;
37911  __ret = (uint8x8_t)(__p0);
37912  return __ret;
37913}
37914#endif
37915
37916#ifdef __LITTLE_ENDIAN__
37917__ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
37918  uint8x8_t __ret;
37919  __ret = (uint8x8_t)(__p0);
37920  return __ret;
37921}
37922#else
37923__ai uint8x8_t vreinterpret_u8_u32(uint32x2_t __p0) {
37924  uint8x8_t __ret;
37925  __ret = (uint8x8_t)(__p0);
37926  return __ret;
37927}
37928#endif
37929
37930#ifdef __LITTLE_ENDIAN__
37931__ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
37932  uint8x8_t __ret;
37933  __ret = (uint8x8_t)(__p0);
37934  return __ret;
37935}
37936#else
37937__ai uint8x8_t vreinterpret_u8_u64(uint64x1_t __p0) {
37938  uint8x8_t __ret;
37939  __ret = (uint8x8_t)(__p0);
37940  return __ret;
37941}
37942#endif
37943
37944#ifdef __LITTLE_ENDIAN__
37945__ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
37946  uint8x8_t __ret;
37947  __ret = (uint8x8_t)(__p0);
37948  return __ret;
37949}
37950#else
37951__ai uint8x8_t vreinterpret_u8_u16(uint16x4_t __p0) {
37952  uint8x8_t __ret;
37953  __ret = (uint8x8_t)(__p0);
37954  return __ret;
37955}
37956#endif
37957
37958#ifdef __LITTLE_ENDIAN__
37959__ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
37960  uint8x8_t __ret;
37961  __ret = (uint8x8_t)(__p0);
37962  return __ret;
37963}
37964#else
37965__ai uint8x8_t vreinterpret_u8_s8(int8x8_t __p0) {
37966  uint8x8_t __ret;
37967  __ret = (uint8x8_t)(__p0);
37968  return __ret;
37969}
37970#endif
37971
37972#ifdef __LITTLE_ENDIAN__
37973__ai uint8x8_t vreinterpret_u8_f64(float64x1_t __p0) {
37974  uint8x8_t __ret;
37975  __ret = (uint8x8_t)(__p0);
37976  return __ret;
37977}
37978#else
37979__ai uint8x8_t vreinterpret_u8_f64(float64x1_t __p0) {
37980  uint8x8_t __ret;
37981  __ret = (uint8x8_t)(__p0);
37982  return __ret;
37983}
37984#endif
37985
37986#ifdef __LITTLE_ENDIAN__
37987__ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
37988  uint8x8_t __ret;
37989  __ret = (uint8x8_t)(__p0);
37990  return __ret;
37991}
37992#else
37993__ai uint8x8_t vreinterpret_u8_f32(float32x2_t __p0) {
37994  uint8x8_t __ret;
37995  __ret = (uint8x8_t)(__p0);
37996  return __ret;
37997}
37998#endif
37999
38000#ifdef __LITTLE_ENDIAN__
38001__ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
38002  uint8x8_t __ret;
38003  __ret = (uint8x8_t)(__p0);
38004  return __ret;
38005}
38006#else
38007__ai uint8x8_t vreinterpret_u8_f16(float16x4_t __p0) {
38008  uint8x8_t __ret;
38009  __ret = (uint8x8_t)(__p0);
38010  return __ret;
38011}
38012#endif
38013
38014#ifdef __LITTLE_ENDIAN__
38015__ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
38016  uint8x8_t __ret;
38017  __ret = (uint8x8_t)(__p0);
38018  return __ret;
38019}
38020#else
38021__ai uint8x8_t vreinterpret_u8_s32(int32x2_t __p0) {
38022  uint8x8_t __ret;
38023  __ret = (uint8x8_t)(__p0);
38024  return __ret;
38025}
38026#endif
38027
38028#ifdef __LITTLE_ENDIAN__
38029__ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
38030  uint8x8_t __ret;
38031  __ret = (uint8x8_t)(__p0);
38032  return __ret;
38033}
38034#else
38035__ai uint8x8_t vreinterpret_u8_s64(int64x1_t __p0) {
38036  uint8x8_t __ret;
38037  __ret = (uint8x8_t)(__p0);
38038  return __ret;
38039}
38040#endif
38041
38042#ifdef __LITTLE_ENDIAN__
38043__ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
38044  uint8x8_t __ret;
38045  __ret = (uint8x8_t)(__p0);
38046  return __ret;
38047}
38048#else
38049__ai uint8x8_t vreinterpret_u8_s16(int16x4_t __p0) {
38050  uint8x8_t __ret;
38051  __ret = (uint8x8_t)(__p0);
38052  return __ret;
38053}
38054#endif
38055
38056#ifdef __LITTLE_ENDIAN__
38057__ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
38058  uint32x2_t __ret;
38059  __ret = (uint32x2_t)(__p0);
38060  return __ret;
38061}
38062#else
38063__ai uint32x2_t vreinterpret_u32_p8(poly8x8_t __p0) {
38064  uint32x2_t __ret;
38065  __ret = (uint32x2_t)(__p0);
38066  return __ret;
38067}
38068#endif
38069
38070#ifdef __LITTLE_ENDIAN__
38071__ai uint32x2_t vreinterpret_u32_p64(poly64x1_t __p0) {
38072  uint32x2_t __ret;
38073  __ret = (uint32x2_t)(__p0);
38074  return __ret;
38075}
38076#else
38077__ai uint32x2_t vreinterpret_u32_p64(poly64x1_t __p0) {
38078  uint32x2_t __ret;
38079  __ret = (uint32x2_t)(__p0);
38080  return __ret;
38081}
38082#endif
38083
38084#ifdef __LITTLE_ENDIAN__
38085__ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
38086  uint32x2_t __ret;
38087  __ret = (uint32x2_t)(__p0);
38088  return __ret;
38089}
38090#else
38091__ai uint32x2_t vreinterpret_u32_p16(poly16x4_t __p0) {
38092  uint32x2_t __ret;
38093  __ret = (uint32x2_t)(__p0);
38094  return __ret;
38095}
38096#endif
38097
38098#ifdef __LITTLE_ENDIAN__
38099__ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
38100  uint32x2_t __ret;
38101  __ret = (uint32x2_t)(__p0);
38102  return __ret;
38103}
38104#else
38105__ai uint32x2_t vreinterpret_u32_u8(uint8x8_t __p0) {
38106  uint32x2_t __ret;
38107  __ret = (uint32x2_t)(__p0);
38108  return __ret;
38109}
38110#endif
38111
38112#ifdef __LITTLE_ENDIAN__
38113__ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
38114  uint32x2_t __ret;
38115  __ret = (uint32x2_t)(__p0);
38116  return __ret;
38117}
38118#else
38119__ai uint32x2_t vreinterpret_u32_u64(uint64x1_t __p0) {
38120  uint32x2_t __ret;
38121  __ret = (uint32x2_t)(__p0);
38122  return __ret;
38123}
38124#endif
38125
38126#ifdef __LITTLE_ENDIAN__
38127__ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
38128  uint32x2_t __ret;
38129  __ret = (uint32x2_t)(__p0);
38130  return __ret;
38131}
38132#else
38133__ai uint32x2_t vreinterpret_u32_u16(uint16x4_t __p0) {
38134  uint32x2_t __ret;
38135  __ret = (uint32x2_t)(__p0);
38136  return __ret;
38137}
38138#endif
38139
38140#ifdef __LITTLE_ENDIAN__
38141__ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
38142  uint32x2_t __ret;
38143  __ret = (uint32x2_t)(__p0);
38144  return __ret;
38145}
38146#else
38147__ai uint32x2_t vreinterpret_u32_s8(int8x8_t __p0) {
38148  uint32x2_t __ret;
38149  __ret = (uint32x2_t)(__p0);
38150  return __ret;
38151}
38152#endif
38153
38154#ifdef __LITTLE_ENDIAN__
38155__ai uint32x2_t vreinterpret_u32_f64(float64x1_t __p0) {
38156  uint32x2_t __ret;
38157  __ret = (uint32x2_t)(__p0);
38158  return __ret;
38159}
38160#else
38161__ai uint32x2_t vreinterpret_u32_f64(float64x1_t __p0) {
38162  uint32x2_t __ret;
38163  __ret = (uint32x2_t)(__p0);
38164  return __ret;
38165}
38166#endif
38167
38168#ifdef __LITTLE_ENDIAN__
38169__ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
38170  uint32x2_t __ret;
38171  __ret = (uint32x2_t)(__p0);
38172  return __ret;
38173}
38174#else
38175__ai uint32x2_t vreinterpret_u32_f32(float32x2_t __p0) {
38176  uint32x2_t __ret;
38177  __ret = (uint32x2_t)(__p0);
38178  return __ret;
38179}
38180#endif
38181
38182#ifdef __LITTLE_ENDIAN__
38183__ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
38184  uint32x2_t __ret;
38185  __ret = (uint32x2_t)(__p0);
38186  return __ret;
38187}
38188#else
38189__ai uint32x2_t vreinterpret_u32_f16(float16x4_t __p0) {
38190  uint32x2_t __ret;
38191  __ret = (uint32x2_t)(__p0);
38192  return __ret;
38193}
38194#endif
38195
38196#ifdef __LITTLE_ENDIAN__
38197__ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
38198  uint32x2_t __ret;
38199  __ret = (uint32x2_t)(__p0);
38200  return __ret;
38201}
38202#else
38203__ai uint32x2_t vreinterpret_u32_s32(int32x2_t __p0) {
38204  uint32x2_t __ret;
38205  __ret = (uint32x2_t)(__p0);
38206  return __ret;
38207}
38208#endif
38209
38210#ifdef __LITTLE_ENDIAN__
38211__ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
38212  uint32x2_t __ret;
38213  __ret = (uint32x2_t)(__p0);
38214  return __ret;
38215}
38216#else
38217__ai uint32x2_t vreinterpret_u32_s64(int64x1_t __p0) {
38218  uint32x2_t __ret;
38219  __ret = (uint32x2_t)(__p0);
38220  return __ret;
38221}
38222#endif
38223
38224#ifdef __LITTLE_ENDIAN__
38225__ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
38226  uint32x2_t __ret;
38227  __ret = (uint32x2_t)(__p0);
38228  return __ret;
38229}
38230#else
38231__ai uint32x2_t vreinterpret_u32_s16(int16x4_t __p0) {
38232  uint32x2_t __ret;
38233  __ret = (uint32x2_t)(__p0);
38234  return __ret;
38235}
38236#endif
38237
38238#ifdef __LITTLE_ENDIAN__
38239__ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
38240  uint64x1_t __ret;
38241  __ret = (uint64x1_t)(__p0);
38242  return __ret;
38243}
38244#else
38245__ai uint64x1_t vreinterpret_u64_p8(poly8x8_t __p0) {
38246  uint64x1_t __ret;
38247  __ret = (uint64x1_t)(__p0);
38248  return __ret;
38249}
38250#endif
38251
38252#ifdef __LITTLE_ENDIAN__
38253__ai uint64x1_t vreinterpret_u64_p64(poly64x1_t __p0) {
38254  uint64x1_t __ret;
38255  __ret = (uint64x1_t)(__p0);
38256  return __ret;
38257}
38258#else
38259__ai uint64x1_t vreinterpret_u64_p64(poly64x1_t __p0) {
38260  uint64x1_t __ret;
38261  __ret = (uint64x1_t)(__p0);
38262  return __ret;
38263}
38264#endif
38265
38266#ifdef __LITTLE_ENDIAN__
38267__ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
38268  uint64x1_t __ret;
38269  __ret = (uint64x1_t)(__p0);
38270  return __ret;
38271}
38272#else
38273__ai uint64x1_t vreinterpret_u64_p16(poly16x4_t __p0) {
38274  uint64x1_t __ret;
38275  __ret = (uint64x1_t)(__p0);
38276  return __ret;
38277}
38278#endif
38279
38280#ifdef __LITTLE_ENDIAN__
38281__ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
38282  uint64x1_t __ret;
38283  __ret = (uint64x1_t)(__p0);
38284  return __ret;
38285}
38286#else
38287__ai uint64x1_t vreinterpret_u64_u8(uint8x8_t __p0) {
38288  uint64x1_t __ret;
38289  __ret = (uint64x1_t)(__p0);
38290  return __ret;
38291}
38292#endif
38293
38294#ifdef __LITTLE_ENDIAN__
38295__ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
38296  uint64x1_t __ret;
38297  __ret = (uint64x1_t)(__p0);
38298  return __ret;
38299}
38300#else
38301__ai uint64x1_t vreinterpret_u64_u32(uint32x2_t __p0) {
38302  uint64x1_t __ret;
38303  __ret = (uint64x1_t)(__p0);
38304  return __ret;
38305}
38306#endif
38307
38308#ifdef __LITTLE_ENDIAN__
38309__ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
38310  uint64x1_t __ret;
38311  __ret = (uint64x1_t)(__p0);
38312  return __ret;
38313}
38314#else
38315__ai uint64x1_t vreinterpret_u64_u16(uint16x4_t __p0) {
38316  uint64x1_t __ret;
38317  __ret = (uint64x1_t)(__p0);
38318  return __ret;
38319}
38320#endif
38321
38322#ifdef __LITTLE_ENDIAN__
38323__ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
38324  uint64x1_t __ret;
38325  __ret = (uint64x1_t)(__p0);
38326  return __ret;
38327}
38328#else
38329__ai uint64x1_t vreinterpret_u64_s8(int8x8_t __p0) {
38330  uint64x1_t __ret;
38331  __ret = (uint64x1_t)(__p0);
38332  return __ret;
38333}
38334#endif
38335
38336#ifdef __LITTLE_ENDIAN__
38337__ai uint64x1_t vreinterpret_u64_f64(float64x1_t __p0) {
38338  uint64x1_t __ret;
38339  __ret = (uint64x1_t)(__p0);
38340  return __ret;
38341}
38342#else
38343__ai uint64x1_t vreinterpret_u64_f64(float64x1_t __p0) {
38344  uint64x1_t __ret;
38345  __ret = (uint64x1_t)(__p0);
38346  return __ret;
38347}
38348#endif
38349
38350#ifdef __LITTLE_ENDIAN__
38351__ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
38352  uint64x1_t __ret;
38353  __ret = (uint64x1_t)(__p0);
38354  return __ret;
38355}
38356#else
38357__ai uint64x1_t vreinterpret_u64_f32(float32x2_t __p0) {
38358  uint64x1_t __ret;
38359  __ret = (uint64x1_t)(__p0);
38360  return __ret;
38361}
38362#endif
38363
38364#ifdef __LITTLE_ENDIAN__
38365__ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
38366  uint64x1_t __ret;
38367  __ret = (uint64x1_t)(__p0);
38368  return __ret;
38369}
38370#else
38371__ai uint64x1_t vreinterpret_u64_f16(float16x4_t __p0) {
38372  uint64x1_t __ret;
38373  __ret = (uint64x1_t)(__p0);
38374  return __ret;
38375}
38376#endif
38377
38378#ifdef __LITTLE_ENDIAN__
38379__ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
38380  uint64x1_t __ret;
38381  __ret = (uint64x1_t)(__p0);
38382  return __ret;
38383}
38384#else
38385__ai uint64x1_t vreinterpret_u64_s32(int32x2_t __p0) {
38386  uint64x1_t __ret;
38387  __ret = (uint64x1_t)(__p0);
38388  return __ret;
38389}
38390#endif
38391
38392#ifdef __LITTLE_ENDIAN__
38393__ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
38394  uint64x1_t __ret;
38395  __ret = (uint64x1_t)(__p0);
38396  return __ret;
38397}
38398#else
38399__ai uint64x1_t vreinterpret_u64_s64(int64x1_t __p0) {
38400  uint64x1_t __ret;
38401  __ret = (uint64x1_t)(__p0);
38402  return __ret;
38403}
38404#endif
38405
38406#ifdef __LITTLE_ENDIAN__
38407__ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
38408  uint64x1_t __ret;
38409  __ret = (uint64x1_t)(__p0);
38410  return __ret;
38411}
38412#else
38413__ai uint64x1_t vreinterpret_u64_s16(int16x4_t __p0) {
38414  uint64x1_t __ret;
38415  __ret = (uint64x1_t)(__p0);
38416  return __ret;
38417}
38418#endif
38419
38420#ifdef __LITTLE_ENDIAN__
38421__ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
38422  uint16x4_t __ret;
38423  __ret = (uint16x4_t)(__p0);
38424  return __ret;
38425}
38426#else
38427__ai uint16x4_t vreinterpret_u16_p8(poly8x8_t __p0) {
38428  uint16x4_t __ret;
38429  __ret = (uint16x4_t)(__p0);
38430  return __ret;
38431}
38432#endif
38433
38434#ifdef __LITTLE_ENDIAN__
38435__ai uint16x4_t vreinterpret_u16_p64(poly64x1_t __p0) {
38436  uint16x4_t __ret;
38437  __ret = (uint16x4_t)(__p0);
38438  return __ret;
38439}
38440#else
38441__ai uint16x4_t vreinterpret_u16_p64(poly64x1_t __p0) {
38442  uint16x4_t __ret;
38443  __ret = (uint16x4_t)(__p0);
38444  return __ret;
38445}
38446#endif
38447
38448#ifdef __LITTLE_ENDIAN__
38449__ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
38450  uint16x4_t __ret;
38451  __ret = (uint16x4_t)(__p0);
38452  return __ret;
38453}
38454#else
38455__ai uint16x4_t vreinterpret_u16_p16(poly16x4_t __p0) {
38456  uint16x4_t __ret;
38457  __ret = (uint16x4_t)(__p0);
38458  return __ret;
38459}
38460#endif
38461
38462#ifdef __LITTLE_ENDIAN__
38463__ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
38464  uint16x4_t __ret;
38465  __ret = (uint16x4_t)(__p0);
38466  return __ret;
38467}
38468#else
38469__ai uint16x4_t vreinterpret_u16_u8(uint8x8_t __p0) {
38470  uint16x4_t __ret;
38471  __ret = (uint16x4_t)(__p0);
38472  return __ret;
38473}
38474#endif
38475
38476#ifdef __LITTLE_ENDIAN__
38477__ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
38478  uint16x4_t __ret;
38479  __ret = (uint16x4_t)(__p0);
38480  return __ret;
38481}
38482#else
38483__ai uint16x4_t vreinterpret_u16_u32(uint32x2_t __p0) {
38484  uint16x4_t __ret;
38485  __ret = (uint16x4_t)(__p0);
38486  return __ret;
38487}
38488#endif
38489
38490#ifdef __LITTLE_ENDIAN__
38491__ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
38492  uint16x4_t __ret;
38493  __ret = (uint16x4_t)(__p0);
38494  return __ret;
38495}
38496#else
38497__ai uint16x4_t vreinterpret_u16_u64(uint64x1_t __p0) {
38498  uint16x4_t __ret;
38499  __ret = (uint16x4_t)(__p0);
38500  return __ret;
38501}
38502#endif
38503
38504#ifdef __LITTLE_ENDIAN__
38505__ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
38506  uint16x4_t __ret;
38507  __ret = (uint16x4_t)(__p0);
38508  return __ret;
38509}
38510#else
38511__ai uint16x4_t vreinterpret_u16_s8(int8x8_t __p0) {
38512  uint16x4_t __ret;
38513  __ret = (uint16x4_t)(__p0);
38514  return __ret;
38515}
38516#endif
38517
38518#ifdef __LITTLE_ENDIAN__
38519__ai uint16x4_t vreinterpret_u16_f64(float64x1_t __p0) {
38520  uint16x4_t __ret;
38521  __ret = (uint16x4_t)(__p0);
38522  return __ret;
38523}
38524#else
38525__ai uint16x4_t vreinterpret_u16_f64(float64x1_t __p0) {
38526  uint16x4_t __ret;
38527  __ret = (uint16x4_t)(__p0);
38528  return __ret;
38529}
38530#endif
38531
38532#ifdef __LITTLE_ENDIAN__
38533__ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
38534  uint16x4_t __ret;
38535  __ret = (uint16x4_t)(__p0);
38536  return __ret;
38537}
38538#else
38539__ai uint16x4_t vreinterpret_u16_f32(float32x2_t __p0) {
38540  uint16x4_t __ret;
38541  __ret = (uint16x4_t)(__p0);
38542  return __ret;
38543}
38544#endif
38545
38546#ifdef __LITTLE_ENDIAN__
38547__ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
38548  uint16x4_t __ret;
38549  __ret = (uint16x4_t)(__p0);
38550  return __ret;
38551}
38552#else
38553__ai uint16x4_t vreinterpret_u16_f16(float16x4_t __p0) {
38554  uint16x4_t __ret;
38555  __ret = (uint16x4_t)(__p0);
38556  return __ret;
38557}
38558#endif
38559
38560#ifdef __LITTLE_ENDIAN__
38561__ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
38562  uint16x4_t __ret;
38563  __ret = (uint16x4_t)(__p0);
38564  return __ret;
38565}
38566#else
38567__ai uint16x4_t vreinterpret_u16_s32(int32x2_t __p0) {
38568  uint16x4_t __ret;
38569  __ret = (uint16x4_t)(__p0);
38570  return __ret;
38571}
38572#endif
38573
38574#ifdef __LITTLE_ENDIAN__
38575__ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
38576  uint16x4_t __ret;
38577  __ret = (uint16x4_t)(__p0);
38578  return __ret;
38579}
38580#else
38581__ai uint16x4_t vreinterpret_u16_s64(int64x1_t __p0) {
38582  uint16x4_t __ret;
38583  __ret = (uint16x4_t)(__p0);
38584  return __ret;
38585}
38586#endif
38587
38588#ifdef __LITTLE_ENDIAN__
38589__ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
38590  uint16x4_t __ret;
38591  __ret = (uint16x4_t)(__p0);
38592  return __ret;
38593}
38594#else
38595__ai uint16x4_t vreinterpret_u16_s16(int16x4_t __p0) {
38596  uint16x4_t __ret;
38597  __ret = (uint16x4_t)(__p0);
38598  return __ret;
38599}
38600#endif
38601
38602#ifdef __LITTLE_ENDIAN__
38603__ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
38604  int8x8_t __ret;
38605  __ret = (int8x8_t)(__p0);
38606  return __ret;
38607}
38608#else
38609__ai int8x8_t vreinterpret_s8_p8(poly8x8_t __p0) {
38610  int8x8_t __ret;
38611  __ret = (int8x8_t)(__p0);
38612  return __ret;
38613}
38614#endif
38615
38616#ifdef __LITTLE_ENDIAN__
38617__ai int8x8_t vreinterpret_s8_p64(poly64x1_t __p0) {
38618  int8x8_t __ret;
38619  __ret = (int8x8_t)(__p0);
38620  return __ret;
38621}
38622#else
38623__ai int8x8_t vreinterpret_s8_p64(poly64x1_t __p0) {
38624  int8x8_t __ret;
38625  __ret = (int8x8_t)(__p0);
38626  return __ret;
38627}
38628#endif
38629
38630#ifdef __LITTLE_ENDIAN__
38631__ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
38632  int8x8_t __ret;
38633  __ret = (int8x8_t)(__p0);
38634  return __ret;
38635}
38636#else
38637__ai int8x8_t vreinterpret_s8_p16(poly16x4_t __p0) {
38638  int8x8_t __ret;
38639  __ret = (int8x8_t)(__p0);
38640  return __ret;
38641}
38642#endif
38643
38644#ifdef __LITTLE_ENDIAN__
38645__ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
38646  int8x8_t __ret;
38647  __ret = (int8x8_t)(__p0);
38648  return __ret;
38649}
38650#else
38651__ai int8x8_t vreinterpret_s8_u8(uint8x8_t __p0) {
38652  int8x8_t __ret;
38653  __ret = (int8x8_t)(__p0);
38654  return __ret;
38655}
38656#endif
38657
38658#ifdef __LITTLE_ENDIAN__
38659__ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
38660  int8x8_t __ret;
38661  __ret = (int8x8_t)(__p0);
38662  return __ret;
38663}
38664#else
38665__ai int8x8_t vreinterpret_s8_u32(uint32x2_t __p0) {
38666  int8x8_t __ret;
38667  __ret = (int8x8_t)(__p0);
38668  return __ret;
38669}
38670#endif
38671
38672#ifdef __LITTLE_ENDIAN__
38673__ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
38674  int8x8_t __ret;
38675  __ret = (int8x8_t)(__p0);
38676  return __ret;
38677}
38678#else
38679__ai int8x8_t vreinterpret_s8_u64(uint64x1_t __p0) {
38680  int8x8_t __ret;
38681  __ret = (int8x8_t)(__p0);
38682  return __ret;
38683}
38684#endif
38685
38686#ifdef __LITTLE_ENDIAN__
38687__ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
38688  int8x8_t __ret;
38689  __ret = (int8x8_t)(__p0);
38690  return __ret;
38691}
38692#else
38693__ai int8x8_t vreinterpret_s8_u16(uint16x4_t __p0) {
38694  int8x8_t __ret;
38695  __ret = (int8x8_t)(__p0);
38696  return __ret;
38697}
38698#endif
38699
38700#ifdef __LITTLE_ENDIAN__
38701__ai int8x8_t vreinterpret_s8_f64(float64x1_t __p0) {
38702  int8x8_t __ret;
38703  __ret = (int8x8_t)(__p0);
38704  return __ret;
38705}
38706#else
38707__ai int8x8_t vreinterpret_s8_f64(float64x1_t __p0) {
38708  int8x8_t __ret;
38709  __ret = (int8x8_t)(__p0);
38710  return __ret;
38711}
38712#endif
38713
38714#ifdef __LITTLE_ENDIAN__
38715__ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
38716  int8x8_t __ret;
38717  __ret = (int8x8_t)(__p0);
38718  return __ret;
38719}
38720#else
38721__ai int8x8_t vreinterpret_s8_f32(float32x2_t __p0) {
38722  int8x8_t __ret;
38723  __ret = (int8x8_t)(__p0);
38724  return __ret;
38725}
38726#endif
38727
38728#ifdef __LITTLE_ENDIAN__
38729__ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
38730  int8x8_t __ret;
38731  __ret = (int8x8_t)(__p0);
38732  return __ret;
38733}
38734#else
38735__ai int8x8_t vreinterpret_s8_f16(float16x4_t __p0) {
38736  int8x8_t __ret;
38737  __ret = (int8x8_t)(__p0);
38738  return __ret;
38739}
38740#endif
38741
38742#ifdef __LITTLE_ENDIAN__
38743__ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
38744  int8x8_t __ret;
38745  __ret = (int8x8_t)(__p0);
38746  return __ret;
38747}
38748#else
38749__ai int8x8_t vreinterpret_s8_s32(int32x2_t __p0) {
38750  int8x8_t __ret;
38751  __ret = (int8x8_t)(__p0);
38752  return __ret;
38753}
38754#endif
38755
38756#ifdef __LITTLE_ENDIAN__
38757__ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
38758  int8x8_t __ret;
38759  __ret = (int8x8_t)(__p0);
38760  return __ret;
38761}
38762#else
38763__ai int8x8_t vreinterpret_s8_s64(int64x1_t __p0) {
38764  int8x8_t __ret;
38765  __ret = (int8x8_t)(__p0);
38766  return __ret;
38767}
38768#endif
38769
38770#ifdef __LITTLE_ENDIAN__
38771__ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
38772  int8x8_t __ret;
38773  __ret = (int8x8_t)(__p0);
38774  return __ret;
38775}
38776#else
38777__ai int8x8_t vreinterpret_s8_s16(int16x4_t __p0) {
38778  int8x8_t __ret;
38779  __ret = (int8x8_t)(__p0);
38780  return __ret;
38781}
38782#endif
38783
38784#ifdef __LITTLE_ENDIAN__
38785__ai float64x1_t vreinterpret_f64_p8(poly8x8_t __p0) {
38786  float64x1_t __ret;
38787  __ret = (float64x1_t)(__p0);
38788  return __ret;
38789}
38790#else
38791__ai float64x1_t vreinterpret_f64_p8(poly8x8_t __p0) {
38792  float64x1_t __ret;
38793  __ret = (float64x1_t)(__p0);
38794  return __ret;
38795}
38796#endif
38797
38798#ifdef __LITTLE_ENDIAN__
38799__ai float64x1_t vreinterpret_f64_p64(poly64x1_t __p0) {
38800  float64x1_t __ret;
38801  __ret = (float64x1_t)(__p0);
38802  return __ret;
38803}
38804#else
38805__ai float64x1_t vreinterpret_f64_p64(poly64x1_t __p0) {
38806  float64x1_t __ret;
38807  __ret = (float64x1_t)(__p0);
38808  return __ret;
38809}
38810#endif
38811
38812#ifdef __LITTLE_ENDIAN__
38813__ai float64x1_t vreinterpret_f64_p16(poly16x4_t __p0) {
38814  float64x1_t __ret;
38815  __ret = (float64x1_t)(__p0);
38816  return __ret;
38817}
38818#else
38819__ai float64x1_t vreinterpret_f64_p16(poly16x4_t __p0) {
38820  float64x1_t __ret;
38821  __ret = (float64x1_t)(__p0);
38822  return __ret;
38823}
38824#endif
38825
38826#ifdef __LITTLE_ENDIAN__
38827__ai float64x1_t vreinterpret_f64_u8(uint8x8_t __p0) {
38828  float64x1_t __ret;
38829  __ret = (float64x1_t)(__p0);
38830  return __ret;
38831}
38832#else
38833__ai float64x1_t vreinterpret_f64_u8(uint8x8_t __p0) {
38834  float64x1_t __ret;
38835  __ret = (float64x1_t)(__p0);
38836  return __ret;
38837}
38838#endif
38839
38840#ifdef __LITTLE_ENDIAN__
38841__ai float64x1_t vreinterpret_f64_u32(uint32x2_t __p0) {
38842  float64x1_t __ret;
38843  __ret = (float64x1_t)(__p0);
38844  return __ret;
38845}
38846#else
38847__ai float64x1_t vreinterpret_f64_u32(uint32x2_t __p0) {
38848  float64x1_t __ret;
38849  __ret = (float64x1_t)(__p0);
38850  return __ret;
38851}
38852#endif
38853
38854#ifdef __LITTLE_ENDIAN__
38855__ai float64x1_t vreinterpret_f64_u64(uint64x1_t __p0) {
38856  float64x1_t __ret;
38857  __ret = (float64x1_t)(__p0);
38858  return __ret;
38859}
38860#else
38861__ai float64x1_t vreinterpret_f64_u64(uint64x1_t __p0) {
38862  float64x1_t __ret;
38863  __ret = (float64x1_t)(__p0);
38864  return __ret;
38865}
38866#endif
38867
38868#ifdef __LITTLE_ENDIAN__
38869__ai float64x1_t vreinterpret_f64_u16(uint16x4_t __p0) {
38870  float64x1_t __ret;
38871  __ret = (float64x1_t)(__p0);
38872  return __ret;
38873}
38874#else
38875__ai float64x1_t vreinterpret_f64_u16(uint16x4_t __p0) {
38876  float64x1_t __ret;
38877  __ret = (float64x1_t)(__p0);
38878  return __ret;
38879}
38880#endif
38881
38882#ifdef __LITTLE_ENDIAN__
38883__ai float64x1_t vreinterpret_f64_s8(int8x8_t __p0) {
38884  float64x1_t __ret;
38885  __ret = (float64x1_t)(__p0);
38886  return __ret;
38887}
38888#else
38889__ai float64x1_t vreinterpret_f64_s8(int8x8_t __p0) {
38890  float64x1_t __ret;
38891  __ret = (float64x1_t)(__p0);
38892  return __ret;
38893}
38894#endif
38895
38896#ifdef __LITTLE_ENDIAN__
38897__ai float64x1_t vreinterpret_f64_f32(float32x2_t __p0) {
38898  float64x1_t __ret;
38899  __ret = (float64x1_t)(__p0);
38900  return __ret;
38901}
38902#else
38903__ai float64x1_t vreinterpret_f64_f32(float32x2_t __p0) {
38904  float64x1_t __ret;
38905  __ret = (float64x1_t)(__p0);
38906  return __ret;
38907}
38908#endif
38909
38910#ifdef __LITTLE_ENDIAN__
38911__ai float64x1_t vreinterpret_f64_f16(float16x4_t __p0) {
38912  float64x1_t __ret;
38913  __ret = (float64x1_t)(__p0);
38914  return __ret;
38915}
38916#else
38917__ai float64x1_t vreinterpret_f64_f16(float16x4_t __p0) {
38918  float64x1_t __ret;
38919  __ret = (float64x1_t)(__p0);
38920  return __ret;
38921}
38922#endif
38923
38924#ifdef __LITTLE_ENDIAN__
38925__ai float64x1_t vreinterpret_f64_s32(int32x2_t __p0) {
38926  float64x1_t __ret;
38927  __ret = (float64x1_t)(__p0);
38928  return __ret;
38929}
38930#else
38931__ai float64x1_t vreinterpret_f64_s32(int32x2_t __p0) {
38932  float64x1_t __ret;
38933  __ret = (float64x1_t)(__p0);
38934  return __ret;
38935}
38936#endif
38937
38938#ifdef __LITTLE_ENDIAN__
38939__ai float64x1_t vreinterpret_f64_s64(int64x1_t __p0) {
38940  float64x1_t __ret;
38941  __ret = (float64x1_t)(__p0);
38942  return __ret;
38943}
38944#else
38945__ai float64x1_t vreinterpret_f64_s64(int64x1_t __p0) {
38946  float64x1_t __ret;
38947  __ret = (float64x1_t)(__p0);
38948  return __ret;
38949}
38950#endif
38951
38952#ifdef __LITTLE_ENDIAN__
38953__ai float64x1_t vreinterpret_f64_s16(int16x4_t __p0) {
38954  float64x1_t __ret;
38955  __ret = (float64x1_t)(__p0);
38956  return __ret;
38957}
38958#else
38959__ai float64x1_t vreinterpret_f64_s16(int16x4_t __p0) {
38960  float64x1_t __ret;
38961  __ret = (float64x1_t)(__p0);
38962  return __ret;
38963}
38964#endif
38965
38966#ifdef __LITTLE_ENDIAN__
38967__ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
38968  float32x2_t __ret;
38969  __ret = (float32x2_t)(__p0);
38970  return __ret;
38971}
38972#else
38973__ai float32x2_t vreinterpret_f32_p8(poly8x8_t __p0) {
38974  float32x2_t __ret;
38975  __ret = (float32x2_t)(__p0);
38976  return __ret;
38977}
38978#endif
38979
38980#ifdef __LITTLE_ENDIAN__
38981__ai float32x2_t vreinterpret_f32_p64(poly64x1_t __p0) {
38982  float32x2_t __ret;
38983  __ret = (float32x2_t)(__p0);
38984  return __ret;
38985}
38986#else
38987__ai float32x2_t vreinterpret_f32_p64(poly64x1_t __p0) {
38988  float32x2_t __ret;
38989  __ret = (float32x2_t)(__p0);
38990  return __ret;
38991}
38992#endif
38993
38994#ifdef __LITTLE_ENDIAN__
38995__ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
38996  float32x2_t __ret;
38997  __ret = (float32x2_t)(__p0);
38998  return __ret;
38999}
39000#else
39001__ai float32x2_t vreinterpret_f32_p16(poly16x4_t __p0) {
39002  float32x2_t __ret;
39003  __ret = (float32x2_t)(__p0);
39004  return __ret;
39005}
39006#endif
39007
39008#ifdef __LITTLE_ENDIAN__
39009__ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
39010  float32x2_t __ret;
39011  __ret = (float32x2_t)(__p0);
39012  return __ret;
39013}
39014#else
39015__ai float32x2_t vreinterpret_f32_u8(uint8x8_t __p0) {
39016  float32x2_t __ret;
39017  __ret = (float32x2_t)(__p0);
39018  return __ret;
39019}
39020#endif
39021
39022#ifdef __LITTLE_ENDIAN__
39023__ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
39024  float32x2_t __ret;
39025  __ret = (float32x2_t)(__p0);
39026  return __ret;
39027}
39028#else
39029__ai float32x2_t vreinterpret_f32_u32(uint32x2_t __p0) {
39030  float32x2_t __ret;
39031  __ret = (float32x2_t)(__p0);
39032  return __ret;
39033}
39034#endif
39035
39036#ifdef __LITTLE_ENDIAN__
39037__ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
39038  float32x2_t __ret;
39039  __ret = (float32x2_t)(__p0);
39040  return __ret;
39041}
39042#else
39043__ai float32x2_t vreinterpret_f32_u64(uint64x1_t __p0) {
39044  float32x2_t __ret;
39045  __ret = (float32x2_t)(__p0);
39046  return __ret;
39047}
39048#endif
39049
39050#ifdef __LITTLE_ENDIAN__
39051__ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
39052  float32x2_t __ret;
39053  __ret = (float32x2_t)(__p0);
39054  return __ret;
39055}
39056#else
39057__ai float32x2_t vreinterpret_f32_u16(uint16x4_t __p0) {
39058  float32x2_t __ret;
39059  __ret = (float32x2_t)(__p0);
39060  return __ret;
39061}
39062#endif
39063
39064#ifdef __LITTLE_ENDIAN__
39065__ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
39066  float32x2_t __ret;
39067  __ret = (float32x2_t)(__p0);
39068  return __ret;
39069}
39070#else
39071__ai float32x2_t vreinterpret_f32_s8(int8x8_t __p0) {
39072  float32x2_t __ret;
39073  __ret = (float32x2_t)(__p0);
39074  return __ret;
39075}
39076#endif
39077
39078#ifdef __LITTLE_ENDIAN__
39079__ai float32x2_t vreinterpret_f32_f64(float64x1_t __p0) {
39080  float32x2_t __ret;
39081  __ret = (float32x2_t)(__p0);
39082  return __ret;
39083}
39084#else
39085__ai float32x2_t vreinterpret_f32_f64(float64x1_t __p0) {
39086  float32x2_t __ret;
39087  __ret = (float32x2_t)(__p0);
39088  return __ret;
39089}
39090#endif
39091
39092#ifdef __LITTLE_ENDIAN__
39093__ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
39094  float32x2_t __ret;
39095  __ret = (float32x2_t)(__p0);
39096  return __ret;
39097}
39098#else
39099__ai float32x2_t vreinterpret_f32_f16(float16x4_t __p0) {
39100  float32x2_t __ret;
39101  __ret = (float32x2_t)(__p0);
39102  return __ret;
39103}
39104#endif
39105
39106#ifdef __LITTLE_ENDIAN__
39107__ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
39108  float32x2_t __ret;
39109  __ret = (float32x2_t)(__p0);
39110  return __ret;
39111}
39112#else
39113__ai float32x2_t vreinterpret_f32_s32(int32x2_t __p0) {
39114  float32x2_t __ret;
39115  __ret = (float32x2_t)(__p0);
39116  return __ret;
39117}
39118#endif
39119
39120#ifdef __LITTLE_ENDIAN__
39121__ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
39122  float32x2_t __ret;
39123  __ret = (float32x2_t)(__p0);
39124  return __ret;
39125}
39126#else
39127__ai float32x2_t vreinterpret_f32_s64(int64x1_t __p0) {
39128  float32x2_t __ret;
39129  __ret = (float32x2_t)(__p0);
39130  return __ret;
39131}
39132#endif
39133
39134#ifdef __LITTLE_ENDIAN__
39135__ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
39136  float32x2_t __ret;
39137  __ret = (float32x2_t)(__p0);
39138  return __ret;
39139}
39140#else
39141__ai float32x2_t vreinterpret_f32_s16(int16x4_t __p0) {
39142  float32x2_t __ret;
39143  __ret = (float32x2_t)(__p0);
39144  return __ret;
39145}
39146#endif
39147
39148#ifdef __LITTLE_ENDIAN__
39149__ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
39150  float16x4_t __ret;
39151  __ret = (float16x4_t)(__p0);
39152  return __ret;
39153}
39154#else
39155__ai float16x4_t vreinterpret_f16_p8(poly8x8_t __p0) {
39156  float16x4_t __ret;
39157  __ret = (float16x4_t)(__p0);
39158  return __ret;
39159}
39160#endif
39161
39162#ifdef __LITTLE_ENDIAN__
39163__ai float16x4_t vreinterpret_f16_p64(poly64x1_t __p0) {
39164  float16x4_t __ret;
39165  __ret = (float16x4_t)(__p0);
39166  return __ret;
39167}
39168#else
39169__ai float16x4_t vreinterpret_f16_p64(poly64x1_t __p0) {
39170  float16x4_t __ret;
39171  __ret = (float16x4_t)(__p0);
39172  return __ret;
39173}
39174#endif
39175
39176#ifdef __LITTLE_ENDIAN__
39177__ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
39178  float16x4_t __ret;
39179  __ret = (float16x4_t)(__p0);
39180  return __ret;
39181}
39182#else
39183__ai float16x4_t vreinterpret_f16_p16(poly16x4_t __p0) {
39184  float16x4_t __ret;
39185  __ret = (float16x4_t)(__p0);
39186  return __ret;
39187}
39188#endif
39189
39190#ifdef __LITTLE_ENDIAN__
39191__ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
39192  float16x4_t __ret;
39193  __ret = (float16x4_t)(__p0);
39194  return __ret;
39195}
39196#else
39197__ai float16x4_t vreinterpret_f16_u8(uint8x8_t __p0) {
39198  float16x4_t __ret;
39199  __ret = (float16x4_t)(__p0);
39200  return __ret;
39201}
39202#endif
39203
39204#ifdef __LITTLE_ENDIAN__
39205__ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
39206  float16x4_t __ret;
39207  __ret = (float16x4_t)(__p0);
39208  return __ret;
39209}
39210#else
39211__ai float16x4_t vreinterpret_f16_u32(uint32x2_t __p0) {
39212  float16x4_t __ret;
39213  __ret = (float16x4_t)(__p0);
39214  return __ret;
39215}
39216#endif
39217
39218#ifdef __LITTLE_ENDIAN__
39219__ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
39220  float16x4_t __ret;
39221  __ret = (float16x4_t)(__p0);
39222  return __ret;
39223}
39224#else
39225__ai float16x4_t vreinterpret_f16_u64(uint64x1_t __p0) {
39226  float16x4_t __ret;
39227  __ret = (float16x4_t)(__p0);
39228  return __ret;
39229}
39230#endif
39231
39232#ifdef __LITTLE_ENDIAN__
39233__ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
39234  float16x4_t __ret;
39235  __ret = (float16x4_t)(__p0);
39236  return __ret;
39237}
39238#else
39239__ai float16x4_t vreinterpret_f16_u16(uint16x4_t __p0) {
39240  float16x4_t __ret;
39241  __ret = (float16x4_t)(__p0);
39242  return __ret;
39243}
39244#endif
39245
39246#ifdef __LITTLE_ENDIAN__
39247__ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
39248  float16x4_t __ret;
39249  __ret = (float16x4_t)(__p0);
39250  return __ret;
39251}
39252#else
39253__ai float16x4_t vreinterpret_f16_s8(int8x8_t __p0) {
39254  float16x4_t __ret;
39255  __ret = (float16x4_t)(__p0);
39256  return __ret;
39257}
39258#endif
39259
39260#ifdef __LITTLE_ENDIAN__
39261__ai float16x4_t vreinterpret_f16_f64(float64x1_t __p0) {
39262  float16x4_t __ret;
39263  __ret = (float16x4_t)(__p0);
39264  return __ret;
39265}
39266#else
39267__ai float16x4_t vreinterpret_f16_f64(float64x1_t __p0) {
39268  float16x4_t __ret;
39269  __ret = (float16x4_t)(__p0);
39270  return __ret;
39271}
39272#endif
39273
39274#ifdef __LITTLE_ENDIAN__
39275__ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
39276  float16x4_t __ret;
39277  __ret = (float16x4_t)(__p0);
39278  return __ret;
39279}
39280#else
39281__ai float16x4_t vreinterpret_f16_f32(float32x2_t __p0) {
39282  float16x4_t __ret;
39283  __ret = (float16x4_t)(__p0);
39284  return __ret;
39285}
39286#endif
39287
39288#ifdef __LITTLE_ENDIAN__
39289__ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
39290  float16x4_t __ret;
39291  __ret = (float16x4_t)(__p0);
39292  return __ret;
39293}
39294#else
39295__ai float16x4_t vreinterpret_f16_s32(int32x2_t __p0) {
39296  float16x4_t __ret;
39297  __ret = (float16x4_t)(__p0);
39298  return __ret;
39299}
39300#endif
39301
39302#ifdef __LITTLE_ENDIAN__
39303__ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
39304  float16x4_t __ret;
39305  __ret = (float16x4_t)(__p0);
39306  return __ret;
39307}
39308#else
39309__ai float16x4_t vreinterpret_f16_s64(int64x1_t __p0) {
39310  float16x4_t __ret;
39311  __ret = (float16x4_t)(__p0);
39312  return __ret;
39313}
39314#endif
39315
39316#ifdef __LITTLE_ENDIAN__
39317__ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
39318  float16x4_t __ret;
39319  __ret = (float16x4_t)(__p0);
39320  return __ret;
39321}
39322#else
39323__ai float16x4_t vreinterpret_f16_s16(int16x4_t __p0) {
39324  float16x4_t __ret;
39325  __ret = (float16x4_t)(__p0);
39326  return __ret;
39327}
39328#endif
39329
39330#ifdef __LITTLE_ENDIAN__
39331__ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
39332  int32x2_t __ret;
39333  __ret = (int32x2_t)(__p0);
39334  return __ret;
39335}
39336#else
39337__ai int32x2_t vreinterpret_s32_p8(poly8x8_t __p0) {
39338  int32x2_t __ret;
39339  __ret = (int32x2_t)(__p0);
39340  return __ret;
39341}
39342#endif
39343
39344#ifdef __LITTLE_ENDIAN__
39345__ai int32x2_t vreinterpret_s32_p64(poly64x1_t __p0) {
39346  int32x2_t __ret;
39347  __ret = (int32x2_t)(__p0);
39348  return __ret;
39349}
39350#else
39351__ai int32x2_t vreinterpret_s32_p64(poly64x1_t __p0) {
39352  int32x2_t __ret;
39353  __ret = (int32x2_t)(__p0);
39354  return __ret;
39355}
39356#endif
39357
39358#ifdef __LITTLE_ENDIAN__
39359__ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
39360  int32x2_t __ret;
39361  __ret = (int32x2_t)(__p0);
39362  return __ret;
39363}
39364#else
39365__ai int32x2_t vreinterpret_s32_p16(poly16x4_t __p0) {
39366  int32x2_t __ret;
39367  __ret = (int32x2_t)(__p0);
39368  return __ret;
39369}
39370#endif
39371
39372#ifdef __LITTLE_ENDIAN__
39373__ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
39374  int32x2_t __ret;
39375  __ret = (int32x2_t)(__p0);
39376  return __ret;
39377}
39378#else
39379__ai int32x2_t vreinterpret_s32_u8(uint8x8_t __p0) {
39380  int32x2_t __ret;
39381  __ret = (int32x2_t)(__p0);
39382  return __ret;
39383}
39384#endif
39385
39386#ifdef __LITTLE_ENDIAN__
39387__ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
39388  int32x2_t __ret;
39389  __ret = (int32x2_t)(__p0);
39390  return __ret;
39391}
39392#else
39393__ai int32x2_t vreinterpret_s32_u32(uint32x2_t __p0) {
39394  int32x2_t __ret;
39395  __ret = (int32x2_t)(__p0);
39396  return __ret;
39397}
39398#endif
39399
39400#ifdef __LITTLE_ENDIAN__
39401__ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
39402  int32x2_t __ret;
39403  __ret = (int32x2_t)(__p0);
39404  return __ret;
39405}
39406#else
39407__ai int32x2_t vreinterpret_s32_u64(uint64x1_t __p0) {
39408  int32x2_t __ret;
39409  __ret = (int32x2_t)(__p0);
39410  return __ret;
39411}
39412#endif
39413
39414#ifdef __LITTLE_ENDIAN__
39415__ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
39416  int32x2_t __ret;
39417  __ret = (int32x2_t)(__p0);
39418  return __ret;
39419}
39420#else
39421__ai int32x2_t vreinterpret_s32_u16(uint16x4_t __p0) {
39422  int32x2_t __ret;
39423  __ret = (int32x2_t)(__p0);
39424  return __ret;
39425}
39426#endif
39427
39428#ifdef __LITTLE_ENDIAN__
39429__ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
39430  int32x2_t __ret;
39431  __ret = (int32x2_t)(__p0);
39432  return __ret;
39433}
39434#else
39435__ai int32x2_t vreinterpret_s32_s8(int8x8_t __p0) {
39436  int32x2_t __ret;
39437  __ret = (int32x2_t)(__p0);
39438  return __ret;
39439}
39440#endif
39441
39442#ifdef __LITTLE_ENDIAN__
39443__ai int32x2_t vreinterpret_s32_f64(float64x1_t __p0) {
39444  int32x2_t __ret;
39445  __ret = (int32x2_t)(__p0);
39446  return __ret;
39447}
39448#else
39449__ai int32x2_t vreinterpret_s32_f64(float64x1_t __p0) {
39450  int32x2_t __ret;
39451  __ret = (int32x2_t)(__p0);
39452  return __ret;
39453}
39454#endif
39455
39456#ifdef __LITTLE_ENDIAN__
39457__ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
39458  int32x2_t __ret;
39459  __ret = (int32x2_t)(__p0);
39460  return __ret;
39461}
39462#else
39463__ai int32x2_t vreinterpret_s32_f32(float32x2_t __p0) {
39464  int32x2_t __ret;
39465  __ret = (int32x2_t)(__p0);
39466  return __ret;
39467}
39468#endif
39469
39470#ifdef __LITTLE_ENDIAN__
39471__ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
39472  int32x2_t __ret;
39473  __ret = (int32x2_t)(__p0);
39474  return __ret;
39475}
39476#else
39477__ai int32x2_t vreinterpret_s32_f16(float16x4_t __p0) {
39478  int32x2_t __ret;
39479  __ret = (int32x2_t)(__p0);
39480  return __ret;
39481}
39482#endif
39483
39484#ifdef __LITTLE_ENDIAN__
39485__ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
39486  int32x2_t __ret;
39487  __ret = (int32x2_t)(__p0);
39488  return __ret;
39489}
39490#else
39491__ai int32x2_t vreinterpret_s32_s64(int64x1_t __p0) {
39492  int32x2_t __ret;
39493  __ret = (int32x2_t)(__p0);
39494  return __ret;
39495}
39496#endif
39497
39498#ifdef __LITTLE_ENDIAN__
39499__ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
39500  int32x2_t __ret;
39501  __ret = (int32x2_t)(__p0);
39502  return __ret;
39503}
39504#else
39505__ai int32x2_t vreinterpret_s32_s16(int16x4_t __p0) {
39506  int32x2_t __ret;
39507  __ret = (int32x2_t)(__p0);
39508  return __ret;
39509}
39510#endif
39511
39512#ifdef __LITTLE_ENDIAN__
39513__ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
39514  int64x1_t __ret;
39515  __ret = (int64x1_t)(__p0);
39516  return __ret;
39517}
39518#else
39519__ai int64x1_t vreinterpret_s64_p8(poly8x8_t __p0) {
39520  int64x1_t __ret;
39521  __ret = (int64x1_t)(__p0);
39522  return __ret;
39523}
39524#endif
39525
39526#ifdef __LITTLE_ENDIAN__
39527__ai int64x1_t vreinterpret_s64_p64(poly64x1_t __p0) {
39528  int64x1_t __ret;
39529  __ret = (int64x1_t)(__p0);
39530  return __ret;
39531}
39532#else
39533__ai int64x1_t vreinterpret_s64_p64(poly64x1_t __p0) {
39534  int64x1_t __ret;
39535  __ret = (int64x1_t)(__p0);
39536  return __ret;
39537}
39538#endif
39539
39540#ifdef __LITTLE_ENDIAN__
39541__ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
39542  int64x1_t __ret;
39543  __ret = (int64x1_t)(__p0);
39544  return __ret;
39545}
39546#else
39547__ai int64x1_t vreinterpret_s64_p16(poly16x4_t __p0) {
39548  int64x1_t __ret;
39549  __ret = (int64x1_t)(__p0);
39550  return __ret;
39551}
39552#endif
39553
39554#ifdef __LITTLE_ENDIAN__
39555__ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
39556  int64x1_t __ret;
39557  __ret = (int64x1_t)(__p0);
39558  return __ret;
39559}
39560#else
39561__ai int64x1_t vreinterpret_s64_u8(uint8x8_t __p0) {
39562  int64x1_t __ret;
39563  __ret = (int64x1_t)(__p0);
39564  return __ret;
39565}
39566#endif
39567
39568#ifdef __LITTLE_ENDIAN__
39569__ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
39570  int64x1_t __ret;
39571  __ret = (int64x1_t)(__p0);
39572  return __ret;
39573}
39574#else
39575__ai int64x1_t vreinterpret_s64_u32(uint32x2_t __p0) {
39576  int64x1_t __ret;
39577  __ret = (int64x1_t)(__p0);
39578  return __ret;
39579}
39580#endif
39581
39582#ifdef __LITTLE_ENDIAN__
39583__ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
39584  int64x1_t __ret;
39585  __ret = (int64x1_t)(__p0);
39586  return __ret;
39587}
39588#else
39589__ai int64x1_t vreinterpret_s64_u64(uint64x1_t __p0) {
39590  int64x1_t __ret;
39591  __ret = (int64x1_t)(__p0);
39592  return __ret;
39593}
39594#endif
39595
39596#ifdef __LITTLE_ENDIAN__
39597__ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
39598  int64x1_t __ret;
39599  __ret = (int64x1_t)(__p0);
39600  return __ret;
39601}
39602#else
39603__ai int64x1_t vreinterpret_s64_u16(uint16x4_t __p0) {
39604  int64x1_t __ret;
39605  __ret = (int64x1_t)(__p0);
39606  return __ret;
39607}
39608#endif
39609
39610#ifdef __LITTLE_ENDIAN__
39611__ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
39612  int64x1_t __ret;
39613  __ret = (int64x1_t)(__p0);
39614  return __ret;
39615}
39616#else
39617__ai int64x1_t vreinterpret_s64_s8(int8x8_t __p0) {
39618  int64x1_t __ret;
39619  __ret = (int64x1_t)(__p0);
39620  return __ret;
39621}
39622#endif
39623
39624#ifdef __LITTLE_ENDIAN__
39625__ai int64x1_t vreinterpret_s64_f64(float64x1_t __p0) {
39626  int64x1_t __ret;
39627  __ret = (int64x1_t)(__p0);
39628  return __ret;
39629}
39630#else
39631__ai int64x1_t vreinterpret_s64_f64(float64x1_t __p0) {
39632  int64x1_t __ret;
39633  __ret = (int64x1_t)(__p0);
39634  return __ret;
39635}
39636#endif
39637
39638#ifdef __LITTLE_ENDIAN__
39639__ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
39640  int64x1_t __ret;
39641  __ret = (int64x1_t)(__p0);
39642  return __ret;
39643}
39644#else
39645__ai int64x1_t vreinterpret_s64_f32(float32x2_t __p0) {
39646  int64x1_t __ret;
39647  __ret = (int64x1_t)(__p0);
39648  return __ret;
39649}
39650#endif
39651
39652#ifdef __LITTLE_ENDIAN__
39653__ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
39654  int64x1_t __ret;
39655  __ret = (int64x1_t)(__p0);
39656  return __ret;
39657}
39658#else
39659__ai int64x1_t vreinterpret_s64_f16(float16x4_t __p0) {
39660  int64x1_t __ret;
39661  __ret = (int64x1_t)(__p0);
39662  return __ret;
39663}
39664#endif
39665
39666#ifdef __LITTLE_ENDIAN__
39667__ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
39668  int64x1_t __ret;
39669  __ret = (int64x1_t)(__p0);
39670  return __ret;
39671}
39672#else
39673__ai int64x1_t vreinterpret_s64_s32(int32x2_t __p0) {
39674  int64x1_t __ret;
39675  __ret = (int64x1_t)(__p0);
39676  return __ret;
39677}
39678#endif
39679
39680#ifdef __LITTLE_ENDIAN__
39681__ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
39682  int64x1_t __ret;
39683  __ret = (int64x1_t)(__p0);
39684  return __ret;
39685}
39686#else
39687__ai int64x1_t vreinterpret_s64_s16(int16x4_t __p0) {
39688  int64x1_t __ret;
39689  __ret = (int64x1_t)(__p0);
39690  return __ret;
39691}
39692#endif
39693
39694#ifdef __LITTLE_ENDIAN__
39695__ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
39696  int16x4_t __ret;
39697  __ret = (int16x4_t)(__p0);
39698  return __ret;
39699}
39700#else
39701__ai int16x4_t vreinterpret_s16_p8(poly8x8_t __p0) {
39702  int16x4_t __ret;
39703  __ret = (int16x4_t)(__p0);
39704  return __ret;
39705}
39706#endif
39707
39708#ifdef __LITTLE_ENDIAN__
39709__ai int16x4_t vreinterpret_s16_p64(poly64x1_t __p0) {
39710  int16x4_t __ret;
39711  __ret = (int16x4_t)(__p0);
39712  return __ret;
39713}
39714#else
39715__ai int16x4_t vreinterpret_s16_p64(poly64x1_t __p0) {
39716  int16x4_t __ret;
39717  __ret = (int16x4_t)(__p0);
39718  return __ret;
39719}
39720#endif
39721
39722#ifdef __LITTLE_ENDIAN__
39723__ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
39724  int16x4_t __ret;
39725  __ret = (int16x4_t)(__p0);
39726  return __ret;
39727}
39728#else
39729__ai int16x4_t vreinterpret_s16_p16(poly16x4_t __p0) {
39730  int16x4_t __ret;
39731  __ret = (int16x4_t)(__p0);
39732  return __ret;
39733}
39734#endif
39735
39736#ifdef __LITTLE_ENDIAN__
39737__ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
39738  int16x4_t __ret;
39739  __ret = (int16x4_t)(__p0);
39740  return __ret;
39741}
39742#else
39743__ai int16x4_t vreinterpret_s16_u8(uint8x8_t __p0) {
39744  int16x4_t __ret;
39745  __ret = (int16x4_t)(__p0);
39746  return __ret;
39747}
39748#endif
39749
39750#ifdef __LITTLE_ENDIAN__
39751__ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
39752  int16x4_t __ret;
39753  __ret = (int16x4_t)(__p0);
39754  return __ret;
39755}
39756#else
39757__ai int16x4_t vreinterpret_s16_u32(uint32x2_t __p0) {
39758  int16x4_t __ret;
39759  __ret = (int16x4_t)(__p0);
39760  return __ret;
39761}
39762#endif
39763
39764#ifdef __LITTLE_ENDIAN__
39765__ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
39766  int16x4_t __ret;
39767  __ret = (int16x4_t)(__p0);
39768  return __ret;
39769}
39770#else
39771__ai int16x4_t vreinterpret_s16_u64(uint64x1_t __p0) {
39772  int16x4_t __ret;
39773  __ret = (int16x4_t)(__p0);
39774  return __ret;
39775}
39776#endif
39777
39778#ifdef __LITTLE_ENDIAN__
39779__ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
39780  int16x4_t __ret;
39781  __ret = (int16x4_t)(__p0);
39782  return __ret;
39783}
39784#else
39785__ai int16x4_t vreinterpret_s16_u16(uint16x4_t __p0) {
39786  int16x4_t __ret;
39787  __ret = (int16x4_t)(__p0);
39788  return __ret;
39789}
39790#endif
39791
39792#ifdef __LITTLE_ENDIAN__
39793__ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
39794  int16x4_t __ret;
39795  __ret = (int16x4_t)(__p0);
39796  return __ret;
39797}
39798#else
39799__ai int16x4_t vreinterpret_s16_s8(int8x8_t __p0) {
39800  int16x4_t __ret;
39801  __ret = (int16x4_t)(__p0);
39802  return __ret;
39803}
39804#endif
39805
39806#ifdef __LITTLE_ENDIAN__
39807__ai int16x4_t vreinterpret_s16_f64(float64x1_t __p0) {
39808  int16x4_t __ret;
39809  __ret = (int16x4_t)(__p0);
39810  return __ret;
39811}
39812#else
39813__ai int16x4_t vreinterpret_s16_f64(float64x1_t __p0) {
39814  int16x4_t __ret;
39815  __ret = (int16x4_t)(__p0);
39816  return __ret;
39817}
39818#endif
39819
39820#ifdef __LITTLE_ENDIAN__
39821__ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
39822  int16x4_t __ret;
39823  __ret = (int16x4_t)(__p0);
39824  return __ret;
39825}
39826#else
39827__ai int16x4_t vreinterpret_s16_f32(float32x2_t __p0) {
39828  int16x4_t __ret;
39829  __ret = (int16x4_t)(__p0);
39830  return __ret;
39831}
39832#endif
39833
39834#ifdef __LITTLE_ENDIAN__
39835__ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
39836  int16x4_t __ret;
39837  __ret = (int16x4_t)(__p0);
39838  return __ret;
39839}
39840#else
39841__ai int16x4_t vreinterpret_s16_f16(float16x4_t __p0) {
39842  int16x4_t __ret;
39843  __ret = (int16x4_t)(__p0);
39844  return __ret;
39845}
39846#endif
39847
39848#ifdef __LITTLE_ENDIAN__
39849__ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
39850  int16x4_t __ret;
39851  __ret = (int16x4_t)(__p0);
39852  return __ret;
39853}
39854#else
39855__ai int16x4_t vreinterpret_s16_s32(int32x2_t __p0) {
39856  int16x4_t __ret;
39857  __ret = (int16x4_t)(__p0);
39858  return __ret;
39859}
39860#endif
39861
39862#ifdef __LITTLE_ENDIAN__
39863__ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
39864  int16x4_t __ret;
39865  __ret = (int16x4_t)(__p0);
39866  return __ret;
39867}
39868#else
39869__ai int16x4_t vreinterpret_s16_s64(int64x1_t __p0) {
39870  int16x4_t __ret;
39871  __ret = (int16x4_t)(__p0);
39872  return __ret;
39873}
39874#endif
39875
39876#endif
39877#if __ARM_ARCH >= 8 && defined(__aarch64__) && defined(__ARM_FEATURE_DIRECTED_ROUNDING)
39878#ifdef __LITTLE_ENDIAN__
39879__ai float64x2_t vrndq_f64(float64x2_t __p0) {
39880  float64x2_t __ret;
39881  __ret = (float64x2_t) __builtin_neon_vrndq_v((int8x16_t)__p0, 42);
39882  return __ret;
39883}
39884#else
39885__ai float64x2_t vrndq_f64(float64x2_t __p0) {
39886  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39887  float64x2_t __ret;
39888  __ret = (float64x2_t) __builtin_neon_vrndq_v((int8x16_t)__rev0, 42);
39889  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39890  return __ret;
39891}
39892#endif
39893
39894#ifdef __LITTLE_ENDIAN__
39895__ai float64x1_t vrnd_f64(float64x1_t __p0) {
39896  float64x1_t __ret;
39897  __ret = (float64x1_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 10);
39898  return __ret;
39899}
39900#else
39901__ai float64x1_t vrnd_f64(float64x1_t __p0) {
39902  float64x1_t __ret;
39903  __ret = (float64x1_t) __builtin_neon_vrnd_v((int8x8_t)__p0, 10);
39904  return __ret;
39905}
39906#endif
39907
39908#ifdef __LITTLE_ENDIAN__
39909__ai float64x2_t vrndaq_f64(float64x2_t __p0) {
39910  float64x2_t __ret;
39911  __ret = (float64x2_t) __builtin_neon_vrndaq_v((int8x16_t)__p0, 42);
39912  return __ret;
39913}
39914#else
39915__ai float64x2_t vrndaq_f64(float64x2_t __p0) {
39916  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39917  float64x2_t __ret;
39918  __ret = (float64x2_t) __builtin_neon_vrndaq_v((int8x16_t)__rev0, 42);
39919  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39920  return __ret;
39921}
39922#endif
39923
39924#ifdef __LITTLE_ENDIAN__
39925__ai float64x1_t vrnda_f64(float64x1_t __p0) {
39926  float64x1_t __ret;
39927  __ret = (float64x1_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 10);
39928  return __ret;
39929}
39930#else
39931__ai float64x1_t vrnda_f64(float64x1_t __p0) {
39932  float64x1_t __ret;
39933  __ret = (float64x1_t) __builtin_neon_vrnda_v((int8x8_t)__p0, 10);
39934  return __ret;
39935}
39936#endif
39937
39938#ifdef __LITTLE_ENDIAN__
39939__ai float64x2_t vrndiq_f64(float64x2_t __p0) {
39940  float64x2_t __ret;
39941  __ret = (float64x2_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 42);
39942  return __ret;
39943}
39944#else
39945__ai float64x2_t vrndiq_f64(float64x2_t __p0) {
39946  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39947  float64x2_t __ret;
39948  __ret = (float64x2_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 42);
39949  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39950  return __ret;
39951}
39952#endif
39953
39954#ifdef __LITTLE_ENDIAN__
39955__ai float32x4_t vrndiq_f32(float32x4_t __p0) {
39956  float32x4_t __ret;
39957  __ret = (float32x4_t) __builtin_neon_vrndiq_v((int8x16_t)__p0, 41);
39958  return __ret;
39959}
39960#else
39961__ai float32x4_t vrndiq_f32(float32x4_t __p0) {
39962  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
39963  float32x4_t __ret;
39964  __ret = (float32x4_t) __builtin_neon_vrndiq_v((int8x16_t)__rev0, 41);
39965  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
39966  return __ret;
39967}
39968#endif
39969
39970#ifdef __LITTLE_ENDIAN__
39971__ai float64x1_t vrndi_f64(float64x1_t __p0) {
39972  float64x1_t __ret;
39973  __ret = (float64x1_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 10);
39974  return __ret;
39975}
39976#else
39977__ai float64x1_t vrndi_f64(float64x1_t __p0) {
39978  float64x1_t __ret;
39979  __ret = (float64x1_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 10);
39980  return __ret;
39981}
39982#endif
39983
39984#ifdef __LITTLE_ENDIAN__
39985__ai float32x2_t vrndi_f32(float32x2_t __p0) {
39986  float32x2_t __ret;
39987  __ret = (float32x2_t) __builtin_neon_vrndi_v((int8x8_t)__p0, 9);
39988  return __ret;
39989}
39990#else
39991__ai float32x2_t vrndi_f32(float32x2_t __p0) {
39992  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
39993  float32x2_t __ret;
39994  __ret = (float32x2_t) __builtin_neon_vrndi_v((int8x8_t)__rev0, 9);
39995  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
39996  return __ret;
39997}
39998#endif
39999
40000#ifdef __LITTLE_ENDIAN__
40001__ai float64x2_t vrndmq_f64(float64x2_t __p0) {
40002  float64x2_t __ret;
40003  __ret = (float64x2_t) __builtin_neon_vrndmq_v((int8x16_t)__p0, 42);
40004  return __ret;
40005}
40006#else
40007__ai float64x2_t vrndmq_f64(float64x2_t __p0) {
40008  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40009  float64x2_t __ret;
40010  __ret = (float64x2_t) __builtin_neon_vrndmq_v((int8x16_t)__rev0, 42);
40011  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40012  return __ret;
40013}
40014#endif
40015
40016#ifdef __LITTLE_ENDIAN__
40017__ai float64x1_t vrndm_f64(float64x1_t __p0) {
40018  float64x1_t __ret;
40019  __ret = (float64x1_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 10);
40020  return __ret;
40021}
40022#else
40023__ai float64x1_t vrndm_f64(float64x1_t __p0) {
40024  float64x1_t __ret;
40025  __ret = (float64x1_t) __builtin_neon_vrndm_v((int8x8_t)__p0, 10);
40026  return __ret;
40027}
40028#endif
40029
40030#ifdef __LITTLE_ENDIAN__
40031__ai float64x2_t vrndnq_f64(float64x2_t __p0) {
40032  float64x2_t __ret;
40033  __ret = (float64x2_t) __builtin_neon_vrndnq_v((int8x16_t)__p0, 42);
40034  return __ret;
40035}
40036#else
40037__ai float64x2_t vrndnq_f64(float64x2_t __p0) {
40038  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40039  float64x2_t __ret;
40040  __ret = (float64x2_t) __builtin_neon_vrndnq_v((int8x16_t)__rev0, 42);
40041  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40042  return __ret;
40043}
40044#endif
40045
40046#ifdef __LITTLE_ENDIAN__
40047__ai float64x1_t vrndn_f64(float64x1_t __p0) {
40048  float64x1_t __ret;
40049  __ret = (float64x1_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 10);
40050  return __ret;
40051}
40052#else
40053__ai float64x1_t vrndn_f64(float64x1_t __p0) {
40054  float64x1_t __ret;
40055  __ret = (float64x1_t) __builtin_neon_vrndn_v((int8x8_t)__p0, 10);
40056  return __ret;
40057}
40058#endif
40059
40060#ifdef __LITTLE_ENDIAN__
40061__ai float64x2_t vrndpq_f64(float64x2_t __p0) {
40062  float64x2_t __ret;
40063  __ret = (float64x2_t) __builtin_neon_vrndpq_v((int8x16_t)__p0, 42);
40064  return __ret;
40065}
40066#else
40067__ai float64x2_t vrndpq_f64(float64x2_t __p0) {
40068  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40069  float64x2_t __ret;
40070  __ret = (float64x2_t) __builtin_neon_vrndpq_v((int8x16_t)__rev0, 42);
40071  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40072  return __ret;
40073}
40074#endif
40075
40076#ifdef __LITTLE_ENDIAN__
40077__ai float64x1_t vrndp_f64(float64x1_t __p0) {
40078  float64x1_t __ret;
40079  __ret = (float64x1_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 10);
40080  return __ret;
40081}
40082#else
40083__ai float64x1_t vrndp_f64(float64x1_t __p0) {
40084  float64x1_t __ret;
40085  __ret = (float64x1_t) __builtin_neon_vrndp_v((int8x8_t)__p0, 10);
40086  return __ret;
40087}
40088#endif
40089
40090#ifdef __LITTLE_ENDIAN__
40091__ai float64x2_t vrndxq_f64(float64x2_t __p0) {
40092  float64x2_t __ret;
40093  __ret = (float64x2_t) __builtin_neon_vrndxq_v((int8x16_t)__p0, 42);
40094  return __ret;
40095}
40096#else
40097__ai float64x2_t vrndxq_f64(float64x2_t __p0) {
40098  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40099  float64x2_t __ret;
40100  __ret = (float64x2_t) __builtin_neon_vrndxq_v((int8x16_t)__rev0, 42);
40101  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40102  return __ret;
40103}
40104#endif
40105
40106#ifdef __LITTLE_ENDIAN__
40107__ai float64x1_t vrndx_f64(float64x1_t __p0) {
40108  float64x1_t __ret;
40109  __ret = (float64x1_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 10);
40110  return __ret;
40111}
40112#else
40113__ai float64x1_t vrndx_f64(float64x1_t __p0) {
40114  float64x1_t __ret;
40115  __ret = (float64x1_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 10);
40116  return __ret;
40117}
40118#endif
40119
40120#endif
40121#if __ARM_ARCH >= 8 && defined(__aarch64__) && defined(__ARM_FEATURE_NUMERIC_MAXMIN)
40122#ifdef __LITTLE_ENDIAN__
40123__ai float64x2_t vmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
40124  float64x2_t __ret;
40125  __ret = (float64x2_t) __builtin_neon_vmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
40126  return __ret;
40127}
40128#else
40129__ai float64x2_t vmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
40130  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40131  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40132  float64x2_t __ret;
40133  __ret = (float64x2_t) __builtin_neon_vmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
40134  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40135  return __ret;
40136}
40137#endif
40138
40139#ifdef __LITTLE_ENDIAN__
40140__ai float64x1_t vmaxnm_f64(float64x1_t __p0, float64x1_t __p1) {
40141  float64x1_t __ret;
40142  __ret = (float64x1_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40143  return __ret;
40144}
40145#else
40146__ai float64x1_t vmaxnm_f64(float64x1_t __p0, float64x1_t __p1) {
40147  float64x1_t __ret;
40148  __ret = (float64x1_t) __builtin_neon_vmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40149  return __ret;
40150}
40151#endif
40152
40153#ifdef __LITTLE_ENDIAN__
40154__ai float64x2_t vminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
40155  float64x2_t __ret;
40156  __ret = (float64x2_t) __builtin_neon_vminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
40157  return __ret;
40158}
40159#else
40160__ai float64x2_t vminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
40161  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40162  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40163  float64x2_t __ret;
40164  __ret = (float64x2_t) __builtin_neon_vminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
40165  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40166  return __ret;
40167}
40168#endif
40169
40170#ifdef __LITTLE_ENDIAN__
40171__ai float64x1_t vminnm_f64(float64x1_t __p0, float64x1_t __p1) {
40172  float64x1_t __ret;
40173  __ret = (float64x1_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40174  return __ret;
40175}
40176#else
40177__ai float64x1_t vminnm_f64(float64x1_t __p0, float64x1_t __p1) {
40178  float64x1_t __ret;
40179  __ret = (float64x1_t) __builtin_neon_vminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40180  return __ret;
40181}
40182#endif
40183
40184#endif
40185#if __ARM_FEATURE_CRYPTO
40186#ifdef __LITTLE_ENDIAN__
40187__ai uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40188  uint8x16_t __ret;
40189  __ret = (uint8x16_t) __builtin_neon_vaesdq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
40190  return __ret;
40191}
40192#else
40193__ai uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40194  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40195  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40196  uint8x16_t __ret;
40197  __ret = (uint8x16_t) __builtin_neon_vaesdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
40198  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40199  return __ret;
40200}
40201#endif
40202
40203#ifdef __LITTLE_ENDIAN__
40204__ai uint8x16_t vaeseq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40205  uint8x16_t __ret;
40206  __ret = (uint8x16_t) __builtin_neon_vaeseq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
40207  return __ret;
40208}
40209#else
40210__ai uint8x16_t vaeseq_u8(uint8x16_t __p0, uint8x16_t __p1) {
40211  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40212  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40213  uint8x16_t __ret;
40214  __ret = (uint8x16_t) __builtin_neon_vaeseq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
40215  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40216  return __ret;
40217}
40218#endif
40219
40220#ifdef __LITTLE_ENDIAN__
40221__ai uint8x16_t vaesimcq_u8(uint8x16_t __p0) {
40222  uint8x16_t __ret;
40223  __ret = (uint8x16_t) __builtin_neon_vaesimcq_v((int8x16_t)__p0, 48);
40224  return __ret;
40225}
40226#else
40227__ai uint8x16_t vaesimcq_u8(uint8x16_t __p0) {
40228  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40229  uint8x16_t __ret;
40230  __ret = (uint8x16_t) __builtin_neon_vaesimcq_v((int8x16_t)__rev0, 48);
40231  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40232  return __ret;
40233}
40234#endif
40235
40236#ifdef __LITTLE_ENDIAN__
40237__ai uint8x16_t vaesmcq_u8(uint8x16_t __p0) {
40238  uint8x16_t __ret;
40239  __ret = (uint8x16_t) __builtin_neon_vaesmcq_v((int8x16_t)__p0, 48);
40240  return __ret;
40241}
40242#else
40243__ai uint8x16_t vaesmcq_u8(uint8x16_t __p0) {
40244  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40245  uint8x16_t __ret;
40246  __ret = (uint8x16_t) __builtin_neon_vaesmcq_v((int8x16_t)__rev0, 48);
40247  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
40248  return __ret;
40249}
40250#endif
40251
40252#ifdef __LITTLE_ENDIAN__
40253__ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40254  uint32x4_t __ret;
40255  __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
40256  return __ret;
40257}
40258#else
40259__ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40260  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40261  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40262  uint32x4_t __ret;
40263  __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
40264  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40265  return __ret;
40266}
40267#endif
40268
40269#ifdef __LITTLE_ENDIAN__
40270__ai uint32_t vsha1h_u32(uint32_t __p0) {
40271  uint32_t __ret;
40272  __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0);
40273  return __ret;
40274}
40275#else
40276__ai uint32_t vsha1h_u32(uint32_t __p0) {
40277  uint32_t __ret;
40278  __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0);
40279  return __ret;
40280}
40281#endif
40282
40283#ifdef __LITTLE_ENDIAN__
40284__ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40285  uint32x4_t __ret;
40286  __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
40287  return __ret;
40288}
40289#else
40290__ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40291  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40292  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40293  uint32x4_t __ret;
40294  __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
40295  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40296  return __ret;
40297}
40298#endif
40299
40300#ifdef __LITTLE_ENDIAN__
40301__ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40302  uint32x4_t __ret;
40303  __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32((int8x16_t)__p0, __p1, (int8x16_t)__p2);
40304  return __ret;
40305}
40306#else
40307__ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) {
40308  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40309  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40310  uint32x4_t __ret;
40311  __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32((int8x16_t)__rev0, __p1, (int8x16_t)__rev2);
40312  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40313  return __ret;
40314}
40315#endif
40316
40317#ifdef __LITTLE_ENDIAN__
40318__ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40319  uint32x4_t __ret;
40320  __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40321  return __ret;
40322}
40323#else
40324__ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40325  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40326  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40327  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40328  uint32x4_t __ret;
40329  __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40330  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40331  return __ret;
40332}
40333#endif
40334
40335#ifdef __LITTLE_ENDIAN__
40336__ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40337  uint32x4_t __ret;
40338  __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
40339  return __ret;
40340}
40341#else
40342__ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40343  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40344  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40345  uint32x4_t __ret;
40346  __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
40347  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40348  return __ret;
40349}
40350#endif
40351
40352#ifdef __LITTLE_ENDIAN__
40353__ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40354  uint32x4_t __ret;
40355  __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40356  return __ret;
40357}
40358#else
40359__ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40360  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40361  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40362  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40363  uint32x4_t __ret;
40364  __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40365  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40366  return __ret;
40367}
40368#endif
40369
40370#ifdef __LITTLE_ENDIAN__
40371__ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40372  uint32x4_t __ret;
40373  __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40374  return __ret;
40375}
40376#else
40377__ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40378  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40379  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40380  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40381  uint32x4_t __ret;
40382  __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40383  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40384  return __ret;
40385}
40386#endif
40387
40388#ifdef __LITTLE_ENDIAN__
40389__ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40390  uint32x4_t __ret;
40391  __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
40392  return __ret;
40393}
40394#else
40395__ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) {
40396  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40397  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40398  uint32x4_t __ret;
40399  __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
40400  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40401  return __ret;
40402}
40403#endif
40404
40405#ifdef __LITTLE_ENDIAN__
40406__ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40407  uint32x4_t __ret;
40408  __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50);
40409  return __ret;
40410}
40411#else
40412__ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
40413  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40414  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40415  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40416  uint32x4_t __ret;
40417  __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50);
40418  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40419  return __ret;
40420}
40421#endif
40422
40423#endif
40424#if defined(__ARM_FEATURE_QRDMX)
40425#ifdef __LITTLE_ENDIAN__
40426__ai int32x4_t vqrdmlahq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
40427  int32x4_t __ret;
40428  __ret = vqaddq_s32(__p0, vqrdmulhq_s32(__p1, __p2));
40429  return __ret;
40430}
40431#else
40432__ai int32x4_t vqrdmlahq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
40433  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40434  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40435  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40436  int32x4_t __ret;
40437  __ret = __noswap_vqaddq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __rev2));
40438  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40439  return __ret;
40440}
40441#endif
40442
40443#ifdef __LITTLE_ENDIAN__
40444__ai int16x8_t vqrdmlahq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
40445  int16x8_t __ret;
40446  __ret = vqaddq_s16(__p0, vqrdmulhq_s16(__p1, __p2));
40447  return __ret;
40448}
40449#else
40450__ai int16x8_t vqrdmlahq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
40451  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40452  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
40453  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
40454  int16x8_t __ret;
40455  __ret = __noswap_vqaddq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __rev2));
40456  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40457  return __ret;
40458}
40459#endif
40460
40461#ifdef __LITTLE_ENDIAN__
40462__ai int32x2_t vqrdmlah_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
40463  int32x2_t __ret;
40464  __ret = vqadd_s32(__p0, vqrdmulh_s32(__p1, __p2));
40465  return __ret;
40466}
40467#else
40468__ai int32x2_t vqrdmlah_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
40469  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40470  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40471  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
40472  int32x2_t __ret;
40473  __ret = __noswap_vqadd_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __rev2));
40474  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40475  return __ret;
40476}
40477#endif
40478
40479#ifdef __LITTLE_ENDIAN__
40480__ai int16x4_t vqrdmlah_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
40481  int16x4_t __ret;
40482  __ret = vqadd_s16(__p0, vqrdmulh_s16(__p1, __p2));
40483  return __ret;
40484}
40485#else
40486__ai int16x4_t vqrdmlah_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
40487  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40488  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40489  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40490  int16x4_t __ret;
40491  __ret = __noswap_vqadd_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __rev2));
40492  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40493  return __ret;
40494}
40495#endif
40496
40497#ifdef __LITTLE_ENDIAN__
40498#define vqrdmlahq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40499  int32x4_t __s0 = __p0; \
40500  int32x4_t __s1 = __p1; \
40501  int32x2_t __s2 = __p2; \
40502  int32x4_t __ret; \
40503  __ret = vqaddq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40504  __ret; \
40505})
40506#else
40507#define vqrdmlahq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40508  int32x4_t __s0 = __p0; \
40509  int32x4_t __s1 = __p1; \
40510  int32x2_t __s2 = __p2; \
40511  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40512  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40513  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
40514  int32x4_t __ret; \
40515  __ret = __noswap_vqaddq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40516  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40517  __ret; \
40518})
40519#endif
40520
40521#ifdef __LITTLE_ENDIAN__
40522#define vqrdmlahq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40523  int16x8_t __s0 = __p0; \
40524  int16x8_t __s1 = __p1; \
40525  int16x4_t __s2 = __p2; \
40526  int16x8_t __ret; \
40527  __ret = vqaddq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40528  __ret; \
40529})
40530#else
40531#define vqrdmlahq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40532  int16x8_t __s0 = __p0; \
40533  int16x8_t __s1 = __p1; \
40534  int16x4_t __s2 = __p2; \
40535  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
40536  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
40537  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40538  int16x8_t __ret; \
40539  __ret = __noswap_vqaddq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40540  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
40541  __ret; \
40542})
40543#endif
40544
40545#ifdef __LITTLE_ENDIAN__
40546#define vqrdmlah_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40547  int32x2_t __s0 = __p0; \
40548  int32x2_t __s1 = __p1; \
40549  int32x2_t __s2 = __p2; \
40550  int32x2_t __ret; \
40551  __ret = vqadd_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
40552  __ret; \
40553})
40554#else
40555#define vqrdmlah_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40556  int32x2_t __s0 = __p0; \
40557  int32x2_t __s1 = __p1; \
40558  int32x2_t __s2 = __p2; \
40559  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
40560  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
40561  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
40562  int32x2_t __ret; \
40563  __ret = __noswap_vqadd_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
40564  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
40565  __ret; \
40566})
40567#endif
40568
40569#ifdef __LITTLE_ENDIAN__
40570#define vqrdmlah_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40571  int16x4_t __s0 = __p0; \
40572  int16x4_t __s1 = __p1; \
40573  int16x4_t __s2 = __p2; \
40574  int16x4_t __ret; \
40575  __ret = vqadd_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40576  __ret; \
40577})
40578#else
40579#define vqrdmlah_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40580  int16x4_t __s0 = __p0; \
40581  int16x4_t __s1 = __p1; \
40582  int16x4_t __s2 = __p2; \
40583  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40584  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40585  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40586  int16x4_t __ret; \
40587  __ret = __noswap_vqadd_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40588  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40589  __ret; \
40590})
40591#endif
40592
40593#ifdef __LITTLE_ENDIAN__
40594__ai int32x4_t vqrdmlshq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
40595  int32x4_t __ret;
40596  __ret = vqsubq_s32(__p0, vqrdmulhq_s32(__p1, __p2));
40597  return __ret;
40598}
40599#else
40600__ai int32x4_t vqrdmlshq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
40601  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40602  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40603  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40604  int32x4_t __ret;
40605  __ret = __noswap_vqsubq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __rev2));
40606  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40607  return __ret;
40608}
40609#endif
40610
40611#ifdef __LITTLE_ENDIAN__
40612__ai int16x8_t vqrdmlshq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
40613  int16x8_t __ret;
40614  __ret = vqsubq_s16(__p0, vqrdmulhq_s16(__p1, __p2));
40615  return __ret;
40616}
40617#else
40618__ai int16x8_t vqrdmlshq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
40619  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
40620  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
40621  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
40622  int16x8_t __ret;
40623  __ret = __noswap_vqsubq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __rev2));
40624  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
40625  return __ret;
40626}
40627#endif
40628
40629#ifdef __LITTLE_ENDIAN__
40630__ai int32x2_t vqrdmlsh_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
40631  int32x2_t __ret;
40632  __ret = vqsub_s32(__p0, vqrdmulh_s32(__p1, __p2));
40633  return __ret;
40634}
40635#else
40636__ai int32x2_t vqrdmlsh_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
40637  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40638  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40639  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
40640  int32x2_t __ret;
40641  __ret = __noswap_vqsub_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __rev2));
40642  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40643  return __ret;
40644}
40645#endif
40646
40647#ifdef __LITTLE_ENDIAN__
40648__ai int16x4_t vqrdmlsh_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
40649  int16x4_t __ret;
40650  __ret = vqsub_s16(__p0, vqrdmulh_s16(__p1, __p2));
40651  return __ret;
40652}
40653#else
40654__ai int16x4_t vqrdmlsh_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
40655  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
40656  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
40657  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
40658  int16x4_t __ret;
40659  __ret = __noswap_vqsub_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __rev2));
40660  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
40661  return __ret;
40662}
40663#endif
40664
40665#ifdef __LITTLE_ENDIAN__
40666#define vqrdmlshq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40667  int32x4_t __s0 = __p0; \
40668  int32x4_t __s1 = __p1; \
40669  int32x2_t __s2 = __p2; \
40670  int32x4_t __ret; \
40671  __ret = vqsubq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40672  __ret; \
40673})
40674#else
40675#define vqrdmlshq_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40676  int32x4_t __s0 = __p0; \
40677  int32x4_t __s1 = __p1; \
40678  int32x2_t __s2 = __p2; \
40679  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40680  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40681  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
40682  int32x4_t __ret; \
40683  __ret = __noswap_vqsubq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40684  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40685  __ret; \
40686})
40687#endif
40688
40689#ifdef __LITTLE_ENDIAN__
40690#define vqrdmlshq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40691  int16x8_t __s0 = __p0; \
40692  int16x8_t __s1 = __p1; \
40693  int16x4_t __s2 = __p2; \
40694  int16x8_t __ret; \
40695  __ret = vqsubq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40696  __ret; \
40697})
40698#else
40699#define vqrdmlshq_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40700  int16x8_t __s0 = __p0; \
40701  int16x8_t __s1 = __p1; \
40702  int16x4_t __s2 = __p2; \
40703  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
40704  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
40705  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40706  int16x8_t __ret; \
40707  __ret = __noswap_vqsubq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40708  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
40709  __ret; \
40710})
40711#endif
40712
40713#ifdef __LITTLE_ENDIAN__
40714#define vqrdmlsh_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40715  int32x2_t __s0 = __p0; \
40716  int32x2_t __s1 = __p1; \
40717  int32x2_t __s2 = __p2; \
40718  int32x2_t __ret; \
40719  __ret = vqsub_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
40720  __ret; \
40721})
40722#else
40723#define vqrdmlsh_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40724  int32x2_t __s0 = __p0; \
40725  int32x2_t __s1 = __p1; \
40726  int32x2_t __s2 = __p2; \
40727  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
40728  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
40729  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
40730  int32x2_t __ret; \
40731  __ret = __noswap_vqsub_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
40732  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
40733  __ret; \
40734})
40735#endif
40736
40737#ifdef __LITTLE_ENDIAN__
40738#define vqrdmlsh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40739  int16x4_t __s0 = __p0; \
40740  int16x4_t __s1 = __p1; \
40741  int16x4_t __s2 = __p2; \
40742  int16x4_t __ret; \
40743  __ret = vqsub_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40744  __ret; \
40745})
40746#else
40747#define vqrdmlsh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40748  int16x4_t __s0 = __p0; \
40749  int16x4_t __s1 = __p1; \
40750  int16x4_t __s2 = __p2; \
40751  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40752  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40753  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40754  int16x4_t __ret; \
40755  __ret = __noswap_vqsub_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40756  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40757  __ret; \
40758})
40759#endif
40760
40761#endif
40762#if defined(__ARM_FEATURE_QRDMX) && defined(__aarch64__)
40763#ifdef __LITTLE_ENDIAN__
40764#define vqrdmlahq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40765  int32x4_t __s0 = __p0; \
40766  int32x4_t __s1 = __p1; \
40767  int32x4_t __s2 = __p2; \
40768  int32x4_t __ret; \
40769  __ret = vqaddq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40770  __ret; \
40771})
40772#else
40773#define vqrdmlahq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40774  int32x4_t __s0 = __p0; \
40775  int32x4_t __s1 = __p1; \
40776  int32x4_t __s2 = __p2; \
40777  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40778  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40779  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40780  int32x4_t __ret; \
40781  __ret = __noswap_vqaddq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40782  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40783  __ret; \
40784})
40785#endif
40786
40787#ifdef __LITTLE_ENDIAN__
40788#define vqrdmlahq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40789  int16x8_t __s0 = __p0; \
40790  int16x8_t __s1 = __p1; \
40791  int16x8_t __s2 = __p2; \
40792  int16x8_t __ret; \
40793  __ret = vqaddq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40794  __ret; \
40795})
40796#else
40797#define vqrdmlahq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40798  int16x8_t __s0 = __p0; \
40799  int16x8_t __s1 = __p1; \
40800  int16x8_t __s2 = __p2; \
40801  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
40802  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
40803  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
40804  int16x8_t __ret; \
40805  __ret = __noswap_vqaddq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40806  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
40807  __ret; \
40808})
40809#endif
40810
40811#ifdef __LITTLE_ENDIAN__
40812#define vqrdmlah_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40813  int32x2_t __s0 = __p0; \
40814  int32x2_t __s1 = __p1; \
40815  int32x4_t __s2 = __p2; \
40816  int32x2_t __ret; \
40817  __ret = vqadd_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
40818  __ret; \
40819})
40820#else
40821#define vqrdmlah_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40822  int32x2_t __s0 = __p0; \
40823  int32x2_t __s1 = __p1; \
40824  int32x4_t __s2 = __p2; \
40825  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
40826  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
40827  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40828  int32x2_t __ret; \
40829  __ret = __noswap_vqadd_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
40830  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
40831  __ret; \
40832})
40833#endif
40834
40835#ifdef __LITTLE_ENDIAN__
40836#define vqrdmlah_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40837  int16x4_t __s0 = __p0; \
40838  int16x4_t __s1 = __p1; \
40839  int16x8_t __s2 = __p2; \
40840  int16x4_t __ret; \
40841  __ret = vqadd_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40842  __ret; \
40843})
40844#else
40845#define vqrdmlah_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40846  int16x4_t __s0 = __p0; \
40847  int16x4_t __s1 = __p1; \
40848  int16x8_t __s2 = __p2; \
40849  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40850  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40851  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
40852  int16x4_t __ret; \
40853  __ret = __noswap_vqadd_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40854  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40855  __ret; \
40856})
40857#endif
40858
40859#ifdef __LITTLE_ENDIAN__
40860#define vqrdmlshq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40861  int32x4_t __s0 = __p0; \
40862  int32x4_t __s1 = __p1; \
40863  int32x4_t __s2 = __p2; \
40864  int32x4_t __ret; \
40865  __ret = vqsubq_s32(__s0, vqrdmulhq_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40866  __ret; \
40867})
40868#else
40869#define vqrdmlshq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40870  int32x4_t __s0 = __p0; \
40871  int32x4_t __s1 = __p1; \
40872  int32x4_t __s2 = __p2; \
40873  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40874  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40875  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40876  int32x4_t __ret; \
40877  __ret = __noswap_vqsubq_s32(__rev0, __noswap_vqrdmulhq_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40878  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40879  __ret; \
40880})
40881#endif
40882
40883#ifdef __LITTLE_ENDIAN__
40884#define vqrdmlshq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40885  int16x8_t __s0 = __p0; \
40886  int16x8_t __s1 = __p1; \
40887  int16x8_t __s2 = __p2; \
40888  int16x8_t __ret; \
40889  __ret = vqsubq_s16(__s0, vqrdmulhq_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40890  __ret; \
40891})
40892#else
40893#define vqrdmlshq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40894  int16x8_t __s0 = __p0; \
40895  int16x8_t __s1 = __p1; \
40896  int16x8_t __s2 = __p2; \
40897  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
40898  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
40899  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
40900  int16x8_t __ret; \
40901  __ret = __noswap_vqsubq_s16(__rev0, __noswap_vqrdmulhq_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3))); \
40902  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
40903  __ret; \
40904})
40905#endif
40906
40907#ifdef __LITTLE_ENDIAN__
40908#define vqrdmlsh_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40909  int32x2_t __s0 = __p0; \
40910  int32x2_t __s1 = __p1; \
40911  int32x4_t __s2 = __p2; \
40912  int32x2_t __ret; \
40913  __ret = vqsub_s32(__s0, vqrdmulh_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3))); \
40914  __ret; \
40915})
40916#else
40917#define vqrdmlsh_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
40918  int32x2_t __s0 = __p0; \
40919  int32x2_t __s1 = __p1; \
40920  int32x4_t __s2 = __p2; \
40921  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
40922  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
40923  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
40924  int32x2_t __ret; \
40925  __ret = __noswap_vqsub_s32(__rev0, __noswap_vqrdmulh_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3))); \
40926  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
40927  __ret; \
40928})
40929#endif
40930
40931#ifdef __LITTLE_ENDIAN__
40932#define vqrdmlsh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40933  int16x4_t __s0 = __p0; \
40934  int16x4_t __s1 = __p1; \
40935  int16x8_t __s2 = __p2; \
40936  int16x4_t __ret; \
40937  __ret = vqsub_s16(__s0, vqrdmulh_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3))); \
40938  __ret; \
40939})
40940#else
40941#define vqrdmlsh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
40942  int16x4_t __s0 = __p0; \
40943  int16x4_t __s1 = __p1; \
40944  int16x8_t __s2 = __p2; \
40945  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
40946  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
40947  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
40948  int16x4_t __ret; \
40949  __ret = __noswap_vqsub_s16(__rev0, __noswap_vqrdmulh_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3))); \
40950  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
40951  __ret; \
40952})
40953#endif
40954
40955#endif
40956#if defined(__aarch64__)
40957#ifdef __LITTLE_ENDIAN__
40958__ai float64x2_t vabdq_f64(float64x2_t __p0, float64x2_t __p1) {
40959  float64x2_t __ret;
40960  __ret = (float64x2_t) __builtin_neon_vabdq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
40961  return __ret;
40962}
40963#else
40964__ai float64x2_t vabdq_f64(float64x2_t __p0, float64x2_t __p1) {
40965  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
40966  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
40967  float64x2_t __ret;
40968  __ret = (float64x2_t) __builtin_neon_vabdq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
40969  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
40970  return __ret;
40971}
40972#endif
40973
40974#ifdef __LITTLE_ENDIAN__
40975__ai float64x1_t vabd_f64(float64x1_t __p0, float64x1_t __p1) {
40976  float64x1_t __ret;
40977  __ret = (float64x1_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40978  return __ret;
40979}
40980#else
40981__ai float64x1_t vabd_f64(float64x1_t __p0, float64x1_t __p1) {
40982  float64x1_t __ret;
40983  __ret = (float64x1_t) __builtin_neon_vabd_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
40984  return __ret;
40985}
40986#endif
40987
40988#ifdef __LITTLE_ENDIAN__
40989__ai float64_t vabdd_f64(float64_t __p0, float64_t __p1) {
40990  float64_t __ret;
40991  __ret = (float64_t) __builtin_neon_vabdd_f64(__p0, __p1);
40992  return __ret;
40993}
40994#else
40995__ai float64_t vabdd_f64(float64_t __p0, float64_t __p1) {
40996  float64_t __ret;
40997  __ret = (float64_t) __builtin_neon_vabdd_f64(__p0, __p1);
40998  return __ret;
40999}
41000#endif
41001
41002#ifdef __LITTLE_ENDIAN__
41003__ai float32_t vabds_f32(float32_t __p0, float32_t __p1) {
41004  float32_t __ret;
41005  __ret = (float32_t) __builtin_neon_vabds_f32(__p0, __p1);
41006  return __ret;
41007}
41008#else
41009__ai float32_t vabds_f32(float32_t __p0, float32_t __p1) {
41010  float32_t __ret;
41011  __ret = (float32_t) __builtin_neon_vabds_f32(__p0, __p1);
41012  return __ret;
41013}
41014#endif
41015
41016#ifdef __LITTLE_ENDIAN__
41017__ai float64x2_t vabsq_f64(float64x2_t __p0) {
41018  float64x2_t __ret;
41019  __ret = (float64x2_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 42);
41020  return __ret;
41021}
41022#else
41023__ai float64x2_t vabsq_f64(float64x2_t __p0) {
41024  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41025  float64x2_t __ret;
41026  __ret = (float64x2_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 42);
41027  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41028  return __ret;
41029}
41030#endif
41031
41032#ifdef __LITTLE_ENDIAN__
41033__ai int64x2_t vabsq_s64(int64x2_t __p0) {
41034  int64x2_t __ret;
41035  __ret = (int64x2_t) __builtin_neon_vabsq_v((int8x16_t)__p0, 35);
41036  return __ret;
41037}
41038#else
41039__ai int64x2_t vabsq_s64(int64x2_t __p0) {
41040  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41041  int64x2_t __ret;
41042  __ret = (int64x2_t) __builtin_neon_vabsq_v((int8x16_t)__rev0, 35);
41043  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41044  return __ret;
41045}
41046#endif
41047
41048#ifdef __LITTLE_ENDIAN__
41049__ai float64x1_t vabs_f64(float64x1_t __p0) {
41050  float64x1_t __ret;
41051  __ret = (float64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 10);
41052  return __ret;
41053}
41054#else
41055__ai float64x1_t vabs_f64(float64x1_t __p0) {
41056  float64x1_t __ret;
41057  __ret = (float64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 10);
41058  return __ret;
41059}
41060#endif
41061
41062#ifdef __LITTLE_ENDIAN__
41063__ai int64x1_t vabs_s64(int64x1_t __p0) {
41064  int64x1_t __ret;
41065  __ret = (int64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 3);
41066  return __ret;
41067}
41068#else
41069__ai int64x1_t vabs_s64(int64x1_t __p0) {
41070  int64x1_t __ret;
41071  __ret = (int64x1_t) __builtin_neon_vabs_v((int8x8_t)__p0, 3);
41072  return __ret;
41073}
41074#endif
41075
41076#ifdef __LITTLE_ENDIAN__
41077__ai int64_t vabsd_s64(int64_t __p0) {
41078  int64_t __ret;
41079  __ret = (int64_t) __builtin_neon_vabsd_s64(__p0);
41080  return __ret;
41081}
41082#else
41083__ai int64_t vabsd_s64(int64_t __p0) {
41084  int64_t __ret;
41085  __ret = (int64_t) __builtin_neon_vabsd_s64(__p0);
41086  return __ret;
41087}
41088#endif
41089
41090#ifdef __LITTLE_ENDIAN__
41091__ai float64x2_t vaddq_f64(float64x2_t __p0, float64x2_t __p1) {
41092  float64x2_t __ret;
41093  __ret = __p0 + __p1;
41094  return __ret;
41095}
41096#else
41097__ai float64x2_t vaddq_f64(float64x2_t __p0, float64x2_t __p1) {
41098  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41099  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41100  float64x2_t __ret;
41101  __ret = __rev0 + __rev1;
41102  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41103  return __ret;
41104}
41105#endif
41106
41107#ifdef __LITTLE_ENDIAN__
41108__ai float64x1_t vadd_f64(float64x1_t __p0, float64x1_t __p1) {
41109  float64x1_t __ret;
41110  __ret = __p0 + __p1;
41111  return __ret;
41112}
41113#else
41114__ai float64x1_t vadd_f64(float64x1_t __p0, float64x1_t __p1) {
41115  float64x1_t __ret;
41116  __ret = __p0 + __p1;
41117  return __ret;
41118}
41119#endif
41120
41121#ifdef __LITTLE_ENDIAN__
41122__ai uint64_t vaddd_u64(uint64_t __p0, uint64_t __p1) {
41123  uint64_t __ret;
41124  __ret = (uint64_t) __builtin_neon_vaddd_u64(__p0, __p1);
41125  return __ret;
41126}
41127#else
41128__ai uint64_t vaddd_u64(uint64_t __p0, uint64_t __p1) {
41129  uint64_t __ret;
41130  __ret = (uint64_t) __builtin_neon_vaddd_u64(__p0, __p1);
41131  return __ret;
41132}
41133#endif
41134
41135#ifdef __LITTLE_ENDIAN__
41136__ai int64_t vaddd_s64(int64_t __p0, int64_t __p1) {
41137  int64_t __ret;
41138  __ret = (int64_t) __builtin_neon_vaddd_s64(__p0, __p1);
41139  return __ret;
41140}
41141#else
41142__ai int64_t vaddd_s64(int64_t __p0, int64_t __p1) {
41143  int64_t __ret;
41144  __ret = (int64_t) __builtin_neon_vaddd_s64(__p0, __p1);
41145  return __ret;
41146}
41147#endif
41148
41149#ifdef __LITTLE_ENDIAN__
41150__ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
41151  uint16x8_t __ret;
41152  __ret = vcombine_u16(__p0, vaddhn_u32(__p1, __p2));
41153  return __ret;
41154}
41155#else
41156__ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
41157  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41158  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41159  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41160  uint16x8_t __ret;
41161  __ret = __noswap_vcombine_u16(__rev0, __noswap_vaddhn_u32(__rev1, __rev2));
41162  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41163  return __ret;
41164}
41165#endif
41166
41167#ifdef __LITTLE_ENDIAN__
41168__ai uint32x4_t vaddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
41169  uint32x4_t __ret;
41170  __ret = vcombine_u32(__p0, vaddhn_u64(__p1, __p2));
41171  return __ret;
41172}
41173#else
41174__ai uint32x4_t vaddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
41175  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41176  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41177  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41178  uint32x4_t __ret;
41179  __ret = __noswap_vcombine_u32(__rev0, __noswap_vaddhn_u64(__rev1, __rev2));
41180  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41181  return __ret;
41182}
41183#endif
41184
41185#ifdef __LITTLE_ENDIAN__
41186__ai uint8x16_t vaddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
41187  uint8x16_t __ret;
41188  __ret = vcombine_u8(__p0, vaddhn_u16(__p1, __p2));
41189  return __ret;
41190}
41191#else
41192__ai uint8x16_t vaddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
41193  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41194  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41195  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
41196  uint8x16_t __ret;
41197  __ret = __noswap_vcombine_u8(__rev0, __noswap_vaddhn_u16(__rev1, __rev2));
41198  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41199  return __ret;
41200}
41201#endif
41202
41203#ifdef __LITTLE_ENDIAN__
41204__ai int16x8_t vaddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
41205  int16x8_t __ret;
41206  __ret = vcombine_s16(__p0, vaddhn_s32(__p1, __p2));
41207  return __ret;
41208}
41209#else
41210__ai int16x8_t vaddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
41211  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41212  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
41213  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
41214  int16x8_t __ret;
41215  __ret = __noswap_vcombine_s16(__rev0, __noswap_vaddhn_s32(__rev1, __rev2));
41216  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
41217  return __ret;
41218}
41219#endif
41220
41221#ifdef __LITTLE_ENDIAN__
41222__ai int32x4_t vaddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
41223  int32x4_t __ret;
41224  __ret = vcombine_s32(__p0, vaddhn_s64(__p1, __p2));
41225  return __ret;
41226}
41227#else
41228__ai int32x4_t vaddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
41229  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41230  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41231  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41232  int32x4_t __ret;
41233  __ret = __noswap_vcombine_s32(__rev0, __noswap_vaddhn_s64(__rev1, __rev2));
41234  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
41235  return __ret;
41236}
41237#endif
41238
41239#ifdef __LITTLE_ENDIAN__
41240__ai int8x16_t vaddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
41241  int8x16_t __ret;
41242  __ret = vcombine_s8(__p0, vaddhn_s16(__p1, __p2));
41243  return __ret;
41244}
41245#else
41246__ai int8x16_t vaddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
41247  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41248  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
41249  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
41250  int8x16_t __ret;
41251  __ret = __noswap_vcombine_s8(__rev0, __noswap_vaddhn_s16(__rev1, __rev2));
41252  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41253  return __ret;
41254}
41255#endif
41256
41257#ifdef __LITTLE_ENDIAN__
41258__ai uint16_t vaddlvq_u8(uint8x16_t __p0) {
41259  uint16_t __ret;
41260  __ret = (uint16_t) __builtin_neon_vaddlvq_u8((int8x16_t)__p0);
41261  return __ret;
41262}
41263#else
41264__ai uint16_t vaddlvq_u8(uint8x16_t __p0) {
41265  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41266  uint16_t __ret;
41267  __ret = (uint16_t) __builtin_neon_vaddlvq_u8((int8x16_t)__rev0);
41268  return __ret;
41269}
41270#endif
41271
41272#ifdef __LITTLE_ENDIAN__
41273__ai uint64_t vaddlvq_u32(uint32x4_t __p0) {
41274  uint64_t __ret;
41275  __ret = (uint64_t) __builtin_neon_vaddlvq_u32((int8x16_t)__p0);
41276  return __ret;
41277}
41278#else
41279__ai uint64_t vaddlvq_u32(uint32x4_t __p0) {
41280  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41281  uint64_t __ret;
41282  __ret = (uint64_t) __builtin_neon_vaddlvq_u32((int8x16_t)__rev0);
41283  return __ret;
41284}
41285#endif
41286
41287#ifdef __LITTLE_ENDIAN__
41288__ai uint32_t vaddlvq_u16(uint16x8_t __p0) {
41289  uint32_t __ret;
41290  __ret = (uint32_t) __builtin_neon_vaddlvq_u16((int8x16_t)__p0);
41291  return __ret;
41292}
41293#else
41294__ai uint32_t vaddlvq_u16(uint16x8_t __p0) {
41295  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41296  uint32_t __ret;
41297  __ret = (uint32_t) __builtin_neon_vaddlvq_u16((int8x16_t)__rev0);
41298  return __ret;
41299}
41300#endif
41301
41302#ifdef __LITTLE_ENDIAN__
41303__ai int16_t vaddlvq_s8(int8x16_t __p0) {
41304  int16_t __ret;
41305  __ret = (int16_t) __builtin_neon_vaddlvq_s8((int8x16_t)__p0);
41306  return __ret;
41307}
41308#else
41309__ai int16_t vaddlvq_s8(int8x16_t __p0) {
41310  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41311  int16_t __ret;
41312  __ret = (int16_t) __builtin_neon_vaddlvq_s8((int8x16_t)__rev0);
41313  return __ret;
41314}
41315#endif
41316
41317#ifdef __LITTLE_ENDIAN__
41318__ai int64_t vaddlvq_s32(int32x4_t __p0) {
41319  int64_t __ret;
41320  __ret = (int64_t) __builtin_neon_vaddlvq_s32((int8x16_t)__p0);
41321  return __ret;
41322}
41323#else
41324__ai int64_t vaddlvq_s32(int32x4_t __p0) {
41325  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41326  int64_t __ret;
41327  __ret = (int64_t) __builtin_neon_vaddlvq_s32((int8x16_t)__rev0);
41328  return __ret;
41329}
41330#endif
41331
41332#ifdef __LITTLE_ENDIAN__
41333__ai int32_t vaddlvq_s16(int16x8_t __p0) {
41334  int32_t __ret;
41335  __ret = (int32_t) __builtin_neon_vaddlvq_s16((int8x16_t)__p0);
41336  return __ret;
41337}
41338#else
41339__ai int32_t vaddlvq_s16(int16x8_t __p0) {
41340  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41341  int32_t __ret;
41342  __ret = (int32_t) __builtin_neon_vaddlvq_s16((int8x16_t)__rev0);
41343  return __ret;
41344}
41345#endif
41346
41347#ifdef __LITTLE_ENDIAN__
41348__ai uint16_t vaddlv_u8(uint8x8_t __p0) {
41349  uint16_t __ret;
41350  __ret = (uint16_t) __builtin_neon_vaddlv_u8((int8x8_t)__p0);
41351  return __ret;
41352}
41353#else
41354__ai uint16_t vaddlv_u8(uint8x8_t __p0) {
41355  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41356  uint16_t __ret;
41357  __ret = (uint16_t) __builtin_neon_vaddlv_u8((int8x8_t)__rev0);
41358  return __ret;
41359}
41360#endif
41361
41362#ifdef __LITTLE_ENDIAN__
41363__ai uint64_t vaddlv_u32(uint32x2_t __p0) {
41364  uint64_t __ret;
41365  __ret = (uint64_t) __builtin_neon_vaddlv_u32((int8x8_t)__p0);
41366  return __ret;
41367}
41368#else
41369__ai uint64_t vaddlv_u32(uint32x2_t __p0) {
41370  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41371  uint64_t __ret;
41372  __ret = (uint64_t) __builtin_neon_vaddlv_u32((int8x8_t)__rev0);
41373  return __ret;
41374}
41375#endif
41376
41377#ifdef __LITTLE_ENDIAN__
41378__ai uint32_t vaddlv_u16(uint16x4_t __p0) {
41379  uint32_t __ret;
41380  __ret = (uint32_t) __builtin_neon_vaddlv_u16((int8x8_t)__p0);
41381  return __ret;
41382}
41383#else
41384__ai uint32_t vaddlv_u16(uint16x4_t __p0) {
41385  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41386  uint32_t __ret;
41387  __ret = (uint32_t) __builtin_neon_vaddlv_u16((int8x8_t)__rev0);
41388  return __ret;
41389}
41390#endif
41391
41392#ifdef __LITTLE_ENDIAN__
41393__ai int16_t vaddlv_s8(int8x8_t __p0) {
41394  int16_t __ret;
41395  __ret = (int16_t) __builtin_neon_vaddlv_s8((int8x8_t)__p0);
41396  return __ret;
41397}
41398#else
41399__ai int16_t vaddlv_s8(int8x8_t __p0) {
41400  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41401  int16_t __ret;
41402  __ret = (int16_t) __builtin_neon_vaddlv_s8((int8x8_t)__rev0);
41403  return __ret;
41404}
41405#endif
41406
41407#ifdef __LITTLE_ENDIAN__
41408__ai int64_t vaddlv_s32(int32x2_t __p0) {
41409  int64_t __ret;
41410  __ret = (int64_t) __builtin_neon_vaddlv_s32((int8x8_t)__p0);
41411  return __ret;
41412}
41413#else
41414__ai int64_t vaddlv_s32(int32x2_t __p0) {
41415  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41416  int64_t __ret;
41417  __ret = (int64_t) __builtin_neon_vaddlv_s32((int8x8_t)__rev0);
41418  return __ret;
41419}
41420#endif
41421
41422#ifdef __LITTLE_ENDIAN__
41423__ai int32_t vaddlv_s16(int16x4_t __p0) {
41424  int32_t __ret;
41425  __ret = (int32_t) __builtin_neon_vaddlv_s16((int8x8_t)__p0);
41426  return __ret;
41427}
41428#else
41429__ai int32_t vaddlv_s16(int16x4_t __p0) {
41430  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41431  int32_t __ret;
41432  __ret = (int32_t) __builtin_neon_vaddlv_s16((int8x8_t)__rev0);
41433  return __ret;
41434}
41435#endif
41436
41437#ifdef __LITTLE_ENDIAN__
41438__ai uint8_t vaddvq_u8(uint8x16_t __p0) {
41439  uint8_t __ret;
41440  __ret = (uint8_t) __builtin_neon_vaddvq_u8((int8x16_t)__p0);
41441  return __ret;
41442}
41443#else
41444__ai uint8_t vaddvq_u8(uint8x16_t __p0) {
41445  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41446  uint8_t __ret;
41447  __ret = (uint8_t) __builtin_neon_vaddvq_u8((int8x16_t)__rev0);
41448  return __ret;
41449}
41450#endif
41451
41452#ifdef __LITTLE_ENDIAN__
41453__ai uint32_t vaddvq_u32(uint32x4_t __p0) {
41454  uint32_t __ret;
41455  __ret = (uint32_t) __builtin_neon_vaddvq_u32((int8x16_t)__p0);
41456  return __ret;
41457}
41458#else
41459__ai uint32_t vaddvq_u32(uint32x4_t __p0) {
41460  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41461  uint32_t __ret;
41462  __ret = (uint32_t) __builtin_neon_vaddvq_u32((int8x16_t)__rev0);
41463  return __ret;
41464}
41465#endif
41466
41467#ifdef __LITTLE_ENDIAN__
41468__ai uint64_t vaddvq_u64(uint64x2_t __p0) {
41469  uint64_t __ret;
41470  __ret = (uint64_t) __builtin_neon_vaddvq_u64((int8x16_t)__p0);
41471  return __ret;
41472}
41473#else
41474__ai uint64_t vaddvq_u64(uint64x2_t __p0) {
41475  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41476  uint64_t __ret;
41477  __ret = (uint64_t) __builtin_neon_vaddvq_u64((int8x16_t)__rev0);
41478  return __ret;
41479}
41480#endif
41481
41482#ifdef __LITTLE_ENDIAN__
41483__ai uint16_t vaddvq_u16(uint16x8_t __p0) {
41484  uint16_t __ret;
41485  __ret = (uint16_t) __builtin_neon_vaddvq_u16((int8x16_t)__p0);
41486  return __ret;
41487}
41488#else
41489__ai uint16_t vaddvq_u16(uint16x8_t __p0) {
41490  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41491  uint16_t __ret;
41492  __ret = (uint16_t) __builtin_neon_vaddvq_u16((int8x16_t)__rev0);
41493  return __ret;
41494}
41495#endif
41496
41497#ifdef __LITTLE_ENDIAN__
41498__ai int8_t vaddvq_s8(int8x16_t __p0) {
41499  int8_t __ret;
41500  __ret = (int8_t) __builtin_neon_vaddvq_s8((int8x16_t)__p0);
41501  return __ret;
41502}
41503#else
41504__ai int8_t vaddvq_s8(int8x16_t __p0) {
41505  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
41506  int8_t __ret;
41507  __ret = (int8_t) __builtin_neon_vaddvq_s8((int8x16_t)__rev0);
41508  return __ret;
41509}
41510#endif
41511
41512#ifdef __LITTLE_ENDIAN__
41513__ai float64_t vaddvq_f64(float64x2_t __p0) {
41514  float64_t __ret;
41515  __ret = (float64_t) __builtin_neon_vaddvq_f64((int8x16_t)__p0);
41516  return __ret;
41517}
41518#else
41519__ai float64_t vaddvq_f64(float64x2_t __p0) {
41520  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41521  float64_t __ret;
41522  __ret = (float64_t) __builtin_neon_vaddvq_f64((int8x16_t)__rev0);
41523  return __ret;
41524}
41525#endif
41526
41527#ifdef __LITTLE_ENDIAN__
41528__ai float32_t vaddvq_f32(float32x4_t __p0) {
41529  float32_t __ret;
41530  __ret = (float32_t) __builtin_neon_vaddvq_f32((int8x16_t)__p0);
41531  return __ret;
41532}
41533#else
41534__ai float32_t vaddvq_f32(float32x4_t __p0) {
41535  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41536  float32_t __ret;
41537  __ret = (float32_t) __builtin_neon_vaddvq_f32((int8x16_t)__rev0);
41538  return __ret;
41539}
41540#endif
41541
41542#ifdef __LITTLE_ENDIAN__
41543__ai int32_t vaddvq_s32(int32x4_t __p0) {
41544  int32_t __ret;
41545  __ret = (int32_t) __builtin_neon_vaddvq_s32((int8x16_t)__p0);
41546  return __ret;
41547}
41548#else
41549__ai int32_t vaddvq_s32(int32x4_t __p0) {
41550  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41551  int32_t __ret;
41552  __ret = (int32_t) __builtin_neon_vaddvq_s32((int8x16_t)__rev0);
41553  return __ret;
41554}
41555#endif
41556
41557#ifdef __LITTLE_ENDIAN__
41558__ai int64_t vaddvq_s64(int64x2_t __p0) {
41559  int64_t __ret;
41560  __ret = (int64_t) __builtin_neon_vaddvq_s64((int8x16_t)__p0);
41561  return __ret;
41562}
41563#else
41564__ai int64_t vaddvq_s64(int64x2_t __p0) {
41565  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41566  int64_t __ret;
41567  __ret = (int64_t) __builtin_neon_vaddvq_s64((int8x16_t)__rev0);
41568  return __ret;
41569}
41570#endif
41571
41572#ifdef __LITTLE_ENDIAN__
41573__ai int16_t vaddvq_s16(int16x8_t __p0) {
41574  int16_t __ret;
41575  __ret = (int16_t) __builtin_neon_vaddvq_s16((int8x16_t)__p0);
41576  return __ret;
41577}
41578#else
41579__ai int16_t vaddvq_s16(int16x8_t __p0) {
41580  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41581  int16_t __ret;
41582  __ret = (int16_t) __builtin_neon_vaddvq_s16((int8x16_t)__rev0);
41583  return __ret;
41584}
41585#endif
41586
41587#ifdef __LITTLE_ENDIAN__
41588__ai uint8_t vaddv_u8(uint8x8_t __p0) {
41589  uint8_t __ret;
41590  __ret = (uint8_t) __builtin_neon_vaddv_u8((int8x8_t)__p0);
41591  return __ret;
41592}
41593#else
41594__ai uint8_t vaddv_u8(uint8x8_t __p0) {
41595  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41596  uint8_t __ret;
41597  __ret = (uint8_t) __builtin_neon_vaddv_u8((int8x8_t)__rev0);
41598  return __ret;
41599}
41600#endif
41601
41602#ifdef __LITTLE_ENDIAN__
41603__ai uint32_t vaddv_u32(uint32x2_t __p0) {
41604  uint32_t __ret;
41605  __ret = (uint32_t) __builtin_neon_vaddv_u32((int8x8_t)__p0);
41606  return __ret;
41607}
41608#else
41609__ai uint32_t vaddv_u32(uint32x2_t __p0) {
41610  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41611  uint32_t __ret;
41612  __ret = (uint32_t) __builtin_neon_vaddv_u32((int8x8_t)__rev0);
41613  return __ret;
41614}
41615#endif
41616
41617#ifdef __LITTLE_ENDIAN__
41618__ai uint16_t vaddv_u16(uint16x4_t __p0) {
41619  uint16_t __ret;
41620  __ret = (uint16_t) __builtin_neon_vaddv_u16((int8x8_t)__p0);
41621  return __ret;
41622}
41623#else
41624__ai uint16_t vaddv_u16(uint16x4_t __p0) {
41625  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41626  uint16_t __ret;
41627  __ret = (uint16_t) __builtin_neon_vaddv_u16((int8x8_t)__rev0);
41628  return __ret;
41629}
41630#endif
41631
41632#ifdef __LITTLE_ENDIAN__
41633__ai int8_t vaddv_s8(int8x8_t __p0) {
41634  int8_t __ret;
41635  __ret = (int8_t) __builtin_neon_vaddv_s8((int8x8_t)__p0);
41636  return __ret;
41637}
41638#else
41639__ai int8_t vaddv_s8(int8x8_t __p0) {
41640  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
41641  int8_t __ret;
41642  __ret = (int8_t) __builtin_neon_vaddv_s8((int8x8_t)__rev0);
41643  return __ret;
41644}
41645#endif
41646
41647#ifdef __LITTLE_ENDIAN__
41648__ai float32_t vaddv_f32(float32x2_t __p0) {
41649  float32_t __ret;
41650  __ret = (float32_t) __builtin_neon_vaddv_f32((int8x8_t)__p0);
41651  return __ret;
41652}
41653#else
41654__ai float32_t vaddv_f32(float32x2_t __p0) {
41655  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41656  float32_t __ret;
41657  __ret = (float32_t) __builtin_neon_vaddv_f32((int8x8_t)__rev0);
41658  return __ret;
41659}
41660#endif
41661
41662#ifdef __LITTLE_ENDIAN__
41663__ai int32_t vaddv_s32(int32x2_t __p0) {
41664  int32_t __ret;
41665  __ret = (int32_t) __builtin_neon_vaddv_s32((int8x8_t)__p0);
41666  return __ret;
41667}
41668#else
41669__ai int32_t vaddv_s32(int32x2_t __p0) {
41670  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41671  int32_t __ret;
41672  __ret = (int32_t) __builtin_neon_vaddv_s32((int8x8_t)__rev0);
41673  return __ret;
41674}
41675#endif
41676
41677#ifdef __LITTLE_ENDIAN__
41678__ai int16_t vaddv_s16(int16x4_t __p0) {
41679  int16_t __ret;
41680  __ret = (int16_t) __builtin_neon_vaddv_s16((int8x8_t)__p0);
41681  return __ret;
41682}
41683#else
41684__ai int16_t vaddv_s16(int16x4_t __p0) {
41685  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
41686  int16_t __ret;
41687  __ret = (int16_t) __builtin_neon_vaddv_s16((int8x8_t)__rev0);
41688  return __ret;
41689}
41690#endif
41691
41692#ifdef __LITTLE_ENDIAN__
41693__ai poly64x1_t vbsl_p64(uint64x1_t __p0, poly64x1_t __p1, poly64x1_t __p2) {
41694  poly64x1_t __ret;
41695  __ret = (poly64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 6);
41696  return __ret;
41697}
41698#else
41699__ai poly64x1_t vbsl_p64(uint64x1_t __p0, poly64x1_t __p1, poly64x1_t __p2) {
41700  poly64x1_t __ret;
41701  __ret = (poly64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 6);
41702  return __ret;
41703}
41704#endif
41705
41706#ifdef __LITTLE_ENDIAN__
41707__ai poly64x2_t vbslq_p64(uint64x2_t __p0, poly64x2_t __p1, poly64x2_t __p2) {
41708  poly64x2_t __ret;
41709  __ret = (poly64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 38);
41710  return __ret;
41711}
41712#else
41713__ai poly64x2_t vbslq_p64(uint64x2_t __p0, poly64x2_t __p1, poly64x2_t __p2) {
41714  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41715  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41716  poly64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41717  poly64x2_t __ret;
41718  __ret = (poly64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 38);
41719  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41720  return __ret;
41721}
41722#endif
41723
41724#ifdef __LITTLE_ENDIAN__
41725__ai float64x2_t vbslq_f64(uint64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
41726  float64x2_t __ret;
41727  __ret = (float64x2_t) __builtin_neon_vbslq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
41728  return __ret;
41729}
41730#else
41731__ai float64x2_t vbslq_f64(uint64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
41732  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41733  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41734  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
41735  float64x2_t __ret;
41736  __ret = (float64x2_t) __builtin_neon_vbslq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
41737  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41738  return __ret;
41739}
41740#endif
41741
41742#ifdef __LITTLE_ENDIAN__
41743__ai float64x1_t vbsl_f64(uint64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
41744  float64x1_t __ret;
41745  __ret = (float64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
41746  return __ret;
41747}
41748#else
41749__ai float64x1_t vbsl_f64(uint64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
41750  float64x1_t __ret;
41751  __ret = (float64x1_t) __builtin_neon_vbsl_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
41752  return __ret;
41753}
41754#endif
41755
41756#ifdef __LITTLE_ENDIAN__
41757__ai uint64x2_t vcageq_f64(float64x2_t __p0, float64x2_t __p1) {
41758  uint64x2_t __ret;
41759  __ret = (uint64x2_t) __builtin_neon_vcageq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
41760  return __ret;
41761}
41762#else
41763__ai uint64x2_t vcageq_f64(float64x2_t __p0, float64x2_t __p1) {
41764  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41765  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41766  uint64x2_t __ret;
41767  __ret = (uint64x2_t) __builtin_neon_vcageq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
41768  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41769  return __ret;
41770}
41771#endif
41772
41773#ifdef __LITTLE_ENDIAN__
41774__ai uint64x1_t vcage_f64(float64x1_t __p0, float64x1_t __p1) {
41775  uint64x1_t __ret;
41776  __ret = (uint64x1_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41777  return __ret;
41778}
41779#else
41780__ai uint64x1_t vcage_f64(float64x1_t __p0, float64x1_t __p1) {
41781  uint64x1_t __ret;
41782  __ret = (uint64x1_t) __builtin_neon_vcage_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41783  return __ret;
41784}
41785#endif
41786
41787#ifdef __LITTLE_ENDIAN__
41788__ai uint64_t vcaged_f64(float64_t __p0, float64_t __p1) {
41789  uint64_t __ret;
41790  __ret = (uint64_t) __builtin_neon_vcaged_f64(__p0, __p1);
41791  return __ret;
41792}
41793#else
41794__ai uint64_t vcaged_f64(float64_t __p0, float64_t __p1) {
41795  uint64_t __ret;
41796  __ret = (uint64_t) __builtin_neon_vcaged_f64(__p0, __p1);
41797  return __ret;
41798}
41799#endif
41800
41801#ifdef __LITTLE_ENDIAN__
41802__ai uint32_t vcages_f32(float32_t __p0, float32_t __p1) {
41803  uint32_t __ret;
41804  __ret = (uint32_t) __builtin_neon_vcages_f32(__p0, __p1);
41805  return __ret;
41806}
41807#else
41808__ai uint32_t vcages_f32(float32_t __p0, float32_t __p1) {
41809  uint32_t __ret;
41810  __ret = (uint32_t) __builtin_neon_vcages_f32(__p0, __p1);
41811  return __ret;
41812}
41813#endif
41814
41815#ifdef __LITTLE_ENDIAN__
41816__ai uint64x2_t vcagtq_f64(float64x2_t __p0, float64x2_t __p1) {
41817  uint64x2_t __ret;
41818  __ret = (uint64x2_t) __builtin_neon_vcagtq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
41819  return __ret;
41820}
41821#else
41822__ai uint64x2_t vcagtq_f64(float64x2_t __p0, float64x2_t __p1) {
41823  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41824  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41825  uint64x2_t __ret;
41826  __ret = (uint64x2_t) __builtin_neon_vcagtq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
41827  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41828  return __ret;
41829}
41830#endif
41831
41832#ifdef __LITTLE_ENDIAN__
41833__ai uint64x1_t vcagt_f64(float64x1_t __p0, float64x1_t __p1) {
41834  uint64x1_t __ret;
41835  __ret = (uint64x1_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41836  return __ret;
41837}
41838#else
41839__ai uint64x1_t vcagt_f64(float64x1_t __p0, float64x1_t __p1) {
41840  uint64x1_t __ret;
41841  __ret = (uint64x1_t) __builtin_neon_vcagt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41842  return __ret;
41843}
41844#endif
41845
41846#ifdef __LITTLE_ENDIAN__
41847__ai uint64_t vcagtd_f64(float64_t __p0, float64_t __p1) {
41848  uint64_t __ret;
41849  __ret = (uint64_t) __builtin_neon_vcagtd_f64(__p0, __p1);
41850  return __ret;
41851}
41852#else
41853__ai uint64_t vcagtd_f64(float64_t __p0, float64_t __p1) {
41854  uint64_t __ret;
41855  __ret = (uint64_t) __builtin_neon_vcagtd_f64(__p0, __p1);
41856  return __ret;
41857}
41858#endif
41859
41860#ifdef __LITTLE_ENDIAN__
41861__ai uint32_t vcagts_f32(float32_t __p0, float32_t __p1) {
41862  uint32_t __ret;
41863  __ret = (uint32_t) __builtin_neon_vcagts_f32(__p0, __p1);
41864  return __ret;
41865}
41866#else
41867__ai uint32_t vcagts_f32(float32_t __p0, float32_t __p1) {
41868  uint32_t __ret;
41869  __ret = (uint32_t) __builtin_neon_vcagts_f32(__p0, __p1);
41870  return __ret;
41871}
41872#endif
41873
41874#ifdef __LITTLE_ENDIAN__
41875__ai uint64x2_t vcaleq_f64(float64x2_t __p0, float64x2_t __p1) {
41876  uint64x2_t __ret;
41877  __ret = (uint64x2_t) __builtin_neon_vcaleq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
41878  return __ret;
41879}
41880#else
41881__ai uint64x2_t vcaleq_f64(float64x2_t __p0, float64x2_t __p1) {
41882  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41883  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41884  uint64x2_t __ret;
41885  __ret = (uint64x2_t) __builtin_neon_vcaleq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
41886  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41887  return __ret;
41888}
41889#endif
41890
41891#ifdef __LITTLE_ENDIAN__
41892__ai uint64x1_t vcale_f64(float64x1_t __p0, float64x1_t __p1) {
41893  uint64x1_t __ret;
41894  __ret = (uint64x1_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41895  return __ret;
41896}
41897#else
41898__ai uint64x1_t vcale_f64(float64x1_t __p0, float64x1_t __p1) {
41899  uint64x1_t __ret;
41900  __ret = (uint64x1_t) __builtin_neon_vcale_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41901  return __ret;
41902}
41903#endif
41904
41905#ifdef __LITTLE_ENDIAN__
41906__ai uint64_t vcaled_f64(float64_t __p0, float64_t __p1) {
41907  uint64_t __ret;
41908  __ret = (uint64_t) __builtin_neon_vcaled_f64(__p0, __p1);
41909  return __ret;
41910}
41911#else
41912__ai uint64_t vcaled_f64(float64_t __p0, float64_t __p1) {
41913  uint64_t __ret;
41914  __ret = (uint64_t) __builtin_neon_vcaled_f64(__p0, __p1);
41915  return __ret;
41916}
41917#endif
41918
41919#ifdef __LITTLE_ENDIAN__
41920__ai uint32_t vcales_f32(float32_t __p0, float32_t __p1) {
41921  uint32_t __ret;
41922  __ret = (uint32_t) __builtin_neon_vcales_f32(__p0, __p1);
41923  return __ret;
41924}
41925#else
41926__ai uint32_t vcales_f32(float32_t __p0, float32_t __p1) {
41927  uint32_t __ret;
41928  __ret = (uint32_t) __builtin_neon_vcales_f32(__p0, __p1);
41929  return __ret;
41930}
41931#endif
41932
41933#ifdef __LITTLE_ENDIAN__
41934__ai uint64x2_t vcaltq_f64(float64x2_t __p0, float64x2_t __p1) {
41935  uint64x2_t __ret;
41936  __ret = (uint64x2_t) __builtin_neon_vcaltq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
41937  return __ret;
41938}
41939#else
41940__ai uint64x2_t vcaltq_f64(float64x2_t __p0, float64x2_t __p1) {
41941  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
41942  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
41943  uint64x2_t __ret;
41944  __ret = (uint64x2_t) __builtin_neon_vcaltq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
41945  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
41946  return __ret;
41947}
41948#endif
41949
41950#ifdef __LITTLE_ENDIAN__
41951__ai uint64x1_t vcalt_f64(float64x1_t __p0, float64x1_t __p1) {
41952  uint64x1_t __ret;
41953  __ret = (uint64x1_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41954  return __ret;
41955}
41956#else
41957__ai uint64x1_t vcalt_f64(float64x1_t __p0, float64x1_t __p1) {
41958  uint64x1_t __ret;
41959  __ret = (uint64x1_t) __builtin_neon_vcalt_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
41960  return __ret;
41961}
41962#endif
41963
41964#ifdef __LITTLE_ENDIAN__
41965__ai uint64_t vcaltd_f64(float64_t __p0, float64_t __p1) {
41966  uint64_t __ret;
41967  __ret = (uint64_t) __builtin_neon_vcaltd_f64(__p0, __p1);
41968  return __ret;
41969}
41970#else
41971__ai uint64_t vcaltd_f64(float64_t __p0, float64_t __p1) {
41972  uint64_t __ret;
41973  __ret = (uint64_t) __builtin_neon_vcaltd_f64(__p0, __p1);
41974  return __ret;
41975}
41976#endif
41977
41978#ifdef __LITTLE_ENDIAN__
41979__ai uint32_t vcalts_f32(float32_t __p0, float32_t __p1) {
41980  uint32_t __ret;
41981  __ret = (uint32_t) __builtin_neon_vcalts_f32(__p0, __p1);
41982  return __ret;
41983}
41984#else
41985__ai uint32_t vcalts_f32(float32_t __p0, float32_t __p1) {
41986  uint32_t __ret;
41987  __ret = (uint32_t) __builtin_neon_vcalts_f32(__p0, __p1);
41988  return __ret;
41989}
41990#endif
41991
41992#ifdef __LITTLE_ENDIAN__
41993__ai uint64x1_t vceq_p64(poly64x1_t __p0, poly64x1_t __p1) {
41994  uint64x1_t __ret;
41995  __ret = (uint64x1_t)(__p0 == __p1);
41996  return __ret;
41997}
41998#else
41999__ai uint64x1_t vceq_p64(poly64x1_t __p0, poly64x1_t __p1) {
42000  uint64x1_t __ret;
42001  __ret = (uint64x1_t)(__p0 == __p1);
42002  return __ret;
42003}
42004#endif
42005
42006#ifdef __LITTLE_ENDIAN__
42007__ai uint64x2_t vceqq_p64(poly64x2_t __p0, poly64x2_t __p1) {
42008  uint64x2_t __ret;
42009  __ret = (uint64x2_t)(__p0 == __p1);
42010  return __ret;
42011}
42012#else
42013__ai uint64x2_t vceqq_p64(poly64x2_t __p0, poly64x2_t __p1) {
42014  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42015  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42016  uint64x2_t __ret;
42017  __ret = (uint64x2_t)(__rev0 == __rev1);
42018  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42019  return __ret;
42020}
42021#endif
42022
42023#ifdef __LITTLE_ENDIAN__
42024__ai uint64x2_t vceqq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42025  uint64x2_t __ret;
42026  __ret = (uint64x2_t)(__p0 == __p1);
42027  return __ret;
42028}
42029#else
42030__ai uint64x2_t vceqq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42031  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42032  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42033  uint64x2_t __ret;
42034  __ret = (uint64x2_t)(__rev0 == __rev1);
42035  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42036  return __ret;
42037}
42038#endif
42039
42040#ifdef __LITTLE_ENDIAN__
42041__ai uint64x2_t vceqq_f64(float64x2_t __p0, float64x2_t __p1) {
42042  uint64x2_t __ret;
42043  __ret = (uint64x2_t)(__p0 == __p1);
42044  return __ret;
42045}
42046#else
42047__ai uint64x2_t vceqq_f64(float64x2_t __p0, float64x2_t __p1) {
42048  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42049  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42050  uint64x2_t __ret;
42051  __ret = (uint64x2_t)(__rev0 == __rev1);
42052  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42053  return __ret;
42054}
42055#endif
42056
42057#ifdef __LITTLE_ENDIAN__
42058__ai uint64x2_t vceqq_s64(int64x2_t __p0, int64x2_t __p1) {
42059  uint64x2_t __ret;
42060  __ret = (uint64x2_t)(__p0 == __p1);
42061  return __ret;
42062}
42063#else
42064__ai uint64x2_t vceqq_s64(int64x2_t __p0, int64x2_t __p1) {
42065  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42066  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42067  uint64x2_t __ret;
42068  __ret = (uint64x2_t)(__rev0 == __rev1);
42069  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42070  return __ret;
42071}
42072#endif
42073
42074#ifdef __LITTLE_ENDIAN__
42075__ai uint64x1_t vceq_u64(uint64x1_t __p0, uint64x1_t __p1) {
42076  uint64x1_t __ret;
42077  __ret = (uint64x1_t)(__p0 == __p1);
42078  return __ret;
42079}
42080#else
42081__ai uint64x1_t vceq_u64(uint64x1_t __p0, uint64x1_t __p1) {
42082  uint64x1_t __ret;
42083  __ret = (uint64x1_t)(__p0 == __p1);
42084  return __ret;
42085}
42086#endif
42087
42088#ifdef __LITTLE_ENDIAN__
42089__ai uint64x1_t vceq_f64(float64x1_t __p0, float64x1_t __p1) {
42090  uint64x1_t __ret;
42091  __ret = (uint64x1_t)(__p0 == __p1);
42092  return __ret;
42093}
42094#else
42095__ai uint64x1_t vceq_f64(float64x1_t __p0, float64x1_t __p1) {
42096  uint64x1_t __ret;
42097  __ret = (uint64x1_t)(__p0 == __p1);
42098  return __ret;
42099}
42100#endif
42101
42102#ifdef __LITTLE_ENDIAN__
42103__ai uint64x1_t vceq_s64(int64x1_t __p0, int64x1_t __p1) {
42104  uint64x1_t __ret;
42105  __ret = (uint64x1_t)(__p0 == __p1);
42106  return __ret;
42107}
42108#else
42109__ai uint64x1_t vceq_s64(int64x1_t __p0, int64x1_t __p1) {
42110  uint64x1_t __ret;
42111  __ret = (uint64x1_t)(__p0 == __p1);
42112  return __ret;
42113}
42114#endif
42115
42116#ifdef __LITTLE_ENDIAN__
42117__ai uint64_t vceqd_u64(uint64_t __p0, uint64_t __p1) {
42118  uint64_t __ret;
42119  __ret = (uint64_t) __builtin_neon_vceqd_u64(__p0, __p1);
42120  return __ret;
42121}
42122#else
42123__ai uint64_t vceqd_u64(uint64_t __p0, uint64_t __p1) {
42124  uint64_t __ret;
42125  __ret = (uint64_t) __builtin_neon_vceqd_u64(__p0, __p1);
42126  return __ret;
42127}
42128#endif
42129
42130#ifdef __LITTLE_ENDIAN__
42131__ai int64_t vceqd_s64(int64_t __p0, int64_t __p1) {
42132  int64_t __ret;
42133  __ret = (int64_t) __builtin_neon_vceqd_s64(__p0, __p1);
42134  return __ret;
42135}
42136#else
42137__ai int64_t vceqd_s64(int64_t __p0, int64_t __p1) {
42138  int64_t __ret;
42139  __ret = (int64_t) __builtin_neon_vceqd_s64(__p0, __p1);
42140  return __ret;
42141}
42142#endif
42143
42144#ifdef __LITTLE_ENDIAN__
42145__ai uint64_t vceqd_f64(float64_t __p0, float64_t __p1) {
42146  uint64_t __ret;
42147  __ret = (uint64_t) __builtin_neon_vceqd_f64(__p0, __p1);
42148  return __ret;
42149}
42150#else
42151__ai uint64_t vceqd_f64(float64_t __p0, float64_t __p1) {
42152  uint64_t __ret;
42153  __ret = (uint64_t) __builtin_neon_vceqd_f64(__p0, __p1);
42154  return __ret;
42155}
42156#endif
42157
42158#ifdef __LITTLE_ENDIAN__
42159__ai uint32_t vceqs_f32(float32_t __p0, float32_t __p1) {
42160  uint32_t __ret;
42161  __ret = (uint32_t) __builtin_neon_vceqs_f32(__p0, __p1);
42162  return __ret;
42163}
42164#else
42165__ai uint32_t vceqs_f32(float32_t __p0, float32_t __p1) {
42166  uint32_t __ret;
42167  __ret = (uint32_t) __builtin_neon_vceqs_f32(__p0, __p1);
42168  return __ret;
42169}
42170#endif
42171
42172#ifdef __LITTLE_ENDIAN__
42173__ai uint8x8_t vceqz_p8(poly8x8_t __p0) {
42174  uint8x8_t __ret;
42175  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
42176  return __ret;
42177}
42178#else
42179__ai uint8x8_t vceqz_p8(poly8x8_t __p0) {
42180  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42181  uint8x8_t __ret;
42182  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
42183  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42184  return __ret;
42185}
42186#endif
42187
42188#ifdef __LITTLE_ENDIAN__
42189__ai uint64x1_t vceqz_p64(poly64x1_t __p0) {
42190  uint64x1_t __ret;
42191  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42192  return __ret;
42193}
42194#else
42195__ai uint64x1_t vceqz_p64(poly64x1_t __p0) {
42196  uint64x1_t __ret;
42197  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42198  return __ret;
42199}
42200#endif
42201
42202#ifdef __LITTLE_ENDIAN__
42203__ai uint16x4_t vceqz_p16(poly16x4_t __p0) {
42204  uint16x4_t __ret;
42205  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
42206  return __ret;
42207}
42208#else
42209__ai uint16x4_t vceqz_p16(poly16x4_t __p0) {
42210  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42211  uint16x4_t __ret;
42212  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
42213  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42214  return __ret;
42215}
42216#endif
42217
42218#ifdef __LITTLE_ENDIAN__
42219__ai uint8x16_t vceqzq_p8(poly8x16_t __p0) {
42220  uint8x16_t __ret;
42221  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
42222  return __ret;
42223}
42224#else
42225__ai uint8x16_t vceqzq_p8(poly8x16_t __p0) {
42226  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42227  uint8x16_t __ret;
42228  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
42229  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42230  return __ret;
42231}
42232#endif
42233
42234#ifdef __LITTLE_ENDIAN__
42235__ai uint64x2_t vceqzq_p64(poly64x2_t __p0) {
42236  uint64x2_t __ret;
42237  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42238  return __ret;
42239}
42240#else
42241__ai uint64x2_t vceqzq_p64(poly64x2_t __p0) {
42242  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42243  uint64x2_t __ret;
42244  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42245  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42246  return __ret;
42247}
42248#endif
42249
42250#ifdef __LITTLE_ENDIAN__
42251__ai uint16x8_t vceqzq_p16(poly16x8_t __p0) {
42252  uint16x8_t __ret;
42253  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
42254  return __ret;
42255}
42256#else
42257__ai uint16x8_t vceqzq_p16(poly16x8_t __p0) {
42258  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42259  uint16x8_t __ret;
42260  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
42261  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42262  return __ret;
42263}
42264#endif
42265
42266#ifdef __LITTLE_ENDIAN__
42267__ai uint8x16_t vceqzq_u8(uint8x16_t __p0) {
42268  uint8x16_t __ret;
42269  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
42270  return __ret;
42271}
42272#else
42273__ai uint8x16_t vceqzq_u8(uint8x16_t __p0) {
42274  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42275  uint8x16_t __ret;
42276  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
42277  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42278  return __ret;
42279}
42280#endif
42281
42282#ifdef __LITTLE_ENDIAN__
42283__ai uint32x4_t vceqzq_u32(uint32x4_t __p0) {
42284  uint32x4_t __ret;
42285  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
42286  return __ret;
42287}
42288#else
42289__ai uint32x4_t vceqzq_u32(uint32x4_t __p0) {
42290  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42291  uint32x4_t __ret;
42292  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
42293  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42294  return __ret;
42295}
42296#endif
42297
42298#ifdef __LITTLE_ENDIAN__
42299__ai uint64x2_t vceqzq_u64(uint64x2_t __p0) {
42300  uint64x2_t __ret;
42301  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42302  return __ret;
42303}
42304#else
42305__ai uint64x2_t vceqzq_u64(uint64x2_t __p0) {
42306  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42307  uint64x2_t __ret;
42308  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42309  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42310  return __ret;
42311}
42312#endif
42313
42314#ifdef __LITTLE_ENDIAN__
42315__ai uint16x8_t vceqzq_u16(uint16x8_t __p0) {
42316  uint16x8_t __ret;
42317  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
42318  return __ret;
42319}
42320#else
42321__ai uint16x8_t vceqzq_u16(uint16x8_t __p0) {
42322  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42323  uint16x8_t __ret;
42324  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
42325  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42326  return __ret;
42327}
42328#endif
42329
42330#ifdef __LITTLE_ENDIAN__
42331__ai uint8x16_t vceqzq_s8(int8x16_t __p0) {
42332  uint8x16_t __ret;
42333  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 48);
42334  return __ret;
42335}
42336#else
42337__ai uint8x16_t vceqzq_s8(int8x16_t __p0) {
42338  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42339  uint8x16_t __ret;
42340  __ret = (uint8x16_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 48);
42341  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42342  return __ret;
42343}
42344#endif
42345
42346#ifdef __LITTLE_ENDIAN__
42347__ai uint64x2_t vceqzq_f64(float64x2_t __p0) {
42348  uint64x2_t __ret;
42349  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42350  return __ret;
42351}
42352#else
42353__ai uint64x2_t vceqzq_f64(float64x2_t __p0) {
42354  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42355  uint64x2_t __ret;
42356  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42357  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42358  return __ret;
42359}
42360#endif
42361
42362#ifdef __LITTLE_ENDIAN__
42363__ai uint32x4_t vceqzq_f32(float32x4_t __p0) {
42364  uint32x4_t __ret;
42365  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
42366  return __ret;
42367}
42368#else
42369__ai uint32x4_t vceqzq_f32(float32x4_t __p0) {
42370  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42371  uint32x4_t __ret;
42372  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
42373  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42374  return __ret;
42375}
42376#endif
42377
42378#ifdef __LITTLE_ENDIAN__
42379__ai uint32x4_t vceqzq_s32(int32x4_t __p0) {
42380  uint32x4_t __ret;
42381  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 50);
42382  return __ret;
42383}
42384#else
42385__ai uint32x4_t vceqzq_s32(int32x4_t __p0) {
42386  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42387  uint32x4_t __ret;
42388  __ret = (uint32x4_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 50);
42389  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42390  return __ret;
42391}
42392#endif
42393
42394#ifdef __LITTLE_ENDIAN__
42395__ai uint64x2_t vceqzq_s64(int64x2_t __p0) {
42396  uint64x2_t __ret;
42397  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 51);
42398  return __ret;
42399}
42400#else
42401__ai uint64x2_t vceqzq_s64(int64x2_t __p0) {
42402  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42403  uint64x2_t __ret;
42404  __ret = (uint64x2_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 51);
42405  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42406  return __ret;
42407}
42408#endif
42409
42410#ifdef __LITTLE_ENDIAN__
42411__ai uint16x8_t vceqzq_s16(int16x8_t __p0) {
42412  uint16x8_t __ret;
42413  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__p0, 49);
42414  return __ret;
42415}
42416#else
42417__ai uint16x8_t vceqzq_s16(int16x8_t __p0) {
42418  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42419  uint16x8_t __ret;
42420  __ret = (uint16x8_t) __builtin_neon_vceqzq_v((int8x16_t)__rev0, 49);
42421  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42422  return __ret;
42423}
42424#endif
42425
42426#ifdef __LITTLE_ENDIAN__
42427__ai uint8x8_t vceqz_u8(uint8x8_t __p0) {
42428  uint8x8_t __ret;
42429  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
42430  return __ret;
42431}
42432#else
42433__ai uint8x8_t vceqz_u8(uint8x8_t __p0) {
42434  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42435  uint8x8_t __ret;
42436  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
42437  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42438  return __ret;
42439}
42440#endif
42441
42442#ifdef __LITTLE_ENDIAN__
42443__ai uint32x2_t vceqz_u32(uint32x2_t __p0) {
42444  uint32x2_t __ret;
42445  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
42446  return __ret;
42447}
42448#else
42449__ai uint32x2_t vceqz_u32(uint32x2_t __p0) {
42450  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42451  uint32x2_t __ret;
42452  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
42453  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42454  return __ret;
42455}
42456#endif
42457
42458#ifdef __LITTLE_ENDIAN__
42459__ai uint64x1_t vceqz_u64(uint64x1_t __p0) {
42460  uint64x1_t __ret;
42461  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42462  return __ret;
42463}
42464#else
42465__ai uint64x1_t vceqz_u64(uint64x1_t __p0) {
42466  uint64x1_t __ret;
42467  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42468  return __ret;
42469}
42470#endif
42471
42472#ifdef __LITTLE_ENDIAN__
42473__ai uint16x4_t vceqz_u16(uint16x4_t __p0) {
42474  uint16x4_t __ret;
42475  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
42476  return __ret;
42477}
42478#else
42479__ai uint16x4_t vceqz_u16(uint16x4_t __p0) {
42480  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42481  uint16x4_t __ret;
42482  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
42483  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42484  return __ret;
42485}
42486#endif
42487
42488#ifdef __LITTLE_ENDIAN__
42489__ai uint8x8_t vceqz_s8(int8x8_t __p0) {
42490  uint8x8_t __ret;
42491  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 16);
42492  return __ret;
42493}
42494#else
42495__ai uint8x8_t vceqz_s8(int8x8_t __p0) {
42496  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42497  uint8x8_t __ret;
42498  __ret = (uint8x8_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 16);
42499  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42500  return __ret;
42501}
42502#endif
42503
42504#ifdef __LITTLE_ENDIAN__
42505__ai uint64x1_t vceqz_f64(float64x1_t __p0) {
42506  uint64x1_t __ret;
42507  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42508  return __ret;
42509}
42510#else
42511__ai uint64x1_t vceqz_f64(float64x1_t __p0) {
42512  uint64x1_t __ret;
42513  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42514  return __ret;
42515}
42516#endif
42517
42518#ifdef __LITTLE_ENDIAN__
42519__ai uint32x2_t vceqz_f32(float32x2_t __p0) {
42520  uint32x2_t __ret;
42521  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
42522  return __ret;
42523}
42524#else
42525__ai uint32x2_t vceqz_f32(float32x2_t __p0) {
42526  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42527  uint32x2_t __ret;
42528  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
42529  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42530  return __ret;
42531}
42532#endif
42533
42534#ifdef __LITTLE_ENDIAN__
42535__ai uint32x2_t vceqz_s32(int32x2_t __p0) {
42536  uint32x2_t __ret;
42537  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 18);
42538  return __ret;
42539}
42540#else
42541__ai uint32x2_t vceqz_s32(int32x2_t __p0) {
42542  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42543  uint32x2_t __ret;
42544  __ret = (uint32x2_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 18);
42545  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42546  return __ret;
42547}
42548#endif
42549
42550#ifdef __LITTLE_ENDIAN__
42551__ai uint64x1_t vceqz_s64(int64x1_t __p0) {
42552  uint64x1_t __ret;
42553  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42554  return __ret;
42555}
42556#else
42557__ai uint64x1_t vceqz_s64(int64x1_t __p0) {
42558  uint64x1_t __ret;
42559  __ret = (uint64x1_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 19);
42560  return __ret;
42561}
42562#endif
42563
42564#ifdef __LITTLE_ENDIAN__
42565__ai uint16x4_t vceqz_s16(int16x4_t __p0) {
42566  uint16x4_t __ret;
42567  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__p0, 17);
42568  return __ret;
42569}
42570#else
42571__ai uint16x4_t vceqz_s16(int16x4_t __p0) {
42572  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42573  uint16x4_t __ret;
42574  __ret = (uint16x4_t) __builtin_neon_vceqz_v((int8x8_t)__rev0, 17);
42575  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42576  return __ret;
42577}
42578#endif
42579
42580#ifdef __LITTLE_ENDIAN__
42581__ai uint64_t vceqzd_u64(uint64_t __p0) {
42582  uint64_t __ret;
42583  __ret = (uint64_t) __builtin_neon_vceqzd_u64(__p0);
42584  return __ret;
42585}
42586#else
42587__ai uint64_t vceqzd_u64(uint64_t __p0) {
42588  uint64_t __ret;
42589  __ret = (uint64_t) __builtin_neon_vceqzd_u64(__p0);
42590  return __ret;
42591}
42592#endif
42593
42594#ifdef __LITTLE_ENDIAN__
42595__ai int64_t vceqzd_s64(int64_t __p0) {
42596  int64_t __ret;
42597  __ret = (int64_t) __builtin_neon_vceqzd_s64(__p0);
42598  return __ret;
42599}
42600#else
42601__ai int64_t vceqzd_s64(int64_t __p0) {
42602  int64_t __ret;
42603  __ret = (int64_t) __builtin_neon_vceqzd_s64(__p0);
42604  return __ret;
42605}
42606#endif
42607
42608#ifdef __LITTLE_ENDIAN__
42609__ai uint64_t vceqzd_f64(float64_t __p0) {
42610  uint64_t __ret;
42611  __ret = (uint64_t) __builtin_neon_vceqzd_f64(__p0);
42612  return __ret;
42613}
42614#else
42615__ai uint64_t vceqzd_f64(float64_t __p0) {
42616  uint64_t __ret;
42617  __ret = (uint64_t) __builtin_neon_vceqzd_f64(__p0);
42618  return __ret;
42619}
42620#endif
42621
42622#ifdef __LITTLE_ENDIAN__
42623__ai uint32_t vceqzs_f32(float32_t __p0) {
42624  uint32_t __ret;
42625  __ret = (uint32_t) __builtin_neon_vceqzs_f32(__p0);
42626  return __ret;
42627}
42628#else
42629__ai uint32_t vceqzs_f32(float32_t __p0) {
42630  uint32_t __ret;
42631  __ret = (uint32_t) __builtin_neon_vceqzs_f32(__p0);
42632  return __ret;
42633}
42634#endif
42635
42636#ifdef __LITTLE_ENDIAN__
42637__ai uint64x2_t vcgeq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42638  uint64x2_t __ret;
42639  __ret = (uint64x2_t)(__p0 >= __p1);
42640  return __ret;
42641}
42642#else
42643__ai uint64x2_t vcgeq_u64(uint64x2_t __p0, uint64x2_t __p1) {
42644  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42645  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42646  uint64x2_t __ret;
42647  __ret = (uint64x2_t)(__rev0 >= __rev1);
42648  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42649  return __ret;
42650}
42651#endif
42652
42653#ifdef __LITTLE_ENDIAN__
42654__ai uint64x2_t vcgeq_f64(float64x2_t __p0, float64x2_t __p1) {
42655  uint64x2_t __ret;
42656  __ret = (uint64x2_t)(__p0 >= __p1);
42657  return __ret;
42658}
42659#else
42660__ai uint64x2_t vcgeq_f64(float64x2_t __p0, float64x2_t __p1) {
42661  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42662  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42663  uint64x2_t __ret;
42664  __ret = (uint64x2_t)(__rev0 >= __rev1);
42665  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42666  return __ret;
42667}
42668#endif
42669
42670#ifdef __LITTLE_ENDIAN__
42671__ai uint64x2_t vcgeq_s64(int64x2_t __p0, int64x2_t __p1) {
42672  uint64x2_t __ret;
42673  __ret = (uint64x2_t)(__p0 >= __p1);
42674  return __ret;
42675}
42676#else
42677__ai uint64x2_t vcgeq_s64(int64x2_t __p0, int64x2_t __p1) {
42678  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42679  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
42680  uint64x2_t __ret;
42681  __ret = (uint64x2_t)(__rev0 >= __rev1);
42682  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42683  return __ret;
42684}
42685#endif
42686
42687#ifdef __LITTLE_ENDIAN__
42688__ai uint64x1_t vcge_u64(uint64x1_t __p0, uint64x1_t __p1) {
42689  uint64x1_t __ret;
42690  __ret = (uint64x1_t)(__p0 >= __p1);
42691  return __ret;
42692}
42693#else
42694__ai uint64x1_t vcge_u64(uint64x1_t __p0, uint64x1_t __p1) {
42695  uint64x1_t __ret;
42696  __ret = (uint64x1_t)(__p0 >= __p1);
42697  return __ret;
42698}
42699#endif
42700
42701#ifdef __LITTLE_ENDIAN__
42702__ai uint64x1_t vcge_f64(float64x1_t __p0, float64x1_t __p1) {
42703  uint64x1_t __ret;
42704  __ret = (uint64x1_t)(__p0 >= __p1);
42705  return __ret;
42706}
42707#else
42708__ai uint64x1_t vcge_f64(float64x1_t __p0, float64x1_t __p1) {
42709  uint64x1_t __ret;
42710  __ret = (uint64x1_t)(__p0 >= __p1);
42711  return __ret;
42712}
42713#endif
42714
42715#ifdef __LITTLE_ENDIAN__
42716__ai uint64x1_t vcge_s64(int64x1_t __p0, int64x1_t __p1) {
42717  uint64x1_t __ret;
42718  __ret = (uint64x1_t)(__p0 >= __p1);
42719  return __ret;
42720}
42721#else
42722__ai uint64x1_t vcge_s64(int64x1_t __p0, int64x1_t __p1) {
42723  uint64x1_t __ret;
42724  __ret = (uint64x1_t)(__p0 >= __p1);
42725  return __ret;
42726}
42727#endif
42728
42729#ifdef __LITTLE_ENDIAN__
42730__ai int64_t vcged_s64(int64_t __p0, int64_t __p1) {
42731  int64_t __ret;
42732  __ret = (int64_t) __builtin_neon_vcged_s64(__p0, __p1);
42733  return __ret;
42734}
42735#else
42736__ai int64_t vcged_s64(int64_t __p0, int64_t __p1) {
42737  int64_t __ret;
42738  __ret = (int64_t) __builtin_neon_vcged_s64(__p0, __p1);
42739  return __ret;
42740}
42741#endif
42742
42743#ifdef __LITTLE_ENDIAN__
42744__ai uint64_t vcged_u64(uint64_t __p0, uint64_t __p1) {
42745  uint64_t __ret;
42746  __ret = (uint64_t) __builtin_neon_vcged_u64(__p0, __p1);
42747  return __ret;
42748}
42749#else
42750__ai uint64_t vcged_u64(uint64_t __p0, uint64_t __p1) {
42751  uint64_t __ret;
42752  __ret = (uint64_t) __builtin_neon_vcged_u64(__p0, __p1);
42753  return __ret;
42754}
42755#endif
42756
42757#ifdef __LITTLE_ENDIAN__
42758__ai uint64_t vcged_f64(float64_t __p0, float64_t __p1) {
42759  uint64_t __ret;
42760  __ret = (uint64_t) __builtin_neon_vcged_f64(__p0, __p1);
42761  return __ret;
42762}
42763#else
42764__ai uint64_t vcged_f64(float64_t __p0, float64_t __p1) {
42765  uint64_t __ret;
42766  __ret = (uint64_t) __builtin_neon_vcged_f64(__p0, __p1);
42767  return __ret;
42768}
42769#endif
42770
42771#ifdef __LITTLE_ENDIAN__
42772__ai uint32_t vcges_f32(float32_t __p0, float32_t __p1) {
42773  uint32_t __ret;
42774  __ret = (uint32_t) __builtin_neon_vcges_f32(__p0, __p1);
42775  return __ret;
42776}
42777#else
42778__ai uint32_t vcges_f32(float32_t __p0, float32_t __p1) {
42779  uint32_t __ret;
42780  __ret = (uint32_t) __builtin_neon_vcges_f32(__p0, __p1);
42781  return __ret;
42782}
42783#endif
42784
42785#ifdef __LITTLE_ENDIAN__
42786__ai uint8x16_t vcgezq_s8(int8x16_t __p0) {
42787  uint8x16_t __ret;
42788  __ret = (uint8x16_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 48);
42789  return __ret;
42790}
42791#else
42792__ai uint8x16_t vcgezq_s8(int8x16_t __p0) {
42793  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42794  uint8x16_t __ret;
42795  __ret = (uint8x16_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 48);
42796  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
42797  return __ret;
42798}
42799#endif
42800
42801#ifdef __LITTLE_ENDIAN__
42802__ai uint64x2_t vcgezq_f64(float64x2_t __p0) {
42803  uint64x2_t __ret;
42804  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 51);
42805  return __ret;
42806}
42807#else
42808__ai uint64x2_t vcgezq_f64(float64x2_t __p0) {
42809  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42810  uint64x2_t __ret;
42811  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 51);
42812  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42813  return __ret;
42814}
42815#endif
42816
42817#ifdef __LITTLE_ENDIAN__
42818__ai uint32x4_t vcgezq_f32(float32x4_t __p0) {
42819  uint32x4_t __ret;
42820  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 50);
42821  return __ret;
42822}
42823#else
42824__ai uint32x4_t vcgezq_f32(float32x4_t __p0) {
42825  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42826  uint32x4_t __ret;
42827  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 50);
42828  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42829  return __ret;
42830}
42831#endif
42832
42833#ifdef __LITTLE_ENDIAN__
42834__ai uint32x4_t vcgezq_s32(int32x4_t __p0) {
42835  uint32x4_t __ret;
42836  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 50);
42837  return __ret;
42838}
42839#else
42840__ai uint32x4_t vcgezq_s32(int32x4_t __p0) {
42841  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42842  uint32x4_t __ret;
42843  __ret = (uint32x4_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 50);
42844  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42845  return __ret;
42846}
42847#endif
42848
42849#ifdef __LITTLE_ENDIAN__
42850__ai uint64x2_t vcgezq_s64(int64x2_t __p0) {
42851  uint64x2_t __ret;
42852  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 51);
42853  return __ret;
42854}
42855#else
42856__ai uint64x2_t vcgezq_s64(int64x2_t __p0) {
42857  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42858  uint64x2_t __ret;
42859  __ret = (uint64x2_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 51);
42860  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42861  return __ret;
42862}
42863#endif
42864
42865#ifdef __LITTLE_ENDIAN__
42866__ai uint16x8_t vcgezq_s16(int16x8_t __p0) {
42867  uint16x8_t __ret;
42868  __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__p0, 49);
42869  return __ret;
42870}
42871#else
42872__ai uint16x8_t vcgezq_s16(int16x8_t __p0) {
42873  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42874  uint16x8_t __ret;
42875  __ret = (uint16x8_t) __builtin_neon_vcgezq_v((int8x16_t)__rev0, 49);
42876  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42877  return __ret;
42878}
42879#endif
42880
42881#ifdef __LITTLE_ENDIAN__
42882__ai uint8x8_t vcgez_s8(int8x8_t __p0) {
42883  uint8x8_t __ret;
42884  __ret = (uint8x8_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 16);
42885  return __ret;
42886}
42887#else
42888__ai uint8x8_t vcgez_s8(int8x8_t __p0) {
42889  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
42890  uint8x8_t __ret;
42891  __ret = (uint8x8_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 16);
42892  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
42893  return __ret;
42894}
42895#endif
42896
42897#ifdef __LITTLE_ENDIAN__
42898__ai uint64x1_t vcgez_f64(float64x1_t __p0) {
42899  uint64x1_t __ret;
42900  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
42901  return __ret;
42902}
42903#else
42904__ai uint64x1_t vcgez_f64(float64x1_t __p0) {
42905  uint64x1_t __ret;
42906  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
42907  return __ret;
42908}
42909#endif
42910
42911#ifdef __LITTLE_ENDIAN__
42912__ai uint32x2_t vcgez_f32(float32x2_t __p0) {
42913  uint32x2_t __ret;
42914  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 18);
42915  return __ret;
42916}
42917#else
42918__ai uint32x2_t vcgez_f32(float32x2_t __p0) {
42919  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42920  uint32x2_t __ret;
42921  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 18);
42922  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42923  return __ret;
42924}
42925#endif
42926
42927#ifdef __LITTLE_ENDIAN__
42928__ai uint32x2_t vcgez_s32(int32x2_t __p0) {
42929  uint32x2_t __ret;
42930  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 18);
42931  return __ret;
42932}
42933#else
42934__ai uint32x2_t vcgez_s32(int32x2_t __p0) {
42935  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
42936  uint32x2_t __ret;
42937  __ret = (uint32x2_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 18);
42938  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
42939  return __ret;
42940}
42941#endif
42942
42943#ifdef __LITTLE_ENDIAN__
42944__ai uint64x1_t vcgez_s64(int64x1_t __p0) {
42945  uint64x1_t __ret;
42946  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
42947  return __ret;
42948}
42949#else
42950__ai uint64x1_t vcgez_s64(int64x1_t __p0) {
42951  uint64x1_t __ret;
42952  __ret = (uint64x1_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 19);
42953  return __ret;
42954}
42955#endif
42956
42957#ifdef __LITTLE_ENDIAN__
42958__ai uint16x4_t vcgez_s16(int16x4_t __p0) {
42959  uint16x4_t __ret;
42960  __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__p0, 17);
42961  return __ret;
42962}
42963#else
42964__ai uint16x4_t vcgez_s16(int16x4_t __p0) {
42965  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
42966  uint16x4_t __ret;
42967  __ret = (uint16x4_t) __builtin_neon_vcgez_v((int8x8_t)__rev0, 17);
42968  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
42969  return __ret;
42970}
42971#endif
42972
42973#ifdef __LITTLE_ENDIAN__
42974__ai int64_t vcgezd_s64(int64_t __p0) {
42975  int64_t __ret;
42976  __ret = (int64_t) __builtin_neon_vcgezd_s64(__p0);
42977  return __ret;
42978}
42979#else
42980__ai int64_t vcgezd_s64(int64_t __p0) {
42981  int64_t __ret;
42982  __ret = (int64_t) __builtin_neon_vcgezd_s64(__p0);
42983  return __ret;
42984}
42985#endif
42986
42987#ifdef __LITTLE_ENDIAN__
42988__ai uint64_t vcgezd_f64(float64_t __p0) {
42989  uint64_t __ret;
42990  __ret = (uint64_t) __builtin_neon_vcgezd_f64(__p0);
42991  return __ret;
42992}
42993#else
42994__ai uint64_t vcgezd_f64(float64_t __p0) {
42995  uint64_t __ret;
42996  __ret = (uint64_t) __builtin_neon_vcgezd_f64(__p0);
42997  return __ret;
42998}
42999#endif
43000
43001#ifdef __LITTLE_ENDIAN__
43002__ai uint32_t vcgezs_f32(float32_t __p0) {
43003  uint32_t __ret;
43004  __ret = (uint32_t) __builtin_neon_vcgezs_f32(__p0);
43005  return __ret;
43006}
43007#else
43008__ai uint32_t vcgezs_f32(float32_t __p0) {
43009  uint32_t __ret;
43010  __ret = (uint32_t) __builtin_neon_vcgezs_f32(__p0);
43011  return __ret;
43012}
43013#endif
43014
43015#ifdef __LITTLE_ENDIAN__
43016__ai uint64x2_t vcgtq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43017  uint64x2_t __ret;
43018  __ret = (uint64x2_t)(__p0 > __p1);
43019  return __ret;
43020}
43021#else
43022__ai uint64x2_t vcgtq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43023  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43024  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43025  uint64x2_t __ret;
43026  __ret = (uint64x2_t)(__rev0 > __rev1);
43027  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43028  return __ret;
43029}
43030#endif
43031
43032#ifdef __LITTLE_ENDIAN__
43033__ai uint64x2_t vcgtq_f64(float64x2_t __p0, float64x2_t __p1) {
43034  uint64x2_t __ret;
43035  __ret = (uint64x2_t)(__p0 > __p1);
43036  return __ret;
43037}
43038#else
43039__ai uint64x2_t vcgtq_f64(float64x2_t __p0, float64x2_t __p1) {
43040  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43041  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43042  uint64x2_t __ret;
43043  __ret = (uint64x2_t)(__rev0 > __rev1);
43044  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43045  return __ret;
43046}
43047#endif
43048
43049#ifdef __LITTLE_ENDIAN__
43050__ai uint64x2_t vcgtq_s64(int64x2_t __p0, int64x2_t __p1) {
43051  uint64x2_t __ret;
43052  __ret = (uint64x2_t)(__p0 > __p1);
43053  return __ret;
43054}
43055#else
43056__ai uint64x2_t vcgtq_s64(int64x2_t __p0, int64x2_t __p1) {
43057  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43058  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43059  uint64x2_t __ret;
43060  __ret = (uint64x2_t)(__rev0 > __rev1);
43061  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43062  return __ret;
43063}
43064#endif
43065
43066#ifdef __LITTLE_ENDIAN__
43067__ai uint64x1_t vcgt_u64(uint64x1_t __p0, uint64x1_t __p1) {
43068  uint64x1_t __ret;
43069  __ret = (uint64x1_t)(__p0 > __p1);
43070  return __ret;
43071}
43072#else
43073__ai uint64x1_t vcgt_u64(uint64x1_t __p0, uint64x1_t __p1) {
43074  uint64x1_t __ret;
43075  __ret = (uint64x1_t)(__p0 > __p1);
43076  return __ret;
43077}
43078#endif
43079
43080#ifdef __LITTLE_ENDIAN__
43081__ai uint64x1_t vcgt_f64(float64x1_t __p0, float64x1_t __p1) {
43082  uint64x1_t __ret;
43083  __ret = (uint64x1_t)(__p0 > __p1);
43084  return __ret;
43085}
43086#else
43087__ai uint64x1_t vcgt_f64(float64x1_t __p0, float64x1_t __p1) {
43088  uint64x1_t __ret;
43089  __ret = (uint64x1_t)(__p0 > __p1);
43090  return __ret;
43091}
43092#endif
43093
43094#ifdef __LITTLE_ENDIAN__
43095__ai uint64x1_t vcgt_s64(int64x1_t __p0, int64x1_t __p1) {
43096  uint64x1_t __ret;
43097  __ret = (uint64x1_t)(__p0 > __p1);
43098  return __ret;
43099}
43100#else
43101__ai uint64x1_t vcgt_s64(int64x1_t __p0, int64x1_t __p1) {
43102  uint64x1_t __ret;
43103  __ret = (uint64x1_t)(__p0 > __p1);
43104  return __ret;
43105}
43106#endif
43107
43108#ifdef __LITTLE_ENDIAN__
43109__ai int64_t vcgtd_s64(int64_t __p0, int64_t __p1) {
43110  int64_t __ret;
43111  __ret = (int64_t) __builtin_neon_vcgtd_s64(__p0, __p1);
43112  return __ret;
43113}
43114#else
43115__ai int64_t vcgtd_s64(int64_t __p0, int64_t __p1) {
43116  int64_t __ret;
43117  __ret = (int64_t) __builtin_neon_vcgtd_s64(__p0, __p1);
43118  return __ret;
43119}
43120#endif
43121
43122#ifdef __LITTLE_ENDIAN__
43123__ai uint64_t vcgtd_u64(uint64_t __p0, uint64_t __p1) {
43124  uint64_t __ret;
43125  __ret = (uint64_t) __builtin_neon_vcgtd_u64(__p0, __p1);
43126  return __ret;
43127}
43128#else
43129__ai uint64_t vcgtd_u64(uint64_t __p0, uint64_t __p1) {
43130  uint64_t __ret;
43131  __ret = (uint64_t) __builtin_neon_vcgtd_u64(__p0, __p1);
43132  return __ret;
43133}
43134#endif
43135
43136#ifdef __LITTLE_ENDIAN__
43137__ai uint64_t vcgtd_f64(float64_t __p0, float64_t __p1) {
43138  uint64_t __ret;
43139  __ret = (uint64_t) __builtin_neon_vcgtd_f64(__p0, __p1);
43140  return __ret;
43141}
43142#else
43143__ai uint64_t vcgtd_f64(float64_t __p0, float64_t __p1) {
43144  uint64_t __ret;
43145  __ret = (uint64_t) __builtin_neon_vcgtd_f64(__p0, __p1);
43146  return __ret;
43147}
43148#endif
43149
43150#ifdef __LITTLE_ENDIAN__
43151__ai uint32_t vcgts_f32(float32_t __p0, float32_t __p1) {
43152  uint32_t __ret;
43153  __ret = (uint32_t) __builtin_neon_vcgts_f32(__p0, __p1);
43154  return __ret;
43155}
43156#else
43157__ai uint32_t vcgts_f32(float32_t __p0, float32_t __p1) {
43158  uint32_t __ret;
43159  __ret = (uint32_t) __builtin_neon_vcgts_f32(__p0, __p1);
43160  return __ret;
43161}
43162#endif
43163
43164#ifdef __LITTLE_ENDIAN__
43165__ai uint8x16_t vcgtzq_s8(int8x16_t __p0) {
43166  uint8x16_t __ret;
43167  __ret = (uint8x16_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 48);
43168  return __ret;
43169}
43170#else
43171__ai uint8x16_t vcgtzq_s8(int8x16_t __p0) {
43172  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43173  uint8x16_t __ret;
43174  __ret = (uint8x16_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 48);
43175  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43176  return __ret;
43177}
43178#endif
43179
43180#ifdef __LITTLE_ENDIAN__
43181__ai uint64x2_t vcgtzq_f64(float64x2_t __p0) {
43182  uint64x2_t __ret;
43183  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 51);
43184  return __ret;
43185}
43186#else
43187__ai uint64x2_t vcgtzq_f64(float64x2_t __p0) {
43188  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43189  uint64x2_t __ret;
43190  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 51);
43191  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43192  return __ret;
43193}
43194#endif
43195
43196#ifdef __LITTLE_ENDIAN__
43197__ai uint32x4_t vcgtzq_f32(float32x4_t __p0) {
43198  uint32x4_t __ret;
43199  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 50);
43200  return __ret;
43201}
43202#else
43203__ai uint32x4_t vcgtzq_f32(float32x4_t __p0) {
43204  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43205  uint32x4_t __ret;
43206  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 50);
43207  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43208  return __ret;
43209}
43210#endif
43211
43212#ifdef __LITTLE_ENDIAN__
43213__ai uint32x4_t vcgtzq_s32(int32x4_t __p0) {
43214  uint32x4_t __ret;
43215  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 50);
43216  return __ret;
43217}
43218#else
43219__ai uint32x4_t vcgtzq_s32(int32x4_t __p0) {
43220  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43221  uint32x4_t __ret;
43222  __ret = (uint32x4_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 50);
43223  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43224  return __ret;
43225}
43226#endif
43227
43228#ifdef __LITTLE_ENDIAN__
43229__ai uint64x2_t vcgtzq_s64(int64x2_t __p0) {
43230  uint64x2_t __ret;
43231  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 51);
43232  return __ret;
43233}
43234#else
43235__ai uint64x2_t vcgtzq_s64(int64x2_t __p0) {
43236  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43237  uint64x2_t __ret;
43238  __ret = (uint64x2_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 51);
43239  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43240  return __ret;
43241}
43242#endif
43243
43244#ifdef __LITTLE_ENDIAN__
43245__ai uint16x8_t vcgtzq_s16(int16x8_t __p0) {
43246  uint16x8_t __ret;
43247  __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__p0, 49);
43248  return __ret;
43249}
43250#else
43251__ai uint16x8_t vcgtzq_s16(int16x8_t __p0) {
43252  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43253  uint16x8_t __ret;
43254  __ret = (uint16x8_t) __builtin_neon_vcgtzq_v((int8x16_t)__rev0, 49);
43255  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43256  return __ret;
43257}
43258#endif
43259
43260#ifdef __LITTLE_ENDIAN__
43261__ai uint8x8_t vcgtz_s8(int8x8_t __p0) {
43262  uint8x8_t __ret;
43263  __ret = (uint8x8_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 16);
43264  return __ret;
43265}
43266#else
43267__ai uint8x8_t vcgtz_s8(int8x8_t __p0) {
43268  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43269  uint8x8_t __ret;
43270  __ret = (uint8x8_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 16);
43271  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43272  return __ret;
43273}
43274#endif
43275
43276#ifdef __LITTLE_ENDIAN__
43277__ai uint64x1_t vcgtz_f64(float64x1_t __p0) {
43278  uint64x1_t __ret;
43279  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
43280  return __ret;
43281}
43282#else
43283__ai uint64x1_t vcgtz_f64(float64x1_t __p0) {
43284  uint64x1_t __ret;
43285  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
43286  return __ret;
43287}
43288#endif
43289
43290#ifdef __LITTLE_ENDIAN__
43291__ai uint32x2_t vcgtz_f32(float32x2_t __p0) {
43292  uint32x2_t __ret;
43293  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 18);
43294  return __ret;
43295}
43296#else
43297__ai uint32x2_t vcgtz_f32(float32x2_t __p0) {
43298  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43299  uint32x2_t __ret;
43300  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 18);
43301  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43302  return __ret;
43303}
43304#endif
43305
43306#ifdef __LITTLE_ENDIAN__
43307__ai uint32x2_t vcgtz_s32(int32x2_t __p0) {
43308  uint32x2_t __ret;
43309  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 18);
43310  return __ret;
43311}
43312#else
43313__ai uint32x2_t vcgtz_s32(int32x2_t __p0) {
43314  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43315  uint32x2_t __ret;
43316  __ret = (uint32x2_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 18);
43317  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43318  return __ret;
43319}
43320#endif
43321
43322#ifdef __LITTLE_ENDIAN__
43323__ai uint64x1_t vcgtz_s64(int64x1_t __p0) {
43324  uint64x1_t __ret;
43325  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
43326  return __ret;
43327}
43328#else
43329__ai uint64x1_t vcgtz_s64(int64x1_t __p0) {
43330  uint64x1_t __ret;
43331  __ret = (uint64x1_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 19);
43332  return __ret;
43333}
43334#endif
43335
43336#ifdef __LITTLE_ENDIAN__
43337__ai uint16x4_t vcgtz_s16(int16x4_t __p0) {
43338  uint16x4_t __ret;
43339  __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__p0, 17);
43340  return __ret;
43341}
43342#else
43343__ai uint16x4_t vcgtz_s16(int16x4_t __p0) {
43344  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43345  uint16x4_t __ret;
43346  __ret = (uint16x4_t) __builtin_neon_vcgtz_v((int8x8_t)__rev0, 17);
43347  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43348  return __ret;
43349}
43350#endif
43351
43352#ifdef __LITTLE_ENDIAN__
43353__ai int64_t vcgtzd_s64(int64_t __p0) {
43354  int64_t __ret;
43355  __ret = (int64_t) __builtin_neon_vcgtzd_s64(__p0);
43356  return __ret;
43357}
43358#else
43359__ai int64_t vcgtzd_s64(int64_t __p0) {
43360  int64_t __ret;
43361  __ret = (int64_t) __builtin_neon_vcgtzd_s64(__p0);
43362  return __ret;
43363}
43364#endif
43365
43366#ifdef __LITTLE_ENDIAN__
43367__ai uint64_t vcgtzd_f64(float64_t __p0) {
43368  uint64_t __ret;
43369  __ret = (uint64_t) __builtin_neon_vcgtzd_f64(__p0);
43370  return __ret;
43371}
43372#else
43373__ai uint64_t vcgtzd_f64(float64_t __p0) {
43374  uint64_t __ret;
43375  __ret = (uint64_t) __builtin_neon_vcgtzd_f64(__p0);
43376  return __ret;
43377}
43378#endif
43379
43380#ifdef __LITTLE_ENDIAN__
43381__ai uint32_t vcgtzs_f32(float32_t __p0) {
43382  uint32_t __ret;
43383  __ret = (uint32_t) __builtin_neon_vcgtzs_f32(__p0);
43384  return __ret;
43385}
43386#else
43387__ai uint32_t vcgtzs_f32(float32_t __p0) {
43388  uint32_t __ret;
43389  __ret = (uint32_t) __builtin_neon_vcgtzs_f32(__p0);
43390  return __ret;
43391}
43392#endif
43393
43394#ifdef __LITTLE_ENDIAN__
43395__ai uint64x2_t vcleq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43396  uint64x2_t __ret;
43397  __ret = (uint64x2_t)(__p0 <= __p1);
43398  return __ret;
43399}
43400#else
43401__ai uint64x2_t vcleq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43402  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43403  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43404  uint64x2_t __ret;
43405  __ret = (uint64x2_t)(__rev0 <= __rev1);
43406  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43407  return __ret;
43408}
43409#endif
43410
43411#ifdef __LITTLE_ENDIAN__
43412__ai uint64x2_t vcleq_f64(float64x2_t __p0, float64x2_t __p1) {
43413  uint64x2_t __ret;
43414  __ret = (uint64x2_t)(__p0 <= __p1);
43415  return __ret;
43416}
43417#else
43418__ai uint64x2_t vcleq_f64(float64x2_t __p0, float64x2_t __p1) {
43419  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43420  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43421  uint64x2_t __ret;
43422  __ret = (uint64x2_t)(__rev0 <= __rev1);
43423  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43424  return __ret;
43425}
43426#endif
43427
43428#ifdef __LITTLE_ENDIAN__
43429__ai uint64x2_t vcleq_s64(int64x2_t __p0, int64x2_t __p1) {
43430  uint64x2_t __ret;
43431  __ret = (uint64x2_t)(__p0 <= __p1);
43432  return __ret;
43433}
43434#else
43435__ai uint64x2_t vcleq_s64(int64x2_t __p0, int64x2_t __p1) {
43436  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43437  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43438  uint64x2_t __ret;
43439  __ret = (uint64x2_t)(__rev0 <= __rev1);
43440  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43441  return __ret;
43442}
43443#endif
43444
43445#ifdef __LITTLE_ENDIAN__
43446__ai uint64x1_t vcle_u64(uint64x1_t __p0, uint64x1_t __p1) {
43447  uint64x1_t __ret;
43448  __ret = (uint64x1_t)(__p0 <= __p1);
43449  return __ret;
43450}
43451#else
43452__ai uint64x1_t vcle_u64(uint64x1_t __p0, uint64x1_t __p1) {
43453  uint64x1_t __ret;
43454  __ret = (uint64x1_t)(__p0 <= __p1);
43455  return __ret;
43456}
43457#endif
43458
43459#ifdef __LITTLE_ENDIAN__
43460__ai uint64x1_t vcle_f64(float64x1_t __p0, float64x1_t __p1) {
43461  uint64x1_t __ret;
43462  __ret = (uint64x1_t)(__p0 <= __p1);
43463  return __ret;
43464}
43465#else
43466__ai uint64x1_t vcle_f64(float64x1_t __p0, float64x1_t __p1) {
43467  uint64x1_t __ret;
43468  __ret = (uint64x1_t)(__p0 <= __p1);
43469  return __ret;
43470}
43471#endif
43472
43473#ifdef __LITTLE_ENDIAN__
43474__ai uint64x1_t vcle_s64(int64x1_t __p0, int64x1_t __p1) {
43475  uint64x1_t __ret;
43476  __ret = (uint64x1_t)(__p0 <= __p1);
43477  return __ret;
43478}
43479#else
43480__ai uint64x1_t vcle_s64(int64x1_t __p0, int64x1_t __p1) {
43481  uint64x1_t __ret;
43482  __ret = (uint64x1_t)(__p0 <= __p1);
43483  return __ret;
43484}
43485#endif
43486
43487#ifdef __LITTLE_ENDIAN__
43488__ai uint64_t vcled_u64(uint64_t __p0, uint64_t __p1) {
43489  uint64_t __ret;
43490  __ret = (uint64_t) __builtin_neon_vcled_u64(__p0, __p1);
43491  return __ret;
43492}
43493#else
43494__ai uint64_t vcled_u64(uint64_t __p0, uint64_t __p1) {
43495  uint64_t __ret;
43496  __ret = (uint64_t) __builtin_neon_vcled_u64(__p0, __p1);
43497  return __ret;
43498}
43499#endif
43500
43501#ifdef __LITTLE_ENDIAN__
43502__ai int64_t vcled_s64(int64_t __p0, int64_t __p1) {
43503  int64_t __ret;
43504  __ret = (int64_t) __builtin_neon_vcled_s64(__p0, __p1);
43505  return __ret;
43506}
43507#else
43508__ai int64_t vcled_s64(int64_t __p0, int64_t __p1) {
43509  int64_t __ret;
43510  __ret = (int64_t) __builtin_neon_vcled_s64(__p0, __p1);
43511  return __ret;
43512}
43513#endif
43514
43515#ifdef __LITTLE_ENDIAN__
43516__ai uint64_t vcled_f64(float64_t __p0, float64_t __p1) {
43517  uint64_t __ret;
43518  __ret = (uint64_t) __builtin_neon_vcled_f64(__p0, __p1);
43519  return __ret;
43520}
43521#else
43522__ai uint64_t vcled_f64(float64_t __p0, float64_t __p1) {
43523  uint64_t __ret;
43524  __ret = (uint64_t) __builtin_neon_vcled_f64(__p0, __p1);
43525  return __ret;
43526}
43527#endif
43528
43529#ifdef __LITTLE_ENDIAN__
43530__ai uint32_t vcles_f32(float32_t __p0, float32_t __p1) {
43531  uint32_t __ret;
43532  __ret = (uint32_t) __builtin_neon_vcles_f32(__p0, __p1);
43533  return __ret;
43534}
43535#else
43536__ai uint32_t vcles_f32(float32_t __p0, float32_t __p1) {
43537  uint32_t __ret;
43538  __ret = (uint32_t) __builtin_neon_vcles_f32(__p0, __p1);
43539  return __ret;
43540}
43541#endif
43542
43543#ifdef __LITTLE_ENDIAN__
43544__ai uint8x16_t vclezq_s8(int8x16_t __p0) {
43545  uint8x16_t __ret;
43546  __ret = (uint8x16_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 48);
43547  return __ret;
43548}
43549#else
43550__ai uint8x16_t vclezq_s8(int8x16_t __p0) {
43551  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43552  uint8x16_t __ret;
43553  __ret = (uint8x16_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 48);
43554  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43555  return __ret;
43556}
43557#endif
43558
43559#ifdef __LITTLE_ENDIAN__
43560__ai uint64x2_t vclezq_f64(float64x2_t __p0) {
43561  uint64x2_t __ret;
43562  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 51);
43563  return __ret;
43564}
43565#else
43566__ai uint64x2_t vclezq_f64(float64x2_t __p0) {
43567  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43568  uint64x2_t __ret;
43569  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 51);
43570  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43571  return __ret;
43572}
43573#endif
43574
43575#ifdef __LITTLE_ENDIAN__
43576__ai uint32x4_t vclezq_f32(float32x4_t __p0) {
43577  uint32x4_t __ret;
43578  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 50);
43579  return __ret;
43580}
43581#else
43582__ai uint32x4_t vclezq_f32(float32x4_t __p0) {
43583  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43584  uint32x4_t __ret;
43585  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 50);
43586  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43587  return __ret;
43588}
43589#endif
43590
43591#ifdef __LITTLE_ENDIAN__
43592__ai uint32x4_t vclezq_s32(int32x4_t __p0) {
43593  uint32x4_t __ret;
43594  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 50);
43595  return __ret;
43596}
43597#else
43598__ai uint32x4_t vclezq_s32(int32x4_t __p0) {
43599  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43600  uint32x4_t __ret;
43601  __ret = (uint32x4_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 50);
43602  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43603  return __ret;
43604}
43605#endif
43606
43607#ifdef __LITTLE_ENDIAN__
43608__ai uint64x2_t vclezq_s64(int64x2_t __p0) {
43609  uint64x2_t __ret;
43610  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 51);
43611  return __ret;
43612}
43613#else
43614__ai uint64x2_t vclezq_s64(int64x2_t __p0) {
43615  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43616  uint64x2_t __ret;
43617  __ret = (uint64x2_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 51);
43618  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43619  return __ret;
43620}
43621#endif
43622
43623#ifdef __LITTLE_ENDIAN__
43624__ai uint16x8_t vclezq_s16(int16x8_t __p0) {
43625  uint16x8_t __ret;
43626  __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__p0, 49);
43627  return __ret;
43628}
43629#else
43630__ai uint16x8_t vclezq_s16(int16x8_t __p0) {
43631  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43632  uint16x8_t __ret;
43633  __ret = (uint16x8_t) __builtin_neon_vclezq_v((int8x16_t)__rev0, 49);
43634  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43635  return __ret;
43636}
43637#endif
43638
43639#ifdef __LITTLE_ENDIAN__
43640__ai uint8x8_t vclez_s8(int8x8_t __p0) {
43641  uint8x8_t __ret;
43642  __ret = (uint8x8_t) __builtin_neon_vclez_v((int8x8_t)__p0, 16);
43643  return __ret;
43644}
43645#else
43646__ai uint8x8_t vclez_s8(int8x8_t __p0) {
43647  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
43648  uint8x8_t __ret;
43649  __ret = (uint8x8_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 16);
43650  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
43651  return __ret;
43652}
43653#endif
43654
43655#ifdef __LITTLE_ENDIAN__
43656__ai uint64x1_t vclez_f64(float64x1_t __p0) {
43657  uint64x1_t __ret;
43658  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
43659  return __ret;
43660}
43661#else
43662__ai uint64x1_t vclez_f64(float64x1_t __p0) {
43663  uint64x1_t __ret;
43664  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
43665  return __ret;
43666}
43667#endif
43668
43669#ifdef __LITTLE_ENDIAN__
43670__ai uint32x2_t vclez_f32(float32x2_t __p0) {
43671  uint32x2_t __ret;
43672  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__p0, 18);
43673  return __ret;
43674}
43675#else
43676__ai uint32x2_t vclez_f32(float32x2_t __p0) {
43677  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43678  uint32x2_t __ret;
43679  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 18);
43680  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43681  return __ret;
43682}
43683#endif
43684
43685#ifdef __LITTLE_ENDIAN__
43686__ai uint32x2_t vclez_s32(int32x2_t __p0) {
43687  uint32x2_t __ret;
43688  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__p0, 18);
43689  return __ret;
43690}
43691#else
43692__ai uint32x2_t vclez_s32(int32x2_t __p0) {
43693  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43694  uint32x2_t __ret;
43695  __ret = (uint32x2_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 18);
43696  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43697  return __ret;
43698}
43699#endif
43700
43701#ifdef __LITTLE_ENDIAN__
43702__ai uint64x1_t vclez_s64(int64x1_t __p0) {
43703  uint64x1_t __ret;
43704  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
43705  return __ret;
43706}
43707#else
43708__ai uint64x1_t vclez_s64(int64x1_t __p0) {
43709  uint64x1_t __ret;
43710  __ret = (uint64x1_t) __builtin_neon_vclez_v((int8x8_t)__p0, 19);
43711  return __ret;
43712}
43713#endif
43714
43715#ifdef __LITTLE_ENDIAN__
43716__ai uint16x4_t vclez_s16(int16x4_t __p0) {
43717  uint16x4_t __ret;
43718  __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__p0, 17);
43719  return __ret;
43720}
43721#else
43722__ai uint16x4_t vclez_s16(int16x4_t __p0) {
43723  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43724  uint16x4_t __ret;
43725  __ret = (uint16x4_t) __builtin_neon_vclez_v((int8x8_t)__rev0, 17);
43726  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43727  return __ret;
43728}
43729#endif
43730
43731#ifdef __LITTLE_ENDIAN__
43732__ai int64_t vclezd_s64(int64_t __p0) {
43733  int64_t __ret;
43734  __ret = (int64_t) __builtin_neon_vclezd_s64(__p0);
43735  return __ret;
43736}
43737#else
43738__ai int64_t vclezd_s64(int64_t __p0) {
43739  int64_t __ret;
43740  __ret = (int64_t) __builtin_neon_vclezd_s64(__p0);
43741  return __ret;
43742}
43743#endif
43744
43745#ifdef __LITTLE_ENDIAN__
43746__ai uint64_t vclezd_f64(float64_t __p0) {
43747  uint64_t __ret;
43748  __ret = (uint64_t) __builtin_neon_vclezd_f64(__p0);
43749  return __ret;
43750}
43751#else
43752__ai uint64_t vclezd_f64(float64_t __p0) {
43753  uint64_t __ret;
43754  __ret = (uint64_t) __builtin_neon_vclezd_f64(__p0);
43755  return __ret;
43756}
43757#endif
43758
43759#ifdef __LITTLE_ENDIAN__
43760__ai uint32_t vclezs_f32(float32_t __p0) {
43761  uint32_t __ret;
43762  __ret = (uint32_t) __builtin_neon_vclezs_f32(__p0);
43763  return __ret;
43764}
43765#else
43766__ai uint32_t vclezs_f32(float32_t __p0) {
43767  uint32_t __ret;
43768  __ret = (uint32_t) __builtin_neon_vclezs_f32(__p0);
43769  return __ret;
43770}
43771#endif
43772
43773#ifdef __LITTLE_ENDIAN__
43774__ai uint64x2_t vcltq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43775  uint64x2_t __ret;
43776  __ret = (uint64x2_t)(__p0 < __p1);
43777  return __ret;
43778}
43779#else
43780__ai uint64x2_t vcltq_u64(uint64x2_t __p0, uint64x2_t __p1) {
43781  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43782  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43783  uint64x2_t __ret;
43784  __ret = (uint64x2_t)(__rev0 < __rev1);
43785  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43786  return __ret;
43787}
43788#endif
43789
43790#ifdef __LITTLE_ENDIAN__
43791__ai uint64x2_t vcltq_f64(float64x2_t __p0, float64x2_t __p1) {
43792  uint64x2_t __ret;
43793  __ret = (uint64x2_t)(__p0 < __p1);
43794  return __ret;
43795}
43796#else
43797__ai uint64x2_t vcltq_f64(float64x2_t __p0, float64x2_t __p1) {
43798  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43799  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43800  uint64x2_t __ret;
43801  __ret = (uint64x2_t)(__rev0 < __rev1);
43802  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43803  return __ret;
43804}
43805#endif
43806
43807#ifdef __LITTLE_ENDIAN__
43808__ai uint64x2_t vcltq_s64(int64x2_t __p0, int64x2_t __p1) {
43809  uint64x2_t __ret;
43810  __ret = (uint64x2_t)(__p0 < __p1);
43811  return __ret;
43812}
43813#else
43814__ai uint64x2_t vcltq_s64(int64x2_t __p0, int64x2_t __p1) {
43815  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43816  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
43817  uint64x2_t __ret;
43818  __ret = (uint64x2_t)(__rev0 < __rev1);
43819  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43820  return __ret;
43821}
43822#endif
43823
43824#ifdef __LITTLE_ENDIAN__
43825__ai uint64x1_t vclt_u64(uint64x1_t __p0, uint64x1_t __p1) {
43826  uint64x1_t __ret;
43827  __ret = (uint64x1_t)(__p0 < __p1);
43828  return __ret;
43829}
43830#else
43831__ai uint64x1_t vclt_u64(uint64x1_t __p0, uint64x1_t __p1) {
43832  uint64x1_t __ret;
43833  __ret = (uint64x1_t)(__p0 < __p1);
43834  return __ret;
43835}
43836#endif
43837
43838#ifdef __LITTLE_ENDIAN__
43839__ai uint64x1_t vclt_f64(float64x1_t __p0, float64x1_t __p1) {
43840  uint64x1_t __ret;
43841  __ret = (uint64x1_t)(__p0 < __p1);
43842  return __ret;
43843}
43844#else
43845__ai uint64x1_t vclt_f64(float64x1_t __p0, float64x1_t __p1) {
43846  uint64x1_t __ret;
43847  __ret = (uint64x1_t)(__p0 < __p1);
43848  return __ret;
43849}
43850#endif
43851
43852#ifdef __LITTLE_ENDIAN__
43853__ai uint64x1_t vclt_s64(int64x1_t __p0, int64x1_t __p1) {
43854  uint64x1_t __ret;
43855  __ret = (uint64x1_t)(__p0 < __p1);
43856  return __ret;
43857}
43858#else
43859__ai uint64x1_t vclt_s64(int64x1_t __p0, int64x1_t __p1) {
43860  uint64x1_t __ret;
43861  __ret = (uint64x1_t)(__p0 < __p1);
43862  return __ret;
43863}
43864#endif
43865
43866#ifdef __LITTLE_ENDIAN__
43867__ai uint64_t vcltd_u64(uint64_t __p0, uint64_t __p1) {
43868  uint64_t __ret;
43869  __ret = (uint64_t) __builtin_neon_vcltd_u64(__p0, __p1);
43870  return __ret;
43871}
43872#else
43873__ai uint64_t vcltd_u64(uint64_t __p0, uint64_t __p1) {
43874  uint64_t __ret;
43875  __ret = (uint64_t) __builtin_neon_vcltd_u64(__p0, __p1);
43876  return __ret;
43877}
43878#endif
43879
43880#ifdef __LITTLE_ENDIAN__
43881__ai int64_t vcltd_s64(int64_t __p0, int64_t __p1) {
43882  int64_t __ret;
43883  __ret = (int64_t) __builtin_neon_vcltd_s64(__p0, __p1);
43884  return __ret;
43885}
43886#else
43887__ai int64_t vcltd_s64(int64_t __p0, int64_t __p1) {
43888  int64_t __ret;
43889  __ret = (int64_t) __builtin_neon_vcltd_s64(__p0, __p1);
43890  return __ret;
43891}
43892#endif
43893
43894#ifdef __LITTLE_ENDIAN__
43895__ai uint64_t vcltd_f64(float64_t __p0, float64_t __p1) {
43896  uint64_t __ret;
43897  __ret = (uint64_t) __builtin_neon_vcltd_f64(__p0, __p1);
43898  return __ret;
43899}
43900#else
43901__ai uint64_t vcltd_f64(float64_t __p0, float64_t __p1) {
43902  uint64_t __ret;
43903  __ret = (uint64_t) __builtin_neon_vcltd_f64(__p0, __p1);
43904  return __ret;
43905}
43906#endif
43907
43908#ifdef __LITTLE_ENDIAN__
43909__ai uint32_t vclts_f32(float32_t __p0, float32_t __p1) {
43910  uint32_t __ret;
43911  __ret = (uint32_t) __builtin_neon_vclts_f32(__p0, __p1);
43912  return __ret;
43913}
43914#else
43915__ai uint32_t vclts_f32(float32_t __p0, float32_t __p1) {
43916  uint32_t __ret;
43917  __ret = (uint32_t) __builtin_neon_vclts_f32(__p0, __p1);
43918  return __ret;
43919}
43920#endif
43921
43922#ifdef __LITTLE_ENDIAN__
43923__ai uint8x16_t vcltzq_s8(int8x16_t __p0) {
43924  uint8x16_t __ret;
43925  __ret = (uint8x16_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 48);
43926  return __ret;
43927}
43928#else
43929__ai uint8x16_t vcltzq_s8(int8x16_t __p0) {
43930  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43931  uint8x16_t __ret;
43932  __ret = (uint8x16_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 48);
43933  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
43934  return __ret;
43935}
43936#endif
43937
43938#ifdef __LITTLE_ENDIAN__
43939__ai uint64x2_t vcltzq_f64(float64x2_t __p0) {
43940  uint64x2_t __ret;
43941  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 51);
43942  return __ret;
43943}
43944#else
43945__ai uint64x2_t vcltzq_f64(float64x2_t __p0) {
43946  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43947  uint64x2_t __ret;
43948  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 51);
43949  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43950  return __ret;
43951}
43952#endif
43953
43954#ifdef __LITTLE_ENDIAN__
43955__ai uint32x4_t vcltzq_f32(float32x4_t __p0) {
43956  uint32x4_t __ret;
43957  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 50);
43958  return __ret;
43959}
43960#else
43961__ai uint32x4_t vcltzq_f32(float32x4_t __p0) {
43962  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43963  uint32x4_t __ret;
43964  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 50);
43965  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43966  return __ret;
43967}
43968#endif
43969
43970#ifdef __LITTLE_ENDIAN__
43971__ai uint32x4_t vcltzq_s32(int32x4_t __p0) {
43972  uint32x4_t __ret;
43973  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 50);
43974  return __ret;
43975}
43976#else
43977__ai uint32x4_t vcltzq_s32(int32x4_t __p0) {
43978  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
43979  uint32x4_t __ret;
43980  __ret = (uint32x4_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 50);
43981  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
43982  return __ret;
43983}
43984#endif
43985
43986#ifdef __LITTLE_ENDIAN__
43987__ai uint64x2_t vcltzq_s64(int64x2_t __p0) {
43988  uint64x2_t __ret;
43989  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 51);
43990  return __ret;
43991}
43992#else
43993__ai uint64x2_t vcltzq_s64(int64x2_t __p0) {
43994  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
43995  uint64x2_t __ret;
43996  __ret = (uint64x2_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 51);
43997  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
43998  return __ret;
43999}
44000#endif
44001
44002#ifdef __LITTLE_ENDIAN__
44003__ai uint16x8_t vcltzq_s16(int16x8_t __p0) {
44004  uint16x8_t __ret;
44005  __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__p0, 49);
44006  return __ret;
44007}
44008#else
44009__ai uint16x8_t vcltzq_s16(int16x8_t __p0) {
44010  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44011  uint16x8_t __ret;
44012  __ret = (uint16x8_t) __builtin_neon_vcltzq_v((int8x16_t)__rev0, 49);
44013  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44014  return __ret;
44015}
44016#endif
44017
44018#ifdef __LITTLE_ENDIAN__
44019__ai uint8x8_t vcltz_s8(int8x8_t __p0) {
44020  uint8x8_t __ret;
44021  __ret = (uint8x8_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 16);
44022  return __ret;
44023}
44024#else
44025__ai uint8x8_t vcltz_s8(int8x8_t __p0) {
44026  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
44027  uint8x8_t __ret;
44028  __ret = (uint8x8_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 16);
44029  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
44030  return __ret;
44031}
44032#endif
44033
44034#ifdef __LITTLE_ENDIAN__
44035__ai uint64x1_t vcltz_f64(float64x1_t __p0) {
44036  uint64x1_t __ret;
44037  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
44038  return __ret;
44039}
44040#else
44041__ai uint64x1_t vcltz_f64(float64x1_t __p0) {
44042  uint64x1_t __ret;
44043  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
44044  return __ret;
44045}
44046#endif
44047
44048#ifdef __LITTLE_ENDIAN__
44049__ai uint32x2_t vcltz_f32(float32x2_t __p0) {
44050  uint32x2_t __ret;
44051  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 18);
44052  return __ret;
44053}
44054#else
44055__ai uint32x2_t vcltz_f32(float32x2_t __p0) {
44056  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44057  uint32x2_t __ret;
44058  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 18);
44059  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44060  return __ret;
44061}
44062#endif
44063
44064#ifdef __LITTLE_ENDIAN__
44065__ai uint32x2_t vcltz_s32(int32x2_t __p0) {
44066  uint32x2_t __ret;
44067  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 18);
44068  return __ret;
44069}
44070#else
44071__ai uint32x2_t vcltz_s32(int32x2_t __p0) {
44072  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
44073  uint32x2_t __ret;
44074  __ret = (uint32x2_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 18);
44075  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44076  return __ret;
44077}
44078#endif
44079
44080#ifdef __LITTLE_ENDIAN__
44081__ai uint64x1_t vcltz_s64(int64x1_t __p0) {
44082  uint64x1_t __ret;
44083  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
44084  return __ret;
44085}
44086#else
44087__ai uint64x1_t vcltz_s64(int64x1_t __p0) {
44088  uint64x1_t __ret;
44089  __ret = (uint64x1_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 19);
44090  return __ret;
44091}
44092#endif
44093
44094#ifdef __LITTLE_ENDIAN__
44095__ai uint16x4_t vcltz_s16(int16x4_t __p0) {
44096  uint16x4_t __ret;
44097  __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__p0, 17);
44098  return __ret;
44099}
44100#else
44101__ai uint16x4_t vcltz_s16(int16x4_t __p0) {
44102  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
44103  uint16x4_t __ret;
44104  __ret = (uint16x4_t) __builtin_neon_vcltz_v((int8x8_t)__rev0, 17);
44105  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
44106  return __ret;
44107}
44108#endif
44109
44110#ifdef __LITTLE_ENDIAN__
44111__ai int64_t vcltzd_s64(int64_t __p0) {
44112  int64_t __ret;
44113  __ret = (int64_t) __builtin_neon_vcltzd_s64(__p0);
44114  return __ret;
44115}
44116#else
44117__ai int64_t vcltzd_s64(int64_t __p0) {
44118  int64_t __ret;
44119  __ret = (int64_t) __builtin_neon_vcltzd_s64(__p0);
44120  return __ret;
44121}
44122#endif
44123
44124#ifdef __LITTLE_ENDIAN__
44125__ai uint64_t vcltzd_f64(float64_t __p0) {
44126  uint64_t __ret;
44127  __ret = (uint64_t) __builtin_neon_vcltzd_f64(__p0);
44128  return __ret;
44129}
44130#else
44131__ai uint64_t vcltzd_f64(float64_t __p0) {
44132  uint64_t __ret;
44133  __ret = (uint64_t) __builtin_neon_vcltzd_f64(__p0);
44134  return __ret;
44135}
44136#endif
44137
44138#ifdef __LITTLE_ENDIAN__
44139__ai uint32_t vcltzs_f32(float32_t __p0) {
44140  uint32_t __ret;
44141  __ret = (uint32_t) __builtin_neon_vcltzs_f32(__p0);
44142  return __ret;
44143}
44144#else
44145__ai uint32_t vcltzs_f32(float32_t __p0) {
44146  uint32_t __ret;
44147  __ret = (uint32_t) __builtin_neon_vcltzs_f32(__p0);
44148  return __ret;
44149}
44150#endif
44151
44152#ifdef __LITTLE_ENDIAN__
44153__ai poly64x2_t vcombine_p64(poly64x1_t __p0, poly64x1_t __p1) {
44154  poly64x2_t __ret;
44155  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
44156  return __ret;
44157}
44158#else
44159__ai poly64x2_t vcombine_p64(poly64x1_t __p0, poly64x1_t __p1) {
44160  poly64x2_t __ret;
44161  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
44162  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44163  return __ret;
44164}
44165#endif
44166
44167#ifdef __LITTLE_ENDIAN__
44168__ai float64x2_t vcombine_f64(float64x1_t __p0, float64x1_t __p1) {
44169  float64x2_t __ret;
44170  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
44171  return __ret;
44172}
44173#else
44174__ai float64x2_t vcombine_f64(float64x1_t __p0, float64x1_t __p1) {
44175  float64x2_t __ret;
44176  __ret = __builtin_shufflevector(__p0, __p1, 0, 1);
44177  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
44178  return __ret;
44179}
44180#endif
44181
44182#ifdef __LITTLE_ENDIAN__
44183#define vcopyq_lane_p8(__p0_0, __p1_0, __p2_0, __p3_0) __extension__ ({ \
44184  poly8x16_t __s0_0 = __p0_0; \
44185  poly8x8_t __s2_0 = __p2_0; \
44186  poly8x16_t __ret_0; \
44187  __ret_0 = vsetq_lane_p8(vget_lane_p8(__s2_0, __p3_0), __s0_0, __p1_0); \
44188  __ret_0; \
44189})
44190#else
44191#define vcopyq_lane_p8(__p0_1, __p1_1, __p2_1, __p3_1) __extension__ ({ \
44192  poly8x16_t __s0_1 = __p0_1; \
44193  poly8x8_t __s2_1 = __p2_1; \
44194  poly8x16_t __rev0_1;  __rev0_1 = __builtin_shufflevector(__s0_1, __s0_1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44195  poly8x8_t __rev2_1;  __rev2_1 = __builtin_shufflevector(__s2_1, __s2_1, 7, 6, 5, 4, 3, 2, 1, 0); \
44196  poly8x16_t __ret_1; \
44197  __ret_1 = __noswap_vsetq_lane_p8(__noswap_vget_lane_p8(__rev2_1, __p3_1), __rev0_1, __p1_1); \
44198  __ret_1 = __builtin_shufflevector(__ret_1, __ret_1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44199  __ret_1; \
44200})
44201#endif
44202
44203#ifdef __LITTLE_ENDIAN__
44204#define vcopyq_lane_p16(__p0_2, __p1_2, __p2_2, __p3_2) __extension__ ({ \
44205  poly16x8_t __s0_2 = __p0_2; \
44206  poly16x4_t __s2_2 = __p2_2; \
44207  poly16x8_t __ret_2; \
44208  __ret_2 = vsetq_lane_p16(vget_lane_p16(__s2_2, __p3_2), __s0_2, __p1_2); \
44209  __ret_2; \
44210})
44211#else
44212#define vcopyq_lane_p16(__p0_3, __p1_3, __p2_3, __p3_3) __extension__ ({ \
44213  poly16x8_t __s0_3 = __p0_3; \
44214  poly16x4_t __s2_3 = __p2_3; \
44215  poly16x8_t __rev0_3;  __rev0_3 = __builtin_shufflevector(__s0_3, __s0_3, 7, 6, 5, 4, 3, 2, 1, 0); \
44216  poly16x4_t __rev2_3;  __rev2_3 = __builtin_shufflevector(__s2_3, __s2_3, 3, 2, 1, 0); \
44217  poly16x8_t __ret_3; \
44218  __ret_3 = __noswap_vsetq_lane_p16(__noswap_vget_lane_p16(__rev2_3, __p3_3), __rev0_3, __p1_3); \
44219  __ret_3 = __builtin_shufflevector(__ret_3, __ret_3, 7, 6, 5, 4, 3, 2, 1, 0); \
44220  __ret_3; \
44221})
44222#endif
44223
44224#ifdef __LITTLE_ENDIAN__
44225#define vcopyq_lane_u8(__p0_4, __p1_4, __p2_4, __p3_4) __extension__ ({ \
44226  uint8x16_t __s0_4 = __p0_4; \
44227  uint8x8_t __s2_4 = __p2_4; \
44228  uint8x16_t __ret_4; \
44229  __ret_4 = vsetq_lane_u8(vget_lane_u8(__s2_4, __p3_4), __s0_4, __p1_4); \
44230  __ret_4; \
44231})
44232#else
44233#define vcopyq_lane_u8(__p0_5, __p1_5, __p2_5, __p3_5) __extension__ ({ \
44234  uint8x16_t __s0_5 = __p0_5; \
44235  uint8x8_t __s2_5 = __p2_5; \
44236  uint8x16_t __rev0_5;  __rev0_5 = __builtin_shufflevector(__s0_5, __s0_5, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44237  uint8x8_t __rev2_5;  __rev2_5 = __builtin_shufflevector(__s2_5, __s2_5, 7, 6, 5, 4, 3, 2, 1, 0); \
44238  uint8x16_t __ret_5; \
44239  __ret_5 = __noswap_vsetq_lane_u8(__noswap_vget_lane_u8(__rev2_5, __p3_5), __rev0_5, __p1_5); \
44240  __ret_5 = __builtin_shufflevector(__ret_5, __ret_5, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44241  __ret_5; \
44242})
44243#endif
44244
44245#ifdef __LITTLE_ENDIAN__
44246#define vcopyq_lane_u32(__p0_6, __p1_6, __p2_6, __p3_6) __extension__ ({ \
44247  uint32x4_t __s0_6 = __p0_6; \
44248  uint32x2_t __s2_6 = __p2_6; \
44249  uint32x4_t __ret_6; \
44250  __ret_6 = vsetq_lane_u32(vget_lane_u32(__s2_6, __p3_6), __s0_6, __p1_6); \
44251  __ret_6; \
44252})
44253#else
44254#define vcopyq_lane_u32(__p0_7, __p1_7, __p2_7, __p3_7) __extension__ ({ \
44255  uint32x4_t __s0_7 = __p0_7; \
44256  uint32x2_t __s2_7 = __p2_7; \
44257  uint32x4_t __rev0_7;  __rev0_7 = __builtin_shufflevector(__s0_7, __s0_7, 3, 2, 1, 0); \
44258  uint32x2_t __rev2_7;  __rev2_7 = __builtin_shufflevector(__s2_7, __s2_7, 1, 0); \
44259  uint32x4_t __ret_7; \
44260  __ret_7 = __noswap_vsetq_lane_u32(__noswap_vget_lane_u32(__rev2_7, __p3_7), __rev0_7, __p1_7); \
44261  __ret_7 = __builtin_shufflevector(__ret_7, __ret_7, 3, 2, 1, 0); \
44262  __ret_7; \
44263})
44264#endif
44265
44266#ifdef __LITTLE_ENDIAN__
44267#define vcopyq_lane_u64(__p0_8, __p1_8, __p2_8, __p3_8) __extension__ ({ \
44268  uint64x2_t __s0_8 = __p0_8; \
44269  uint64x1_t __s2_8 = __p2_8; \
44270  uint64x2_t __ret_8; \
44271  __ret_8 = vsetq_lane_u64(vget_lane_u64(__s2_8, __p3_8), __s0_8, __p1_8); \
44272  __ret_8; \
44273})
44274#else
44275#define vcopyq_lane_u64(__p0_9, __p1_9, __p2_9, __p3_9) __extension__ ({ \
44276  uint64x2_t __s0_9 = __p0_9; \
44277  uint64x1_t __s2_9 = __p2_9; \
44278  uint64x2_t __rev0_9;  __rev0_9 = __builtin_shufflevector(__s0_9, __s0_9, 1, 0); \
44279  uint64x2_t __ret_9; \
44280  __ret_9 = __noswap_vsetq_lane_u64(__noswap_vget_lane_u64(__s2_9, __p3_9), __rev0_9, __p1_9); \
44281  __ret_9 = __builtin_shufflevector(__ret_9, __ret_9, 1, 0); \
44282  __ret_9; \
44283})
44284#endif
44285
44286#ifdef __LITTLE_ENDIAN__
44287#define vcopyq_lane_u16(__p0_10, __p1_10, __p2_10, __p3_10) __extension__ ({ \
44288  uint16x8_t __s0_10 = __p0_10; \
44289  uint16x4_t __s2_10 = __p2_10; \
44290  uint16x8_t __ret_10; \
44291  __ret_10 = vsetq_lane_u16(vget_lane_u16(__s2_10, __p3_10), __s0_10, __p1_10); \
44292  __ret_10; \
44293})
44294#else
44295#define vcopyq_lane_u16(__p0_11, __p1_11, __p2_11, __p3_11) __extension__ ({ \
44296  uint16x8_t __s0_11 = __p0_11; \
44297  uint16x4_t __s2_11 = __p2_11; \
44298  uint16x8_t __rev0_11;  __rev0_11 = __builtin_shufflevector(__s0_11, __s0_11, 7, 6, 5, 4, 3, 2, 1, 0); \
44299  uint16x4_t __rev2_11;  __rev2_11 = __builtin_shufflevector(__s2_11, __s2_11, 3, 2, 1, 0); \
44300  uint16x8_t __ret_11; \
44301  __ret_11 = __noswap_vsetq_lane_u16(__noswap_vget_lane_u16(__rev2_11, __p3_11), __rev0_11, __p1_11); \
44302  __ret_11 = __builtin_shufflevector(__ret_11, __ret_11, 7, 6, 5, 4, 3, 2, 1, 0); \
44303  __ret_11; \
44304})
44305#endif
44306
44307#ifdef __LITTLE_ENDIAN__
44308#define vcopyq_lane_s8(__p0_12, __p1_12, __p2_12, __p3_12) __extension__ ({ \
44309  int8x16_t __s0_12 = __p0_12; \
44310  int8x8_t __s2_12 = __p2_12; \
44311  int8x16_t __ret_12; \
44312  __ret_12 = vsetq_lane_s8(vget_lane_s8(__s2_12, __p3_12), __s0_12, __p1_12); \
44313  __ret_12; \
44314})
44315#else
44316#define vcopyq_lane_s8(__p0_13, __p1_13, __p2_13, __p3_13) __extension__ ({ \
44317  int8x16_t __s0_13 = __p0_13; \
44318  int8x8_t __s2_13 = __p2_13; \
44319  int8x16_t __rev0_13;  __rev0_13 = __builtin_shufflevector(__s0_13, __s0_13, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44320  int8x8_t __rev2_13;  __rev2_13 = __builtin_shufflevector(__s2_13, __s2_13, 7, 6, 5, 4, 3, 2, 1, 0); \
44321  int8x16_t __ret_13; \
44322  __ret_13 = __noswap_vsetq_lane_s8(__noswap_vget_lane_s8(__rev2_13, __p3_13), __rev0_13, __p1_13); \
44323  __ret_13 = __builtin_shufflevector(__ret_13, __ret_13, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44324  __ret_13; \
44325})
44326#endif
44327
44328#ifdef __LITTLE_ENDIAN__
44329#define vcopyq_lane_f32(__p0_14, __p1_14, __p2_14, __p3_14) __extension__ ({ \
44330  float32x4_t __s0_14 = __p0_14; \
44331  float32x2_t __s2_14 = __p2_14; \
44332  float32x4_t __ret_14; \
44333  __ret_14 = vsetq_lane_f32(vget_lane_f32(__s2_14, __p3_14), __s0_14, __p1_14); \
44334  __ret_14; \
44335})
44336#else
44337#define vcopyq_lane_f32(__p0_15, __p1_15, __p2_15, __p3_15) __extension__ ({ \
44338  float32x4_t __s0_15 = __p0_15; \
44339  float32x2_t __s2_15 = __p2_15; \
44340  float32x4_t __rev0_15;  __rev0_15 = __builtin_shufflevector(__s0_15, __s0_15, 3, 2, 1, 0); \
44341  float32x2_t __rev2_15;  __rev2_15 = __builtin_shufflevector(__s2_15, __s2_15, 1, 0); \
44342  float32x4_t __ret_15; \
44343  __ret_15 = __noswap_vsetq_lane_f32(__noswap_vget_lane_f32(__rev2_15, __p3_15), __rev0_15, __p1_15); \
44344  __ret_15 = __builtin_shufflevector(__ret_15, __ret_15, 3, 2, 1, 0); \
44345  __ret_15; \
44346})
44347#endif
44348
44349#ifdef __LITTLE_ENDIAN__
44350#define vcopyq_lane_s32(__p0_16, __p1_16, __p2_16, __p3_16) __extension__ ({ \
44351  int32x4_t __s0_16 = __p0_16; \
44352  int32x2_t __s2_16 = __p2_16; \
44353  int32x4_t __ret_16; \
44354  __ret_16 = vsetq_lane_s32(vget_lane_s32(__s2_16, __p3_16), __s0_16, __p1_16); \
44355  __ret_16; \
44356})
44357#else
44358#define vcopyq_lane_s32(__p0_17, __p1_17, __p2_17, __p3_17) __extension__ ({ \
44359  int32x4_t __s0_17 = __p0_17; \
44360  int32x2_t __s2_17 = __p2_17; \
44361  int32x4_t __rev0_17;  __rev0_17 = __builtin_shufflevector(__s0_17, __s0_17, 3, 2, 1, 0); \
44362  int32x2_t __rev2_17;  __rev2_17 = __builtin_shufflevector(__s2_17, __s2_17, 1, 0); \
44363  int32x4_t __ret_17; \
44364  __ret_17 = __noswap_vsetq_lane_s32(__noswap_vget_lane_s32(__rev2_17, __p3_17), __rev0_17, __p1_17); \
44365  __ret_17 = __builtin_shufflevector(__ret_17, __ret_17, 3, 2, 1, 0); \
44366  __ret_17; \
44367})
44368#endif
44369
44370#ifdef __LITTLE_ENDIAN__
44371#define vcopyq_lane_s64(__p0_18, __p1_18, __p2_18, __p3_18) __extension__ ({ \
44372  int64x2_t __s0_18 = __p0_18; \
44373  int64x1_t __s2_18 = __p2_18; \
44374  int64x2_t __ret_18; \
44375  __ret_18 = vsetq_lane_s64(vget_lane_s64(__s2_18, __p3_18), __s0_18, __p1_18); \
44376  __ret_18; \
44377})
44378#else
44379#define vcopyq_lane_s64(__p0_19, __p1_19, __p2_19, __p3_19) __extension__ ({ \
44380  int64x2_t __s0_19 = __p0_19; \
44381  int64x1_t __s2_19 = __p2_19; \
44382  int64x2_t __rev0_19;  __rev0_19 = __builtin_shufflevector(__s0_19, __s0_19, 1, 0); \
44383  int64x2_t __ret_19; \
44384  __ret_19 = __noswap_vsetq_lane_s64(__noswap_vget_lane_s64(__s2_19, __p3_19), __rev0_19, __p1_19); \
44385  __ret_19 = __builtin_shufflevector(__ret_19, __ret_19, 1, 0); \
44386  __ret_19; \
44387})
44388#endif
44389
44390#ifdef __LITTLE_ENDIAN__
44391#define vcopyq_lane_s16(__p0_20, __p1_20, __p2_20, __p3_20) __extension__ ({ \
44392  int16x8_t __s0_20 = __p0_20; \
44393  int16x4_t __s2_20 = __p2_20; \
44394  int16x8_t __ret_20; \
44395  __ret_20 = vsetq_lane_s16(vget_lane_s16(__s2_20, __p3_20), __s0_20, __p1_20); \
44396  __ret_20; \
44397})
44398#else
44399#define vcopyq_lane_s16(__p0_21, __p1_21, __p2_21, __p3_21) __extension__ ({ \
44400  int16x8_t __s0_21 = __p0_21; \
44401  int16x4_t __s2_21 = __p2_21; \
44402  int16x8_t __rev0_21;  __rev0_21 = __builtin_shufflevector(__s0_21, __s0_21, 7, 6, 5, 4, 3, 2, 1, 0); \
44403  int16x4_t __rev2_21;  __rev2_21 = __builtin_shufflevector(__s2_21, __s2_21, 3, 2, 1, 0); \
44404  int16x8_t __ret_21; \
44405  __ret_21 = __noswap_vsetq_lane_s16(__noswap_vget_lane_s16(__rev2_21, __p3_21), __rev0_21, __p1_21); \
44406  __ret_21 = __builtin_shufflevector(__ret_21, __ret_21, 7, 6, 5, 4, 3, 2, 1, 0); \
44407  __ret_21; \
44408})
44409#endif
44410
44411#ifdef __LITTLE_ENDIAN__
44412#define vcopy_lane_p8(__p0_22, __p1_22, __p2_22, __p3_22) __extension__ ({ \
44413  poly8x8_t __s0_22 = __p0_22; \
44414  poly8x8_t __s2_22 = __p2_22; \
44415  poly8x8_t __ret_22; \
44416  __ret_22 = vset_lane_p8(vget_lane_p8(__s2_22, __p3_22), __s0_22, __p1_22); \
44417  __ret_22; \
44418})
44419#else
44420#define vcopy_lane_p8(__p0_23, __p1_23, __p2_23, __p3_23) __extension__ ({ \
44421  poly8x8_t __s0_23 = __p0_23; \
44422  poly8x8_t __s2_23 = __p2_23; \
44423  poly8x8_t __rev0_23;  __rev0_23 = __builtin_shufflevector(__s0_23, __s0_23, 7, 6, 5, 4, 3, 2, 1, 0); \
44424  poly8x8_t __rev2_23;  __rev2_23 = __builtin_shufflevector(__s2_23, __s2_23, 7, 6, 5, 4, 3, 2, 1, 0); \
44425  poly8x8_t __ret_23; \
44426  __ret_23 = __noswap_vset_lane_p8(__noswap_vget_lane_p8(__rev2_23, __p3_23), __rev0_23, __p1_23); \
44427  __ret_23 = __builtin_shufflevector(__ret_23, __ret_23, 7, 6, 5, 4, 3, 2, 1, 0); \
44428  __ret_23; \
44429})
44430#endif
44431
44432#ifdef __LITTLE_ENDIAN__
44433#define vcopy_lane_p16(__p0_24, __p1_24, __p2_24, __p3_24) __extension__ ({ \
44434  poly16x4_t __s0_24 = __p0_24; \
44435  poly16x4_t __s2_24 = __p2_24; \
44436  poly16x4_t __ret_24; \
44437  __ret_24 = vset_lane_p16(vget_lane_p16(__s2_24, __p3_24), __s0_24, __p1_24); \
44438  __ret_24; \
44439})
44440#else
44441#define vcopy_lane_p16(__p0_25, __p1_25, __p2_25, __p3_25) __extension__ ({ \
44442  poly16x4_t __s0_25 = __p0_25; \
44443  poly16x4_t __s2_25 = __p2_25; \
44444  poly16x4_t __rev0_25;  __rev0_25 = __builtin_shufflevector(__s0_25, __s0_25, 3, 2, 1, 0); \
44445  poly16x4_t __rev2_25;  __rev2_25 = __builtin_shufflevector(__s2_25, __s2_25, 3, 2, 1, 0); \
44446  poly16x4_t __ret_25; \
44447  __ret_25 = __noswap_vset_lane_p16(__noswap_vget_lane_p16(__rev2_25, __p3_25), __rev0_25, __p1_25); \
44448  __ret_25 = __builtin_shufflevector(__ret_25, __ret_25, 3, 2, 1, 0); \
44449  __ret_25; \
44450})
44451#endif
44452
44453#ifdef __LITTLE_ENDIAN__
44454#define vcopy_lane_u8(__p0_26, __p1_26, __p2_26, __p3_26) __extension__ ({ \
44455  uint8x8_t __s0_26 = __p0_26; \
44456  uint8x8_t __s2_26 = __p2_26; \
44457  uint8x8_t __ret_26; \
44458  __ret_26 = vset_lane_u8(vget_lane_u8(__s2_26, __p3_26), __s0_26, __p1_26); \
44459  __ret_26; \
44460})
44461#else
44462#define vcopy_lane_u8(__p0_27, __p1_27, __p2_27, __p3_27) __extension__ ({ \
44463  uint8x8_t __s0_27 = __p0_27; \
44464  uint8x8_t __s2_27 = __p2_27; \
44465  uint8x8_t __rev0_27;  __rev0_27 = __builtin_shufflevector(__s0_27, __s0_27, 7, 6, 5, 4, 3, 2, 1, 0); \
44466  uint8x8_t __rev2_27;  __rev2_27 = __builtin_shufflevector(__s2_27, __s2_27, 7, 6, 5, 4, 3, 2, 1, 0); \
44467  uint8x8_t __ret_27; \
44468  __ret_27 = __noswap_vset_lane_u8(__noswap_vget_lane_u8(__rev2_27, __p3_27), __rev0_27, __p1_27); \
44469  __ret_27 = __builtin_shufflevector(__ret_27, __ret_27, 7, 6, 5, 4, 3, 2, 1, 0); \
44470  __ret_27; \
44471})
44472#endif
44473
44474#ifdef __LITTLE_ENDIAN__
44475#define vcopy_lane_u32(__p0_28, __p1_28, __p2_28, __p3_28) __extension__ ({ \
44476  uint32x2_t __s0_28 = __p0_28; \
44477  uint32x2_t __s2_28 = __p2_28; \
44478  uint32x2_t __ret_28; \
44479  __ret_28 = vset_lane_u32(vget_lane_u32(__s2_28, __p3_28), __s0_28, __p1_28); \
44480  __ret_28; \
44481})
44482#else
44483#define vcopy_lane_u32(__p0_29, __p1_29, __p2_29, __p3_29) __extension__ ({ \
44484  uint32x2_t __s0_29 = __p0_29; \
44485  uint32x2_t __s2_29 = __p2_29; \
44486  uint32x2_t __rev0_29;  __rev0_29 = __builtin_shufflevector(__s0_29, __s0_29, 1, 0); \
44487  uint32x2_t __rev2_29;  __rev2_29 = __builtin_shufflevector(__s2_29, __s2_29, 1, 0); \
44488  uint32x2_t __ret_29; \
44489  __ret_29 = __noswap_vset_lane_u32(__noswap_vget_lane_u32(__rev2_29, __p3_29), __rev0_29, __p1_29); \
44490  __ret_29 = __builtin_shufflevector(__ret_29, __ret_29, 1, 0); \
44491  __ret_29; \
44492})
44493#endif
44494
44495#ifdef __LITTLE_ENDIAN__
44496#define vcopy_lane_u64(__p0_30, __p1_30, __p2_30, __p3_30) __extension__ ({ \
44497  uint64x1_t __s0_30 = __p0_30; \
44498  uint64x1_t __s2_30 = __p2_30; \
44499  uint64x1_t __ret_30; \
44500  __ret_30 = vset_lane_u64(vget_lane_u64(__s2_30, __p3_30), __s0_30, __p1_30); \
44501  __ret_30; \
44502})
44503#else
44504#define vcopy_lane_u64(__p0_31, __p1_31, __p2_31, __p3_31) __extension__ ({ \
44505  uint64x1_t __s0_31 = __p0_31; \
44506  uint64x1_t __s2_31 = __p2_31; \
44507  uint64x1_t __ret_31; \
44508  __ret_31 = __noswap_vset_lane_u64(__noswap_vget_lane_u64(__s2_31, __p3_31), __s0_31, __p1_31); \
44509  __ret_31; \
44510})
44511#endif
44512
44513#ifdef __LITTLE_ENDIAN__
44514#define vcopy_lane_u16(__p0_32, __p1_32, __p2_32, __p3_32) __extension__ ({ \
44515  uint16x4_t __s0_32 = __p0_32; \
44516  uint16x4_t __s2_32 = __p2_32; \
44517  uint16x4_t __ret_32; \
44518  __ret_32 = vset_lane_u16(vget_lane_u16(__s2_32, __p3_32), __s0_32, __p1_32); \
44519  __ret_32; \
44520})
44521#else
44522#define vcopy_lane_u16(__p0_33, __p1_33, __p2_33, __p3_33) __extension__ ({ \
44523  uint16x4_t __s0_33 = __p0_33; \
44524  uint16x4_t __s2_33 = __p2_33; \
44525  uint16x4_t __rev0_33;  __rev0_33 = __builtin_shufflevector(__s0_33, __s0_33, 3, 2, 1, 0); \
44526  uint16x4_t __rev2_33;  __rev2_33 = __builtin_shufflevector(__s2_33, __s2_33, 3, 2, 1, 0); \
44527  uint16x4_t __ret_33; \
44528  __ret_33 = __noswap_vset_lane_u16(__noswap_vget_lane_u16(__rev2_33, __p3_33), __rev0_33, __p1_33); \
44529  __ret_33 = __builtin_shufflevector(__ret_33, __ret_33, 3, 2, 1, 0); \
44530  __ret_33; \
44531})
44532#endif
44533
44534#ifdef __LITTLE_ENDIAN__
44535#define vcopy_lane_s8(__p0_34, __p1_34, __p2_34, __p3_34) __extension__ ({ \
44536  int8x8_t __s0_34 = __p0_34; \
44537  int8x8_t __s2_34 = __p2_34; \
44538  int8x8_t __ret_34; \
44539  __ret_34 = vset_lane_s8(vget_lane_s8(__s2_34, __p3_34), __s0_34, __p1_34); \
44540  __ret_34; \
44541})
44542#else
44543#define vcopy_lane_s8(__p0_35, __p1_35, __p2_35, __p3_35) __extension__ ({ \
44544  int8x8_t __s0_35 = __p0_35; \
44545  int8x8_t __s2_35 = __p2_35; \
44546  int8x8_t __rev0_35;  __rev0_35 = __builtin_shufflevector(__s0_35, __s0_35, 7, 6, 5, 4, 3, 2, 1, 0); \
44547  int8x8_t __rev2_35;  __rev2_35 = __builtin_shufflevector(__s2_35, __s2_35, 7, 6, 5, 4, 3, 2, 1, 0); \
44548  int8x8_t __ret_35; \
44549  __ret_35 = __noswap_vset_lane_s8(__noswap_vget_lane_s8(__rev2_35, __p3_35), __rev0_35, __p1_35); \
44550  __ret_35 = __builtin_shufflevector(__ret_35, __ret_35, 7, 6, 5, 4, 3, 2, 1, 0); \
44551  __ret_35; \
44552})
44553#endif
44554
44555#ifdef __LITTLE_ENDIAN__
44556#define vcopy_lane_f32(__p0_36, __p1_36, __p2_36, __p3_36) __extension__ ({ \
44557  float32x2_t __s0_36 = __p0_36; \
44558  float32x2_t __s2_36 = __p2_36; \
44559  float32x2_t __ret_36; \
44560  __ret_36 = vset_lane_f32(vget_lane_f32(__s2_36, __p3_36), __s0_36, __p1_36); \
44561  __ret_36; \
44562})
44563#else
44564#define vcopy_lane_f32(__p0_37, __p1_37, __p2_37, __p3_37) __extension__ ({ \
44565  float32x2_t __s0_37 = __p0_37; \
44566  float32x2_t __s2_37 = __p2_37; \
44567  float32x2_t __rev0_37;  __rev0_37 = __builtin_shufflevector(__s0_37, __s0_37, 1, 0); \
44568  float32x2_t __rev2_37;  __rev2_37 = __builtin_shufflevector(__s2_37, __s2_37, 1, 0); \
44569  float32x2_t __ret_37; \
44570  __ret_37 = __noswap_vset_lane_f32(__noswap_vget_lane_f32(__rev2_37, __p3_37), __rev0_37, __p1_37); \
44571  __ret_37 = __builtin_shufflevector(__ret_37, __ret_37, 1, 0); \
44572  __ret_37; \
44573})
44574#endif
44575
44576#ifdef __LITTLE_ENDIAN__
44577#define vcopy_lane_s32(__p0_38, __p1_38, __p2_38, __p3_38) __extension__ ({ \
44578  int32x2_t __s0_38 = __p0_38; \
44579  int32x2_t __s2_38 = __p2_38; \
44580  int32x2_t __ret_38; \
44581  __ret_38 = vset_lane_s32(vget_lane_s32(__s2_38, __p3_38), __s0_38, __p1_38); \
44582  __ret_38; \
44583})
44584#else
44585#define vcopy_lane_s32(__p0_39, __p1_39, __p2_39, __p3_39) __extension__ ({ \
44586  int32x2_t __s0_39 = __p0_39; \
44587  int32x2_t __s2_39 = __p2_39; \
44588  int32x2_t __rev0_39;  __rev0_39 = __builtin_shufflevector(__s0_39, __s0_39, 1, 0); \
44589  int32x2_t __rev2_39;  __rev2_39 = __builtin_shufflevector(__s2_39, __s2_39, 1, 0); \
44590  int32x2_t __ret_39; \
44591  __ret_39 = __noswap_vset_lane_s32(__noswap_vget_lane_s32(__rev2_39, __p3_39), __rev0_39, __p1_39); \
44592  __ret_39 = __builtin_shufflevector(__ret_39, __ret_39, 1, 0); \
44593  __ret_39; \
44594})
44595#endif
44596
44597#ifdef __LITTLE_ENDIAN__
44598#define vcopy_lane_s64(__p0_40, __p1_40, __p2_40, __p3_40) __extension__ ({ \
44599  int64x1_t __s0_40 = __p0_40; \
44600  int64x1_t __s2_40 = __p2_40; \
44601  int64x1_t __ret_40; \
44602  __ret_40 = vset_lane_s64(vget_lane_s64(__s2_40, __p3_40), __s0_40, __p1_40); \
44603  __ret_40; \
44604})
44605#else
44606#define vcopy_lane_s64(__p0_41, __p1_41, __p2_41, __p3_41) __extension__ ({ \
44607  int64x1_t __s0_41 = __p0_41; \
44608  int64x1_t __s2_41 = __p2_41; \
44609  int64x1_t __ret_41; \
44610  __ret_41 = __noswap_vset_lane_s64(__noswap_vget_lane_s64(__s2_41, __p3_41), __s0_41, __p1_41); \
44611  __ret_41; \
44612})
44613#endif
44614
44615#ifdef __LITTLE_ENDIAN__
44616#define vcopy_lane_s16(__p0_42, __p1_42, __p2_42, __p3_42) __extension__ ({ \
44617  int16x4_t __s0_42 = __p0_42; \
44618  int16x4_t __s2_42 = __p2_42; \
44619  int16x4_t __ret_42; \
44620  __ret_42 = vset_lane_s16(vget_lane_s16(__s2_42, __p3_42), __s0_42, __p1_42); \
44621  __ret_42; \
44622})
44623#else
44624#define vcopy_lane_s16(__p0_43, __p1_43, __p2_43, __p3_43) __extension__ ({ \
44625  int16x4_t __s0_43 = __p0_43; \
44626  int16x4_t __s2_43 = __p2_43; \
44627  int16x4_t __rev0_43;  __rev0_43 = __builtin_shufflevector(__s0_43, __s0_43, 3, 2, 1, 0); \
44628  int16x4_t __rev2_43;  __rev2_43 = __builtin_shufflevector(__s2_43, __s2_43, 3, 2, 1, 0); \
44629  int16x4_t __ret_43; \
44630  __ret_43 = __noswap_vset_lane_s16(__noswap_vget_lane_s16(__rev2_43, __p3_43), __rev0_43, __p1_43); \
44631  __ret_43 = __builtin_shufflevector(__ret_43, __ret_43, 3, 2, 1, 0); \
44632  __ret_43; \
44633})
44634#endif
44635
44636#ifdef __LITTLE_ENDIAN__
44637#define vcopyq_laneq_p8(__p0_44, __p1_44, __p2_44, __p3_44) __extension__ ({ \
44638  poly8x16_t __s0_44 = __p0_44; \
44639  poly8x16_t __s2_44 = __p2_44; \
44640  poly8x16_t __ret_44; \
44641  __ret_44 = vsetq_lane_p8(vgetq_lane_p8(__s2_44, __p3_44), __s0_44, __p1_44); \
44642  __ret_44; \
44643})
44644#else
44645#define vcopyq_laneq_p8(__p0_45, __p1_45, __p2_45, __p3_45) __extension__ ({ \
44646  poly8x16_t __s0_45 = __p0_45; \
44647  poly8x16_t __s2_45 = __p2_45; \
44648  poly8x16_t __rev0_45;  __rev0_45 = __builtin_shufflevector(__s0_45, __s0_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44649  poly8x16_t __rev2_45;  __rev2_45 = __builtin_shufflevector(__s2_45, __s2_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44650  poly8x16_t __ret_45; \
44651  __ret_45 = __noswap_vsetq_lane_p8(__noswap_vgetq_lane_p8(__rev2_45, __p3_45), __rev0_45, __p1_45); \
44652  __ret_45 = __builtin_shufflevector(__ret_45, __ret_45, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44653  __ret_45; \
44654})
44655#endif
44656
44657#ifdef __LITTLE_ENDIAN__
44658#define vcopyq_laneq_p16(__p0_46, __p1_46, __p2_46, __p3_46) __extension__ ({ \
44659  poly16x8_t __s0_46 = __p0_46; \
44660  poly16x8_t __s2_46 = __p2_46; \
44661  poly16x8_t __ret_46; \
44662  __ret_46 = vsetq_lane_p16(vgetq_lane_p16(__s2_46, __p3_46), __s0_46, __p1_46); \
44663  __ret_46; \
44664})
44665#else
44666#define vcopyq_laneq_p16(__p0_47, __p1_47, __p2_47, __p3_47) __extension__ ({ \
44667  poly16x8_t __s0_47 = __p0_47; \
44668  poly16x8_t __s2_47 = __p2_47; \
44669  poly16x8_t __rev0_47;  __rev0_47 = __builtin_shufflevector(__s0_47, __s0_47, 7, 6, 5, 4, 3, 2, 1, 0); \
44670  poly16x8_t __rev2_47;  __rev2_47 = __builtin_shufflevector(__s2_47, __s2_47, 7, 6, 5, 4, 3, 2, 1, 0); \
44671  poly16x8_t __ret_47; \
44672  __ret_47 = __noswap_vsetq_lane_p16(__noswap_vgetq_lane_p16(__rev2_47, __p3_47), __rev0_47, __p1_47); \
44673  __ret_47 = __builtin_shufflevector(__ret_47, __ret_47, 7, 6, 5, 4, 3, 2, 1, 0); \
44674  __ret_47; \
44675})
44676#endif
44677
44678#ifdef __LITTLE_ENDIAN__
44679#define vcopyq_laneq_u8(__p0_48, __p1_48, __p2_48, __p3_48) __extension__ ({ \
44680  uint8x16_t __s0_48 = __p0_48; \
44681  uint8x16_t __s2_48 = __p2_48; \
44682  uint8x16_t __ret_48; \
44683  __ret_48 = vsetq_lane_u8(vgetq_lane_u8(__s2_48, __p3_48), __s0_48, __p1_48); \
44684  __ret_48; \
44685})
44686#else
44687#define vcopyq_laneq_u8(__p0_49, __p1_49, __p2_49, __p3_49) __extension__ ({ \
44688  uint8x16_t __s0_49 = __p0_49; \
44689  uint8x16_t __s2_49 = __p2_49; \
44690  uint8x16_t __rev0_49;  __rev0_49 = __builtin_shufflevector(__s0_49, __s0_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44691  uint8x16_t __rev2_49;  __rev2_49 = __builtin_shufflevector(__s2_49, __s2_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44692  uint8x16_t __ret_49; \
44693  __ret_49 = __noswap_vsetq_lane_u8(__noswap_vgetq_lane_u8(__rev2_49, __p3_49), __rev0_49, __p1_49); \
44694  __ret_49 = __builtin_shufflevector(__ret_49, __ret_49, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44695  __ret_49; \
44696})
44697#endif
44698
44699#ifdef __LITTLE_ENDIAN__
44700#define vcopyq_laneq_u32(__p0_50, __p1_50, __p2_50, __p3_50) __extension__ ({ \
44701  uint32x4_t __s0_50 = __p0_50; \
44702  uint32x4_t __s2_50 = __p2_50; \
44703  uint32x4_t __ret_50; \
44704  __ret_50 = vsetq_lane_u32(vgetq_lane_u32(__s2_50, __p3_50), __s0_50, __p1_50); \
44705  __ret_50; \
44706})
44707#else
44708#define vcopyq_laneq_u32(__p0_51, __p1_51, __p2_51, __p3_51) __extension__ ({ \
44709  uint32x4_t __s0_51 = __p0_51; \
44710  uint32x4_t __s2_51 = __p2_51; \
44711  uint32x4_t __rev0_51;  __rev0_51 = __builtin_shufflevector(__s0_51, __s0_51, 3, 2, 1, 0); \
44712  uint32x4_t __rev2_51;  __rev2_51 = __builtin_shufflevector(__s2_51, __s2_51, 3, 2, 1, 0); \
44713  uint32x4_t __ret_51; \
44714  __ret_51 = __noswap_vsetq_lane_u32(__noswap_vgetq_lane_u32(__rev2_51, __p3_51), __rev0_51, __p1_51); \
44715  __ret_51 = __builtin_shufflevector(__ret_51, __ret_51, 3, 2, 1, 0); \
44716  __ret_51; \
44717})
44718#endif
44719
44720#ifdef __LITTLE_ENDIAN__
44721#define vcopyq_laneq_u64(__p0_52, __p1_52, __p2_52, __p3_52) __extension__ ({ \
44722  uint64x2_t __s0_52 = __p0_52; \
44723  uint64x2_t __s2_52 = __p2_52; \
44724  uint64x2_t __ret_52; \
44725  __ret_52 = vsetq_lane_u64(vgetq_lane_u64(__s2_52, __p3_52), __s0_52, __p1_52); \
44726  __ret_52; \
44727})
44728#else
44729#define vcopyq_laneq_u64(__p0_53, __p1_53, __p2_53, __p3_53) __extension__ ({ \
44730  uint64x2_t __s0_53 = __p0_53; \
44731  uint64x2_t __s2_53 = __p2_53; \
44732  uint64x2_t __rev0_53;  __rev0_53 = __builtin_shufflevector(__s0_53, __s0_53, 1, 0); \
44733  uint64x2_t __rev2_53;  __rev2_53 = __builtin_shufflevector(__s2_53, __s2_53, 1, 0); \
44734  uint64x2_t __ret_53; \
44735  __ret_53 = __noswap_vsetq_lane_u64(__noswap_vgetq_lane_u64(__rev2_53, __p3_53), __rev0_53, __p1_53); \
44736  __ret_53 = __builtin_shufflevector(__ret_53, __ret_53, 1, 0); \
44737  __ret_53; \
44738})
44739#endif
44740
44741#ifdef __LITTLE_ENDIAN__
44742#define vcopyq_laneq_u16(__p0_54, __p1_54, __p2_54, __p3_54) __extension__ ({ \
44743  uint16x8_t __s0_54 = __p0_54; \
44744  uint16x8_t __s2_54 = __p2_54; \
44745  uint16x8_t __ret_54; \
44746  __ret_54 = vsetq_lane_u16(vgetq_lane_u16(__s2_54, __p3_54), __s0_54, __p1_54); \
44747  __ret_54; \
44748})
44749#else
44750#define vcopyq_laneq_u16(__p0_55, __p1_55, __p2_55, __p3_55) __extension__ ({ \
44751  uint16x8_t __s0_55 = __p0_55; \
44752  uint16x8_t __s2_55 = __p2_55; \
44753  uint16x8_t __rev0_55;  __rev0_55 = __builtin_shufflevector(__s0_55, __s0_55, 7, 6, 5, 4, 3, 2, 1, 0); \
44754  uint16x8_t __rev2_55;  __rev2_55 = __builtin_shufflevector(__s2_55, __s2_55, 7, 6, 5, 4, 3, 2, 1, 0); \
44755  uint16x8_t __ret_55; \
44756  __ret_55 = __noswap_vsetq_lane_u16(__noswap_vgetq_lane_u16(__rev2_55, __p3_55), __rev0_55, __p1_55); \
44757  __ret_55 = __builtin_shufflevector(__ret_55, __ret_55, 7, 6, 5, 4, 3, 2, 1, 0); \
44758  __ret_55; \
44759})
44760#endif
44761
44762#ifdef __LITTLE_ENDIAN__
44763#define vcopyq_laneq_s8(__p0_56, __p1_56, __p2_56, __p3_56) __extension__ ({ \
44764  int8x16_t __s0_56 = __p0_56; \
44765  int8x16_t __s2_56 = __p2_56; \
44766  int8x16_t __ret_56; \
44767  __ret_56 = vsetq_lane_s8(vgetq_lane_s8(__s2_56, __p3_56), __s0_56, __p1_56); \
44768  __ret_56; \
44769})
44770#else
44771#define vcopyq_laneq_s8(__p0_57, __p1_57, __p2_57, __p3_57) __extension__ ({ \
44772  int8x16_t __s0_57 = __p0_57; \
44773  int8x16_t __s2_57 = __p2_57; \
44774  int8x16_t __rev0_57;  __rev0_57 = __builtin_shufflevector(__s0_57, __s0_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44775  int8x16_t __rev2_57;  __rev2_57 = __builtin_shufflevector(__s2_57, __s2_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44776  int8x16_t __ret_57; \
44777  __ret_57 = __noswap_vsetq_lane_s8(__noswap_vgetq_lane_s8(__rev2_57, __p3_57), __rev0_57, __p1_57); \
44778  __ret_57 = __builtin_shufflevector(__ret_57, __ret_57, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44779  __ret_57; \
44780})
44781#endif
44782
44783#ifdef __LITTLE_ENDIAN__
44784#define vcopyq_laneq_f32(__p0_58, __p1_58, __p2_58, __p3_58) __extension__ ({ \
44785  float32x4_t __s0_58 = __p0_58; \
44786  float32x4_t __s2_58 = __p2_58; \
44787  float32x4_t __ret_58; \
44788  __ret_58 = vsetq_lane_f32(vgetq_lane_f32(__s2_58, __p3_58), __s0_58, __p1_58); \
44789  __ret_58; \
44790})
44791#else
44792#define vcopyq_laneq_f32(__p0_59, __p1_59, __p2_59, __p3_59) __extension__ ({ \
44793  float32x4_t __s0_59 = __p0_59; \
44794  float32x4_t __s2_59 = __p2_59; \
44795  float32x4_t __rev0_59;  __rev0_59 = __builtin_shufflevector(__s0_59, __s0_59, 3, 2, 1, 0); \
44796  float32x4_t __rev2_59;  __rev2_59 = __builtin_shufflevector(__s2_59, __s2_59, 3, 2, 1, 0); \
44797  float32x4_t __ret_59; \
44798  __ret_59 = __noswap_vsetq_lane_f32(__noswap_vgetq_lane_f32(__rev2_59, __p3_59), __rev0_59, __p1_59); \
44799  __ret_59 = __builtin_shufflevector(__ret_59, __ret_59, 3, 2, 1, 0); \
44800  __ret_59; \
44801})
44802#endif
44803
44804#ifdef __LITTLE_ENDIAN__
44805#define vcopyq_laneq_s32(__p0_60, __p1_60, __p2_60, __p3_60) __extension__ ({ \
44806  int32x4_t __s0_60 = __p0_60; \
44807  int32x4_t __s2_60 = __p2_60; \
44808  int32x4_t __ret_60; \
44809  __ret_60 = vsetq_lane_s32(vgetq_lane_s32(__s2_60, __p3_60), __s0_60, __p1_60); \
44810  __ret_60; \
44811})
44812#else
44813#define vcopyq_laneq_s32(__p0_61, __p1_61, __p2_61, __p3_61) __extension__ ({ \
44814  int32x4_t __s0_61 = __p0_61; \
44815  int32x4_t __s2_61 = __p2_61; \
44816  int32x4_t __rev0_61;  __rev0_61 = __builtin_shufflevector(__s0_61, __s0_61, 3, 2, 1, 0); \
44817  int32x4_t __rev2_61;  __rev2_61 = __builtin_shufflevector(__s2_61, __s2_61, 3, 2, 1, 0); \
44818  int32x4_t __ret_61; \
44819  __ret_61 = __noswap_vsetq_lane_s32(__noswap_vgetq_lane_s32(__rev2_61, __p3_61), __rev0_61, __p1_61); \
44820  __ret_61 = __builtin_shufflevector(__ret_61, __ret_61, 3, 2, 1, 0); \
44821  __ret_61; \
44822})
44823#endif
44824
44825#ifdef __LITTLE_ENDIAN__
44826#define vcopyq_laneq_s64(__p0_62, __p1_62, __p2_62, __p3_62) __extension__ ({ \
44827  int64x2_t __s0_62 = __p0_62; \
44828  int64x2_t __s2_62 = __p2_62; \
44829  int64x2_t __ret_62; \
44830  __ret_62 = vsetq_lane_s64(vgetq_lane_s64(__s2_62, __p3_62), __s0_62, __p1_62); \
44831  __ret_62; \
44832})
44833#else
44834#define vcopyq_laneq_s64(__p0_63, __p1_63, __p2_63, __p3_63) __extension__ ({ \
44835  int64x2_t __s0_63 = __p0_63; \
44836  int64x2_t __s2_63 = __p2_63; \
44837  int64x2_t __rev0_63;  __rev0_63 = __builtin_shufflevector(__s0_63, __s0_63, 1, 0); \
44838  int64x2_t __rev2_63;  __rev2_63 = __builtin_shufflevector(__s2_63, __s2_63, 1, 0); \
44839  int64x2_t __ret_63; \
44840  __ret_63 = __noswap_vsetq_lane_s64(__noswap_vgetq_lane_s64(__rev2_63, __p3_63), __rev0_63, __p1_63); \
44841  __ret_63 = __builtin_shufflevector(__ret_63, __ret_63, 1, 0); \
44842  __ret_63; \
44843})
44844#endif
44845
44846#ifdef __LITTLE_ENDIAN__
44847#define vcopyq_laneq_s16(__p0_64, __p1_64, __p2_64, __p3_64) __extension__ ({ \
44848  int16x8_t __s0_64 = __p0_64; \
44849  int16x8_t __s2_64 = __p2_64; \
44850  int16x8_t __ret_64; \
44851  __ret_64 = vsetq_lane_s16(vgetq_lane_s16(__s2_64, __p3_64), __s0_64, __p1_64); \
44852  __ret_64; \
44853})
44854#else
44855#define vcopyq_laneq_s16(__p0_65, __p1_65, __p2_65, __p3_65) __extension__ ({ \
44856  int16x8_t __s0_65 = __p0_65; \
44857  int16x8_t __s2_65 = __p2_65; \
44858  int16x8_t __rev0_65;  __rev0_65 = __builtin_shufflevector(__s0_65, __s0_65, 7, 6, 5, 4, 3, 2, 1, 0); \
44859  int16x8_t __rev2_65;  __rev2_65 = __builtin_shufflevector(__s2_65, __s2_65, 7, 6, 5, 4, 3, 2, 1, 0); \
44860  int16x8_t __ret_65; \
44861  __ret_65 = __noswap_vsetq_lane_s16(__noswap_vgetq_lane_s16(__rev2_65, __p3_65), __rev0_65, __p1_65); \
44862  __ret_65 = __builtin_shufflevector(__ret_65, __ret_65, 7, 6, 5, 4, 3, 2, 1, 0); \
44863  __ret_65; \
44864})
44865#endif
44866
44867#ifdef __LITTLE_ENDIAN__
44868#define vcopy_laneq_p8(__p0_66, __p1_66, __p2_66, __p3_66) __extension__ ({ \
44869  poly8x8_t __s0_66 = __p0_66; \
44870  poly8x16_t __s2_66 = __p2_66; \
44871  poly8x8_t __ret_66; \
44872  __ret_66 = vset_lane_p8(vgetq_lane_p8(__s2_66, __p3_66), __s0_66, __p1_66); \
44873  __ret_66; \
44874})
44875#else
44876#define vcopy_laneq_p8(__p0_67, __p1_67, __p2_67, __p3_67) __extension__ ({ \
44877  poly8x8_t __s0_67 = __p0_67; \
44878  poly8x16_t __s2_67 = __p2_67; \
44879  poly8x8_t __rev0_67;  __rev0_67 = __builtin_shufflevector(__s0_67, __s0_67, 7, 6, 5, 4, 3, 2, 1, 0); \
44880  poly8x16_t __rev2_67;  __rev2_67 = __builtin_shufflevector(__s2_67, __s2_67, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44881  poly8x8_t __ret_67; \
44882  __ret_67 = __noswap_vset_lane_p8(__noswap_vgetq_lane_p8(__rev2_67, __p3_67), __rev0_67, __p1_67); \
44883  __ret_67 = __builtin_shufflevector(__ret_67, __ret_67, 7, 6, 5, 4, 3, 2, 1, 0); \
44884  __ret_67; \
44885})
44886#endif
44887
44888#ifdef __LITTLE_ENDIAN__
44889#define vcopy_laneq_p16(__p0_68, __p1_68, __p2_68, __p3_68) __extension__ ({ \
44890  poly16x4_t __s0_68 = __p0_68; \
44891  poly16x8_t __s2_68 = __p2_68; \
44892  poly16x4_t __ret_68; \
44893  __ret_68 = vset_lane_p16(vgetq_lane_p16(__s2_68, __p3_68), __s0_68, __p1_68); \
44894  __ret_68; \
44895})
44896#else
44897#define vcopy_laneq_p16(__p0_69, __p1_69, __p2_69, __p3_69) __extension__ ({ \
44898  poly16x4_t __s0_69 = __p0_69; \
44899  poly16x8_t __s2_69 = __p2_69; \
44900  poly16x4_t __rev0_69;  __rev0_69 = __builtin_shufflevector(__s0_69, __s0_69, 3, 2, 1, 0); \
44901  poly16x8_t __rev2_69;  __rev2_69 = __builtin_shufflevector(__s2_69, __s2_69, 7, 6, 5, 4, 3, 2, 1, 0); \
44902  poly16x4_t __ret_69; \
44903  __ret_69 = __noswap_vset_lane_p16(__noswap_vgetq_lane_p16(__rev2_69, __p3_69), __rev0_69, __p1_69); \
44904  __ret_69 = __builtin_shufflevector(__ret_69, __ret_69, 3, 2, 1, 0); \
44905  __ret_69; \
44906})
44907#endif
44908
44909#ifdef __LITTLE_ENDIAN__
44910#define vcopy_laneq_u8(__p0_70, __p1_70, __p2_70, __p3_70) __extension__ ({ \
44911  uint8x8_t __s0_70 = __p0_70; \
44912  uint8x16_t __s2_70 = __p2_70; \
44913  uint8x8_t __ret_70; \
44914  __ret_70 = vset_lane_u8(vgetq_lane_u8(__s2_70, __p3_70), __s0_70, __p1_70); \
44915  __ret_70; \
44916})
44917#else
44918#define vcopy_laneq_u8(__p0_71, __p1_71, __p2_71, __p3_71) __extension__ ({ \
44919  uint8x8_t __s0_71 = __p0_71; \
44920  uint8x16_t __s2_71 = __p2_71; \
44921  uint8x8_t __rev0_71;  __rev0_71 = __builtin_shufflevector(__s0_71, __s0_71, 7, 6, 5, 4, 3, 2, 1, 0); \
44922  uint8x16_t __rev2_71;  __rev2_71 = __builtin_shufflevector(__s2_71, __s2_71, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
44923  uint8x8_t __ret_71; \
44924  __ret_71 = __noswap_vset_lane_u8(__noswap_vgetq_lane_u8(__rev2_71, __p3_71), __rev0_71, __p1_71); \
44925  __ret_71 = __builtin_shufflevector(__ret_71, __ret_71, 7, 6, 5, 4, 3, 2, 1, 0); \
44926  __ret_71; \
44927})
44928#endif
44929
44930#ifdef __LITTLE_ENDIAN__
44931#define vcopy_laneq_u32(__p0_72, __p1_72, __p2_72, __p3_72) __extension__ ({ \
44932  uint32x2_t __s0_72 = __p0_72; \
44933  uint32x4_t __s2_72 = __p2_72; \
44934  uint32x2_t __ret_72; \
44935  __ret_72 = vset_lane_u32(vgetq_lane_u32(__s2_72, __p3_72), __s0_72, __p1_72); \
44936  __ret_72; \
44937})
44938#else
44939#define vcopy_laneq_u32(__p0_73, __p1_73, __p2_73, __p3_73) __extension__ ({ \
44940  uint32x2_t __s0_73 = __p0_73; \
44941  uint32x4_t __s2_73 = __p2_73; \
44942  uint32x2_t __rev0_73;  __rev0_73 = __builtin_shufflevector(__s0_73, __s0_73, 1, 0); \
44943  uint32x4_t __rev2_73;  __rev2_73 = __builtin_shufflevector(__s2_73, __s2_73, 3, 2, 1, 0); \
44944  uint32x2_t __ret_73; \
44945  __ret_73 = __noswap_vset_lane_u32(__noswap_vgetq_lane_u32(__rev2_73, __p3_73), __rev0_73, __p1_73); \
44946  __ret_73 = __builtin_shufflevector(__ret_73, __ret_73, 1, 0); \
44947  __ret_73; \
44948})
44949#endif
44950
44951#ifdef __LITTLE_ENDIAN__
44952#define vcopy_laneq_u64(__p0_74, __p1_74, __p2_74, __p3_74) __extension__ ({ \
44953  uint64x1_t __s0_74 = __p0_74; \
44954  uint64x2_t __s2_74 = __p2_74; \
44955  uint64x1_t __ret_74; \
44956  __ret_74 = vset_lane_u64(vgetq_lane_u64(__s2_74, __p3_74), __s0_74, __p1_74); \
44957  __ret_74; \
44958})
44959#else
44960#define vcopy_laneq_u64(__p0_75, __p1_75, __p2_75, __p3_75) __extension__ ({ \
44961  uint64x1_t __s0_75 = __p0_75; \
44962  uint64x2_t __s2_75 = __p2_75; \
44963  uint64x2_t __rev2_75;  __rev2_75 = __builtin_shufflevector(__s2_75, __s2_75, 1, 0); \
44964  uint64x1_t __ret_75; \
44965  __ret_75 = __noswap_vset_lane_u64(__noswap_vgetq_lane_u64(__rev2_75, __p3_75), __s0_75, __p1_75); \
44966  __ret_75; \
44967})
44968#endif
44969
44970#ifdef __LITTLE_ENDIAN__
44971#define vcopy_laneq_u16(__p0_76, __p1_76, __p2_76, __p3_76) __extension__ ({ \
44972  uint16x4_t __s0_76 = __p0_76; \
44973  uint16x8_t __s2_76 = __p2_76; \
44974  uint16x4_t __ret_76; \
44975  __ret_76 = vset_lane_u16(vgetq_lane_u16(__s2_76, __p3_76), __s0_76, __p1_76); \
44976  __ret_76; \
44977})
44978#else
44979#define vcopy_laneq_u16(__p0_77, __p1_77, __p2_77, __p3_77) __extension__ ({ \
44980  uint16x4_t __s0_77 = __p0_77; \
44981  uint16x8_t __s2_77 = __p2_77; \
44982  uint16x4_t __rev0_77;  __rev0_77 = __builtin_shufflevector(__s0_77, __s0_77, 3, 2, 1, 0); \
44983  uint16x8_t __rev2_77;  __rev2_77 = __builtin_shufflevector(__s2_77, __s2_77, 7, 6, 5, 4, 3, 2, 1, 0); \
44984  uint16x4_t __ret_77; \
44985  __ret_77 = __noswap_vset_lane_u16(__noswap_vgetq_lane_u16(__rev2_77, __p3_77), __rev0_77, __p1_77); \
44986  __ret_77 = __builtin_shufflevector(__ret_77, __ret_77, 3, 2, 1, 0); \
44987  __ret_77; \
44988})
44989#endif
44990
44991#ifdef __LITTLE_ENDIAN__
44992#define vcopy_laneq_s8(__p0_78, __p1_78, __p2_78, __p3_78) __extension__ ({ \
44993  int8x8_t __s0_78 = __p0_78; \
44994  int8x16_t __s2_78 = __p2_78; \
44995  int8x8_t __ret_78; \
44996  __ret_78 = vset_lane_s8(vgetq_lane_s8(__s2_78, __p3_78), __s0_78, __p1_78); \
44997  __ret_78; \
44998})
44999#else
45000#define vcopy_laneq_s8(__p0_79, __p1_79, __p2_79, __p3_79) __extension__ ({ \
45001  int8x8_t __s0_79 = __p0_79; \
45002  int8x16_t __s2_79 = __p2_79; \
45003  int8x8_t __rev0_79;  __rev0_79 = __builtin_shufflevector(__s0_79, __s0_79, 7, 6, 5, 4, 3, 2, 1, 0); \
45004  int8x16_t __rev2_79;  __rev2_79 = __builtin_shufflevector(__s2_79, __s2_79, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
45005  int8x8_t __ret_79; \
45006  __ret_79 = __noswap_vset_lane_s8(__noswap_vgetq_lane_s8(__rev2_79, __p3_79), __rev0_79, __p1_79); \
45007  __ret_79 = __builtin_shufflevector(__ret_79, __ret_79, 7, 6, 5, 4, 3, 2, 1, 0); \
45008  __ret_79; \
45009})
45010#endif
45011
45012#ifdef __LITTLE_ENDIAN__
45013#define vcopy_laneq_f32(__p0_80, __p1_80, __p2_80, __p3_80) __extension__ ({ \
45014  float32x2_t __s0_80 = __p0_80; \
45015  float32x4_t __s2_80 = __p2_80; \
45016  float32x2_t __ret_80; \
45017  __ret_80 = vset_lane_f32(vgetq_lane_f32(__s2_80, __p3_80), __s0_80, __p1_80); \
45018  __ret_80; \
45019})
45020#else
45021#define vcopy_laneq_f32(__p0_81, __p1_81, __p2_81, __p3_81) __extension__ ({ \
45022  float32x2_t __s0_81 = __p0_81; \
45023  float32x4_t __s2_81 = __p2_81; \
45024  float32x2_t __rev0_81;  __rev0_81 = __builtin_shufflevector(__s0_81, __s0_81, 1, 0); \
45025  float32x4_t __rev2_81;  __rev2_81 = __builtin_shufflevector(__s2_81, __s2_81, 3, 2, 1, 0); \
45026  float32x2_t __ret_81; \
45027  __ret_81 = __noswap_vset_lane_f32(__noswap_vgetq_lane_f32(__rev2_81, __p3_81), __rev0_81, __p1_81); \
45028  __ret_81 = __builtin_shufflevector(__ret_81, __ret_81, 1, 0); \
45029  __ret_81; \
45030})
45031#endif
45032
45033#ifdef __LITTLE_ENDIAN__
45034#define vcopy_laneq_s32(__p0_82, __p1_82, __p2_82, __p3_82) __extension__ ({ \
45035  int32x2_t __s0_82 = __p0_82; \
45036  int32x4_t __s2_82 = __p2_82; \
45037  int32x2_t __ret_82; \
45038  __ret_82 = vset_lane_s32(vgetq_lane_s32(__s2_82, __p3_82), __s0_82, __p1_82); \
45039  __ret_82; \
45040})
45041#else
45042#define vcopy_laneq_s32(__p0_83, __p1_83, __p2_83, __p3_83) __extension__ ({ \
45043  int32x2_t __s0_83 = __p0_83; \
45044  int32x4_t __s2_83 = __p2_83; \
45045  int32x2_t __rev0_83;  __rev0_83 = __builtin_shufflevector(__s0_83, __s0_83, 1, 0); \
45046  int32x4_t __rev2_83;  __rev2_83 = __builtin_shufflevector(__s2_83, __s2_83, 3, 2, 1, 0); \
45047  int32x2_t __ret_83; \
45048  __ret_83 = __noswap_vset_lane_s32(__noswap_vgetq_lane_s32(__rev2_83, __p3_83), __rev0_83, __p1_83); \
45049  __ret_83 = __builtin_shufflevector(__ret_83, __ret_83, 1, 0); \
45050  __ret_83; \
45051})
45052#endif
45053
45054#ifdef __LITTLE_ENDIAN__
45055#define vcopy_laneq_s64(__p0_84, __p1_84, __p2_84, __p3_84) __extension__ ({ \
45056  int64x1_t __s0_84 = __p0_84; \
45057  int64x2_t __s2_84 = __p2_84; \
45058  int64x1_t __ret_84; \
45059  __ret_84 = vset_lane_s64(vgetq_lane_s64(__s2_84, __p3_84), __s0_84, __p1_84); \
45060  __ret_84; \
45061})
45062#else
45063#define vcopy_laneq_s64(__p0_85, __p1_85, __p2_85, __p3_85) __extension__ ({ \
45064  int64x1_t __s0_85 = __p0_85; \
45065  int64x2_t __s2_85 = __p2_85; \
45066  int64x2_t __rev2_85;  __rev2_85 = __builtin_shufflevector(__s2_85, __s2_85, 1, 0); \
45067  int64x1_t __ret_85; \
45068  __ret_85 = __noswap_vset_lane_s64(__noswap_vgetq_lane_s64(__rev2_85, __p3_85), __s0_85, __p1_85); \
45069  __ret_85; \
45070})
45071#endif
45072
45073#ifdef __LITTLE_ENDIAN__
45074#define vcopy_laneq_s16(__p0_86, __p1_86, __p2_86, __p3_86) __extension__ ({ \
45075  int16x4_t __s0_86 = __p0_86; \
45076  int16x8_t __s2_86 = __p2_86; \
45077  int16x4_t __ret_86; \
45078  __ret_86 = vset_lane_s16(vgetq_lane_s16(__s2_86, __p3_86), __s0_86, __p1_86); \
45079  __ret_86; \
45080})
45081#else
45082#define vcopy_laneq_s16(__p0_87, __p1_87, __p2_87, __p3_87) __extension__ ({ \
45083  int16x4_t __s0_87 = __p0_87; \
45084  int16x8_t __s2_87 = __p2_87; \
45085  int16x4_t __rev0_87;  __rev0_87 = __builtin_shufflevector(__s0_87, __s0_87, 3, 2, 1, 0); \
45086  int16x8_t __rev2_87;  __rev2_87 = __builtin_shufflevector(__s2_87, __s2_87, 7, 6, 5, 4, 3, 2, 1, 0); \
45087  int16x4_t __ret_87; \
45088  __ret_87 = __noswap_vset_lane_s16(__noswap_vgetq_lane_s16(__rev2_87, __p3_87), __rev0_87, __p1_87); \
45089  __ret_87 = __builtin_shufflevector(__ret_87, __ret_87, 3, 2, 1, 0); \
45090  __ret_87; \
45091})
45092#endif
45093
45094#ifdef __LITTLE_ENDIAN__
45095__ai poly64x1_t vcreate_p64(uint64_t __p0) {
45096  poly64x1_t __ret;
45097  __ret = (poly64x1_t)(__p0);
45098  return __ret;
45099}
45100#else
45101__ai poly64x1_t vcreate_p64(uint64_t __p0) {
45102  poly64x1_t __ret;
45103  __ret = (poly64x1_t)(__p0);
45104  return __ret;
45105}
45106#endif
45107
45108#ifdef __LITTLE_ENDIAN__
45109__ai float64x1_t vcreate_f64(uint64_t __p0) {
45110  float64x1_t __ret;
45111  __ret = (float64x1_t)(__p0);
45112  return __ret;
45113}
45114#else
45115__ai float64x1_t vcreate_f64(uint64_t __p0) {
45116  float64x1_t __ret;
45117  __ret = (float64x1_t)(__p0);
45118  return __ret;
45119}
45120#endif
45121
45122#ifdef __LITTLE_ENDIAN__
45123__ai float32_t vcvts_f32_s32(int32_t __p0) {
45124  float32_t __ret;
45125  __ret = (float32_t) __builtin_neon_vcvts_f32_s32(__p0);
45126  return __ret;
45127}
45128#else
45129__ai float32_t vcvts_f32_s32(int32_t __p0) {
45130  float32_t __ret;
45131  __ret = (float32_t) __builtin_neon_vcvts_f32_s32(__p0);
45132  return __ret;
45133}
45134#endif
45135
45136#ifdef __LITTLE_ENDIAN__
45137__ai float32_t vcvts_f32_u32(uint32_t __p0) {
45138  float32_t __ret;
45139  __ret = (float32_t) __builtin_neon_vcvts_f32_u32(__p0);
45140  return __ret;
45141}
45142#else
45143__ai float32_t vcvts_f32_u32(uint32_t __p0) {
45144  float32_t __ret;
45145  __ret = (float32_t) __builtin_neon_vcvts_f32_u32(__p0);
45146  return __ret;
45147}
45148#endif
45149
45150#ifdef __LITTLE_ENDIAN__
45151__ai float32x2_t vcvt_f32_f64(float64x2_t __p0) {
45152  float32x2_t __ret;
45153  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__p0, 9);
45154  return __ret;
45155}
45156#else
45157__ai float32x2_t vcvt_f32_f64(float64x2_t __p0) {
45158  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45159  float32x2_t __ret;
45160  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__rev0, 9);
45161  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45162  return __ret;
45163}
45164__ai float32x2_t __noswap_vcvt_f32_f64(float64x2_t __p0) {
45165  float32x2_t __ret;
45166  __ret = (float32x2_t) __builtin_neon_vcvt_f32_f64((int8x16_t)__p0, 9);
45167  return __ret;
45168}
45169#endif
45170
45171#ifdef __LITTLE_ENDIAN__
45172__ai float64_t vcvtd_f64_s64(int64_t __p0) {
45173  float64_t __ret;
45174  __ret = (float64_t) __builtin_neon_vcvtd_f64_s64(__p0);
45175  return __ret;
45176}
45177#else
45178__ai float64_t vcvtd_f64_s64(int64_t __p0) {
45179  float64_t __ret;
45180  __ret = (float64_t) __builtin_neon_vcvtd_f64_s64(__p0);
45181  return __ret;
45182}
45183#endif
45184
45185#ifdef __LITTLE_ENDIAN__
45186__ai float64_t vcvtd_f64_u64(uint64_t __p0) {
45187  float64_t __ret;
45188  __ret = (float64_t) __builtin_neon_vcvtd_f64_u64(__p0);
45189  return __ret;
45190}
45191#else
45192__ai float64_t vcvtd_f64_u64(uint64_t __p0) {
45193  float64_t __ret;
45194  __ret = (float64_t) __builtin_neon_vcvtd_f64_u64(__p0);
45195  return __ret;
45196}
45197#endif
45198
45199#ifdef __LITTLE_ENDIAN__
45200__ai float64x2_t vcvtq_f64_u64(uint64x2_t __p0) {
45201  float64x2_t __ret;
45202  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__p0, 51);
45203  return __ret;
45204}
45205#else
45206__ai float64x2_t vcvtq_f64_u64(uint64x2_t __p0) {
45207  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45208  float64x2_t __ret;
45209  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__rev0, 51);
45210  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45211  return __ret;
45212}
45213#endif
45214
45215#ifdef __LITTLE_ENDIAN__
45216__ai float64x2_t vcvtq_f64_s64(int64x2_t __p0) {
45217  float64x2_t __ret;
45218  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__p0, 35);
45219  return __ret;
45220}
45221#else
45222__ai float64x2_t vcvtq_f64_s64(int64x2_t __p0) {
45223  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45224  float64x2_t __ret;
45225  __ret = (float64x2_t) __builtin_neon_vcvtq_f64_v((int8x16_t)__rev0, 35);
45226  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45227  return __ret;
45228}
45229#endif
45230
45231#ifdef __LITTLE_ENDIAN__
45232__ai float64x1_t vcvt_f64_u64(uint64x1_t __p0) {
45233  float64x1_t __ret;
45234  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 19);
45235  return __ret;
45236}
45237#else
45238__ai float64x1_t vcvt_f64_u64(uint64x1_t __p0) {
45239  float64x1_t __ret;
45240  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 19);
45241  return __ret;
45242}
45243#endif
45244
45245#ifdef __LITTLE_ENDIAN__
45246__ai float64x1_t vcvt_f64_s64(int64x1_t __p0) {
45247  float64x1_t __ret;
45248  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 3);
45249  return __ret;
45250}
45251#else
45252__ai float64x1_t vcvt_f64_s64(int64x1_t __p0) {
45253  float64x1_t __ret;
45254  __ret = (float64x1_t) __builtin_neon_vcvt_f64_v((int8x8_t)__p0, 3);
45255  return __ret;
45256}
45257#endif
45258
45259#ifdef __LITTLE_ENDIAN__
45260__ai float64x2_t vcvt_f64_f32(float32x2_t __p0) {
45261  float64x2_t __ret;
45262  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__p0, 42);
45263  return __ret;
45264}
45265#else
45266__ai float64x2_t vcvt_f64_f32(float32x2_t __p0) {
45267  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45268  float64x2_t __ret;
45269  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__rev0, 42);
45270  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45271  return __ret;
45272}
45273__ai float64x2_t __noswap_vcvt_f64_f32(float32x2_t __p0) {
45274  float64x2_t __ret;
45275  __ret = (float64x2_t) __builtin_neon_vcvt_f64_f32((int8x8_t)__p0, 42);
45276  return __ret;
45277}
45278#endif
45279
45280#ifdef __LITTLE_ENDIAN__
45281__ai float16x8_t vcvt_high_f16_f32(float16x4_t __p0, float32x4_t __p1) {
45282  float16x8_t __ret;
45283  __ret = vcombine_f16(__p0, vcvt_f16_f32(__p1));
45284  return __ret;
45285}
45286#else
45287__ai float16x8_t vcvt_high_f16_f32(float16x4_t __p0, float32x4_t __p1) {
45288  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45289  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
45290  float16x8_t __ret;
45291  __ret = __noswap_vcombine_f16(__rev0, __noswap_vcvt_f16_f32(__rev1));
45292  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
45293  return __ret;
45294}
45295#endif
45296
45297#ifdef __LITTLE_ENDIAN__
45298__ai float32x4_t vcvt_high_f32_f16(float16x8_t __p0) {
45299  float32x4_t __ret;
45300  __ret = vcvt_f32_f16(vget_high_f16(__p0));
45301  return __ret;
45302}
45303#else
45304__ai float32x4_t vcvt_high_f32_f16(float16x8_t __p0) {
45305  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
45306  float32x4_t __ret;
45307  __ret = __noswap_vcvt_f32_f16(__noswap_vget_high_f16(__rev0));
45308  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
45309  return __ret;
45310}
45311#endif
45312
45313#ifdef __LITTLE_ENDIAN__
45314__ai float32x4_t vcvt_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
45315  float32x4_t __ret;
45316  __ret = vcombine_f32(__p0, vcvt_f32_f64(__p1));
45317  return __ret;
45318}
45319#else
45320__ai float32x4_t vcvt_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
45321  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45322  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45323  float32x4_t __ret;
45324  __ret = __noswap_vcombine_f32(__rev0, __noswap_vcvt_f32_f64(__rev1));
45325  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
45326  return __ret;
45327}
45328#endif
45329
45330#ifdef __LITTLE_ENDIAN__
45331__ai float64x2_t vcvt_high_f64_f32(float32x4_t __p0) {
45332  float64x2_t __ret;
45333  __ret = vcvt_f64_f32(vget_high_f32(__p0));
45334  return __ret;
45335}
45336#else
45337__ai float64x2_t vcvt_high_f64_f32(float32x4_t __p0) {
45338  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
45339  float64x2_t __ret;
45340  __ret = __noswap_vcvt_f64_f32(__noswap_vget_high_f32(__rev0));
45341  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45342  return __ret;
45343}
45344#endif
45345
45346#ifdef __LITTLE_ENDIAN__
45347#define vcvts_n_f32_u32(__p0, __p1) __extension__ ({ \
45348  uint32_t __s0 = __p0; \
45349  float32_t __ret; \
45350  __ret = (float32_t) __builtin_neon_vcvts_n_f32_u32(__s0, __p1); \
45351  __ret; \
45352})
45353#else
45354#define vcvts_n_f32_u32(__p0, __p1) __extension__ ({ \
45355  uint32_t __s0 = __p0; \
45356  float32_t __ret; \
45357  __ret = (float32_t) __builtin_neon_vcvts_n_f32_u32(__s0, __p1); \
45358  __ret; \
45359})
45360#endif
45361
45362#ifdef __LITTLE_ENDIAN__
45363#define vcvts_n_f32_s32(__p0, __p1) __extension__ ({ \
45364  int32_t __s0 = __p0; \
45365  float32_t __ret; \
45366  __ret = (float32_t) __builtin_neon_vcvts_n_f32_s32(__s0, __p1); \
45367  __ret; \
45368})
45369#else
45370#define vcvts_n_f32_s32(__p0, __p1) __extension__ ({ \
45371  int32_t __s0 = __p0; \
45372  float32_t __ret; \
45373  __ret = (float32_t) __builtin_neon_vcvts_n_f32_s32(__s0, __p1); \
45374  __ret; \
45375})
45376#endif
45377
45378#ifdef __LITTLE_ENDIAN__
45379#define vcvtq_n_f64_u64(__p0, __p1) __extension__ ({ \
45380  uint64x2_t __s0 = __p0; \
45381  float64x2_t __ret; \
45382  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__s0, __p1, 51); \
45383  __ret; \
45384})
45385#else
45386#define vcvtq_n_f64_u64(__p0, __p1) __extension__ ({ \
45387  uint64x2_t __s0 = __p0; \
45388  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45389  float64x2_t __ret; \
45390  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__rev0, __p1, 51); \
45391  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45392  __ret; \
45393})
45394#endif
45395
45396#ifdef __LITTLE_ENDIAN__
45397#define vcvtq_n_f64_s64(__p0, __p1) __extension__ ({ \
45398  int64x2_t __s0 = __p0; \
45399  float64x2_t __ret; \
45400  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__s0, __p1, 35); \
45401  __ret; \
45402})
45403#else
45404#define vcvtq_n_f64_s64(__p0, __p1) __extension__ ({ \
45405  int64x2_t __s0 = __p0; \
45406  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45407  float64x2_t __ret; \
45408  __ret = (float64x2_t) __builtin_neon_vcvtq_n_f64_v((int8x16_t)__rev0, __p1, 35); \
45409  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45410  __ret; \
45411})
45412#endif
45413
45414#ifdef __LITTLE_ENDIAN__
45415#define vcvt_n_f64_u64(__p0, __p1) __extension__ ({ \
45416  uint64x1_t __s0 = __p0; \
45417  float64x1_t __ret; \
45418  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 19); \
45419  __ret; \
45420})
45421#else
45422#define vcvt_n_f64_u64(__p0, __p1) __extension__ ({ \
45423  uint64x1_t __s0 = __p0; \
45424  float64x1_t __ret; \
45425  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 19); \
45426  __ret; \
45427})
45428#endif
45429
45430#ifdef __LITTLE_ENDIAN__
45431#define vcvt_n_f64_s64(__p0, __p1) __extension__ ({ \
45432  int64x1_t __s0 = __p0; \
45433  float64x1_t __ret; \
45434  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 3); \
45435  __ret; \
45436})
45437#else
45438#define vcvt_n_f64_s64(__p0, __p1) __extension__ ({ \
45439  int64x1_t __s0 = __p0; \
45440  float64x1_t __ret; \
45441  __ret = (float64x1_t) __builtin_neon_vcvt_n_f64_v((int8x8_t)__s0, __p1, 3); \
45442  __ret; \
45443})
45444#endif
45445
45446#ifdef __LITTLE_ENDIAN__
45447#define vcvtd_n_f64_u64(__p0, __p1) __extension__ ({ \
45448  uint64_t __s0 = __p0; \
45449  float64_t __ret; \
45450  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_u64(__s0, __p1); \
45451  __ret; \
45452})
45453#else
45454#define vcvtd_n_f64_u64(__p0, __p1) __extension__ ({ \
45455  uint64_t __s0 = __p0; \
45456  float64_t __ret; \
45457  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_u64(__s0, __p1); \
45458  __ret; \
45459})
45460#endif
45461
45462#ifdef __LITTLE_ENDIAN__
45463#define vcvtd_n_f64_s64(__p0, __p1) __extension__ ({ \
45464  int64_t __s0 = __p0; \
45465  float64_t __ret; \
45466  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_s64(__s0, __p1); \
45467  __ret; \
45468})
45469#else
45470#define vcvtd_n_f64_s64(__p0, __p1) __extension__ ({ \
45471  int64_t __s0 = __p0; \
45472  float64_t __ret; \
45473  __ret = (float64_t) __builtin_neon_vcvtd_n_f64_s64(__s0, __p1); \
45474  __ret; \
45475})
45476#endif
45477
45478#ifdef __LITTLE_ENDIAN__
45479#define vcvts_n_s32_f32(__p0, __p1) __extension__ ({ \
45480  float32_t __s0 = __p0; \
45481  int32_t __ret; \
45482  __ret = (int32_t) __builtin_neon_vcvts_n_s32_f32(__s0, __p1); \
45483  __ret; \
45484})
45485#else
45486#define vcvts_n_s32_f32(__p0, __p1) __extension__ ({ \
45487  float32_t __s0 = __p0; \
45488  int32_t __ret; \
45489  __ret = (int32_t) __builtin_neon_vcvts_n_s32_f32(__s0, __p1); \
45490  __ret; \
45491})
45492#endif
45493
45494#ifdef __LITTLE_ENDIAN__
45495#define vcvtq_n_s64_f64(__p0, __p1) __extension__ ({ \
45496  float64x2_t __s0 = __p0; \
45497  int64x2_t __ret; \
45498  __ret = (int64x2_t) __builtin_neon_vcvtq_n_s64_v((int8x16_t)__s0, __p1, 35); \
45499  __ret; \
45500})
45501#else
45502#define vcvtq_n_s64_f64(__p0, __p1) __extension__ ({ \
45503  float64x2_t __s0 = __p0; \
45504  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45505  int64x2_t __ret; \
45506  __ret = (int64x2_t) __builtin_neon_vcvtq_n_s64_v((int8x16_t)__rev0, __p1, 35); \
45507  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45508  __ret; \
45509})
45510#endif
45511
45512#ifdef __LITTLE_ENDIAN__
45513#define vcvt_n_s64_f64(__p0, __p1) __extension__ ({ \
45514  float64x1_t __s0 = __p0; \
45515  int64x1_t __ret; \
45516  __ret = (int64x1_t) __builtin_neon_vcvt_n_s64_v((int8x8_t)__s0, __p1, 3); \
45517  __ret; \
45518})
45519#else
45520#define vcvt_n_s64_f64(__p0, __p1) __extension__ ({ \
45521  float64x1_t __s0 = __p0; \
45522  int64x1_t __ret; \
45523  __ret = (int64x1_t) __builtin_neon_vcvt_n_s64_v((int8x8_t)__s0, __p1, 3); \
45524  __ret; \
45525})
45526#endif
45527
45528#ifdef __LITTLE_ENDIAN__
45529#define vcvtd_n_s64_f64(__p0, __p1) __extension__ ({ \
45530  float64_t __s0 = __p0; \
45531  int64_t __ret; \
45532  __ret = (int64_t) __builtin_neon_vcvtd_n_s64_f64(__s0, __p1); \
45533  __ret; \
45534})
45535#else
45536#define vcvtd_n_s64_f64(__p0, __p1) __extension__ ({ \
45537  float64_t __s0 = __p0; \
45538  int64_t __ret; \
45539  __ret = (int64_t) __builtin_neon_vcvtd_n_s64_f64(__s0, __p1); \
45540  __ret; \
45541})
45542#endif
45543
45544#ifdef __LITTLE_ENDIAN__
45545#define vcvts_n_u32_f32(__p0, __p1) __extension__ ({ \
45546  float32_t __s0 = __p0; \
45547  uint32_t __ret; \
45548  __ret = (uint32_t) __builtin_neon_vcvts_n_u32_f32(__s0, __p1); \
45549  __ret; \
45550})
45551#else
45552#define vcvts_n_u32_f32(__p0, __p1) __extension__ ({ \
45553  float32_t __s0 = __p0; \
45554  uint32_t __ret; \
45555  __ret = (uint32_t) __builtin_neon_vcvts_n_u32_f32(__s0, __p1); \
45556  __ret; \
45557})
45558#endif
45559
45560#ifdef __LITTLE_ENDIAN__
45561#define vcvtq_n_u64_f64(__p0, __p1) __extension__ ({ \
45562  float64x2_t __s0 = __p0; \
45563  uint64x2_t __ret; \
45564  __ret = (uint64x2_t) __builtin_neon_vcvtq_n_u64_v((int8x16_t)__s0, __p1, 51); \
45565  __ret; \
45566})
45567#else
45568#define vcvtq_n_u64_f64(__p0, __p1) __extension__ ({ \
45569  float64x2_t __s0 = __p0; \
45570  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
45571  uint64x2_t __ret; \
45572  __ret = (uint64x2_t) __builtin_neon_vcvtq_n_u64_v((int8x16_t)__rev0, __p1, 51); \
45573  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
45574  __ret; \
45575})
45576#endif
45577
45578#ifdef __LITTLE_ENDIAN__
45579#define vcvt_n_u64_f64(__p0, __p1) __extension__ ({ \
45580  float64x1_t __s0 = __p0; \
45581  uint64x1_t __ret; \
45582  __ret = (uint64x1_t) __builtin_neon_vcvt_n_u64_v((int8x8_t)__s0, __p1, 19); \
45583  __ret; \
45584})
45585#else
45586#define vcvt_n_u64_f64(__p0, __p1) __extension__ ({ \
45587  float64x1_t __s0 = __p0; \
45588  uint64x1_t __ret; \
45589  __ret = (uint64x1_t) __builtin_neon_vcvt_n_u64_v((int8x8_t)__s0, __p1, 19); \
45590  __ret; \
45591})
45592#endif
45593
45594#ifdef __LITTLE_ENDIAN__
45595#define vcvtd_n_u64_f64(__p0, __p1) __extension__ ({ \
45596  float64_t __s0 = __p0; \
45597  uint64_t __ret; \
45598  __ret = (uint64_t) __builtin_neon_vcvtd_n_u64_f64(__s0, __p1); \
45599  __ret; \
45600})
45601#else
45602#define vcvtd_n_u64_f64(__p0, __p1) __extension__ ({ \
45603  float64_t __s0 = __p0; \
45604  uint64_t __ret; \
45605  __ret = (uint64_t) __builtin_neon_vcvtd_n_u64_f64(__s0, __p1); \
45606  __ret; \
45607})
45608#endif
45609
45610#ifdef __LITTLE_ENDIAN__
45611__ai int32_t vcvts_s32_f32(float32_t __p0) {
45612  int32_t __ret;
45613  __ret = (int32_t) __builtin_neon_vcvts_s32_f32(__p0);
45614  return __ret;
45615}
45616#else
45617__ai int32_t vcvts_s32_f32(float32_t __p0) {
45618  int32_t __ret;
45619  __ret = (int32_t) __builtin_neon_vcvts_s32_f32(__p0);
45620  return __ret;
45621}
45622#endif
45623
45624#ifdef __LITTLE_ENDIAN__
45625__ai int64_t vcvtd_s64_f64(float64_t __p0) {
45626  int64_t __ret;
45627  __ret = (int64_t) __builtin_neon_vcvtd_s64_f64(__p0);
45628  return __ret;
45629}
45630#else
45631__ai int64_t vcvtd_s64_f64(float64_t __p0) {
45632  int64_t __ret;
45633  __ret = (int64_t) __builtin_neon_vcvtd_s64_f64(__p0);
45634  return __ret;
45635}
45636#endif
45637
45638#ifdef __LITTLE_ENDIAN__
45639__ai int64x2_t vcvtq_s64_f64(float64x2_t __p0) {
45640  int64x2_t __ret;
45641  __ret = (int64x2_t) __builtin_neon_vcvtq_s64_v((int8x16_t)__p0, 35);
45642  return __ret;
45643}
45644#else
45645__ai int64x2_t vcvtq_s64_f64(float64x2_t __p0) {
45646  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45647  int64x2_t __ret;
45648  __ret = (int64x2_t) __builtin_neon_vcvtq_s64_v((int8x16_t)__rev0, 35);
45649  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45650  return __ret;
45651}
45652#endif
45653
45654#ifdef __LITTLE_ENDIAN__
45655__ai int64x1_t vcvt_s64_f64(float64x1_t __p0) {
45656  int64x1_t __ret;
45657  __ret = (int64x1_t) __builtin_neon_vcvt_s64_v((int8x8_t)__p0, 3);
45658  return __ret;
45659}
45660#else
45661__ai int64x1_t vcvt_s64_f64(float64x1_t __p0) {
45662  int64x1_t __ret;
45663  __ret = (int64x1_t) __builtin_neon_vcvt_s64_v((int8x8_t)__p0, 3);
45664  return __ret;
45665}
45666#endif
45667
45668#ifdef __LITTLE_ENDIAN__
45669__ai uint32_t vcvts_u32_f32(float32_t __p0) {
45670  uint32_t __ret;
45671  __ret = (uint32_t) __builtin_neon_vcvts_u32_f32(__p0);
45672  return __ret;
45673}
45674#else
45675__ai uint32_t vcvts_u32_f32(float32_t __p0) {
45676  uint32_t __ret;
45677  __ret = (uint32_t) __builtin_neon_vcvts_u32_f32(__p0);
45678  return __ret;
45679}
45680#endif
45681
45682#ifdef __LITTLE_ENDIAN__
45683__ai uint64_t vcvtd_u64_f64(float64_t __p0) {
45684  uint64_t __ret;
45685  __ret = (uint64_t) __builtin_neon_vcvtd_u64_f64(__p0);
45686  return __ret;
45687}
45688#else
45689__ai uint64_t vcvtd_u64_f64(float64_t __p0) {
45690  uint64_t __ret;
45691  __ret = (uint64_t) __builtin_neon_vcvtd_u64_f64(__p0);
45692  return __ret;
45693}
45694#endif
45695
45696#ifdef __LITTLE_ENDIAN__
45697__ai uint64x2_t vcvtq_u64_f64(float64x2_t __p0) {
45698  uint64x2_t __ret;
45699  __ret = (uint64x2_t) __builtin_neon_vcvtq_u64_v((int8x16_t)__p0, 51);
45700  return __ret;
45701}
45702#else
45703__ai uint64x2_t vcvtq_u64_f64(float64x2_t __p0) {
45704  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45705  uint64x2_t __ret;
45706  __ret = (uint64x2_t) __builtin_neon_vcvtq_u64_v((int8x16_t)__rev0, 51);
45707  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45708  return __ret;
45709}
45710#endif
45711
45712#ifdef __LITTLE_ENDIAN__
45713__ai uint64x1_t vcvt_u64_f64(float64x1_t __p0) {
45714  uint64x1_t __ret;
45715  __ret = (uint64x1_t) __builtin_neon_vcvt_u64_v((int8x8_t)__p0, 19);
45716  return __ret;
45717}
45718#else
45719__ai uint64x1_t vcvt_u64_f64(float64x1_t __p0) {
45720  uint64x1_t __ret;
45721  __ret = (uint64x1_t) __builtin_neon_vcvt_u64_v((int8x8_t)__p0, 19);
45722  return __ret;
45723}
45724#endif
45725
45726#ifdef __LITTLE_ENDIAN__
45727__ai int32_t vcvtas_s32_f32(float32_t __p0) {
45728  int32_t __ret;
45729  __ret = (int32_t) __builtin_neon_vcvtas_s32_f32(__p0);
45730  return __ret;
45731}
45732#else
45733__ai int32_t vcvtas_s32_f32(float32_t __p0) {
45734  int32_t __ret;
45735  __ret = (int32_t) __builtin_neon_vcvtas_s32_f32(__p0);
45736  return __ret;
45737}
45738#endif
45739
45740#ifdef __LITTLE_ENDIAN__
45741__ai int64_t vcvtad_s64_f64(float64_t __p0) {
45742  int64_t __ret;
45743  __ret = (int64_t) __builtin_neon_vcvtad_s64_f64(__p0);
45744  return __ret;
45745}
45746#else
45747__ai int64_t vcvtad_s64_f64(float64_t __p0) {
45748  int64_t __ret;
45749  __ret = (int64_t) __builtin_neon_vcvtad_s64_f64(__p0);
45750  return __ret;
45751}
45752#endif
45753
45754#ifdef __LITTLE_ENDIAN__
45755__ai uint32_t vcvtas_u32_f32(float32_t __p0) {
45756  uint32_t __ret;
45757  __ret = (uint32_t) __builtin_neon_vcvtas_u32_f32(__p0);
45758  return __ret;
45759}
45760#else
45761__ai uint32_t vcvtas_u32_f32(float32_t __p0) {
45762  uint32_t __ret;
45763  __ret = (uint32_t) __builtin_neon_vcvtas_u32_f32(__p0);
45764  return __ret;
45765}
45766#endif
45767
45768#ifdef __LITTLE_ENDIAN__
45769__ai uint64_t vcvtad_u64_f64(float64_t __p0) {
45770  uint64_t __ret;
45771  __ret = (uint64_t) __builtin_neon_vcvtad_u64_f64(__p0);
45772  return __ret;
45773}
45774#else
45775__ai uint64_t vcvtad_u64_f64(float64_t __p0) {
45776  uint64_t __ret;
45777  __ret = (uint64_t) __builtin_neon_vcvtad_u64_f64(__p0);
45778  return __ret;
45779}
45780#endif
45781
45782#ifdef __LITTLE_ENDIAN__
45783__ai int32_t vcvtms_s32_f32(float32_t __p0) {
45784  int32_t __ret;
45785  __ret = (int32_t) __builtin_neon_vcvtms_s32_f32(__p0);
45786  return __ret;
45787}
45788#else
45789__ai int32_t vcvtms_s32_f32(float32_t __p0) {
45790  int32_t __ret;
45791  __ret = (int32_t) __builtin_neon_vcvtms_s32_f32(__p0);
45792  return __ret;
45793}
45794#endif
45795
45796#ifdef __LITTLE_ENDIAN__
45797__ai int64_t vcvtmd_s64_f64(float64_t __p0) {
45798  int64_t __ret;
45799  __ret = (int64_t) __builtin_neon_vcvtmd_s64_f64(__p0);
45800  return __ret;
45801}
45802#else
45803__ai int64_t vcvtmd_s64_f64(float64_t __p0) {
45804  int64_t __ret;
45805  __ret = (int64_t) __builtin_neon_vcvtmd_s64_f64(__p0);
45806  return __ret;
45807}
45808#endif
45809
45810#ifdef __LITTLE_ENDIAN__
45811__ai uint32_t vcvtms_u32_f32(float32_t __p0) {
45812  uint32_t __ret;
45813  __ret = (uint32_t) __builtin_neon_vcvtms_u32_f32(__p0);
45814  return __ret;
45815}
45816#else
45817__ai uint32_t vcvtms_u32_f32(float32_t __p0) {
45818  uint32_t __ret;
45819  __ret = (uint32_t) __builtin_neon_vcvtms_u32_f32(__p0);
45820  return __ret;
45821}
45822#endif
45823
45824#ifdef __LITTLE_ENDIAN__
45825__ai uint64_t vcvtmd_u64_f64(float64_t __p0) {
45826  uint64_t __ret;
45827  __ret = (uint64_t) __builtin_neon_vcvtmd_u64_f64(__p0);
45828  return __ret;
45829}
45830#else
45831__ai uint64_t vcvtmd_u64_f64(float64_t __p0) {
45832  uint64_t __ret;
45833  __ret = (uint64_t) __builtin_neon_vcvtmd_u64_f64(__p0);
45834  return __ret;
45835}
45836#endif
45837
45838#ifdef __LITTLE_ENDIAN__
45839__ai int32_t vcvtns_s32_f32(float32_t __p0) {
45840  int32_t __ret;
45841  __ret = (int32_t) __builtin_neon_vcvtns_s32_f32(__p0);
45842  return __ret;
45843}
45844#else
45845__ai int32_t vcvtns_s32_f32(float32_t __p0) {
45846  int32_t __ret;
45847  __ret = (int32_t) __builtin_neon_vcvtns_s32_f32(__p0);
45848  return __ret;
45849}
45850#endif
45851
45852#ifdef __LITTLE_ENDIAN__
45853__ai int64_t vcvtnd_s64_f64(float64_t __p0) {
45854  int64_t __ret;
45855  __ret = (int64_t) __builtin_neon_vcvtnd_s64_f64(__p0);
45856  return __ret;
45857}
45858#else
45859__ai int64_t vcvtnd_s64_f64(float64_t __p0) {
45860  int64_t __ret;
45861  __ret = (int64_t) __builtin_neon_vcvtnd_s64_f64(__p0);
45862  return __ret;
45863}
45864#endif
45865
45866#ifdef __LITTLE_ENDIAN__
45867__ai uint32_t vcvtns_u32_f32(float32_t __p0) {
45868  uint32_t __ret;
45869  __ret = (uint32_t) __builtin_neon_vcvtns_u32_f32(__p0);
45870  return __ret;
45871}
45872#else
45873__ai uint32_t vcvtns_u32_f32(float32_t __p0) {
45874  uint32_t __ret;
45875  __ret = (uint32_t) __builtin_neon_vcvtns_u32_f32(__p0);
45876  return __ret;
45877}
45878#endif
45879
45880#ifdef __LITTLE_ENDIAN__
45881__ai uint64_t vcvtnd_u64_f64(float64_t __p0) {
45882  uint64_t __ret;
45883  __ret = (uint64_t) __builtin_neon_vcvtnd_u64_f64(__p0);
45884  return __ret;
45885}
45886#else
45887__ai uint64_t vcvtnd_u64_f64(float64_t __p0) {
45888  uint64_t __ret;
45889  __ret = (uint64_t) __builtin_neon_vcvtnd_u64_f64(__p0);
45890  return __ret;
45891}
45892#endif
45893
45894#ifdef __LITTLE_ENDIAN__
45895__ai int32_t vcvtps_s32_f32(float32_t __p0) {
45896  int32_t __ret;
45897  __ret = (int32_t) __builtin_neon_vcvtps_s32_f32(__p0);
45898  return __ret;
45899}
45900#else
45901__ai int32_t vcvtps_s32_f32(float32_t __p0) {
45902  int32_t __ret;
45903  __ret = (int32_t) __builtin_neon_vcvtps_s32_f32(__p0);
45904  return __ret;
45905}
45906#endif
45907
45908#ifdef __LITTLE_ENDIAN__
45909__ai int64_t vcvtpd_s64_f64(float64_t __p0) {
45910  int64_t __ret;
45911  __ret = (int64_t) __builtin_neon_vcvtpd_s64_f64(__p0);
45912  return __ret;
45913}
45914#else
45915__ai int64_t vcvtpd_s64_f64(float64_t __p0) {
45916  int64_t __ret;
45917  __ret = (int64_t) __builtin_neon_vcvtpd_s64_f64(__p0);
45918  return __ret;
45919}
45920#endif
45921
45922#ifdef __LITTLE_ENDIAN__
45923__ai uint32_t vcvtps_u32_f32(float32_t __p0) {
45924  uint32_t __ret;
45925  __ret = (uint32_t) __builtin_neon_vcvtps_u32_f32(__p0);
45926  return __ret;
45927}
45928#else
45929__ai uint32_t vcvtps_u32_f32(float32_t __p0) {
45930  uint32_t __ret;
45931  __ret = (uint32_t) __builtin_neon_vcvtps_u32_f32(__p0);
45932  return __ret;
45933}
45934#endif
45935
45936#ifdef __LITTLE_ENDIAN__
45937__ai uint64_t vcvtpd_u64_f64(float64_t __p0) {
45938  uint64_t __ret;
45939  __ret = (uint64_t) __builtin_neon_vcvtpd_u64_f64(__p0);
45940  return __ret;
45941}
45942#else
45943__ai uint64_t vcvtpd_u64_f64(float64_t __p0) {
45944  uint64_t __ret;
45945  __ret = (uint64_t) __builtin_neon_vcvtpd_u64_f64(__p0);
45946  return __ret;
45947}
45948#endif
45949
45950#ifdef __LITTLE_ENDIAN__
45951__ai float32_t vcvtxd_f32_f64(float64_t __p0) {
45952  float32_t __ret;
45953  __ret = (float32_t) __builtin_neon_vcvtxd_f32_f64(__p0);
45954  return __ret;
45955}
45956#else
45957__ai float32_t vcvtxd_f32_f64(float64_t __p0) {
45958  float32_t __ret;
45959  __ret = (float32_t) __builtin_neon_vcvtxd_f32_f64(__p0);
45960  return __ret;
45961}
45962#endif
45963
45964#ifdef __LITTLE_ENDIAN__
45965__ai float32x2_t vcvtx_f32_f64(float64x2_t __p0) {
45966  float32x2_t __ret;
45967  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__p0, 42);
45968  return __ret;
45969}
45970#else
45971__ai float32x2_t vcvtx_f32_f64(float64x2_t __p0) {
45972  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45973  float32x2_t __ret;
45974  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__rev0, 42);
45975  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
45976  return __ret;
45977}
45978__ai float32x2_t __noswap_vcvtx_f32_f64(float64x2_t __p0) {
45979  float32x2_t __ret;
45980  __ret = (float32x2_t) __builtin_neon_vcvtx_f32_v((int8x16_t)__p0, 42);
45981  return __ret;
45982}
45983#endif
45984
45985#ifdef __LITTLE_ENDIAN__
45986__ai float32x4_t vcvtx_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
45987  float32x4_t __ret;
45988  __ret = vcombine_f32(__p0, vcvtx_f32_f64(__p1));
45989  return __ret;
45990}
45991#else
45992__ai float32x4_t vcvtx_high_f32_f64(float32x2_t __p0, float64x2_t __p1) {
45993  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
45994  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
45995  float32x4_t __ret;
45996  __ret = __noswap_vcombine_f32(__rev0, __noswap_vcvtx_f32_f64(__rev1));
45997  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
45998  return __ret;
45999}
46000#endif
46001
46002#ifdef __LITTLE_ENDIAN__
46003__ai float64x2_t vdivq_f64(float64x2_t __p0, float64x2_t __p1) {
46004  float64x2_t __ret;
46005  __ret = __p0 / __p1;
46006  return __ret;
46007}
46008#else
46009__ai float64x2_t vdivq_f64(float64x2_t __p0, float64x2_t __p1) {
46010  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46011  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46012  float64x2_t __ret;
46013  __ret = __rev0 / __rev1;
46014  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46015  return __ret;
46016}
46017#endif
46018
46019#ifdef __LITTLE_ENDIAN__
46020__ai float32x4_t vdivq_f32(float32x4_t __p0, float32x4_t __p1) {
46021  float32x4_t __ret;
46022  __ret = __p0 / __p1;
46023  return __ret;
46024}
46025#else
46026__ai float32x4_t vdivq_f32(float32x4_t __p0, float32x4_t __p1) {
46027  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
46028  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
46029  float32x4_t __ret;
46030  __ret = __rev0 / __rev1;
46031  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
46032  return __ret;
46033}
46034#endif
46035
46036#ifdef __LITTLE_ENDIAN__
46037__ai float64x1_t vdiv_f64(float64x1_t __p0, float64x1_t __p1) {
46038  float64x1_t __ret;
46039  __ret = __p0 / __p1;
46040  return __ret;
46041}
46042#else
46043__ai float64x1_t vdiv_f64(float64x1_t __p0, float64x1_t __p1) {
46044  float64x1_t __ret;
46045  __ret = __p0 / __p1;
46046  return __ret;
46047}
46048#endif
46049
46050#ifdef __LITTLE_ENDIAN__
46051__ai float32x2_t vdiv_f32(float32x2_t __p0, float32x2_t __p1) {
46052  float32x2_t __ret;
46053  __ret = __p0 / __p1;
46054  return __ret;
46055}
46056#else
46057__ai float32x2_t vdiv_f32(float32x2_t __p0, float32x2_t __p1) {
46058  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
46059  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
46060  float32x2_t __ret;
46061  __ret = __rev0 / __rev1;
46062  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
46063  return __ret;
46064}
46065#endif
46066
46067#ifdef __LITTLE_ENDIAN__
46068#define vdupb_lane_p8(__p0, __p1) __extension__ ({ \
46069  poly8x8_t __s0 = __p0; \
46070  poly8_t __ret; \
46071  __ret = (poly8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
46072  __ret; \
46073})
46074#else
46075#define vdupb_lane_p8(__p0, __p1) __extension__ ({ \
46076  poly8x8_t __s0 = __p0; \
46077  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46078  poly8_t __ret; \
46079  __ret = (poly8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
46080  __ret; \
46081})
46082#endif
46083
46084#ifdef __LITTLE_ENDIAN__
46085#define vduph_lane_p16(__p0, __p1) __extension__ ({ \
46086  poly16x4_t __s0 = __p0; \
46087  poly16_t __ret; \
46088  __ret = (poly16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
46089  __ret; \
46090})
46091#else
46092#define vduph_lane_p16(__p0, __p1) __extension__ ({ \
46093  poly16x4_t __s0 = __p0; \
46094  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46095  poly16_t __ret; \
46096  __ret = (poly16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
46097  __ret; \
46098})
46099#endif
46100
46101#ifdef __LITTLE_ENDIAN__
46102#define vdupb_lane_u8(__p0, __p1) __extension__ ({ \
46103  uint8x8_t __s0 = __p0; \
46104  uint8_t __ret; \
46105  __ret = (uint8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
46106  __ret; \
46107})
46108#else
46109#define vdupb_lane_u8(__p0, __p1) __extension__ ({ \
46110  uint8x8_t __s0 = __p0; \
46111  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46112  uint8_t __ret; \
46113  __ret = (uint8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
46114  __ret; \
46115})
46116#endif
46117
46118#ifdef __LITTLE_ENDIAN__
46119#define vdups_lane_u32(__p0, __p1) __extension__ ({ \
46120  uint32x2_t __s0 = __p0; \
46121  uint32_t __ret; \
46122  __ret = (uint32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__s0, __p1); \
46123  __ret; \
46124})
46125#else
46126#define vdups_lane_u32(__p0, __p1) __extension__ ({ \
46127  uint32x2_t __s0 = __p0; \
46128  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46129  uint32_t __ret; \
46130  __ret = (uint32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__rev0, __p1); \
46131  __ret; \
46132})
46133#endif
46134
46135#ifdef __LITTLE_ENDIAN__
46136#define vdupd_lane_u64(__p0, __p1) __extension__ ({ \
46137  uint64x1_t __s0 = __p0; \
46138  uint64_t __ret; \
46139  __ret = (uint64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
46140  __ret; \
46141})
46142#else
46143#define vdupd_lane_u64(__p0, __p1) __extension__ ({ \
46144  uint64x1_t __s0 = __p0; \
46145  uint64_t __ret; \
46146  __ret = (uint64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
46147  __ret; \
46148})
46149#endif
46150
46151#ifdef __LITTLE_ENDIAN__
46152#define vduph_lane_u16(__p0, __p1) __extension__ ({ \
46153  uint16x4_t __s0 = __p0; \
46154  uint16_t __ret; \
46155  __ret = (uint16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
46156  __ret; \
46157})
46158#else
46159#define vduph_lane_u16(__p0, __p1) __extension__ ({ \
46160  uint16x4_t __s0 = __p0; \
46161  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46162  uint16_t __ret; \
46163  __ret = (uint16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
46164  __ret; \
46165})
46166#endif
46167
46168#ifdef __LITTLE_ENDIAN__
46169#define vdupb_lane_s8(__p0, __p1) __extension__ ({ \
46170  int8x8_t __s0 = __p0; \
46171  int8_t __ret; \
46172  __ret = (int8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__s0, __p1); \
46173  __ret; \
46174})
46175#else
46176#define vdupb_lane_s8(__p0, __p1) __extension__ ({ \
46177  int8x8_t __s0 = __p0; \
46178  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46179  int8_t __ret; \
46180  __ret = (int8_t) __builtin_neon_vdupb_lane_i8((int8x8_t)__rev0, __p1); \
46181  __ret; \
46182})
46183#endif
46184
46185#ifdef __LITTLE_ENDIAN__
46186#define vdupd_lane_f64(__p0, __p1) __extension__ ({ \
46187  float64x1_t __s0 = __p0; \
46188  float64_t __ret; \
46189  __ret = (float64_t) __builtin_neon_vdupd_lane_f64((int8x8_t)__s0, __p1); \
46190  __ret; \
46191})
46192#else
46193#define vdupd_lane_f64(__p0, __p1) __extension__ ({ \
46194  float64x1_t __s0 = __p0; \
46195  float64_t __ret; \
46196  __ret = (float64_t) __builtin_neon_vdupd_lane_f64((int8x8_t)__s0, __p1); \
46197  __ret; \
46198})
46199#endif
46200
46201#ifdef __LITTLE_ENDIAN__
46202#define vdups_lane_f32(__p0, __p1) __extension__ ({ \
46203  float32x2_t __s0 = __p0; \
46204  float32_t __ret; \
46205  __ret = (float32_t) __builtin_neon_vdups_lane_f32((int8x8_t)__s0, __p1); \
46206  __ret; \
46207})
46208#else
46209#define vdups_lane_f32(__p0, __p1) __extension__ ({ \
46210  float32x2_t __s0 = __p0; \
46211  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46212  float32_t __ret; \
46213  __ret = (float32_t) __builtin_neon_vdups_lane_f32((int8x8_t)__rev0, __p1); \
46214  __ret; \
46215})
46216#endif
46217
46218#ifdef __LITTLE_ENDIAN__
46219#define vdups_lane_s32(__p0, __p1) __extension__ ({ \
46220  int32x2_t __s0 = __p0; \
46221  int32_t __ret; \
46222  __ret = (int32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__s0, __p1); \
46223  __ret; \
46224})
46225#else
46226#define vdups_lane_s32(__p0, __p1) __extension__ ({ \
46227  int32x2_t __s0 = __p0; \
46228  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46229  int32_t __ret; \
46230  __ret = (int32_t) __builtin_neon_vdups_lane_i32((int8x8_t)__rev0, __p1); \
46231  __ret; \
46232})
46233#endif
46234
46235#ifdef __LITTLE_ENDIAN__
46236#define vdupd_lane_s64(__p0, __p1) __extension__ ({ \
46237  int64x1_t __s0 = __p0; \
46238  int64_t __ret; \
46239  __ret = (int64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
46240  __ret; \
46241})
46242#else
46243#define vdupd_lane_s64(__p0, __p1) __extension__ ({ \
46244  int64x1_t __s0 = __p0; \
46245  int64_t __ret; \
46246  __ret = (int64_t) __builtin_neon_vdupd_lane_i64((int8x8_t)__s0, __p1); \
46247  __ret; \
46248})
46249#endif
46250
46251#ifdef __LITTLE_ENDIAN__
46252#define vduph_lane_s16(__p0, __p1) __extension__ ({ \
46253  int16x4_t __s0 = __p0; \
46254  int16_t __ret; \
46255  __ret = (int16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__s0, __p1); \
46256  __ret; \
46257})
46258#else
46259#define vduph_lane_s16(__p0, __p1) __extension__ ({ \
46260  int16x4_t __s0 = __p0; \
46261  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46262  int16_t __ret; \
46263  __ret = (int16_t) __builtin_neon_vduph_lane_i16((int8x8_t)__rev0, __p1); \
46264  __ret; \
46265})
46266#endif
46267
46268#ifdef __LITTLE_ENDIAN__
46269#define vdup_lane_p64(__p0, __p1) __extension__ ({ \
46270  poly64x1_t __s0 = __p0; \
46271  poly64x1_t __ret; \
46272  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46273  __ret; \
46274})
46275#else
46276#define vdup_lane_p64(__p0, __p1) __extension__ ({ \
46277  poly64x1_t __s0 = __p0; \
46278  poly64x1_t __ret; \
46279  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46280  __ret; \
46281})
46282#endif
46283
46284#ifdef __LITTLE_ENDIAN__
46285#define vdupq_lane_p64(__p0, __p1) __extension__ ({ \
46286  poly64x1_t __s0 = __p0; \
46287  poly64x2_t __ret; \
46288  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46289  __ret; \
46290})
46291#else
46292#define vdupq_lane_p64(__p0, __p1) __extension__ ({ \
46293  poly64x1_t __s0 = __p0; \
46294  poly64x2_t __ret; \
46295  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46296  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46297  __ret; \
46298})
46299#endif
46300
46301#ifdef __LITTLE_ENDIAN__
46302#define vdupq_lane_f64(__p0, __p1) __extension__ ({ \
46303  float64x1_t __s0 = __p0; \
46304  float64x2_t __ret; \
46305  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46306  __ret; \
46307})
46308#else
46309#define vdupq_lane_f64(__p0, __p1) __extension__ ({ \
46310  float64x1_t __s0 = __p0; \
46311  float64x2_t __ret; \
46312  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46313  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46314  __ret; \
46315})
46316#endif
46317
46318#ifdef __LITTLE_ENDIAN__
46319#define vdupq_lane_f16(__p0, __p1) __extension__ ({ \
46320  float16x4_t __s0 = __p0; \
46321  float16x8_t __ret; \
46322  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46323  __ret; \
46324})
46325#else
46326#define vdupq_lane_f16(__p0, __p1) __extension__ ({ \
46327  float16x4_t __s0 = __p0; \
46328  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46329  float16x8_t __ret; \
46330  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46331  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46332  __ret; \
46333})
46334#endif
46335
46336#ifdef __LITTLE_ENDIAN__
46337#define vdup_lane_f64(__p0, __p1) __extension__ ({ \
46338  float64x1_t __s0 = __p0; \
46339  float64x1_t __ret; \
46340  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46341  __ret; \
46342})
46343#else
46344#define vdup_lane_f64(__p0, __p1) __extension__ ({ \
46345  float64x1_t __s0 = __p0; \
46346  float64x1_t __ret; \
46347  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46348  __ret; \
46349})
46350#endif
46351
46352#ifdef __LITTLE_ENDIAN__
46353#define vdup_lane_f16(__p0, __p1) __extension__ ({ \
46354  float16x4_t __s0 = __p0; \
46355  float16x4_t __ret; \
46356  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46357  __ret; \
46358})
46359#else
46360#define vdup_lane_f16(__p0, __p1) __extension__ ({ \
46361  float16x4_t __s0 = __p0; \
46362  float16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46363  float16x4_t __ret; \
46364  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46365  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46366  __ret; \
46367})
46368#endif
46369
46370#ifdef __LITTLE_ENDIAN__
46371#define vdupb_laneq_p8(__p0, __p1) __extension__ ({ \
46372  poly8x16_t __s0 = __p0; \
46373  poly8_t __ret; \
46374  __ret = (poly8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
46375  __ret; \
46376})
46377#else
46378#define vdupb_laneq_p8(__p0, __p1) __extension__ ({ \
46379  poly8x16_t __s0 = __p0; \
46380  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46381  poly8_t __ret; \
46382  __ret = (poly8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
46383  __ret; \
46384})
46385#endif
46386
46387#ifdef __LITTLE_ENDIAN__
46388#define vduph_laneq_p16(__p0, __p1) __extension__ ({ \
46389  poly16x8_t __s0 = __p0; \
46390  poly16_t __ret; \
46391  __ret = (poly16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
46392  __ret; \
46393})
46394#else
46395#define vduph_laneq_p16(__p0, __p1) __extension__ ({ \
46396  poly16x8_t __s0 = __p0; \
46397  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46398  poly16_t __ret; \
46399  __ret = (poly16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
46400  __ret; \
46401})
46402#endif
46403
46404#ifdef __LITTLE_ENDIAN__
46405#define vdupb_laneq_u8(__p0, __p1) __extension__ ({ \
46406  uint8x16_t __s0 = __p0; \
46407  uint8_t __ret; \
46408  __ret = (uint8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
46409  __ret; \
46410})
46411#else
46412#define vdupb_laneq_u8(__p0, __p1) __extension__ ({ \
46413  uint8x16_t __s0 = __p0; \
46414  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46415  uint8_t __ret; \
46416  __ret = (uint8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
46417  __ret; \
46418})
46419#endif
46420
46421#ifdef __LITTLE_ENDIAN__
46422#define vdups_laneq_u32(__p0, __p1) __extension__ ({ \
46423  uint32x4_t __s0 = __p0; \
46424  uint32_t __ret; \
46425  __ret = (uint32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__s0, __p1); \
46426  __ret; \
46427})
46428#else
46429#define vdups_laneq_u32(__p0, __p1) __extension__ ({ \
46430  uint32x4_t __s0 = __p0; \
46431  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46432  uint32_t __ret; \
46433  __ret = (uint32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__rev0, __p1); \
46434  __ret; \
46435})
46436#endif
46437
46438#ifdef __LITTLE_ENDIAN__
46439#define vdupd_laneq_u64(__p0, __p1) __extension__ ({ \
46440  uint64x2_t __s0 = __p0; \
46441  uint64_t __ret; \
46442  __ret = (uint64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__s0, __p1); \
46443  __ret; \
46444})
46445#else
46446#define vdupd_laneq_u64(__p0, __p1) __extension__ ({ \
46447  uint64x2_t __s0 = __p0; \
46448  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46449  uint64_t __ret; \
46450  __ret = (uint64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__rev0, __p1); \
46451  __ret; \
46452})
46453#endif
46454
46455#ifdef __LITTLE_ENDIAN__
46456#define vduph_laneq_u16(__p0, __p1) __extension__ ({ \
46457  uint16x8_t __s0 = __p0; \
46458  uint16_t __ret; \
46459  __ret = (uint16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
46460  __ret; \
46461})
46462#else
46463#define vduph_laneq_u16(__p0, __p1) __extension__ ({ \
46464  uint16x8_t __s0 = __p0; \
46465  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46466  uint16_t __ret; \
46467  __ret = (uint16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
46468  __ret; \
46469})
46470#endif
46471
46472#ifdef __LITTLE_ENDIAN__
46473#define vdupb_laneq_s8(__p0, __p1) __extension__ ({ \
46474  int8x16_t __s0 = __p0; \
46475  int8_t __ret; \
46476  __ret = (int8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__s0, __p1); \
46477  __ret; \
46478})
46479#else
46480#define vdupb_laneq_s8(__p0, __p1) __extension__ ({ \
46481  int8x16_t __s0 = __p0; \
46482  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46483  int8_t __ret; \
46484  __ret = (int8_t) __builtin_neon_vdupb_laneq_i8((int8x16_t)__rev0, __p1); \
46485  __ret; \
46486})
46487#endif
46488
46489#ifdef __LITTLE_ENDIAN__
46490#define vdupd_laneq_f64(__p0, __p1) __extension__ ({ \
46491  float64x2_t __s0 = __p0; \
46492  float64_t __ret; \
46493  __ret = (float64_t) __builtin_neon_vdupd_laneq_f64((int8x16_t)__s0, __p1); \
46494  __ret; \
46495})
46496#else
46497#define vdupd_laneq_f64(__p0, __p1) __extension__ ({ \
46498  float64x2_t __s0 = __p0; \
46499  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46500  float64_t __ret; \
46501  __ret = (float64_t) __builtin_neon_vdupd_laneq_f64((int8x16_t)__rev0, __p1); \
46502  __ret; \
46503})
46504#endif
46505
46506#ifdef __LITTLE_ENDIAN__
46507#define vdups_laneq_f32(__p0, __p1) __extension__ ({ \
46508  float32x4_t __s0 = __p0; \
46509  float32_t __ret; \
46510  __ret = (float32_t) __builtin_neon_vdups_laneq_f32((int8x16_t)__s0, __p1); \
46511  __ret; \
46512})
46513#else
46514#define vdups_laneq_f32(__p0, __p1) __extension__ ({ \
46515  float32x4_t __s0 = __p0; \
46516  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46517  float32_t __ret; \
46518  __ret = (float32_t) __builtin_neon_vdups_laneq_f32((int8x16_t)__rev0, __p1); \
46519  __ret; \
46520})
46521#endif
46522
46523#ifdef __LITTLE_ENDIAN__
46524#define vdups_laneq_s32(__p0, __p1) __extension__ ({ \
46525  int32x4_t __s0 = __p0; \
46526  int32_t __ret; \
46527  __ret = (int32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__s0, __p1); \
46528  __ret; \
46529})
46530#else
46531#define vdups_laneq_s32(__p0, __p1) __extension__ ({ \
46532  int32x4_t __s0 = __p0; \
46533  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46534  int32_t __ret; \
46535  __ret = (int32_t) __builtin_neon_vdups_laneq_i32((int8x16_t)__rev0, __p1); \
46536  __ret; \
46537})
46538#endif
46539
46540#ifdef __LITTLE_ENDIAN__
46541#define vdupd_laneq_s64(__p0, __p1) __extension__ ({ \
46542  int64x2_t __s0 = __p0; \
46543  int64_t __ret; \
46544  __ret = (int64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__s0, __p1); \
46545  __ret; \
46546})
46547#else
46548#define vdupd_laneq_s64(__p0, __p1) __extension__ ({ \
46549  int64x2_t __s0 = __p0; \
46550  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46551  int64_t __ret; \
46552  __ret = (int64_t) __builtin_neon_vdupd_laneq_i64((int8x16_t)__rev0, __p1); \
46553  __ret; \
46554})
46555#endif
46556
46557#ifdef __LITTLE_ENDIAN__
46558#define vduph_laneq_s16(__p0, __p1) __extension__ ({ \
46559  int16x8_t __s0 = __p0; \
46560  int16_t __ret; \
46561  __ret = (int16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__s0, __p1); \
46562  __ret; \
46563})
46564#else
46565#define vduph_laneq_s16(__p0, __p1) __extension__ ({ \
46566  int16x8_t __s0 = __p0; \
46567  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46568  int16_t __ret; \
46569  __ret = (int16_t) __builtin_neon_vduph_laneq_i16((int8x16_t)__rev0, __p1); \
46570  __ret; \
46571})
46572#endif
46573
46574#ifdef __LITTLE_ENDIAN__
46575#define vdup_laneq_p8(__p0, __p1) __extension__ ({ \
46576  poly8x16_t __s0 = __p0; \
46577  poly8x8_t __ret; \
46578  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46579  __ret; \
46580})
46581#else
46582#define vdup_laneq_p8(__p0, __p1) __extension__ ({ \
46583  poly8x16_t __s0 = __p0; \
46584  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46585  poly8x8_t __ret; \
46586  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46587  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46588  __ret; \
46589})
46590#endif
46591
46592#ifdef __LITTLE_ENDIAN__
46593#define vdup_laneq_p64(__p0, __p1) __extension__ ({ \
46594  poly64x2_t __s0 = __p0; \
46595  poly64x1_t __ret; \
46596  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46597  __ret; \
46598})
46599#else
46600#define vdup_laneq_p64(__p0, __p1) __extension__ ({ \
46601  poly64x2_t __s0 = __p0; \
46602  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46603  poly64x1_t __ret; \
46604  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
46605  __ret; \
46606})
46607#endif
46608
46609#ifdef __LITTLE_ENDIAN__
46610#define vdup_laneq_p16(__p0, __p1) __extension__ ({ \
46611  poly16x8_t __s0 = __p0; \
46612  poly16x4_t __ret; \
46613  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46614  __ret; \
46615})
46616#else
46617#define vdup_laneq_p16(__p0, __p1) __extension__ ({ \
46618  poly16x8_t __s0 = __p0; \
46619  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46620  poly16x4_t __ret; \
46621  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46622  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46623  __ret; \
46624})
46625#endif
46626
46627#ifdef __LITTLE_ENDIAN__
46628#define vdupq_laneq_p8(__p0, __p1) __extension__ ({ \
46629  poly8x16_t __s0 = __p0; \
46630  poly8x16_t __ret; \
46631  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46632  __ret; \
46633})
46634#else
46635#define vdupq_laneq_p8(__p0, __p1) __extension__ ({ \
46636  poly8x16_t __s0 = __p0; \
46637  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46638  poly8x16_t __ret; \
46639  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46640  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46641  __ret; \
46642})
46643#endif
46644
46645#ifdef __LITTLE_ENDIAN__
46646#define vdupq_laneq_p64(__p0, __p1) __extension__ ({ \
46647  poly64x2_t __s0 = __p0; \
46648  poly64x2_t __ret; \
46649  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46650  __ret; \
46651})
46652#else
46653#define vdupq_laneq_p64(__p0, __p1) __extension__ ({ \
46654  poly64x2_t __s0 = __p0; \
46655  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46656  poly64x2_t __ret; \
46657  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46658  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46659  __ret; \
46660})
46661#endif
46662
46663#ifdef __LITTLE_ENDIAN__
46664#define vdupq_laneq_p16(__p0, __p1) __extension__ ({ \
46665  poly16x8_t __s0 = __p0; \
46666  poly16x8_t __ret; \
46667  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46668  __ret; \
46669})
46670#else
46671#define vdupq_laneq_p16(__p0, __p1) __extension__ ({ \
46672  poly16x8_t __s0 = __p0; \
46673  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46674  poly16x8_t __ret; \
46675  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46676  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46677  __ret; \
46678})
46679#endif
46680
46681#ifdef __LITTLE_ENDIAN__
46682#define vdupq_laneq_u8(__p0, __p1) __extension__ ({ \
46683  uint8x16_t __s0 = __p0; \
46684  uint8x16_t __ret; \
46685  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46686  __ret; \
46687})
46688#else
46689#define vdupq_laneq_u8(__p0, __p1) __extension__ ({ \
46690  uint8x16_t __s0 = __p0; \
46691  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46692  uint8x16_t __ret; \
46693  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46694  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46695  __ret; \
46696})
46697#endif
46698
46699#ifdef __LITTLE_ENDIAN__
46700#define vdupq_laneq_u32(__p0, __p1) __extension__ ({ \
46701  uint32x4_t __s0 = __p0; \
46702  uint32x4_t __ret; \
46703  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46704  __ret; \
46705})
46706#else
46707#define vdupq_laneq_u32(__p0, __p1) __extension__ ({ \
46708  uint32x4_t __s0 = __p0; \
46709  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46710  uint32x4_t __ret; \
46711  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46712  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46713  __ret; \
46714})
46715#endif
46716
46717#ifdef __LITTLE_ENDIAN__
46718#define vdupq_laneq_u64(__p0, __p1) __extension__ ({ \
46719  uint64x2_t __s0 = __p0; \
46720  uint64x2_t __ret; \
46721  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46722  __ret; \
46723})
46724#else
46725#define vdupq_laneq_u64(__p0, __p1) __extension__ ({ \
46726  uint64x2_t __s0 = __p0; \
46727  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46728  uint64x2_t __ret; \
46729  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46730  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46731  __ret; \
46732})
46733#endif
46734
46735#ifdef __LITTLE_ENDIAN__
46736#define vdupq_laneq_u16(__p0, __p1) __extension__ ({ \
46737  uint16x8_t __s0 = __p0; \
46738  uint16x8_t __ret; \
46739  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46740  __ret; \
46741})
46742#else
46743#define vdupq_laneq_u16(__p0, __p1) __extension__ ({ \
46744  uint16x8_t __s0 = __p0; \
46745  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46746  uint16x8_t __ret; \
46747  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46748  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46749  __ret; \
46750})
46751#endif
46752
46753#ifdef __LITTLE_ENDIAN__
46754#define vdupq_laneq_s8(__p0, __p1) __extension__ ({ \
46755  int8x16_t __s0 = __p0; \
46756  int8x16_t __ret; \
46757  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46758  __ret; \
46759})
46760#else
46761#define vdupq_laneq_s8(__p0, __p1) __extension__ ({ \
46762  int8x16_t __s0 = __p0; \
46763  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46764  int8x16_t __ret; \
46765  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46766  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46767  __ret; \
46768})
46769#endif
46770
46771#ifdef __LITTLE_ENDIAN__
46772#define vdupq_laneq_f64(__p0, __p1) __extension__ ({ \
46773  float64x2_t __s0 = __p0; \
46774  float64x2_t __ret; \
46775  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46776  __ret; \
46777})
46778#else
46779#define vdupq_laneq_f64(__p0, __p1) __extension__ ({ \
46780  float64x2_t __s0 = __p0; \
46781  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46782  float64x2_t __ret; \
46783  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46784  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46785  __ret; \
46786})
46787#endif
46788
46789#ifdef __LITTLE_ENDIAN__
46790#define vdupq_laneq_f32(__p0, __p1) __extension__ ({ \
46791  float32x4_t __s0 = __p0; \
46792  float32x4_t __ret; \
46793  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46794  __ret; \
46795})
46796#else
46797#define vdupq_laneq_f32(__p0, __p1) __extension__ ({ \
46798  float32x4_t __s0 = __p0; \
46799  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46800  float32x4_t __ret; \
46801  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46802  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46803  __ret; \
46804})
46805#endif
46806
46807#ifdef __LITTLE_ENDIAN__
46808#define vdupq_laneq_f16(__p0, __p1) __extension__ ({ \
46809  float16x8_t __s0 = __p0; \
46810  float16x8_t __ret; \
46811  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46812  __ret; \
46813})
46814#else
46815#define vdupq_laneq_f16(__p0, __p1) __extension__ ({ \
46816  float16x8_t __s0 = __p0; \
46817  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46818  float16x8_t __ret; \
46819  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46820  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46821  __ret; \
46822})
46823#endif
46824
46825#ifdef __LITTLE_ENDIAN__
46826#define vdupq_laneq_s32(__p0, __p1) __extension__ ({ \
46827  int32x4_t __s0 = __p0; \
46828  int32x4_t __ret; \
46829  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46830  __ret; \
46831})
46832#else
46833#define vdupq_laneq_s32(__p0, __p1) __extension__ ({ \
46834  int32x4_t __s0 = __p0; \
46835  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46836  int32x4_t __ret; \
46837  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46838  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46839  __ret; \
46840})
46841#endif
46842
46843#ifdef __LITTLE_ENDIAN__
46844#define vdupq_laneq_s64(__p0, __p1) __extension__ ({ \
46845  int64x2_t __s0 = __p0; \
46846  int64x2_t __ret; \
46847  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46848  __ret; \
46849})
46850#else
46851#define vdupq_laneq_s64(__p0, __p1) __extension__ ({ \
46852  int64x2_t __s0 = __p0; \
46853  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46854  int64x2_t __ret; \
46855  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46856  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46857  __ret; \
46858})
46859#endif
46860
46861#ifdef __LITTLE_ENDIAN__
46862#define vdupq_laneq_s16(__p0, __p1) __extension__ ({ \
46863  int16x8_t __s0 = __p0; \
46864  int16x8_t __ret; \
46865  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46866  __ret; \
46867})
46868#else
46869#define vdupq_laneq_s16(__p0, __p1) __extension__ ({ \
46870  int16x8_t __s0 = __p0; \
46871  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46872  int16x8_t __ret; \
46873  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46874  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46875  __ret; \
46876})
46877#endif
46878
46879#ifdef __LITTLE_ENDIAN__
46880#define vdup_laneq_u8(__p0, __p1) __extension__ ({ \
46881  uint8x16_t __s0 = __p0; \
46882  uint8x8_t __ret; \
46883  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46884  __ret; \
46885})
46886#else
46887#define vdup_laneq_u8(__p0, __p1) __extension__ ({ \
46888  uint8x16_t __s0 = __p0; \
46889  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46890  uint8x8_t __ret; \
46891  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46892  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46893  __ret; \
46894})
46895#endif
46896
46897#ifdef __LITTLE_ENDIAN__
46898#define vdup_laneq_u32(__p0, __p1) __extension__ ({ \
46899  uint32x4_t __s0 = __p0; \
46900  uint32x2_t __ret; \
46901  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46902  __ret; \
46903})
46904#else
46905#define vdup_laneq_u32(__p0, __p1) __extension__ ({ \
46906  uint32x4_t __s0 = __p0; \
46907  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46908  uint32x2_t __ret; \
46909  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46910  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46911  __ret; \
46912})
46913#endif
46914
46915#ifdef __LITTLE_ENDIAN__
46916#define vdup_laneq_u64(__p0, __p1) __extension__ ({ \
46917  uint64x2_t __s0 = __p0; \
46918  uint64x1_t __ret; \
46919  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46920  __ret; \
46921})
46922#else
46923#define vdup_laneq_u64(__p0, __p1) __extension__ ({ \
46924  uint64x2_t __s0 = __p0; \
46925  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46926  uint64x1_t __ret; \
46927  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
46928  __ret; \
46929})
46930#endif
46931
46932#ifdef __LITTLE_ENDIAN__
46933#define vdup_laneq_u16(__p0, __p1) __extension__ ({ \
46934  uint16x8_t __s0 = __p0; \
46935  uint16x4_t __ret; \
46936  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
46937  __ret; \
46938})
46939#else
46940#define vdup_laneq_u16(__p0, __p1) __extension__ ({ \
46941  uint16x8_t __s0 = __p0; \
46942  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
46943  uint16x4_t __ret; \
46944  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
46945  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
46946  __ret; \
46947})
46948#endif
46949
46950#ifdef __LITTLE_ENDIAN__
46951#define vdup_laneq_s8(__p0, __p1) __extension__ ({ \
46952  int8x16_t __s0 = __p0; \
46953  int8x8_t __ret; \
46954  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46955  __ret; \
46956})
46957#else
46958#define vdup_laneq_s8(__p0, __p1) __extension__ ({ \
46959  int8x16_t __s0 = __p0; \
46960  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
46961  int8x8_t __ret; \
46962  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1, __p1, __p1, __p1, __p1); \
46963  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
46964  __ret; \
46965})
46966#endif
46967
46968#ifdef __LITTLE_ENDIAN__
46969#define vdup_laneq_f64(__p0, __p1) __extension__ ({ \
46970  float64x2_t __s0 = __p0; \
46971  float64x1_t __ret; \
46972  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
46973  __ret; \
46974})
46975#else
46976#define vdup_laneq_f64(__p0, __p1) __extension__ ({ \
46977  float64x2_t __s0 = __p0; \
46978  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
46979  float64x1_t __ret; \
46980  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
46981  __ret; \
46982})
46983#endif
46984
46985#ifdef __LITTLE_ENDIAN__
46986#define vdup_laneq_f32(__p0, __p1) __extension__ ({ \
46987  float32x4_t __s0 = __p0; \
46988  float32x2_t __ret; \
46989  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
46990  __ret; \
46991})
46992#else
46993#define vdup_laneq_f32(__p0, __p1) __extension__ ({ \
46994  float32x4_t __s0 = __p0; \
46995  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
46996  float32x2_t __ret; \
46997  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
46998  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
46999  __ret; \
47000})
47001#endif
47002
47003#ifdef __LITTLE_ENDIAN__
47004#define vdup_laneq_f16(__p0, __p1) __extension__ ({ \
47005  float16x8_t __s0 = __p0; \
47006  float16x4_t __ret; \
47007  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
47008  __ret; \
47009})
47010#else
47011#define vdup_laneq_f16(__p0, __p1) __extension__ ({ \
47012  float16x8_t __s0 = __p0; \
47013  float16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
47014  float16x4_t __ret; \
47015  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
47016  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
47017  __ret; \
47018})
47019#endif
47020
47021#ifdef __LITTLE_ENDIAN__
47022#define vdup_laneq_s32(__p0, __p1) __extension__ ({ \
47023  int32x4_t __s0 = __p0; \
47024  int32x2_t __ret; \
47025  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1); \
47026  __ret; \
47027})
47028#else
47029#define vdup_laneq_s32(__p0, __p1) __extension__ ({ \
47030  int32x4_t __s0 = __p0; \
47031  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
47032  int32x2_t __ret; \
47033  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1); \
47034  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47035  __ret; \
47036})
47037#endif
47038
47039#ifdef __LITTLE_ENDIAN__
47040#define vdup_laneq_s64(__p0, __p1) __extension__ ({ \
47041  int64x2_t __s0 = __p0; \
47042  int64x1_t __ret; \
47043  __ret = __builtin_shufflevector(__s0, __s0, __p1); \
47044  __ret; \
47045})
47046#else
47047#define vdup_laneq_s64(__p0, __p1) __extension__ ({ \
47048  int64x2_t __s0 = __p0; \
47049  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47050  int64x1_t __ret; \
47051  __ret = __builtin_shufflevector(__rev0, __rev0, __p1); \
47052  __ret; \
47053})
47054#endif
47055
47056#ifdef __LITTLE_ENDIAN__
47057#define vdup_laneq_s16(__p0, __p1) __extension__ ({ \
47058  int16x8_t __s0 = __p0; \
47059  int16x4_t __ret; \
47060  __ret = __builtin_shufflevector(__s0, __s0, __p1, __p1, __p1, __p1); \
47061  __ret; \
47062})
47063#else
47064#define vdup_laneq_s16(__p0, __p1) __extension__ ({ \
47065  int16x8_t __s0 = __p0; \
47066  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
47067  int16x4_t __ret; \
47068  __ret = __builtin_shufflevector(__rev0, __rev0, __p1, __p1, __p1, __p1); \
47069  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
47070  __ret; \
47071})
47072#endif
47073
47074#ifdef __LITTLE_ENDIAN__
47075__ai poly64x1_t vdup_n_p64(poly64_t __p0) {
47076  poly64x1_t __ret;
47077  __ret = (poly64x1_t) {__p0};
47078  return __ret;
47079}
47080#else
47081__ai poly64x1_t vdup_n_p64(poly64_t __p0) {
47082  poly64x1_t __ret;
47083  __ret = (poly64x1_t) {__p0};
47084  return __ret;
47085}
47086#endif
47087
47088#ifdef __LITTLE_ENDIAN__
47089__ai poly64x2_t vdupq_n_p64(poly64_t __p0) {
47090  poly64x2_t __ret;
47091  __ret = (poly64x2_t) {__p0, __p0};
47092  return __ret;
47093}
47094#else
47095__ai poly64x2_t vdupq_n_p64(poly64_t __p0) {
47096  poly64x2_t __ret;
47097  __ret = (poly64x2_t) {__p0, __p0};
47098  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47099  return __ret;
47100}
47101#endif
47102
47103#ifdef __LITTLE_ENDIAN__
47104__ai float64x2_t vdupq_n_f64(float64_t __p0) {
47105  float64x2_t __ret;
47106  __ret = (float64x2_t) {__p0, __p0};
47107  return __ret;
47108}
47109#else
47110__ai float64x2_t vdupq_n_f64(float64_t __p0) {
47111  float64x2_t __ret;
47112  __ret = (float64x2_t) {__p0, __p0};
47113  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47114  return __ret;
47115}
47116#endif
47117
47118#ifdef __LITTLE_ENDIAN__
47119__ai float64x1_t vdup_n_f64(float64_t __p0) {
47120  float64x1_t __ret;
47121  __ret = (float64x1_t) {__p0};
47122  return __ret;
47123}
47124#else
47125__ai float64x1_t vdup_n_f64(float64_t __p0) {
47126  float64x1_t __ret;
47127  __ret = (float64x1_t) {__p0};
47128  return __ret;
47129}
47130#endif
47131
47132#ifdef __LITTLE_ENDIAN__
47133#define vext_p64(__p0, __p1, __p2) __extension__ ({ \
47134  poly64x1_t __s0 = __p0; \
47135  poly64x1_t __s1 = __p1; \
47136  poly64x1_t __ret; \
47137  __ret = (poly64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
47138  __ret; \
47139})
47140#else
47141#define vext_p64(__p0, __p1, __p2) __extension__ ({ \
47142  poly64x1_t __s0 = __p0; \
47143  poly64x1_t __s1 = __p1; \
47144  poly64x1_t __ret; \
47145  __ret = (poly64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
47146  __ret; \
47147})
47148#endif
47149
47150#ifdef __LITTLE_ENDIAN__
47151#define vextq_p64(__p0, __p1, __p2) __extension__ ({ \
47152  poly64x2_t __s0 = __p0; \
47153  poly64x2_t __s1 = __p1; \
47154  poly64x2_t __ret; \
47155  __ret = (poly64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
47156  __ret; \
47157})
47158#else
47159#define vextq_p64(__p0, __p1, __p2) __extension__ ({ \
47160  poly64x2_t __s0 = __p0; \
47161  poly64x2_t __s1 = __p1; \
47162  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47163  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47164  poly64x2_t __ret; \
47165  __ret = (poly64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
47166  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47167  __ret; \
47168})
47169#endif
47170
47171#ifdef __LITTLE_ENDIAN__
47172#define vextq_f64(__p0, __p1, __p2) __extension__ ({ \
47173  float64x2_t __s0 = __p0; \
47174  float64x2_t __s1 = __p1; \
47175  float64x2_t __ret; \
47176  __ret = (float64x2_t) __builtin_neon_vextq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 42); \
47177  __ret; \
47178})
47179#else
47180#define vextq_f64(__p0, __p1, __p2) __extension__ ({ \
47181  float64x2_t __s0 = __p0; \
47182  float64x2_t __s1 = __p1; \
47183  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47184  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47185  float64x2_t __ret; \
47186  __ret = (float64x2_t) __builtin_neon_vextq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 42); \
47187  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47188  __ret; \
47189})
47190#endif
47191
47192#ifdef __LITTLE_ENDIAN__
47193#define vext_f64(__p0, __p1, __p2) __extension__ ({ \
47194  float64x1_t __s0 = __p0; \
47195  float64x1_t __s1 = __p1; \
47196  float64x1_t __ret; \
47197  __ret = (float64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
47198  __ret; \
47199})
47200#else
47201#define vext_f64(__p0, __p1, __p2) __extension__ ({ \
47202  float64x1_t __s0 = __p0; \
47203  float64x1_t __s1 = __p1; \
47204  float64x1_t __ret; \
47205  __ret = (float64x1_t) __builtin_neon_vext_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
47206  __ret; \
47207})
47208#endif
47209
47210#ifdef __LITTLE_ENDIAN__
47211__ai float64x2_t vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47212  float64x2_t __ret;
47213  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
47214  return __ret;
47215}
47216#else
47217__ai float64x2_t vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47218  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47219  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47220  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
47221  float64x2_t __ret;
47222  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
47223  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47224  return __ret;
47225}
47226__ai float64x2_t __noswap_vfmaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47227  float64x2_t __ret;
47228  __ret = (float64x2_t) __builtin_neon_vfmaq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
47229  return __ret;
47230}
47231#endif
47232
47233#ifdef __LITTLE_ENDIAN__
47234__ai float64x1_t vfma_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
47235  float64x1_t __ret;
47236  __ret = (float64x1_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
47237  return __ret;
47238}
47239#else
47240__ai float64x1_t vfma_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
47241  float64x1_t __ret;
47242  __ret = (float64x1_t) __builtin_neon_vfma_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
47243  return __ret;
47244}
47245#endif
47246
47247#ifdef __LITTLE_ENDIAN__
47248#define vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47249  float64_t __s0 = __p0; \
47250  float64_t __s1 = __p1; \
47251  float64x1_t __s2 = __p2; \
47252  float64_t __ret; \
47253  __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
47254  __ret; \
47255})
47256#else
47257#define vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47258  float64_t __s0 = __p0; \
47259  float64_t __s1 = __p1; \
47260  float64x1_t __s2 = __p2; \
47261  float64_t __ret; \
47262  __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
47263  __ret; \
47264})
47265#define __noswap_vfmad_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47266  float64_t __s0 = __p0; \
47267  float64_t __s1 = __p1; \
47268  float64x1_t __s2 = __p2; \
47269  float64_t __ret; \
47270  __ret = (float64_t) __builtin_neon_vfmad_lane_f64(__s0, __s1, (int8x8_t)__s2, __p3); \
47271  __ret; \
47272})
47273#endif
47274
47275#ifdef __LITTLE_ENDIAN__
47276#define vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47277  float32_t __s0 = __p0; \
47278  float32_t __s1 = __p1; \
47279  float32x2_t __s2 = __p2; \
47280  float32_t __ret; \
47281  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__s2, __p3); \
47282  __ret; \
47283})
47284#else
47285#define vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47286  float32_t __s0 = __p0; \
47287  float32_t __s1 = __p1; \
47288  float32x2_t __s2 = __p2; \
47289  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47290  float32_t __ret; \
47291  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__rev2, __p3); \
47292  __ret; \
47293})
47294#define __noswap_vfmas_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47295  float32_t __s0 = __p0; \
47296  float32_t __s1 = __p1; \
47297  float32x2_t __s2 = __p2; \
47298  float32_t __ret; \
47299  __ret = (float32_t) __builtin_neon_vfmas_lane_f32(__s0, __s1, (int8x8_t)__s2, __p3); \
47300  __ret; \
47301})
47302#endif
47303
47304#ifdef __LITTLE_ENDIAN__
47305#define vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47306  float64x2_t __s0 = __p0; \
47307  float64x2_t __s1 = __p1; \
47308  float64x1_t __s2 = __p2; \
47309  float64x2_t __ret; \
47310  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 42); \
47311  __ret; \
47312})
47313#else
47314#define vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47315  float64x2_t __s0 = __p0; \
47316  float64x2_t __s1 = __p1; \
47317  float64x1_t __s2 = __p2; \
47318  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47319  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47320  float64x2_t __ret; \
47321  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__s2, __p3, 42); \
47322  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47323  __ret; \
47324})
47325#define __noswap_vfmaq_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47326  float64x2_t __s0 = __p0; \
47327  float64x2_t __s1 = __p1; \
47328  float64x1_t __s2 = __p2; \
47329  float64x2_t __ret; \
47330  __ret = (float64x2_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 42); \
47331  __ret; \
47332})
47333#endif
47334
47335#ifdef __LITTLE_ENDIAN__
47336#define vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47337  float32x4_t __s0 = __p0; \
47338  float32x4_t __s1 = __p1; \
47339  float32x2_t __s2 = __p2; \
47340  float32x4_t __ret; \
47341  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 41); \
47342  __ret; \
47343})
47344#else
47345#define vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47346  float32x4_t __s0 = __p0; \
47347  float32x4_t __s1 = __p1; \
47348  float32x2_t __s2 = __p2; \
47349  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
47350  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
47351  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47352  float32x4_t __ret; \
47353  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, __p3, 41); \
47354  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
47355  __ret; \
47356})
47357#define __noswap_vfmaq_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47358  float32x4_t __s0 = __p0; \
47359  float32x4_t __s1 = __p1; \
47360  float32x2_t __s2 = __p2; \
47361  float32x4_t __ret; \
47362  __ret = (float32x4_t) __builtin_neon_vfmaq_lane_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x8_t)__s2, __p3, 41); \
47363  __ret; \
47364})
47365#endif
47366
47367#ifdef __LITTLE_ENDIAN__
47368#define vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47369  float64x1_t __s0 = __p0; \
47370  float64x1_t __s1 = __p1; \
47371  float64x1_t __s2 = __p2; \
47372  float64x1_t __ret; \
47373  __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
47374  __ret; \
47375})
47376#else
47377#define vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47378  float64x1_t __s0 = __p0; \
47379  float64x1_t __s1 = __p1; \
47380  float64x1_t __s2 = __p2; \
47381  float64x1_t __ret; \
47382  __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
47383  __ret; \
47384})
47385#define __noswap_vfma_lane_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47386  float64x1_t __s0 = __p0; \
47387  float64x1_t __s1 = __p1; \
47388  float64x1_t __s2 = __p2; \
47389  float64x1_t __ret; \
47390  __ret = (float64x1_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 10); \
47391  __ret; \
47392})
47393#endif
47394
47395#ifdef __LITTLE_ENDIAN__
47396#define vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47397  float32x2_t __s0 = __p0; \
47398  float32x2_t __s1 = __p1; \
47399  float32x2_t __s2 = __p2; \
47400  float32x2_t __ret; \
47401  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 9); \
47402  __ret; \
47403})
47404#else
47405#define vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47406  float32x2_t __s0 = __p0; \
47407  float32x2_t __s1 = __p1; \
47408  float32x2_t __s2 = __p2; \
47409  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47410  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47411  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47412  float32x2_t __ret; \
47413  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, __p3, 9); \
47414  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47415  __ret; \
47416})
47417#define __noswap_vfma_lane_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47418  float32x2_t __s0 = __p0; \
47419  float32x2_t __s1 = __p1; \
47420  float32x2_t __s2 = __p2; \
47421  float32x2_t __ret; \
47422  __ret = (float32x2_t) __builtin_neon_vfma_lane_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x8_t)__s2, __p3, 9); \
47423  __ret; \
47424})
47425#endif
47426
47427#ifdef __LITTLE_ENDIAN__
47428#define vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47429  float64_t __s0 = __p0; \
47430  float64_t __s1 = __p1; \
47431  float64x2_t __s2 = __p2; \
47432  float64_t __ret; \
47433  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__s2, __p3); \
47434  __ret; \
47435})
47436#else
47437#define vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47438  float64_t __s0 = __p0; \
47439  float64_t __s1 = __p1; \
47440  float64x2_t __s2 = __p2; \
47441  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47442  float64_t __ret; \
47443  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__rev2, __p3); \
47444  __ret; \
47445})
47446#define __noswap_vfmad_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47447  float64_t __s0 = __p0; \
47448  float64_t __s1 = __p1; \
47449  float64x2_t __s2 = __p2; \
47450  float64_t __ret; \
47451  __ret = (float64_t) __builtin_neon_vfmad_laneq_f64(__s0, __s1, (int8x16_t)__s2, __p3); \
47452  __ret; \
47453})
47454#endif
47455
47456#ifdef __LITTLE_ENDIAN__
47457#define vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47458  float32_t __s0 = __p0; \
47459  float32_t __s1 = __p1; \
47460  float32x4_t __s2 = __p2; \
47461  float32_t __ret; \
47462  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__s2, __p3); \
47463  __ret; \
47464})
47465#else
47466#define vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47467  float32_t __s0 = __p0; \
47468  float32_t __s1 = __p1; \
47469  float32x4_t __s2 = __p2; \
47470  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
47471  float32_t __ret; \
47472  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__rev2, __p3); \
47473  __ret; \
47474})
47475#define __noswap_vfmas_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47476  float32_t __s0 = __p0; \
47477  float32_t __s1 = __p1; \
47478  float32x4_t __s2 = __p2; \
47479  float32_t __ret; \
47480  __ret = (float32_t) __builtin_neon_vfmas_laneq_f32(__s0, __s1, (int8x16_t)__s2, __p3); \
47481  __ret; \
47482})
47483#endif
47484
47485#ifdef __LITTLE_ENDIAN__
47486#define vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47487  float64x2_t __s0 = __p0; \
47488  float64x2_t __s1 = __p1; \
47489  float64x2_t __s2 = __p2; \
47490  float64x2_t __ret; \
47491  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 42); \
47492  __ret; \
47493})
47494#else
47495#define vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47496  float64x2_t __s0 = __p0; \
47497  float64x2_t __s1 = __p1; \
47498  float64x2_t __s2 = __p2; \
47499  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47500  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47501  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47502  float64x2_t __ret; \
47503  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 42); \
47504  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47505  __ret; \
47506})
47507#define __noswap_vfmaq_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47508  float64x2_t __s0 = __p0; \
47509  float64x2_t __s1 = __p1; \
47510  float64x2_t __s2 = __p2; \
47511  float64x2_t __ret; \
47512  __ret = (float64x2_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 42); \
47513  __ret; \
47514})
47515#endif
47516
47517#ifdef __LITTLE_ENDIAN__
47518#define vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47519  float32x4_t __s0 = __p0; \
47520  float32x4_t __s1 = __p1; \
47521  float32x4_t __s2 = __p2; \
47522  float32x4_t __ret; \
47523  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 41); \
47524  __ret; \
47525})
47526#else
47527#define vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47528  float32x4_t __s0 = __p0; \
47529  float32x4_t __s1 = __p1; \
47530  float32x4_t __s2 = __p2; \
47531  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
47532  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
47533  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
47534  float32x4_t __ret; \
47535  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 41); \
47536  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
47537  __ret; \
47538})
47539#define __noswap_vfmaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47540  float32x4_t __s0 = __p0; \
47541  float32x4_t __s1 = __p1; \
47542  float32x4_t __s2 = __p2; \
47543  float32x4_t __ret; \
47544  __ret = (float32x4_t) __builtin_neon_vfmaq_laneq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 41); \
47545  __ret; \
47546})
47547#endif
47548
47549#ifdef __LITTLE_ENDIAN__
47550#define vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47551  float64x1_t __s0 = __p0; \
47552  float64x1_t __s1 = __p1; \
47553  float64x2_t __s2 = __p2; \
47554  float64x1_t __ret; \
47555  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 10); \
47556  __ret; \
47557})
47558#else
47559#define vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47560  float64x1_t __s0 = __p0; \
47561  float64x1_t __s1 = __p1; \
47562  float64x2_t __s2 = __p2; \
47563  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
47564  float64x1_t __ret; \
47565  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__rev2, __p3, 10); \
47566  __ret; \
47567})
47568#define __noswap_vfma_laneq_f64(__p0, __p1, __p2, __p3) __extension__ ({ \
47569  float64x1_t __s0 = __p0; \
47570  float64x1_t __s1 = __p1; \
47571  float64x2_t __s2 = __p2; \
47572  float64x1_t __ret; \
47573  __ret = (float64x1_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 10); \
47574  __ret; \
47575})
47576#endif
47577
47578#ifdef __LITTLE_ENDIAN__
47579#define vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47580  float32x2_t __s0 = __p0; \
47581  float32x2_t __s1 = __p1; \
47582  float32x4_t __s2 = __p2; \
47583  float32x2_t __ret; \
47584  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 9); \
47585  __ret; \
47586})
47587#else
47588#define vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47589  float32x2_t __s0 = __p0; \
47590  float32x2_t __s1 = __p1; \
47591  float32x4_t __s2 = __p2; \
47592  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
47593  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
47594  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
47595  float32x2_t __ret; \
47596  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x16_t)__rev2, __p3, 9); \
47597  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
47598  __ret; \
47599})
47600#define __noswap_vfma_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
47601  float32x2_t __s0 = __p0; \
47602  float32x2_t __s1 = __p1; \
47603  float32x4_t __s2 = __p2; \
47604  float32x2_t __ret; \
47605  __ret = (float32x2_t) __builtin_neon_vfma_laneq_v((int8x8_t)__s0, (int8x8_t)__s1, (int8x16_t)__s2, __p3, 9); \
47606  __ret; \
47607})
47608#endif
47609
47610#ifdef __LITTLE_ENDIAN__
47611__ai float64x2_t vfmaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
47612  float64x2_t __ret;
47613  __ret = vfmaq_f64(__p0, __p1, (float64x2_t) {__p2, __p2});
47614  return __ret;
47615}
47616#else
47617__ai float64x2_t vfmaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
47618  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47619  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47620  float64x2_t __ret;
47621  __ret = __noswap_vfmaq_f64(__rev0, __rev1, (float64x2_t) {__p2, __p2});
47622  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47623  return __ret;
47624}
47625#endif
47626
47627#ifdef __LITTLE_ENDIAN__
47628__ai float32x4_t vfmaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
47629  float32x4_t __ret;
47630  __ret = vfmaq_f32(__p0, __p1, (float32x4_t) {__p2, __p2, __p2, __p2});
47631  return __ret;
47632}
47633#else
47634__ai float32x4_t vfmaq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
47635  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
47636  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
47637  float32x4_t __ret;
47638  __ret = __noswap_vfmaq_f32(__rev0, __rev1, (float32x4_t) {__p2, __p2, __p2, __p2});
47639  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
47640  return __ret;
47641}
47642#endif
47643
47644#ifdef __LITTLE_ENDIAN__
47645__ai float32x2_t vfma_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
47646  float32x2_t __ret;
47647  __ret = vfma_f32(__p0, __p1, (float32x2_t) {__p2, __p2});
47648  return __ret;
47649}
47650#else
47651__ai float32x2_t vfma_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
47652  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47653  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47654  float32x2_t __ret;
47655  __ret = __noswap_vfma_f32(__rev0, __rev1, (float32x2_t) {__p2, __p2});
47656  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47657  return __ret;
47658}
47659#endif
47660
47661#ifdef __LITTLE_ENDIAN__
47662__ai float64x2_t vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47663  float64x2_t __ret;
47664  __ret = (float64x2_t) __builtin_neon_vfmsq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
47665  return __ret;
47666}
47667#else
47668__ai float64x2_t vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47669  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47670  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47671  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
47672  float64x2_t __ret;
47673  __ret = (float64x2_t) __builtin_neon_vfmsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 42);
47674  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47675  return __ret;
47676}
47677__ai float64x2_t __noswap_vfmsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
47678  float64x2_t __ret;
47679  __ret = (float64x2_t) __builtin_neon_vfmsq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 42);
47680  return __ret;
47681}
47682#endif
47683
47684#ifdef __LITTLE_ENDIAN__
47685__ai float32x4_t vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
47686  float32x4_t __ret;
47687  __ret = (float32x4_t) __builtin_neon_vfmsq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
47688  return __ret;
47689}
47690#else
47691__ai float32x4_t vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
47692  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
47693  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
47694  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
47695  float32x4_t __ret;
47696  __ret = (float32x4_t) __builtin_neon_vfmsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 41);
47697  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
47698  return __ret;
47699}
47700__ai float32x4_t __noswap_vfmsq_f32(float32x4_t __p0, float32x4_t __p1, float32x4_t __p2) {
47701  float32x4_t __ret;
47702  __ret = (float32x4_t) __builtin_neon_vfmsq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 41);
47703  return __ret;
47704}
47705#endif
47706
47707#ifdef __LITTLE_ENDIAN__
47708__ai float64x1_t vfms_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
47709  float64x1_t __ret;
47710  __ret = (float64x1_t) __builtin_neon_vfms_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
47711  return __ret;
47712}
47713#else
47714__ai float64x1_t vfms_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
47715  float64x1_t __ret;
47716  __ret = (float64x1_t) __builtin_neon_vfms_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 10);
47717  return __ret;
47718}
47719#endif
47720
47721#ifdef __LITTLE_ENDIAN__
47722__ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
47723  float32x2_t __ret;
47724  __ret = (float32x2_t) __builtin_neon_vfms_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
47725  return __ret;
47726}
47727#else
47728__ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
47729  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
47730  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
47731  float32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
47732  float32x2_t __ret;
47733  __ret = (float32x2_t) __builtin_neon_vfms_v((int8x8_t)__rev0, (int8x8_t)__rev1, (int8x8_t)__rev2, 9);
47734  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
47735  return __ret;
47736}
47737__ai float32x2_t __noswap_vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) {
47738  float32x2_t __ret;
47739  __ret = (float32x2_t) __builtin_neon_vfms_v((int8x8_t)__p0, (int8x8_t)__p1, (int8x8_t)__p2, 9);
47740  return __ret;
47741}
47742#endif
47743
47744#ifdef __LITTLE_ENDIAN__
47745#define vfmsd_lane_f64(__p0_88, __p1_88, __p2_88, __p3_88) __extension__ ({ \
47746  float64_t __s0_88 = __p0_88; \
47747  float64_t __s1_88 = __p1_88; \
47748  float64x1_t __s2_88 = __p2_88; \
47749  float64_t __ret_88; \
47750  __ret_88 = vfmad_lane_f64(__s0_88, __s1_88, -__s2_88, __p3_88); \
47751  __ret_88; \
47752})
47753#else
47754#define vfmsd_lane_f64(__p0_89, __p1_89, __p2_89, __p3_89) __extension__ ({ \
47755  float64_t __s0_89 = __p0_89; \
47756  float64_t __s1_89 = __p1_89; \
47757  float64x1_t __s2_89 = __p2_89; \
47758  float64_t __ret_89; \
47759  __ret_89 = __noswap_vfmad_lane_f64(__s0_89, __s1_89, -__s2_89, __p3_89); \
47760  __ret_89; \
47761})
47762#endif
47763
47764#ifdef __LITTLE_ENDIAN__
47765#define vfmss_lane_f32(__p0_90, __p1_90, __p2_90, __p3_90) __extension__ ({ \
47766  float32_t __s0_90 = __p0_90; \
47767  float32_t __s1_90 = __p1_90; \
47768  float32x2_t __s2_90 = __p2_90; \
47769  float32_t __ret_90; \
47770  __ret_90 = vfmas_lane_f32(__s0_90, __s1_90, -__s2_90, __p3_90); \
47771  __ret_90; \
47772})
47773#else
47774#define vfmss_lane_f32(__p0_91, __p1_91, __p2_91, __p3_91) __extension__ ({ \
47775  float32_t __s0_91 = __p0_91; \
47776  float32_t __s1_91 = __p1_91; \
47777  float32x2_t __s2_91 = __p2_91; \
47778  float32x2_t __rev2_91;  __rev2_91 = __builtin_shufflevector(__s2_91, __s2_91, 1, 0); \
47779  float32_t __ret_91; \
47780  __ret_91 = __noswap_vfmas_lane_f32(__s0_91, __s1_91, -__rev2_91, __p3_91); \
47781  __ret_91; \
47782})
47783#endif
47784
47785#ifdef __LITTLE_ENDIAN__
47786#define vfmsq_lane_f64(__p0_92, __p1_92, __p2_92, __p3_92) __extension__ ({ \
47787  float64x2_t __s0_92 = __p0_92; \
47788  float64x2_t __s1_92 = __p1_92; \
47789  float64x1_t __s2_92 = __p2_92; \
47790  float64x2_t __ret_92; \
47791  __ret_92 = vfmaq_lane_f64(__s0_92, __s1_92, -__s2_92, __p3_92); \
47792  __ret_92; \
47793})
47794#else
47795#define vfmsq_lane_f64(__p0_93, __p1_93, __p2_93, __p3_93) __extension__ ({ \
47796  float64x2_t __s0_93 = __p0_93; \
47797  float64x2_t __s1_93 = __p1_93; \
47798  float64x1_t __s2_93 = __p2_93; \
47799  float64x2_t __rev0_93;  __rev0_93 = __builtin_shufflevector(__s0_93, __s0_93, 1, 0); \
47800  float64x2_t __rev1_93;  __rev1_93 = __builtin_shufflevector(__s1_93, __s1_93, 1, 0); \
47801  float64x2_t __ret_93; \
47802  __ret_93 = __noswap_vfmaq_lane_f64(__rev0_93, __rev1_93, -__s2_93, __p3_93); \
47803  __ret_93 = __builtin_shufflevector(__ret_93, __ret_93, 1, 0); \
47804  __ret_93; \
47805})
47806#endif
47807
47808#ifdef __LITTLE_ENDIAN__
47809#define vfmsq_lane_f32(__p0_94, __p1_94, __p2_94, __p3_94) __extension__ ({ \
47810  float32x4_t __s0_94 = __p0_94; \
47811  float32x4_t __s1_94 = __p1_94; \
47812  float32x2_t __s2_94 = __p2_94; \
47813  float32x4_t __ret_94; \
47814  __ret_94 = vfmaq_lane_f32(__s0_94, __s1_94, -__s2_94, __p3_94); \
47815  __ret_94; \
47816})
47817#else
47818#define vfmsq_lane_f32(__p0_95, __p1_95, __p2_95, __p3_95) __extension__ ({ \
47819  float32x4_t __s0_95 = __p0_95; \
47820  float32x4_t __s1_95 = __p1_95; \
47821  float32x2_t __s2_95 = __p2_95; \
47822  float32x4_t __rev0_95;  __rev0_95 = __builtin_shufflevector(__s0_95, __s0_95, 3, 2, 1, 0); \
47823  float32x4_t __rev1_95;  __rev1_95 = __builtin_shufflevector(__s1_95, __s1_95, 3, 2, 1, 0); \
47824  float32x2_t __rev2_95;  __rev2_95 = __builtin_shufflevector(__s2_95, __s2_95, 1, 0); \
47825  float32x4_t __ret_95; \
47826  __ret_95 = __noswap_vfmaq_lane_f32(__rev0_95, __rev1_95, -__rev2_95, __p3_95); \
47827  __ret_95 = __builtin_shufflevector(__ret_95, __ret_95, 3, 2, 1, 0); \
47828  __ret_95; \
47829})
47830#endif
47831
47832#ifdef __LITTLE_ENDIAN__
47833#define vfms_lane_f64(__p0_96, __p1_96, __p2_96, __p3_96) __extension__ ({ \
47834  float64x1_t __s0_96 = __p0_96; \
47835  float64x1_t __s1_96 = __p1_96; \
47836  float64x1_t __s2_96 = __p2_96; \
47837  float64x1_t __ret_96; \
47838  __ret_96 = vfma_lane_f64(__s0_96, __s1_96, -__s2_96, __p3_96); \
47839  __ret_96; \
47840})
47841#else
47842#define vfms_lane_f64(__p0_97, __p1_97, __p2_97, __p3_97) __extension__ ({ \
47843  float64x1_t __s0_97 = __p0_97; \
47844  float64x1_t __s1_97 = __p1_97; \
47845  float64x1_t __s2_97 = __p2_97; \
47846  float64x1_t __ret_97; \
47847  __ret_97 = __noswap_vfma_lane_f64(__s0_97, __s1_97, -__s2_97, __p3_97); \
47848  __ret_97; \
47849})
47850#endif
47851
47852#ifdef __LITTLE_ENDIAN__
47853#define vfms_lane_f32(__p0_98, __p1_98, __p2_98, __p3_98) __extension__ ({ \
47854  float32x2_t __s0_98 = __p0_98; \
47855  float32x2_t __s1_98 = __p1_98; \
47856  float32x2_t __s2_98 = __p2_98; \
47857  float32x2_t __ret_98; \
47858  __ret_98 = vfma_lane_f32(__s0_98, __s1_98, -__s2_98, __p3_98); \
47859  __ret_98; \
47860})
47861#else
47862#define vfms_lane_f32(__p0_99, __p1_99, __p2_99, __p3_99) __extension__ ({ \
47863  float32x2_t __s0_99 = __p0_99; \
47864  float32x2_t __s1_99 = __p1_99; \
47865  float32x2_t __s2_99 = __p2_99; \
47866  float32x2_t __rev0_99;  __rev0_99 = __builtin_shufflevector(__s0_99, __s0_99, 1, 0); \
47867  float32x2_t __rev1_99;  __rev1_99 = __builtin_shufflevector(__s1_99, __s1_99, 1, 0); \
47868  float32x2_t __rev2_99;  __rev2_99 = __builtin_shufflevector(__s2_99, __s2_99, 1, 0); \
47869  float32x2_t __ret_99; \
47870  __ret_99 = __noswap_vfma_lane_f32(__rev0_99, __rev1_99, -__rev2_99, __p3_99); \
47871  __ret_99 = __builtin_shufflevector(__ret_99, __ret_99, 1, 0); \
47872  __ret_99; \
47873})
47874#endif
47875
47876#ifdef __LITTLE_ENDIAN__
47877#define vfmsd_laneq_f64(__p0_100, __p1_100, __p2_100, __p3_100) __extension__ ({ \
47878  float64_t __s0_100 = __p0_100; \
47879  float64_t __s1_100 = __p1_100; \
47880  float64x2_t __s2_100 = __p2_100; \
47881  float64_t __ret_100; \
47882  __ret_100 = vfmad_laneq_f64(__s0_100, __s1_100, -__s2_100, __p3_100); \
47883  __ret_100; \
47884})
47885#else
47886#define vfmsd_laneq_f64(__p0_101, __p1_101, __p2_101, __p3_101) __extension__ ({ \
47887  float64_t __s0_101 = __p0_101; \
47888  float64_t __s1_101 = __p1_101; \
47889  float64x2_t __s2_101 = __p2_101; \
47890  float64x2_t __rev2_101;  __rev2_101 = __builtin_shufflevector(__s2_101, __s2_101, 1, 0); \
47891  float64_t __ret_101; \
47892  __ret_101 = __noswap_vfmad_laneq_f64(__s0_101, __s1_101, -__rev2_101, __p3_101); \
47893  __ret_101; \
47894})
47895#endif
47896
47897#ifdef __LITTLE_ENDIAN__
47898#define vfmss_laneq_f32(__p0_102, __p1_102, __p2_102, __p3_102) __extension__ ({ \
47899  float32_t __s0_102 = __p0_102; \
47900  float32_t __s1_102 = __p1_102; \
47901  float32x4_t __s2_102 = __p2_102; \
47902  float32_t __ret_102; \
47903  __ret_102 = vfmas_laneq_f32(__s0_102, __s1_102, -__s2_102, __p3_102); \
47904  __ret_102; \
47905})
47906#else
47907#define vfmss_laneq_f32(__p0_103, __p1_103, __p2_103, __p3_103) __extension__ ({ \
47908  float32_t __s0_103 = __p0_103; \
47909  float32_t __s1_103 = __p1_103; \
47910  float32x4_t __s2_103 = __p2_103; \
47911  float32x4_t __rev2_103;  __rev2_103 = __builtin_shufflevector(__s2_103, __s2_103, 3, 2, 1, 0); \
47912  float32_t __ret_103; \
47913  __ret_103 = __noswap_vfmas_laneq_f32(__s0_103, __s1_103, -__rev2_103, __p3_103); \
47914  __ret_103; \
47915})
47916#endif
47917
47918#ifdef __LITTLE_ENDIAN__
47919#define vfmsq_laneq_f64(__p0_104, __p1_104, __p2_104, __p3_104) __extension__ ({ \
47920  float64x2_t __s0_104 = __p0_104; \
47921  float64x2_t __s1_104 = __p1_104; \
47922  float64x2_t __s2_104 = __p2_104; \
47923  float64x2_t __ret_104; \
47924  __ret_104 = vfmaq_laneq_f64(__s0_104, __s1_104, -__s2_104, __p3_104); \
47925  __ret_104; \
47926})
47927#else
47928#define vfmsq_laneq_f64(__p0_105, __p1_105, __p2_105, __p3_105) __extension__ ({ \
47929  float64x2_t __s0_105 = __p0_105; \
47930  float64x2_t __s1_105 = __p1_105; \
47931  float64x2_t __s2_105 = __p2_105; \
47932  float64x2_t __rev0_105;  __rev0_105 = __builtin_shufflevector(__s0_105, __s0_105, 1, 0); \
47933  float64x2_t __rev1_105;  __rev1_105 = __builtin_shufflevector(__s1_105, __s1_105, 1, 0); \
47934  float64x2_t __rev2_105;  __rev2_105 = __builtin_shufflevector(__s2_105, __s2_105, 1, 0); \
47935  float64x2_t __ret_105; \
47936  __ret_105 = __noswap_vfmaq_laneq_f64(__rev0_105, __rev1_105, -__rev2_105, __p3_105); \
47937  __ret_105 = __builtin_shufflevector(__ret_105, __ret_105, 1, 0); \
47938  __ret_105; \
47939})
47940#endif
47941
47942#ifdef __LITTLE_ENDIAN__
47943#define vfmsq_laneq_f32(__p0_106, __p1_106, __p2_106, __p3_106) __extension__ ({ \
47944  float32x4_t __s0_106 = __p0_106; \
47945  float32x4_t __s1_106 = __p1_106; \
47946  float32x4_t __s2_106 = __p2_106; \
47947  float32x4_t __ret_106; \
47948  __ret_106 = vfmaq_laneq_f32(__s0_106, __s1_106, -__s2_106, __p3_106); \
47949  __ret_106; \
47950})
47951#else
47952#define vfmsq_laneq_f32(__p0_107, __p1_107, __p2_107, __p3_107) __extension__ ({ \
47953  float32x4_t __s0_107 = __p0_107; \
47954  float32x4_t __s1_107 = __p1_107; \
47955  float32x4_t __s2_107 = __p2_107; \
47956  float32x4_t __rev0_107;  __rev0_107 = __builtin_shufflevector(__s0_107, __s0_107, 3, 2, 1, 0); \
47957  float32x4_t __rev1_107;  __rev1_107 = __builtin_shufflevector(__s1_107, __s1_107, 3, 2, 1, 0); \
47958  float32x4_t __rev2_107;  __rev2_107 = __builtin_shufflevector(__s2_107, __s2_107, 3, 2, 1, 0); \
47959  float32x4_t __ret_107; \
47960  __ret_107 = __noswap_vfmaq_laneq_f32(__rev0_107, __rev1_107, -__rev2_107, __p3_107); \
47961  __ret_107 = __builtin_shufflevector(__ret_107, __ret_107, 3, 2, 1, 0); \
47962  __ret_107; \
47963})
47964#endif
47965
47966#ifdef __LITTLE_ENDIAN__
47967#define vfms_laneq_f64(__p0_108, __p1_108, __p2_108, __p3_108) __extension__ ({ \
47968  float64x1_t __s0_108 = __p0_108; \
47969  float64x1_t __s1_108 = __p1_108; \
47970  float64x2_t __s2_108 = __p2_108; \
47971  float64x1_t __ret_108; \
47972  __ret_108 = vfma_laneq_f64(__s0_108, __s1_108, -__s2_108, __p3_108); \
47973  __ret_108; \
47974})
47975#else
47976#define vfms_laneq_f64(__p0_109, __p1_109, __p2_109, __p3_109) __extension__ ({ \
47977  float64x1_t __s0_109 = __p0_109; \
47978  float64x1_t __s1_109 = __p1_109; \
47979  float64x2_t __s2_109 = __p2_109; \
47980  float64x2_t __rev2_109;  __rev2_109 = __builtin_shufflevector(__s2_109, __s2_109, 1, 0); \
47981  float64x1_t __ret_109; \
47982  __ret_109 = __noswap_vfma_laneq_f64(__s0_109, __s1_109, -__rev2_109, __p3_109); \
47983  __ret_109; \
47984})
47985#endif
47986
47987#ifdef __LITTLE_ENDIAN__
47988#define vfms_laneq_f32(__p0_110, __p1_110, __p2_110, __p3_110) __extension__ ({ \
47989  float32x2_t __s0_110 = __p0_110; \
47990  float32x2_t __s1_110 = __p1_110; \
47991  float32x4_t __s2_110 = __p2_110; \
47992  float32x2_t __ret_110; \
47993  __ret_110 = vfma_laneq_f32(__s0_110, __s1_110, -__s2_110, __p3_110); \
47994  __ret_110; \
47995})
47996#else
47997#define vfms_laneq_f32(__p0_111, __p1_111, __p2_111, __p3_111) __extension__ ({ \
47998  float32x2_t __s0_111 = __p0_111; \
47999  float32x2_t __s1_111 = __p1_111; \
48000  float32x4_t __s2_111 = __p2_111; \
48001  float32x2_t __rev0_111;  __rev0_111 = __builtin_shufflevector(__s0_111, __s0_111, 1, 0); \
48002  float32x2_t __rev1_111;  __rev1_111 = __builtin_shufflevector(__s1_111, __s1_111, 1, 0); \
48003  float32x4_t __rev2_111;  __rev2_111 = __builtin_shufflevector(__s2_111, __s2_111, 3, 2, 1, 0); \
48004  float32x2_t __ret_111; \
48005  __ret_111 = __noswap_vfma_laneq_f32(__rev0_111, __rev1_111, -__rev2_111, __p3_111); \
48006  __ret_111 = __builtin_shufflevector(__ret_111, __ret_111, 1, 0); \
48007  __ret_111; \
48008})
48009#endif
48010
48011#ifdef __LITTLE_ENDIAN__
48012__ai float64x2_t vfmsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
48013  float64x2_t __ret;
48014  __ret = vfmsq_f64(__p0, __p1, (float64x2_t) {__p2, __p2});
48015  return __ret;
48016}
48017#else
48018__ai float64x2_t vfmsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
48019  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48020  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
48021  float64x2_t __ret;
48022  __ret = __noswap_vfmsq_f64(__rev0, __rev1, (float64x2_t) {__p2, __p2});
48023  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48024  return __ret;
48025}
48026#endif
48027
48028#ifdef __LITTLE_ENDIAN__
48029__ai float32x4_t vfmsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
48030  float32x4_t __ret;
48031  __ret = vfmsq_f32(__p0, __p1, (float32x4_t) {__p2, __p2, __p2, __p2});
48032  return __ret;
48033}
48034#else
48035__ai float32x4_t vfmsq_n_f32(float32x4_t __p0, float32x4_t __p1, float32_t __p2) {
48036  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
48037  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
48038  float32x4_t __ret;
48039  __ret = __noswap_vfmsq_f32(__rev0, __rev1, (float32x4_t) {__p2, __p2, __p2, __p2});
48040  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
48041  return __ret;
48042}
48043#endif
48044
48045#ifdef __LITTLE_ENDIAN__
48046__ai float32x2_t vfms_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
48047  float32x2_t __ret;
48048  __ret = vfms_f32(__p0, __p1, (float32x2_t) {__p2, __p2});
48049  return __ret;
48050}
48051#else
48052__ai float32x2_t vfms_n_f32(float32x2_t __p0, float32x2_t __p1, float32_t __p2) {
48053  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48054  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
48055  float32x2_t __ret;
48056  __ret = __noswap_vfms_f32(__rev0, __rev1, (float32x2_t) {__p2, __p2});
48057  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
48058  return __ret;
48059}
48060#endif
48061
48062#ifdef __LITTLE_ENDIAN__
48063__ai poly64x1_t vget_high_p64(poly64x2_t __p0) {
48064  poly64x1_t __ret;
48065  __ret = __builtin_shufflevector(__p0, __p0, 1);
48066  return __ret;
48067}
48068#else
48069__ai poly64x1_t vget_high_p64(poly64x2_t __p0) {
48070  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48071  poly64x1_t __ret;
48072  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
48073  return __ret;
48074}
48075__ai poly64x1_t __noswap_vget_high_p64(poly64x2_t __p0) {
48076  poly64x1_t __ret;
48077  __ret = __builtin_shufflevector(__p0, __p0, 1);
48078  return __ret;
48079}
48080#endif
48081
48082#ifdef __LITTLE_ENDIAN__
48083__ai float64x1_t vget_high_f64(float64x2_t __p0) {
48084  float64x1_t __ret;
48085  __ret = __builtin_shufflevector(__p0, __p0, 1);
48086  return __ret;
48087}
48088#else
48089__ai float64x1_t vget_high_f64(float64x2_t __p0) {
48090  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48091  float64x1_t __ret;
48092  __ret = __builtin_shufflevector(__rev0, __rev0, 1);
48093  return __ret;
48094}
48095#endif
48096
48097#ifdef __LITTLE_ENDIAN__
48098#define vget_lane_p64(__p0, __p1) __extension__ ({ \
48099  poly64x1_t __s0 = __p0; \
48100  poly64_t __ret; \
48101  __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
48102  __ret; \
48103})
48104#else
48105#define vget_lane_p64(__p0, __p1) __extension__ ({ \
48106  poly64x1_t __s0 = __p0; \
48107  poly64_t __ret; \
48108  __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
48109  __ret; \
48110})
48111#define __noswap_vget_lane_p64(__p0, __p1) __extension__ ({ \
48112  poly64x1_t __s0 = __p0; \
48113  poly64_t __ret; \
48114  __ret = (poly64_t) __builtin_neon_vget_lane_i64((int8x8_t)__s0, __p1); \
48115  __ret; \
48116})
48117#endif
48118
48119#ifdef __LITTLE_ENDIAN__
48120#define vgetq_lane_p64(__p0, __p1) __extension__ ({ \
48121  poly64x2_t __s0 = __p0; \
48122  poly64_t __ret; \
48123  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
48124  __ret; \
48125})
48126#else
48127#define vgetq_lane_p64(__p0, __p1) __extension__ ({ \
48128  poly64x2_t __s0 = __p0; \
48129  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
48130  poly64_t __ret; \
48131  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__rev0, __p1); \
48132  __ret; \
48133})
48134#define __noswap_vgetq_lane_p64(__p0, __p1) __extension__ ({ \
48135  poly64x2_t __s0 = __p0; \
48136  poly64_t __ret; \
48137  __ret = (poly64_t) __builtin_neon_vgetq_lane_i64((int8x16_t)__s0, __p1); \
48138  __ret; \
48139})
48140#endif
48141
48142#ifdef __LITTLE_ENDIAN__
48143#define vgetq_lane_f64(__p0, __p1) __extension__ ({ \
48144  float64x2_t __s0 = __p0; \
48145  float64_t __ret; \
48146  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__s0, __p1); \
48147  __ret; \
48148})
48149#else
48150#define vgetq_lane_f64(__p0, __p1) __extension__ ({ \
48151  float64x2_t __s0 = __p0; \
48152  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
48153  float64_t __ret; \
48154  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__rev0, __p1); \
48155  __ret; \
48156})
48157#define __noswap_vgetq_lane_f64(__p0, __p1) __extension__ ({ \
48158  float64x2_t __s0 = __p0; \
48159  float64_t __ret; \
48160  __ret = (float64_t) __builtin_neon_vgetq_lane_f64((int8x16_t)__s0, __p1); \
48161  __ret; \
48162})
48163#endif
48164
48165#ifdef __LITTLE_ENDIAN__
48166#define vget_lane_f64(__p0, __p1) __extension__ ({ \
48167  float64x1_t __s0 = __p0; \
48168  float64_t __ret; \
48169  __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
48170  __ret; \
48171})
48172#else
48173#define vget_lane_f64(__p0, __p1) __extension__ ({ \
48174  float64x1_t __s0 = __p0; \
48175  float64_t __ret; \
48176  __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
48177  __ret; \
48178})
48179#define __noswap_vget_lane_f64(__p0, __p1) __extension__ ({ \
48180  float64x1_t __s0 = __p0; \
48181  float64_t __ret; \
48182  __ret = (float64_t) __builtin_neon_vget_lane_f64((int8x8_t)__s0, __p1); \
48183  __ret; \
48184})
48185#endif
48186
48187#ifdef __LITTLE_ENDIAN__
48188__ai poly64x1_t vget_low_p64(poly64x2_t __p0) {
48189  poly64x1_t __ret;
48190  __ret = __builtin_shufflevector(__p0, __p0, 0);
48191  return __ret;
48192}
48193#else
48194__ai poly64x1_t vget_low_p64(poly64x2_t __p0) {
48195  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48196  poly64x1_t __ret;
48197  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
48198  return __ret;
48199}
48200#endif
48201
48202#ifdef __LITTLE_ENDIAN__
48203__ai float64x1_t vget_low_f64(float64x2_t __p0) {
48204  float64x1_t __ret;
48205  __ret = __builtin_shufflevector(__p0, __p0, 0);
48206  return __ret;
48207}
48208#else
48209__ai float64x1_t vget_low_f64(float64x2_t __p0) {
48210  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
48211  float64x1_t __ret;
48212  __ret = __builtin_shufflevector(__rev0, __rev0, 0);
48213  return __ret;
48214}
48215#endif
48216
48217#ifdef __LITTLE_ENDIAN__
48218#define vld1_p64(__p0) __extension__ ({ \
48219  poly64x1_t __ret; \
48220  __ret = (poly64x1_t) __builtin_neon_vld1_v(__p0, 6); \
48221  __ret; \
48222})
48223#else
48224#define vld1_p64(__p0) __extension__ ({ \
48225  poly64x1_t __ret; \
48226  __ret = (poly64x1_t) __builtin_neon_vld1_v(__p0, 6); \
48227  __ret; \
48228})
48229#endif
48230
48231#ifdef __LITTLE_ENDIAN__
48232#define vld1q_p64(__p0) __extension__ ({ \
48233  poly64x2_t __ret; \
48234  __ret = (poly64x2_t) __builtin_neon_vld1q_v(__p0, 38); \
48235  __ret; \
48236})
48237#else
48238#define vld1q_p64(__p0) __extension__ ({ \
48239  poly64x2_t __ret; \
48240  __ret = (poly64x2_t) __builtin_neon_vld1q_v(__p0, 38); \
48241  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48242  __ret; \
48243})
48244#endif
48245
48246#ifdef __LITTLE_ENDIAN__
48247#define vld1q_f64(__p0) __extension__ ({ \
48248  float64x2_t __ret; \
48249  __ret = (float64x2_t) __builtin_neon_vld1q_v(__p0, 42); \
48250  __ret; \
48251})
48252#else
48253#define vld1q_f64(__p0) __extension__ ({ \
48254  float64x2_t __ret; \
48255  __ret = (float64x2_t) __builtin_neon_vld1q_v(__p0, 42); \
48256  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48257  __ret; \
48258})
48259#endif
48260
48261#ifdef __LITTLE_ENDIAN__
48262#define vld1_f64(__p0) __extension__ ({ \
48263  float64x1_t __ret; \
48264  __ret = (float64x1_t) __builtin_neon_vld1_v(__p0, 10); \
48265  __ret; \
48266})
48267#else
48268#define vld1_f64(__p0) __extension__ ({ \
48269  float64x1_t __ret; \
48270  __ret = (float64x1_t) __builtin_neon_vld1_v(__p0, 10); \
48271  __ret; \
48272})
48273#endif
48274
48275#ifdef __LITTLE_ENDIAN__
48276#define vld1_dup_p64(__p0) __extension__ ({ \
48277  poly64x1_t __ret; \
48278  __ret = (poly64x1_t) __builtin_neon_vld1_dup_v(__p0, 6); \
48279  __ret; \
48280})
48281#else
48282#define vld1_dup_p64(__p0) __extension__ ({ \
48283  poly64x1_t __ret; \
48284  __ret = (poly64x1_t) __builtin_neon_vld1_dup_v(__p0, 6); \
48285  __ret; \
48286})
48287#endif
48288
48289#ifdef __LITTLE_ENDIAN__
48290#define vld1q_dup_p64(__p0) __extension__ ({ \
48291  poly64x2_t __ret; \
48292  __ret = (poly64x2_t) __builtin_neon_vld1q_dup_v(__p0, 38); \
48293  __ret; \
48294})
48295#else
48296#define vld1q_dup_p64(__p0) __extension__ ({ \
48297  poly64x2_t __ret; \
48298  __ret = (poly64x2_t) __builtin_neon_vld1q_dup_v(__p0, 38); \
48299  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48300  __ret; \
48301})
48302#endif
48303
48304#ifdef __LITTLE_ENDIAN__
48305#define vld1q_dup_f64(__p0) __extension__ ({ \
48306  float64x2_t __ret; \
48307  __ret = (float64x2_t) __builtin_neon_vld1q_dup_v(__p0, 42); \
48308  __ret; \
48309})
48310#else
48311#define vld1q_dup_f64(__p0) __extension__ ({ \
48312  float64x2_t __ret; \
48313  __ret = (float64x2_t) __builtin_neon_vld1q_dup_v(__p0, 42); \
48314  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48315  __ret; \
48316})
48317#endif
48318
48319#ifdef __LITTLE_ENDIAN__
48320#define vld1_dup_f64(__p0) __extension__ ({ \
48321  float64x1_t __ret; \
48322  __ret = (float64x1_t) __builtin_neon_vld1_dup_v(__p0, 10); \
48323  __ret; \
48324})
48325#else
48326#define vld1_dup_f64(__p0) __extension__ ({ \
48327  float64x1_t __ret; \
48328  __ret = (float64x1_t) __builtin_neon_vld1_dup_v(__p0, 10); \
48329  __ret; \
48330})
48331#endif
48332
48333#ifdef __LITTLE_ENDIAN__
48334#define vld1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48335  poly64x1_t __s1 = __p1; \
48336  poly64x1_t __ret; \
48337  __ret = (poly64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
48338  __ret; \
48339})
48340#else
48341#define vld1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48342  poly64x1_t __s1 = __p1; \
48343  poly64x1_t __ret; \
48344  __ret = (poly64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
48345  __ret; \
48346})
48347#endif
48348
48349#ifdef __LITTLE_ENDIAN__
48350#define vld1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48351  poly64x2_t __s1 = __p1; \
48352  poly64x2_t __ret; \
48353  __ret = (poly64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 38); \
48354  __ret; \
48355})
48356#else
48357#define vld1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
48358  poly64x2_t __s1 = __p1; \
48359  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
48360  poly64x2_t __ret; \
48361  __ret = (poly64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 38); \
48362  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48363  __ret; \
48364})
48365#endif
48366
48367#ifdef __LITTLE_ENDIAN__
48368#define vld1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48369  float64x2_t __s1 = __p1; \
48370  float64x2_t __ret; \
48371  __ret = (float64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__s1, __p2, 42); \
48372  __ret; \
48373})
48374#else
48375#define vld1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48376  float64x2_t __s1 = __p1; \
48377  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
48378  float64x2_t __ret; \
48379  __ret = (float64x2_t) __builtin_neon_vld1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 42); \
48380  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
48381  __ret; \
48382})
48383#endif
48384
48385#ifdef __LITTLE_ENDIAN__
48386#define vld1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48387  float64x1_t __s1 = __p1; \
48388  float64x1_t __ret; \
48389  __ret = (float64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
48390  __ret; \
48391})
48392#else
48393#define vld1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
48394  float64x1_t __s1 = __p1; \
48395  float64x1_t __ret; \
48396  __ret = (float64x1_t) __builtin_neon_vld1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
48397  __ret; \
48398})
48399#endif
48400
48401#ifdef __LITTLE_ENDIAN__
48402#define vld1_p8_x2(__p0) __extension__ ({ \
48403  poly8x8x2_t __ret; \
48404  __builtin_neon_vld1_x2_v(&__ret, __p0, 4); \
48405  __ret; \
48406})
48407#else
48408#define vld1_p8_x2(__p0) __extension__ ({ \
48409  poly8x8x2_t __ret; \
48410  __builtin_neon_vld1_x2_v(&__ret, __p0, 4); \
48411 \
48412  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48413  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48414  __ret; \
48415})
48416#endif
48417
48418#ifdef __LITTLE_ENDIAN__
48419#define vld1_p64_x2(__p0) __extension__ ({ \
48420  poly64x1x2_t __ret; \
48421  __builtin_neon_vld1_x2_v(&__ret, __p0, 6); \
48422  __ret; \
48423})
48424#else
48425#define vld1_p64_x2(__p0) __extension__ ({ \
48426  poly64x1x2_t __ret; \
48427  __builtin_neon_vld1_x2_v(&__ret, __p0, 6); \
48428  __ret; \
48429})
48430#endif
48431
48432#ifdef __LITTLE_ENDIAN__
48433#define vld1_p16_x2(__p0) __extension__ ({ \
48434  poly16x4x2_t __ret; \
48435  __builtin_neon_vld1_x2_v(&__ret, __p0, 5); \
48436  __ret; \
48437})
48438#else
48439#define vld1_p16_x2(__p0) __extension__ ({ \
48440  poly16x4x2_t __ret; \
48441  __builtin_neon_vld1_x2_v(&__ret, __p0, 5); \
48442 \
48443  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48444  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48445  __ret; \
48446})
48447#endif
48448
48449#ifdef __LITTLE_ENDIAN__
48450#define vld1q_p8_x2(__p0) __extension__ ({ \
48451  poly8x16x2_t __ret; \
48452  __builtin_neon_vld1q_x2_v(&__ret, __p0, 36); \
48453  __ret; \
48454})
48455#else
48456#define vld1q_p8_x2(__p0) __extension__ ({ \
48457  poly8x16x2_t __ret; \
48458  __builtin_neon_vld1q_x2_v(&__ret, __p0, 36); \
48459 \
48460  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48461  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48462  __ret; \
48463})
48464#endif
48465
48466#ifdef __LITTLE_ENDIAN__
48467#define vld1q_p64_x2(__p0) __extension__ ({ \
48468  poly64x2x2_t __ret; \
48469  __builtin_neon_vld1q_x2_v(&__ret, __p0, 38); \
48470  __ret; \
48471})
48472#else
48473#define vld1q_p64_x2(__p0) __extension__ ({ \
48474  poly64x2x2_t __ret; \
48475  __builtin_neon_vld1q_x2_v(&__ret, __p0, 38); \
48476 \
48477  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48478  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48479  __ret; \
48480})
48481#endif
48482
48483#ifdef __LITTLE_ENDIAN__
48484#define vld1q_p16_x2(__p0) __extension__ ({ \
48485  poly16x8x2_t __ret; \
48486  __builtin_neon_vld1q_x2_v(&__ret, __p0, 37); \
48487  __ret; \
48488})
48489#else
48490#define vld1q_p16_x2(__p0) __extension__ ({ \
48491  poly16x8x2_t __ret; \
48492  __builtin_neon_vld1q_x2_v(&__ret, __p0, 37); \
48493 \
48494  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48495  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48496  __ret; \
48497})
48498#endif
48499
48500#ifdef __LITTLE_ENDIAN__
48501#define vld1q_u8_x2(__p0) __extension__ ({ \
48502  uint8x16x2_t __ret; \
48503  __builtin_neon_vld1q_x2_v(&__ret, __p0, 48); \
48504  __ret; \
48505})
48506#else
48507#define vld1q_u8_x2(__p0) __extension__ ({ \
48508  uint8x16x2_t __ret; \
48509  __builtin_neon_vld1q_x2_v(&__ret, __p0, 48); \
48510 \
48511  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48512  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48513  __ret; \
48514})
48515#endif
48516
48517#ifdef __LITTLE_ENDIAN__
48518#define vld1q_u32_x2(__p0) __extension__ ({ \
48519  uint32x4x2_t __ret; \
48520  __builtin_neon_vld1q_x2_v(&__ret, __p0, 50); \
48521  __ret; \
48522})
48523#else
48524#define vld1q_u32_x2(__p0) __extension__ ({ \
48525  uint32x4x2_t __ret; \
48526  __builtin_neon_vld1q_x2_v(&__ret, __p0, 50); \
48527 \
48528  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48529  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48530  __ret; \
48531})
48532#endif
48533
48534#ifdef __LITTLE_ENDIAN__
48535#define vld1q_u64_x2(__p0) __extension__ ({ \
48536  uint64x2x2_t __ret; \
48537  __builtin_neon_vld1q_x2_v(&__ret, __p0, 51); \
48538  __ret; \
48539})
48540#else
48541#define vld1q_u64_x2(__p0) __extension__ ({ \
48542  uint64x2x2_t __ret; \
48543  __builtin_neon_vld1q_x2_v(&__ret, __p0, 51); \
48544 \
48545  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48546  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48547  __ret; \
48548})
48549#endif
48550
48551#ifdef __LITTLE_ENDIAN__
48552#define vld1q_u16_x2(__p0) __extension__ ({ \
48553  uint16x8x2_t __ret; \
48554  __builtin_neon_vld1q_x2_v(&__ret, __p0, 49); \
48555  __ret; \
48556})
48557#else
48558#define vld1q_u16_x2(__p0) __extension__ ({ \
48559  uint16x8x2_t __ret; \
48560  __builtin_neon_vld1q_x2_v(&__ret, __p0, 49); \
48561 \
48562  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48563  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48564  __ret; \
48565})
48566#endif
48567
48568#ifdef __LITTLE_ENDIAN__
48569#define vld1q_s8_x2(__p0) __extension__ ({ \
48570  int8x16x2_t __ret; \
48571  __builtin_neon_vld1q_x2_v(&__ret, __p0, 32); \
48572  __ret; \
48573})
48574#else
48575#define vld1q_s8_x2(__p0) __extension__ ({ \
48576  int8x16x2_t __ret; \
48577  __builtin_neon_vld1q_x2_v(&__ret, __p0, 32); \
48578 \
48579  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48580  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48581  __ret; \
48582})
48583#endif
48584
48585#ifdef __LITTLE_ENDIAN__
48586#define vld1q_f64_x2(__p0) __extension__ ({ \
48587  float64x2x2_t __ret; \
48588  __builtin_neon_vld1q_x2_v(&__ret, __p0, 42); \
48589  __ret; \
48590})
48591#else
48592#define vld1q_f64_x2(__p0) __extension__ ({ \
48593  float64x2x2_t __ret; \
48594  __builtin_neon_vld1q_x2_v(&__ret, __p0, 42); \
48595 \
48596  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48597  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48598  __ret; \
48599})
48600#endif
48601
48602#ifdef __LITTLE_ENDIAN__
48603#define vld1q_f32_x2(__p0) __extension__ ({ \
48604  float32x4x2_t __ret; \
48605  __builtin_neon_vld1q_x2_v(&__ret, __p0, 41); \
48606  __ret; \
48607})
48608#else
48609#define vld1q_f32_x2(__p0) __extension__ ({ \
48610  float32x4x2_t __ret; \
48611  __builtin_neon_vld1q_x2_v(&__ret, __p0, 41); \
48612 \
48613  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48614  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48615  __ret; \
48616})
48617#endif
48618
48619#ifdef __LITTLE_ENDIAN__
48620#define vld1q_f16_x2(__p0) __extension__ ({ \
48621  float16x8x2_t __ret; \
48622  __builtin_neon_vld1q_x2_v(&__ret, __p0, 40); \
48623  __ret; \
48624})
48625#else
48626#define vld1q_f16_x2(__p0) __extension__ ({ \
48627  float16x8x2_t __ret; \
48628  __builtin_neon_vld1q_x2_v(&__ret, __p0, 40); \
48629 \
48630  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48631  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48632  __ret; \
48633})
48634#endif
48635
48636#ifdef __LITTLE_ENDIAN__
48637#define vld1q_s32_x2(__p0) __extension__ ({ \
48638  int32x4x2_t __ret; \
48639  __builtin_neon_vld1q_x2_v(&__ret, __p0, 34); \
48640  __ret; \
48641})
48642#else
48643#define vld1q_s32_x2(__p0) __extension__ ({ \
48644  int32x4x2_t __ret; \
48645  __builtin_neon_vld1q_x2_v(&__ret, __p0, 34); \
48646 \
48647  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48648  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48649  __ret; \
48650})
48651#endif
48652
48653#ifdef __LITTLE_ENDIAN__
48654#define vld1q_s64_x2(__p0) __extension__ ({ \
48655  int64x2x2_t __ret; \
48656  __builtin_neon_vld1q_x2_v(&__ret, __p0, 35); \
48657  __ret; \
48658})
48659#else
48660#define vld1q_s64_x2(__p0) __extension__ ({ \
48661  int64x2x2_t __ret; \
48662  __builtin_neon_vld1q_x2_v(&__ret, __p0, 35); \
48663 \
48664  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48665  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48666  __ret; \
48667})
48668#endif
48669
48670#ifdef __LITTLE_ENDIAN__
48671#define vld1q_s16_x2(__p0) __extension__ ({ \
48672  int16x8x2_t __ret; \
48673  __builtin_neon_vld1q_x2_v(&__ret, __p0, 33); \
48674  __ret; \
48675})
48676#else
48677#define vld1q_s16_x2(__p0) __extension__ ({ \
48678  int16x8x2_t __ret; \
48679  __builtin_neon_vld1q_x2_v(&__ret, __p0, 33); \
48680 \
48681  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48682  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48683  __ret; \
48684})
48685#endif
48686
48687#ifdef __LITTLE_ENDIAN__
48688#define vld1_u8_x2(__p0) __extension__ ({ \
48689  uint8x8x2_t __ret; \
48690  __builtin_neon_vld1_x2_v(&__ret, __p0, 16); \
48691  __ret; \
48692})
48693#else
48694#define vld1_u8_x2(__p0) __extension__ ({ \
48695  uint8x8x2_t __ret; \
48696  __builtin_neon_vld1_x2_v(&__ret, __p0, 16); \
48697 \
48698  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48699  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48700  __ret; \
48701})
48702#endif
48703
48704#ifdef __LITTLE_ENDIAN__
48705#define vld1_u32_x2(__p0) __extension__ ({ \
48706  uint32x2x2_t __ret; \
48707  __builtin_neon_vld1_x2_v(&__ret, __p0, 18); \
48708  __ret; \
48709})
48710#else
48711#define vld1_u32_x2(__p0) __extension__ ({ \
48712  uint32x2x2_t __ret; \
48713  __builtin_neon_vld1_x2_v(&__ret, __p0, 18); \
48714 \
48715  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48716  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48717  __ret; \
48718})
48719#endif
48720
48721#ifdef __LITTLE_ENDIAN__
48722#define vld1_u64_x2(__p0) __extension__ ({ \
48723  uint64x1x2_t __ret; \
48724  __builtin_neon_vld1_x2_v(&__ret, __p0, 19); \
48725  __ret; \
48726})
48727#else
48728#define vld1_u64_x2(__p0) __extension__ ({ \
48729  uint64x1x2_t __ret; \
48730  __builtin_neon_vld1_x2_v(&__ret, __p0, 19); \
48731  __ret; \
48732})
48733#endif
48734
48735#ifdef __LITTLE_ENDIAN__
48736#define vld1_u16_x2(__p0) __extension__ ({ \
48737  uint16x4x2_t __ret; \
48738  __builtin_neon_vld1_x2_v(&__ret, __p0, 17); \
48739  __ret; \
48740})
48741#else
48742#define vld1_u16_x2(__p0) __extension__ ({ \
48743  uint16x4x2_t __ret; \
48744  __builtin_neon_vld1_x2_v(&__ret, __p0, 17); \
48745 \
48746  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48747  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48748  __ret; \
48749})
48750#endif
48751
48752#ifdef __LITTLE_ENDIAN__
48753#define vld1_s8_x2(__p0) __extension__ ({ \
48754  int8x8x2_t __ret; \
48755  __builtin_neon_vld1_x2_v(&__ret, __p0, 0); \
48756  __ret; \
48757})
48758#else
48759#define vld1_s8_x2(__p0) __extension__ ({ \
48760  int8x8x2_t __ret; \
48761  __builtin_neon_vld1_x2_v(&__ret, __p0, 0); \
48762 \
48763  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48764  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48765  __ret; \
48766})
48767#endif
48768
48769#ifdef __LITTLE_ENDIAN__
48770#define vld1_f64_x2(__p0) __extension__ ({ \
48771  float64x1x2_t __ret; \
48772  __builtin_neon_vld1_x2_v(&__ret, __p0, 10); \
48773  __ret; \
48774})
48775#else
48776#define vld1_f64_x2(__p0) __extension__ ({ \
48777  float64x1x2_t __ret; \
48778  __builtin_neon_vld1_x2_v(&__ret, __p0, 10); \
48779  __ret; \
48780})
48781#endif
48782
48783#ifdef __LITTLE_ENDIAN__
48784#define vld1_f32_x2(__p0) __extension__ ({ \
48785  float32x2x2_t __ret; \
48786  __builtin_neon_vld1_x2_v(&__ret, __p0, 9); \
48787  __ret; \
48788})
48789#else
48790#define vld1_f32_x2(__p0) __extension__ ({ \
48791  float32x2x2_t __ret; \
48792  __builtin_neon_vld1_x2_v(&__ret, __p0, 9); \
48793 \
48794  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48795  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48796  __ret; \
48797})
48798#endif
48799
48800#ifdef __LITTLE_ENDIAN__
48801#define vld1_f16_x2(__p0) __extension__ ({ \
48802  float16x4x2_t __ret; \
48803  __builtin_neon_vld1_x2_v(&__ret, __p0, 8); \
48804  __ret; \
48805})
48806#else
48807#define vld1_f16_x2(__p0) __extension__ ({ \
48808  float16x4x2_t __ret; \
48809  __builtin_neon_vld1_x2_v(&__ret, __p0, 8); \
48810 \
48811  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48812  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48813  __ret; \
48814})
48815#endif
48816
48817#ifdef __LITTLE_ENDIAN__
48818#define vld1_s32_x2(__p0) __extension__ ({ \
48819  int32x2x2_t __ret; \
48820  __builtin_neon_vld1_x2_v(&__ret, __p0, 2); \
48821  __ret; \
48822})
48823#else
48824#define vld1_s32_x2(__p0) __extension__ ({ \
48825  int32x2x2_t __ret; \
48826  __builtin_neon_vld1_x2_v(&__ret, __p0, 2); \
48827 \
48828  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48829  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48830  __ret; \
48831})
48832#endif
48833
48834#ifdef __LITTLE_ENDIAN__
48835#define vld1_s64_x2(__p0) __extension__ ({ \
48836  int64x1x2_t __ret; \
48837  __builtin_neon_vld1_x2_v(&__ret, __p0, 3); \
48838  __ret; \
48839})
48840#else
48841#define vld1_s64_x2(__p0) __extension__ ({ \
48842  int64x1x2_t __ret; \
48843  __builtin_neon_vld1_x2_v(&__ret, __p0, 3); \
48844  __ret; \
48845})
48846#endif
48847
48848#ifdef __LITTLE_ENDIAN__
48849#define vld1_s16_x2(__p0) __extension__ ({ \
48850  int16x4x2_t __ret; \
48851  __builtin_neon_vld1_x2_v(&__ret, __p0, 1); \
48852  __ret; \
48853})
48854#else
48855#define vld1_s16_x2(__p0) __extension__ ({ \
48856  int16x4x2_t __ret; \
48857  __builtin_neon_vld1_x2_v(&__ret, __p0, 1); \
48858 \
48859  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48860  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48861  __ret; \
48862})
48863#endif
48864
48865#ifdef __LITTLE_ENDIAN__
48866#define vld1_p8_x3(__p0) __extension__ ({ \
48867  poly8x8x3_t __ret; \
48868  __builtin_neon_vld1_x3_v(&__ret, __p0, 4); \
48869  __ret; \
48870})
48871#else
48872#define vld1_p8_x3(__p0) __extension__ ({ \
48873  poly8x8x3_t __ret; \
48874  __builtin_neon_vld1_x3_v(&__ret, __p0, 4); \
48875 \
48876  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48877  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48878  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48879  __ret; \
48880})
48881#endif
48882
48883#ifdef __LITTLE_ENDIAN__
48884#define vld1_p64_x3(__p0) __extension__ ({ \
48885  poly64x1x3_t __ret; \
48886  __builtin_neon_vld1_x3_v(&__ret, __p0, 6); \
48887  __ret; \
48888})
48889#else
48890#define vld1_p64_x3(__p0) __extension__ ({ \
48891  poly64x1x3_t __ret; \
48892  __builtin_neon_vld1_x3_v(&__ret, __p0, 6); \
48893  __ret; \
48894})
48895#endif
48896
48897#ifdef __LITTLE_ENDIAN__
48898#define vld1_p16_x3(__p0) __extension__ ({ \
48899  poly16x4x3_t __ret; \
48900  __builtin_neon_vld1_x3_v(&__ret, __p0, 5); \
48901  __ret; \
48902})
48903#else
48904#define vld1_p16_x3(__p0) __extension__ ({ \
48905  poly16x4x3_t __ret; \
48906  __builtin_neon_vld1_x3_v(&__ret, __p0, 5); \
48907 \
48908  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48909  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
48910  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
48911  __ret; \
48912})
48913#endif
48914
48915#ifdef __LITTLE_ENDIAN__
48916#define vld1q_p8_x3(__p0) __extension__ ({ \
48917  poly8x16x3_t __ret; \
48918  __builtin_neon_vld1q_x3_v(&__ret, __p0, 36); \
48919  __ret; \
48920})
48921#else
48922#define vld1q_p8_x3(__p0) __extension__ ({ \
48923  poly8x16x3_t __ret; \
48924  __builtin_neon_vld1q_x3_v(&__ret, __p0, 36); \
48925 \
48926  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48927  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48928  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48929  __ret; \
48930})
48931#endif
48932
48933#ifdef __LITTLE_ENDIAN__
48934#define vld1q_p64_x3(__p0) __extension__ ({ \
48935  poly64x2x3_t __ret; \
48936  __builtin_neon_vld1q_x3_v(&__ret, __p0, 38); \
48937  __ret; \
48938})
48939#else
48940#define vld1q_p64_x3(__p0) __extension__ ({ \
48941  poly64x2x3_t __ret; \
48942  __builtin_neon_vld1q_x3_v(&__ret, __p0, 38); \
48943 \
48944  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
48945  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
48946  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
48947  __ret; \
48948})
48949#endif
48950
48951#ifdef __LITTLE_ENDIAN__
48952#define vld1q_p16_x3(__p0) __extension__ ({ \
48953  poly16x8x3_t __ret; \
48954  __builtin_neon_vld1q_x3_v(&__ret, __p0, 37); \
48955  __ret; \
48956})
48957#else
48958#define vld1q_p16_x3(__p0) __extension__ ({ \
48959  poly16x8x3_t __ret; \
48960  __builtin_neon_vld1q_x3_v(&__ret, __p0, 37); \
48961 \
48962  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
48963  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
48964  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
48965  __ret; \
48966})
48967#endif
48968
48969#ifdef __LITTLE_ENDIAN__
48970#define vld1q_u8_x3(__p0) __extension__ ({ \
48971  uint8x16x3_t __ret; \
48972  __builtin_neon_vld1q_x3_v(&__ret, __p0, 48); \
48973  __ret; \
48974})
48975#else
48976#define vld1q_u8_x3(__p0) __extension__ ({ \
48977  uint8x16x3_t __ret; \
48978  __builtin_neon_vld1q_x3_v(&__ret, __p0, 48); \
48979 \
48980  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48981  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48982  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
48983  __ret; \
48984})
48985#endif
48986
48987#ifdef __LITTLE_ENDIAN__
48988#define vld1q_u32_x3(__p0) __extension__ ({ \
48989  uint32x4x3_t __ret; \
48990  __builtin_neon_vld1q_x3_v(&__ret, __p0, 50); \
48991  __ret; \
48992})
48993#else
48994#define vld1q_u32_x3(__p0) __extension__ ({ \
48995  uint32x4x3_t __ret; \
48996  __builtin_neon_vld1q_x3_v(&__ret, __p0, 50); \
48997 \
48998  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
48999  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49000  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49001  __ret; \
49002})
49003#endif
49004
49005#ifdef __LITTLE_ENDIAN__
49006#define vld1q_u64_x3(__p0) __extension__ ({ \
49007  uint64x2x3_t __ret; \
49008  __builtin_neon_vld1q_x3_v(&__ret, __p0, 51); \
49009  __ret; \
49010})
49011#else
49012#define vld1q_u64_x3(__p0) __extension__ ({ \
49013  uint64x2x3_t __ret; \
49014  __builtin_neon_vld1q_x3_v(&__ret, __p0, 51); \
49015 \
49016  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49017  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49018  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49019  __ret; \
49020})
49021#endif
49022
49023#ifdef __LITTLE_ENDIAN__
49024#define vld1q_u16_x3(__p0) __extension__ ({ \
49025  uint16x8x3_t __ret; \
49026  __builtin_neon_vld1q_x3_v(&__ret, __p0, 49); \
49027  __ret; \
49028})
49029#else
49030#define vld1q_u16_x3(__p0) __extension__ ({ \
49031  uint16x8x3_t __ret; \
49032  __builtin_neon_vld1q_x3_v(&__ret, __p0, 49); \
49033 \
49034  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49035  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49036  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49037  __ret; \
49038})
49039#endif
49040
49041#ifdef __LITTLE_ENDIAN__
49042#define vld1q_s8_x3(__p0) __extension__ ({ \
49043  int8x16x3_t __ret; \
49044  __builtin_neon_vld1q_x3_v(&__ret, __p0, 32); \
49045  __ret; \
49046})
49047#else
49048#define vld1q_s8_x3(__p0) __extension__ ({ \
49049  int8x16x3_t __ret; \
49050  __builtin_neon_vld1q_x3_v(&__ret, __p0, 32); \
49051 \
49052  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49053  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49054  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49055  __ret; \
49056})
49057#endif
49058
49059#ifdef __LITTLE_ENDIAN__
49060#define vld1q_f64_x3(__p0) __extension__ ({ \
49061  float64x2x3_t __ret; \
49062  __builtin_neon_vld1q_x3_v(&__ret, __p0, 42); \
49063  __ret; \
49064})
49065#else
49066#define vld1q_f64_x3(__p0) __extension__ ({ \
49067  float64x2x3_t __ret; \
49068  __builtin_neon_vld1q_x3_v(&__ret, __p0, 42); \
49069 \
49070  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49071  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49072  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49073  __ret; \
49074})
49075#endif
49076
49077#ifdef __LITTLE_ENDIAN__
49078#define vld1q_f32_x3(__p0) __extension__ ({ \
49079  float32x4x3_t __ret; \
49080  __builtin_neon_vld1q_x3_v(&__ret, __p0, 41); \
49081  __ret; \
49082})
49083#else
49084#define vld1q_f32_x3(__p0) __extension__ ({ \
49085  float32x4x3_t __ret; \
49086  __builtin_neon_vld1q_x3_v(&__ret, __p0, 41); \
49087 \
49088  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49089  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49090  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49091  __ret; \
49092})
49093#endif
49094
49095#ifdef __LITTLE_ENDIAN__
49096#define vld1q_f16_x3(__p0) __extension__ ({ \
49097  float16x8x3_t __ret; \
49098  __builtin_neon_vld1q_x3_v(&__ret, __p0, 40); \
49099  __ret; \
49100})
49101#else
49102#define vld1q_f16_x3(__p0) __extension__ ({ \
49103  float16x8x3_t __ret; \
49104  __builtin_neon_vld1q_x3_v(&__ret, __p0, 40); \
49105 \
49106  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49107  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49108  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49109  __ret; \
49110})
49111#endif
49112
49113#ifdef __LITTLE_ENDIAN__
49114#define vld1q_s32_x3(__p0) __extension__ ({ \
49115  int32x4x3_t __ret; \
49116  __builtin_neon_vld1q_x3_v(&__ret, __p0, 34); \
49117  __ret; \
49118})
49119#else
49120#define vld1q_s32_x3(__p0) __extension__ ({ \
49121  int32x4x3_t __ret; \
49122  __builtin_neon_vld1q_x3_v(&__ret, __p0, 34); \
49123 \
49124  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49125  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49126  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49127  __ret; \
49128})
49129#endif
49130
49131#ifdef __LITTLE_ENDIAN__
49132#define vld1q_s64_x3(__p0) __extension__ ({ \
49133  int64x2x3_t __ret; \
49134  __builtin_neon_vld1q_x3_v(&__ret, __p0, 35); \
49135  __ret; \
49136})
49137#else
49138#define vld1q_s64_x3(__p0) __extension__ ({ \
49139  int64x2x3_t __ret; \
49140  __builtin_neon_vld1q_x3_v(&__ret, __p0, 35); \
49141 \
49142  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49143  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49144  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49145  __ret; \
49146})
49147#endif
49148
49149#ifdef __LITTLE_ENDIAN__
49150#define vld1q_s16_x3(__p0) __extension__ ({ \
49151  int16x8x3_t __ret; \
49152  __builtin_neon_vld1q_x3_v(&__ret, __p0, 33); \
49153  __ret; \
49154})
49155#else
49156#define vld1q_s16_x3(__p0) __extension__ ({ \
49157  int16x8x3_t __ret; \
49158  __builtin_neon_vld1q_x3_v(&__ret, __p0, 33); \
49159 \
49160  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49161  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49162  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49163  __ret; \
49164})
49165#endif
49166
49167#ifdef __LITTLE_ENDIAN__
49168#define vld1_u8_x3(__p0) __extension__ ({ \
49169  uint8x8x3_t __ret; \
49170  __builtin_neon_vld1_x3_v(&__ret, __p0, 16); \
49171  __ret; \
49172})
49173#else
49174#define vld1_u8_x3(__p0) __extension__ ({ \
49175  uint8x8x3_t __ret; \
49176  __builtin_neon_vld1_x3_v(&__ret, __p0, 16); \
49177 \
49178  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49179  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49180  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49181  __ret; \
49182})
49183#endif
49184
49185#ifdef __LITTLE_ENDIAN__
49186#define vld1_u32_x3(__p0) __extension__ ({ \
49187  uint32x2x3_t __ret; \
49188  __builtin_neon_vld1_x3_v(&__ret, __p0, 18); \
49189  __ret; \
49190})
49191#else
49192#define vld1_u32_x3(__p0) __extension__ ({ \
49193  uint32x2x3_t __ret; \
49194  __builtin_neon_vld1_x3_v(&__ret, __p0, 18); \
49195 \
49196  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49197  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49198  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49199  __ret; \
49200})
49201#endif
49202
49203#ifdef __LITTLE_ENDIAN__
49204#define vld1_u64_x3(__p0) __extension__ ({ \
49205  uint64x1x3_t __ret; \
49206  __builtin_neon_vld1_x3_v(&__ret, __p0, 19); \
49207  __ret; \
49208})
49209#else
49210#define vld1_u64_x3(__p0) __extension__ ({ \
49211  uint64x1x3_t __ret; \
49212  __builtin_neon_vld1_x3_v(&__ret, __p0, 19); \
49213  __ret; \
49214})
49215#endif
49216
49217#ifdef __LITTLE_ENDIAN__
49218#define vld1_u16_x3(__p0) __extension__ ({ \
49219  uint16x4x3_t __ret; \
49220  __builtin_neon_vld1_x3_v(&__ret, __p0, 17); \
49221  __ret; \
49222})
49223#else
49224#define vld1_u16_x3(__p0) __extension__ ({ \
49225  uint16x4x3_t __ret; \
49226  __builtin_neon_vld1_x3_v(&__ret, __p0, 17); \
49227 \
49228  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49229  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49230  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49231  __ret; \
49232})
49233#endif
49234
49235#ifdef __LITTLE_ENDIAN__
49236#define vld1_s8_x3(__p0) __extension__ ({ \
49237  int8x8x3_t __ret; \
49238  __builtin_neon_vld1_x3_v(&__ret, __p0, 0); \
49239  __ret; \
49240})
49241#else
49242#define vld1_s8_x3(__p0) __extension__ ({ \
49243  int8x8x3_t __ret; \
49244  __builtin_neon_vld1_x3_v(&__ret, __p0, 0); \
49245 \
49246  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49247  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49248  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49249  __ret; \
49250})
49251#endif
49252
49253#ifdef __LITTLE_ENDIAN__
49254#define vld1_f64_x3(__p0) __extension__ ({ \
49255  float64x1x3_t __ret; \
49256  __builtin_neon_vld1_x3_v(&__ret, __p0, 10); \
49257  __ret; \
49258})
49259#else
49260#define vld1_f64_x3(__p0) __extension__ ({ \
49261  float64x1x3_t __ret; \
49262  __builtin_neon_vld1_x3_v(&__ret, __p0, 10); \
49263  __ret; \
49264})
49265#endif
49266
49267#ifdef __LITTLE_ENDIAN__
49268#define vld1_f32_x3(__p0) __extension__ ({ \
49269  float32x2x3_t __ret; \
49270  __builtin_neon_vld1_x3_v(&__ret, __p0, 9); \
49271  __ret; \
49272})
49273#else
49274#define vld1_f32_x3(__p0) __extension__ ({ \
49275  float32x2x3_t __ret; \
49276  __builtin_neon_vld1_x3_v(&__ret, __p0, 9); \
49277 \
49278  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49279  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49280  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49281  __ret; \
49282})
49283#endif
49284
49285#ifdef __LITTLE_ENDIAN__
49286#define vld1_f16_x3(__p0) __extension__ ({ \
49287  float16x4x3_t __ret; \
49288  __builtin_neon_vld1_x3_v(&__ret, __p0, 8); \
49289  __ret; \
49290})
49291#else
49292#define vld1_f16_x3(__p0) __extension__ ({ \
49293  float16x4x3_t __ret; \
49294  __builtin_neon_vld1_x3_v(&__ret, __p0, 8); \
49295 \
49296  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49297  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49298  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49299  __ret; \
49300})
49301#endif
49302
49303#ifdef __LITTLE_ENDIAN__
49304#define vld1_s32_x3(__p0) __extension__ ({ \
49305  int32x2x3_t __ret; \
49306  __builtin_neon_vld1_x3_v(&__ret, __p0, 2); \
49307  __ret; \
49308})
49309#else
49310#define vld1_s32_x3(__p0) __extension__ ({ \
49311  int32x2x3_t __ret; \
49312  __builtin_neon_vld1_x3_v(&__ret, __p0, 2); \
49313 \
49314  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49315  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49316  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49317  __ret; \
49318})
49319#endif
49320
49321#ifdef __LITTLE_ENDIAN__
49322#define vld1_s64_x3(__p0) __extension__ ({ \
49323  int64x1x3_t __ret; \
49324  __builtin_neon_vld1_x3_v(&__ret, __p0, 3); \
49325  __ret; \
49326})
49327#else
49328#define vld1_s64_x3(__p0) __extension__ ({ \
49329  int64x1x3_t __ret; \
49330  __builtin_neon_vld1_x3_v(&__ret, __p0, 3); \
49331  __ret; \
49332})
49333#endif
49334
49335#ifdef __LITTLE_ENDIAN__
49336#define vld1_s16_x3(__p0) __extension__ ({ \
49337  int16x4x3_t __ret; \
49338  __builtin_neon_vld1_x3_v(&__ret, __p0, 1); \
49339  __ret; \
49340})
49341#else
49342#define vld1_s16_x3(__p0) __extension__ ({ \
49343  int16x4x3_t __ret; \
49344  __builtin_neon_vld1_x3_v(&__ret, __p0, 1); \
49345 \
49346  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49347  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49348  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49349  __ret; \
49350})
49351#endif
49352
49353#ifdef __LITTLE_ENDIAN__
49354#define vld1_p8_x4(__p0) __extension__ ({ \
49355  poly8x8x4_t __ret; \
49356  __builtin_neon_vld1_x4_v(&__ret, __p0, 4); \
49357  __ret; \
49358})
49359#else
49360#define vld1_p8_x4(__p0) __extension__ ({ \
49361  poly8x8x4_t __ret; \
49362  __builtin_neon_vld1_x4_v(&__ret, __p0, 4); \
49363 \
49364  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49365  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49366  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49367  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49368  __ret; \
49369})
49370#endif
49371
49372#ifdef __LITTLE_ENDIAN__
49373#define vld1_p64_x4(__p0) __extension__ ({ \
49374  poly64x1x4_t __ret; \
49375  __builtin_neon_vld1_x4_v(&__ret, __p0, 6); \
49376  __ret; \
49377})
49378#else
49379#define vld1_p64_x4(__p0) __extension__ ({ \
49380  poly64x1x4_t __ret; \
49381  __builtin_neon_vld1_x4_v(&__ret, __p0, 6); \
49382  __ret; \
49383})
49384#endif
49385
49386#ifdef __LITTLE_ENDIAN__
49387#define vld1_p16_x4(__p0) __extension__ ({ \
49388  poly16x4x4_t __ret; \
49389  __builtin_neon_vld1_x4_v(&__ret, __p0, 5); \
49390  __ret; \
49391})
49392#else
49393#define vld1_p16_x4(__p0) __extension__ ({ \
49394  poly16x4x4_t __ret; \
49395  __builtin_neon_vld1_x4_v(&__ret, __p0, 5); \
49396 \
49397  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49398  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49399  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49400  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49401  __ret; \
49402})
49403#endif
49404
49405#ifdef __LITTLE_ENDIAN__
49406#define vld1q_p8_x4(__p0) __extension__ ({ \
49407  poly8x16x4_t __ret; \
49408  __builtin_neon_vld1q_x4_v(&__ret, __p0, 36); \
49409  __ret; \
49410})
49411#else
49412#define vld1q_p8_x4(__p0) __extension__ ({ \
49413  poly8x16x4_t __ret; \
49414  __builtin_neon_vld1q_x4_v(&__ret, __p0, 36); \
49415 \
49416  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49417  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49418  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49419  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49420  __ret; \
49421})
49422#endif
49423
49424#ifdef __LITTLE_ENDIAN__
49425#define vld1q_p64_x4(__p0) __extension__ ({ \
49426  poly64x2x4_t __ret; \
49427  __builtin_neon_vld1q_x4_v(&__ret, __p0, 38); \
49428  __ret; \
49429})
49430#else
49431#define vld1q_p64_x4(__p0) __extension__ ({ \
49432  poly64x2x4_t __ret; \
49433  __builtin_neon_vld1q_x4_v(&__ret, __p0, 38); \
49434 \
49435  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49436  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49437  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49438  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49439  __ret; \
49440})
49441#endif
49442
49443#ifdef __LITTLE_ENDIAN__
49444#define vld1q_p16_x4(__p0) __extension__ ({ \
49445  poly16x8x4_t __ret; \
49446  __builtin_neon_vld1q_x4_v(&__ret, __p0, 37); \
49447  __ret; \
49448})
49449#else
49450#define vld1q_p16_x4(__p0) __extension__ ({ \
49451  poly16x8x4_t __ret; \
49452  __builtin_neon_vld1q_x4_v(&__ret, __p0, 37); \
49453 \
49454  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49455  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49456  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49457  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49458  __ret; \
49459})
49460#endif
49461
49462#ifdef __LITTLE_ENDIAN__
49463#define vld1q_u8_x4(__p0) __extension__ ({ \
49464  uint8x16x4_t __ret; \
49465  __builtin_neon_vld1q_x4_v(&__ret, __p0, 48); \
49466  __ret; \
49467})
49468#else
49469#define vld1q_u8_x4(__p0) __extension__ ({ \
49470  uint8x16x4_t __ret; \
49471  __builtin_neon_vld1q_x4_v(&__ret, __p0, 48); \
49472 \
49473  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49474  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49475  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49476  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49477  __ret; \
49478})
49479#endif
49480
49481#ifdef __LITTLE_ENDIAN__
49482#define vld1q_u32_x4(__p0) __extension__ ({ \
49483  uint32x4x4_t __ret; \
49484  __builtin_neon_vld1q_x4_v(&__ret, __p0, 50); \
49485  __ret; \
49486})
49487#else
49488#define vld1q_u32_x4(__p0) __extension__ ({ \
49489  uint32x4x4_t __ret; \
49490  __builtin_neon_vld1q_x4_v(&__ret, __p0, 50); \
49491 \
49492  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49493  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49494  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49495  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49496  __ret; \
49497})
49498#endif
49499
49500#ifdef __LITTLE_ENDIAN__
49501#define vld1q_u64_x4(__p0) __extension__ ({ \
49502  uint64x2x4_t __ret; \
49503  __builtin_neon_vld1q_x4_v(&__ret, __p0, 51); \
49504  __ret; \
49505})
49506#else
49507#define vld1q_u64_x4(__p0) __extension__ ({ \
49508  uint64x2x4_t __ret; \
49509  __builtin_neon_vld1q_x4_v(&__ret, __p0, 51); \
49510 \
49511  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49512  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49513  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49514  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49515  __ret; \
49516})
49517#endif
49518
49519#ifdef __LITTLE_ENDIAN__
49520#define vld1q_u16_x4(__p0) __extension__ ({ \
49521  uint16x8x4_t __ret; \
49522  __builtin_neon_vld1q_x4_v(&__ret, __p0, 49); \
49523  __ret; \
49524})
49525#else
49526#define vld1q_u16_x4(__p0) __extension__ ({ \
49527  uint16x8x4_t __ret; \
49528  __builtin_neon_vld1q_x4_v(&__ret, __p0, 49); \
49529 \
49530  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49531  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49532  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49533  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49534  __ret; \
49535})
49536#endif
49537
49538#ifdef __LITTLE_ENDIAN__
49539#define vld1q_s8_x4(__p0) __extension__ ({ \
49540  int8x16x4_t __ret; \
49541  __builtin_neon_vld1q_x4_v(&__ret, __p0, 32); \
49542  __ret; \
49543})
49544#else
49545#define vld1q_s8_x4(__p0) __extension__ ({ \
49546  int8x16x4_t __ret; \
49547  __builtin_neon_vld1q_x4_v(&__ret, __p0, 32); \
49548 \
49549  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49550  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49551  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49552  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49553  __ret; \
49554})
49555#endif
49556
49557#ifdef __LITTLE_ENDIAN__
49558#define vld1q_f64_x4(__p0) __extension__ ({ \
49559  float64x2x4_t __ret; \
49560  __builtin_neon_vld1q_x4_v(&__ret, __p0, 42); \
49561  __ret; \
49562})
49563#else
49564#define vld1q_f64_x4(__p0) __extension__ ({ \
49565  float64x2x4_t __ret; \
49566  __builtin_neon_vld1q_x4_v(&__ret, __p0, 42); \
49567 \
49568  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49569  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49570  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49571  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49572  __ret; \
49573})
49574#endif
49575
49576#ifdef __LITTLE_ENDIAN__
49577#define vld1q_f32_x4(__p0) __extension__ ({ \
49578  float32x4x4_t __ret; \
49579  __builtin_neon_vld1q_x4_v(&__ret, __p0, 41); \
49580  __ret; \
49581})
49582#else
49583#define vld1q_f32_x4(__p0) __extension__ ({ \
49584  float32x4x4_t __ret; \
49585  __builtin_neon_vld1q_x4_v(&__ret, __p0, 41); \
49586 \
49587  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49588  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49589  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49590  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49591  __ret; \
49592})
49593#endif
49594
49595#ifdef __LITTLE_ENDIAN__
49596#define vld1q_f16_x4(__p0) __extension__ ({ \
49597  float16x8x4_t __ret; \
49598  __builtin_neon_vld1q_x4_v(&__ret, __p0, 40); \
49599  __ret; \
49600})
49601#else
49602#define vld1q_f16_x4(__p0) __extension__ ({ \
49603  float16x8x4_t __ret; \
49604  __builtin_neon_vld1q_x4_v(&__ret, __p0, 40); \
49605 \
49606  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49607  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49608  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49609  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49610  __ret; \
49611})
49612#endif
49613
49614#ifdef __LITTLE_ENDIAN__
49615#define vld1q_s32_x4(__p0) __extension__ ({ \
49616  int32x4x4_t __ret; \
49617  __builtin_neon_vld1q_x4_v(&__ret, __p0, 34); \
49618  __ret; \
49619})
49620#else
49621#define vld1q_s32_x4(__p0) __extension__ ({ \
49622  int32x4x4_t __ret; \
49623  __builtin_neon_vld1q_x4_v(&__ret, __p0, 34); \
49624 \
49625  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49626  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49627  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49628  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49629  __ret; \
49630})
49631#endif
49632
49633#ifdef __LITTLE_ENDIAN__
49634#define vld1q_s64_x4(__p0) __extension__ ({ \
49635  int64x2x4_t __ret; \
49636  __builtin_neon_vld1q_x4_v(&__ret, __p0, 35); \
49637  __ret; \
49638})
49639#else
49640#define vld1q_s64_x4(__p0) __extension__ ({ \
49641  int64x2x4_t __ret; \
49642  __builtin_neon_vld1q_x4_v(&__ret, __p0, 35); \
49643 \
49644  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49645  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49646  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49647  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49648  __ret; \
49649})
49650#endif
49651
49652#ifdef __LITTLE_ENDIAN__
49653#define vld1q_s16_x4(__p0) __extension__ ({ \
49654  int16x8x4_t __ret; \
49655  __builtin_neon_vld1q_x4_v(&__ret, __p0, 33); \
49656  __ret; \
49657})
49658#else
49659#define vld1q_s16_x4(__p0) __extension__ ({ \
49660  int16x8x4_t __ret; \
49661  __builtin_neon_vld1q_x4_v(&__ret, __p0, 33); \
49662 \
49663  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49664  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49665  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49666  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49667  __ret; \
49668})
49669#endif
49670
49671#ifdef __LITTLE_ENDIAN__
49672#define vld1_u8_x4(__p0) __extension__ ({ \
49673  uint8x8x4_t __ret; \
49674  __builtin_neon_vld1_x4_v(&__ret, __p0, 16); \
49675  __ret; \
49676})
49677#else
49678#define vld1_u8_x4(__p0) __extension__ ({ \
49679  uint8x8x4_t __ret; \
49680  __builtin_neon_vld1_x4_v(&__ret, __p0, 16); \
49681 \
49682  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49683  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49684  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49685  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49686  __ret; \
49687})
49688#endif
49689
49690#ifdef __LITTLE_ENDIAN__
49691#define vld1_u32_x4(__p0) __extension__ ({ \
49692  uint32x2x4_t __ret; \
49693  __builtin_neon_vld1_x4_v(&__ret, __p0, 18); \
49694  __ret; \
49695})
49696#else
49697#define vld1_u32_x4(__p0) __extension__ ({ \
49698  uint32x2x4_t __ret; \
49699  __builtin_neon_vld1_x4_v(&__ret, __p0, 18); \
49700 \
49701  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49702  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49703  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49704  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49705  __ret; \
49706})
49707#endif
49708
49709#ifdef __LITTLE_ENDIAN__
49710#define vld1_u64_x4(__p0) __extension__ ({ \
49711  uint64x1x4_t __ret; \
49712  __builtin_neon_vld1_x4_v(&__ret, __p0, 19); \
49713  __ret; \
49714})
49715#else
49716#define vld1_u64_x4(__p0) __extension__ ({ \
49717  uint64x1x4_t __ret; \
49718  __builtin_neon_vld1_x4_v(&__ret, __p0, 19); \
49719  __ret; \
49720})
49721#endif
49722
49723#ifdef __LITTLE_ENDIAN__
49724#define vld1_u16_x4(__p0) __extension__ ({ \
49725  uint16x4x4_t __ret; \
49726  __builtin_neon_vld1_x4_v(&__ret, __p0, 17); \
49727  __ret; \
49728})
49729#else
49730#define vld1_u16_x4(__p0) __extension__ ({ \
49731  uint16x4x4_t __ret; \
49732  __builtin_neon_vld1_x4_v(&__ret, __p0, 17); \
49733 \
49734  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49735  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49736  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49737  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49738  __ret; \
49739})
49740#endif
49741
49742#ifdef __LITTLE_ENDIAN__
49743#define vld1_s8_x4(__p0) __extension__ ({ \
49744  int8x8x4_t __ret; \
49745  __builtin_neon_vld1_x4_v(&__ret, __p0, 0); \
49746  __ret; \
49747})
49748#else
49749#define vld1_s8_x4(__p0) __extension__ ({ \
49750  int8x8x4_t __ret; \
49751  __builtin_neon_vld1_x4_v(&__ret, __p0, 0); \
49752 \
49753  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
49754  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
49755  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
49756  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
49757  __ret; \
49758})
49759#endif
49760
49761#ifdef __LITTLE_ENDIAN__
49762#define vld1_f64_x4(__p0) __extension__ ({ \
49763  float64x1x4_t __ret; \
49764  __builtin_neon_vld1_x4_v(&__ret, __p0, 10); \
49765  __ret; \
49766})
49767#else
49768#define vld1_f64_x4(__p0) __extension__ ({ \
49769  float64x1x4_t __ret; \
49770  __builtin_neon_vld1_x4_v(&__ret, __p0, 10); \
49771  __ret; \
49772})
49773#endif
49774
49775#ifdef __LITTLE_ENDIAN__
49776#define vld1_f32_x4(__p0) __extension__ ({ \
49777  float32x2x4_t __ret; \
49778  __builtin_neon_vld1_x4_v(&__ret, __p0, 9); \
49779  __ret; \
49780})
49781#else
49782#define vld1_f32_x4(__p0) __extension__ ({ \
49783  float32x2x4_t __ret; \
49784  __builtin_neon_vld1_x4_v(&__ret, __p0, 9); \
49785 \
49786  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49787  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49788  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49789  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49790  __ret; \
49791})
49792#endif
49793
49794#ifdef __LITTLE_ENDIAN__
49795#define vld1_f16_x4(__p0) __extension__ ({ \
49796  float16x4x4_t __ret; \
49797  __builtin_neon_vld1_x4_v(&__ret, __p0, 8); \
49798  __ret; \
49799})
49800#else
49801#define vld1_f16_x4(__p0) __extension__ ({ \
49802  float16x4x4_t __ret; \
49803  __builtin_neon_vld1_x4_v(&__ret, __p0, 8); \
49804 \
49805  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49806  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49807  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49808  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49809  __ret; \
49810})
49811#endif
49812
49813#ifdef __LITTLE_ENDIAN__
49814#define vld1_s32_x4(__p0) __extension__ ({ \
49815  int32x2x4_t __ret; \
49816  __builtin_neon_vld1_x4_v(&__ret, __p0, 2); \
49817  __ret; \
49818})
49819#else
49820#define vld1_s32_x4(__p0) __extension__ ({ \
49821  int32x2x4_t __ret; \
49822  __builtin_neon_vld1_x4_v(&__ret, __p0, 2); \
49823 \
49824  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49825  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49826  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
49827  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
49828  __ret; \
49829})
49830#endif
49831
49832#ifdef __LITTLE_ENDIAN__
49833#define vld1_s64_x4(__p0) __extension__ ({ \
49834  int64x1x4_t __ret; \
49835  __builtin_neon_vld1_x4_v(&__ret, __p0, 3); \
49836  __ret; \
49837})
49838#else
49839#define vld1_s64_x4(__p0) __extension__ ({ \
49840  int64x1x4_t __ret; \
49841  __builtin_neon_vld1_x4_v(&__ret, __p0, 3); \
49842  __ret; \
49843})
49844#endif
49845
49846#ifdef __LITTLE_ENDIAN__
49847#define vld1_s16_x4(__p0) __extension__ ({ \
49848  int16x4x4_t __ret; \
49849  __builtin_neon_vld1_x4_v(&__ret, __p0, 1); \
49850  __ret; \
49851})
49852#else
49853#define vld1_s16_x4(__p0) __extension__ ({ \
49854  int16x4x4_t __ret; \
49855  __builtin_neon_vld1_x4_v(&__ret, __p0, 1); \
49856 \
49857  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
49858  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
49859  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
49860  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
49861  __ret; \
49862})
49863#endif
49864
49865#ifdef __LITTLE_ENDIAN__
49866#define vld2_p64(__p0) __extension__ ({ \
49867  poly64x1x2_t __ret; \
49868  __builtin_neon_vld2_v(&__ret, __p0, 6); \
49869  __ret; \
49870})
49871#else
49872#define vld2_p64(__p0) __extension__ ({ \
49873  poly64x1x2_t __ret; \
49874  __builtin_neon_vld2_v(&__ret, __p0, 6); \
49875  __ret; \
49876})
49877#endif
49878
49879#ifdef __LITTLE_ENDIAN__
49880#define vld2q_p64(__p0) __extension__ ({ \
49881  poly64x2x2_t __ret; \
49882  __builtin_neon_vld2q_v(&__ret, __p0, 38); \
49883  __ret; \
49884})
49885#else
49886#define vld2q_p64(__p0) __extension__ ({ \
49887  poly64x2x2_t __ret; \
49888  __builtin_neon_vld2q_v(&__ret, __p0, 38); \
49889 \
49890  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49891  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49892  __ret; \
49893})
49894#endif
49895
49896#ifdef __LITTLE_ENDIAN__
49897#define vld2q_u64(__p0) __extension__ ({ \
49898  uint64x2x2_t __ret; \
49899  __builtin_neon_vld2q_v(&__ret, __p0, 51); \
49900  __ret; \
49901})
49902#else
49903#define vld2q_u64(__p0) __extension__ ({ \
49904  uint64x2x2_t __ret; \
49905  __builtin_neon_vld2q_v(&__ret, __p0, 51); \
49906 \
49907  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49908  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49909  __ret; \
49910})
49911#endif
49912
49913#ifdef __LITTLE_ENDIAN__
49914#define vld2q_f64(__p0) __extension__ ({ \
49915  float64x2x2_t __ret; \
49916  __builtin_neon_vld2q_v(&__ret, __p0, 42); \
49917  __ret; \
49918})
49919#else
49920#define vld2q_f64(__p0) __extension__ ({ \
49921  float64x2x2_t __ret; \
49922  __builtin_neon_vld2q_v(&__ret, __p0, 42); \
49923 \
49924  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49925  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49926  __ret; \
49927})
49928#endif
49929
49930#ifdef __LITTLE_ENDIAN__
49931#define vld2q_s64(__p0) __extension__ ({ \
49932  int64x2x2_t __ret; \
49933  __builtin_neon_vld2q_v(&__ret, __p0, 35); \
49934  __ret; \
49935})
49936#else
49937#define vld2q_s64(__p0) __extension__ ({ \
49938  int64x2x2_t __ret; \
49939  __builtin_neon_vld2q_v(&__ret, __p0, 35); \
49940 \
49941  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
49942  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
49943  __ret; \
49944})
49945#endif
49946
49947#ifdef __LITTLE_ENDIAN__
49948#define vld2_f64(__p0) __extension__ ({ \
49949  float64x1x2_t __ret; \
49950  __builtin_neon_vld2_v(&__ret, __p0, 10); \
49951  __ret; \
49952})
49953#else
49954#define vld2_f64(__p0) __extension__ ({ \
49955  float64x1x2_t __ret; \
49956  __builtin_neon_vld2_v(&__ret, __p0, 10); \
49957  __ret; \
49958})
49959#endif
49960
49961#ifdef __LITTLE_ENDIAN__
49962#define vld2_dup_p64(__p0) __extension__ ({ \
49963  poly64x1x2_t __ret; \
49964  __builtin_neon_vld2_dup_v(&__ret, __p0, 6); \
49965  __ret; \
49966})
49967#else
49968#define vld2_dup_p64(__p0) __extension__ ({ \
49969  poly64x1x2_t __ret; \
49970  __builtin_neon_vld2_dup_v(&__ret, __p0, 6); \
49971  __ret; \
49972})
49973#endif
49974
49975#ifdef __LITTLE_ENDIAN__
49976#define vld2q_dup_p8(__p0) __extension__ ({ \
49977  poly8x16x2_t __ret; \
49978  __builtin_neon_vld2q_dup_v(&__ret, __p0, 36); \
49979  __ret; \
49980})
49981#else
49982#define vld2q_dup_p8(__p0) __extension__ ({ \
49983  poly8x16x2_t __ret; \
49984  __builtin_neon_vld2q_dup_v(&__ret, __p0, 36); \
49985 \
49986  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49987  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
49988  __ret; \
49989})
49990#endif
49991
49992#ifdef __LITTLE_ENDIAN__
49993#define vld2q_dup_p64(__p0) __extension__ ({ \
49994  poly64x2x2_t __ret; \
49995  __builtin_neon_vld2q_dup_v(&__ret, __p0, 38); \
49996  __ret; \
49997})
49998#else
49999#define vld2q_dup_p64(__p0) __extension__ ({ \
50000  poly64x2x2_t __ret; \
50001  __builtin_neon_vld2q_dup_v(&__ret, __p0, 38); \
50002 \
50003  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50004  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50005  __ret; \
50006})
50007#endif
50008
50009#ifdef __LITTLE_ENDIAN__
50010#define vld2q_dup_p16(__p0) __extension__ ({ \
50011  poly16x8x2_t __ret; \
50012  __builtin_neon_vld2q_dup_v(&__ret, __p0, 37); \
50013  __ret; \
50014})
50015#else
50016#define vld2q_dup_p16(__p0) __extension__ ({ \
50017  poly16x8x2_t __ret; \
50018  __builtin_neon_vld2q_dup_v(&__ret, __p0, 37); \
50019 \
50020  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50021  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50022  __ret; \
50023})
50024#endif
50025
50026#ifdef __LITTLE_ENDIAN__
50027#define vld2q_dup_u8(__p0) __extension__ ({ \
50028  uint8x16x2_t __ret; \
50029  __builtin_neon_vld2q_dup_v(&__ret, __p0, 48); \
50030  __ret; \
50031})
50032#else
50033#define vld2q_dup_u8(__p0) __extension__ ({ \
50034  uint8x16x2_t __ret; \
50035  __builtin_neon_vld2q_dup_v(&__ret, __p0, 48); \
50036 \
50037  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50038  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50039  __ret; \
50040})
50041#endif
50042
50043#ifdef __LITTLE_ENDIAN__
50044#define vld2q_dup_u32(__p0) __extension__ ({ \
50045  uint32x4x2_t __ret; \
50046  __builtin_neon_vld2q_dup_v(&__ret, __p0, 50); \
50047  __ret; \
50048})
50049#else
50050#define vld2q_dup_u32(__p0) __extension__ ({ \
50051  uint32x4x2_t __ret; \
50052  __builtin_neon_vld2q_dup_v(&__ret, __p0, 50); \
50053 \
50054  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50055  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50056  __ret; \
50057})
50058#endif
50059
50060#ifdef __LITTLE_ENDIAN__
50061#define vld2q_dup_u64(__p0) __extension__ ({ \
50062  uint64x2x2_t __ret; \
50063  __builtin_neon_vld2q_dup_v(&__ret, __p0, 51); \
50064  __ret; \
50065})
50066#else
50067#define vld2q_dup_u64(__p0) __extension__ ({ \
50068  uint64x2x2_t __ret; \
50069  __builtin_neon_vld2q_dup_v(&__ret, __p0, 51); \
50070 \
50071  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50072  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50073  __ret; \
50074})
50075#endif
50076
50077#ifdef __LITTLE_ENDIAN__
50078#define vld2q_dup_u16(__p0) __extension__ ({ \
50079  uint16x8x2_t __ret; \
50080  __builtin_neon_vld2q_dup_v(&__ret, __p0, 49); \
50081  __ret; \
50082})
50083#else
50084#define vld2q_dup_u16(__p0) __extension__ ({ \
50085  uint16x8x2_t __ret; \
50086  __builtin_neon_vld2q_dup_v(&__ret, __p0, 49); \
50087 \
50088  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50089  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50090  __ret; \
50091})
50092#endif
50093
50094#ifdef __LITTLE_ENDIAN__
50095#define vld2q_dup_s8(__p0) __extension__ ({ \
50096  int8x16x2_t __ret; \
50097  __builtin_neon_vld2q_dup_v(&__ret, __p0, 32); \
50098  __ret; \
50099})
50100#else
50101#define vld2q_dup_s8(__p0) __extension__ ({ \
50102  int8x16x2_t __ret; \
50103  __builtin_neon_vld2q_dup_v(&__ret, __p0, 32); \
50104 \
50105  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50106  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50107  __ret; \
50108})
50109#endif
50110
50111#ifdef __LITTLE_ENDIAN__
50112#define vld2q_dup_f64(__p0) __extension__ ({ \
50113  float64x2x2_t __ret; \
50114  __builtin_neon_vld2q_dup_v(&__ret, __p0, 42); \
50115  __ret; \
50116})
50117#else
50118#define vld2q_dup_f64(__p0) __extension__ ({ \
50119  float64x2x2_t __ret; \
50120  __builtin_neon_vld2q_dup_v(&__ret, __p0, 42); \
50121 \
50122  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50123  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50124  __ret; \
50125})
50126#endif
50127
50128#ifdef __LITTLE_ENDIAN__
50129#define vld2q_dup_f32(__p0) __extension__ ({ \
50130  float32x4x2_t __ret; \
50131  __builtin_neon_vld2q_dup_v(&__ret, __p0, 41); \
50132  __ret; \
50133})
50134#else
50135#define vld2q_dup_f32(__p0) __extension__ ({ \
50136  float32x4x2_t __ret; \
50137  __builtin_neon_vld2q_dup_v(&__ret, __p0, 41); \
50138 \
50139  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50140  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50141  __ret; \
50142})
50143#endif
50144
50145#ifdef __LITTLE_ENDIAN__
50146#define vld2q_dup_f16(__p0) __extension__ ({ \
50147  float16x8x2_t __ret; \
50148  __builtin_neon_vld2q_dup_v(&__ret, __p0, 40); \
50149  __ret; \
50150})
50151#else
50152#define vld2q_dup_f16(__p0) __extension__ ({ \
50153  float16x8x2_t __ret; \
50154  __builtin_neon_vld2q_dup_v(&__ret, __p0, 40); \
50155 \
50156  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50157  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50158  __ret; \
50159})
50160#endif
50161
50162#ifdef __LITTLE_ENDIAN__
50163#define vld2q_dup_s32(__p0) __extension__ ({ \
50164  int32x4x2_t __ret; \
50165  __builtin_neon_vld2q_dup_v(&__ret, __p0, 34); \
50166  __ret; \
50167})
50168#else
50169#define vld2q_dup_s32(__p0) __extension__ ({ \
50170  int32x4x2_t __ret; \
50171  __builtin_neon_vld2q_dup_v(&__ret, __p0, 34); \
50172 \
50173  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50174  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50175  __ret; \
50176})
50177#endif
50178
50179#ifdef __LITTLE_ENDIAN__
50180#define vld2q_dup_s64(__p0) __extension__ ({ \
50181  int64x2x2_t __ret; \
50182  __builtin_neon_vld2q_dup_v(&__ret, __p0, 35); \
50183  __ret; \
50184})
50185#else
50186#define vld2q_dup_s64(__p0) __extension__ ({ \
50187  int64x2x2_t __ret; \
50188  __builtin_neon_vld2q_dup_v(&__ret, __p0, 35); \
50189 \
50190  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50191  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50192  __ret; \
50193})
50194#endif
50195
50196#ifdef __LITTLE_ENDIAN__
50197#define vld2q_dup_s16(__p0) __extension__ ({ \
50198  int16x8x2_t __ret; \
50199  __builtin_neon_vld2q_dup_v(&__ret, __p0, 33); \
50200  __ret; \
50201})
50202#else
50203#define vld2q_dup_s16(__p0) __extension__ ({ \
50204  int16x8x2_t __ret; \
50205  __builtin_neon_vld2q_dup_v(&__ret, __p0, 33); \
50206 \
50207  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50208  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50209  __ret; \
50210})
50211#endif
50212
50213#ifdef __LITTLE_ENDIAN__
50214#define vld2_dup_f64(__p0) __extension__ ({ \
50215  float64x1x2_t __ret; \
50216  __builtin_neon_vld2_dup_v(&__ret, __p0, 10); \
50217  __ret; \
50218})
50219#else
50220#define vld2_dup_f64(__p0) __extension__ ({ \
50221  float64x1x2_t __ret; \
50222  __builtin_neon_vld2_dup_v(&__ret, __p0, 10); \
50223  __ret; \
50224})
50225#endif
50226
50227#ifdef __LITTLE_ENDIAN__
50228#define vld2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50229  poly64x1x2_t __s1 = __p1; \
50230  poly64x1x2_t __ret; \
50231  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
50232  __ret; \
50233})
50234#else
50235#define vld2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50236  poly64x1x2_t __s1 = __p1; \
50237  poly64x1x2_t __ret; \
50238  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
50239  __ret; \
50240})
50241#endif
50242
50243#ifdef __LITTLE_ENDIAN__
50244#define vld2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
50245  poly8x16x2_t __s1 = __p1; \
50246  poly8x16x2_t __ret; \
50247  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 36); \
50248  __ret; \
50249})
50250#else
50251#define vld2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
50252  poly8x16x2_t __s1 = __p1; \
50253  poly8x16x2_t __rev1; \
50254  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50255  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50256  poly8x16x2_t __ret; \
50257  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 36); \
50258 \
50259  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50260  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50261  __ret; \
50262})
50263#endif
50264
50265#ifdef __LITTLE_ENDIAN__
50266#define vld2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50267  poly64x2x2_t __s1 = __p1; \
50268  poly64x2x2_t __ret; \
50269  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 38); \
50270  __ret; \
50271})
50272#else
50273#define vld2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50274  poly64x2x2_t __s1 = __p1; \
50275  poly64x2x2_t __rev1; \
50276  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50277  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50278  poly64x2x2_t __ret; \
50279  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 38); \
50280 \
50281  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50282  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50283  __ret; \
50284})
50285#endif
50286
50287#ifdef __LITTLE_ENDIAN__
50288#define vld2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
50289  uint8x16x2_t __s1 = __p1; \
50290  uint8x16x2_t __ret; \
50291  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 48); \
50292  __ret; \
50293})
50294#else
50295#define vld2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
50296  uint8x16x2_t __s1 = __p1; \
50297  uint8x16x2_t __rev1; \
50298  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50299  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50300  uint8x16x2_t __ret; \
50301  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 48); \
50302 \
50303  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50304  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50305  __ret; \
50306})
50307#endif
50308
50309#ifdef __LITTLE_ENDIAN__
50310#define vld2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50311  uint64x2x2_t __s1 = __p1; \
50312  uint64x2x2_t __ret; \
50313  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 51); \
50314  __ret; \
50315})
50316#else
50317#define vld2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50318  uint64x2x2_t __s1 = __p1; \
50319  uint64x2x2_t __rev1; \
50320  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50321  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50322  uint64x2x2_t __ret; \
50323  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 51); \
50324 \
50325  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50326  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50327  __ret; \
50328})
50329#endif
50330
50331#ifdef __LITTLE_ENDIAN__
50332#define vld2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
50333  int8x16x2_t __s1 = __p1; \
50334  int8x16x2_t __ret; \
50335  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 32); \
50336  __ret; \
50337})
50338#else
50339#define vld2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
50340  int8x16x2_t __s1 = __p1; \
50341  int8x16x2_t __rev1; \
50342  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50343  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50344  int8x16x2_t __ret; \
50345  __builtin_neon_vld2q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 32); \
50346 \
50347  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50348  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50349  __ret; \
50350})
50351#endif
50352
50353#ifdef __LITTLE_ENDIAN__
50354#define vld2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50355  float64x2x2_t __s1 = __p1; \
50356  float64x2x2_t __ret; \
50357  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 42); \
50358  __ret; \
50359})
50360#else
50361#define vld2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50362  float64x2x2_t __s1 = __p1; \
50363  float64x2x2_t __rev1; \
50364  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50365  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50366  float64x2x2_t __ret; \
50367  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 42); \
50368 \
50369  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50370  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50371  __ret; \
50372})
50373#endif
50374
50375#ifdef __LITTLE_ENDIAN__
50376#define vld2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50377  int64x2x2_t __s1 = __p1; \
50378  int64x2x2_t __ret; \
50379  __builtin_neon_vld2q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 35); \
50380  __ret; \
50381})
50382#else
50383#define vld2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50384  int64x2x2_t __s1 = __p1; \
50385  int64x2x2_t __rev1; \
50386  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50387  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50388  int64x2x2_t __ret; \
50389  __builtin_neon_vld2q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __p2, 35); \
50390 \
50391  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50392  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50393  __ret; \
50394})
50395#endif
50396
50397#ifdef __LITTLE_ENDIAN__
50398#define vld2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50399  uint64x1x2_t __s1 = __p1; \
50400  uint64x1x2_t __ret; \
50401  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
50402  __ret; \
50403})
50404#else
50405#define vld2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50406  uint64x1x2_t __s1 = __p1; \
50407  uint64x1x2_t __ret; \
50408  __builtin_neon_vld2_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
50409  __ret; \
50410})
50411#endif
50412
50413#ifdef __LITTLE_ENDIAN__
50414#define vld2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50415  float64x1x2_t __s1 = __p1; \
50416  float64x1x2_t __ret; \
50417  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 10); \
50418  __ret; \
50419})
50420#else
50421#define vld2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50422  float64x1x2_t __s1 = __p1; \
50423  float64x1x2_t __ret; \
50424  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 10); \
50425  __ret; \
50426})
50427#endif
50428
50429#ifdef __LITTLE_ENDIAN__
50430#define vld2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50431  int64x1x2_t __s1 = __p1; \
50432  int64x1x2_t __ret; \
50433  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 3); \
50434  __ret; \
50435})
50436#else
50437#define vld2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50438  int64x1x2_t __s1 = __p1; \
50439  int64x1x2_t __ret; \
50440  __builtin_neon_vld2_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __p2, 3); \
50441  __ret; \
50442})
50443#endif
50444
50445#ifdef __LITTLE_ENDIAN__
50446#define vld3_p64(__p0) __extension__ ({ \
50447  poly64x1x3_t __ret; \
50448  __builtin_neon_vld3_v(&__ret, __p0, 6); \
50449  __ret; \
50450})
50451#else
50452#define vld3_p64(__p0) __extension__ ({ \
50453  poly64x1x3_t __ret; \
50454  __builtin_neon_vld3_v(&__ret, __p0, 6); \
50455  __ret; \
50456})
50457#endif
50458
50459#ifdef __LITTLE_ENDIAN__
50460#define vld3q_p64(__p0) __extension__ ({ \
50461  poly64x2x3_t __ret; \
50462  __builtin_neon_vld3q_v(&__ret, __p0, 38); \
50463  __ret; \
50464})
50465#else
50466#define vld3q_p64(__p0) __extension__ ({ \
50467  poly64x2x3_t __ret; \
50468  __builtin_neon_vld3q_v(&__ret, __p0, 38); \
50469 \
50470  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50471  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50472  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50473  __ret; \
50474})
50475#endif
50476
50477#ifdef __LITTLE_ENDIAN__
50478#define vld3q_u64(__p0) __extension__ ({ \
50479  uint64x2x3_t __ret; \
50480  __builtin_neon_vld3q_v(&__ret, __p0, 51); \
50481  __ret; \
50482})
50483#else
50484#define vld3q_u64(__p0) __extension__ ({ \
50485  uint64x2x3_t __ret; \
50486  __builtin_neon_vld3q_v(&__ret, __p0, 51); \
50487 \
50488  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50489  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50490  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50491  __ret; \
50492})
50493#endif
50494
50495#ifdef __LITTLE_ENDIAN__
50496#define vld3q_f64(__p0) __extension__ ({ \
50497  float64x2x3_t __ret; \
50498  __builtin_neon_vld3q_v(&__ret, __p0, 42); \
50499  __ret; \
50500})
50501#else
50502#define vld3q_f64(__p0) __extension__ ({ \
50503  float64x2x3_t __ret; \
50504  __builtin_neon_vld3q_v(&__ret, __p0, 42); \
50505 \
50506  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50507  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50508  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50509  __ret; \
50510})
50511#endif
50512
50513#ifdef __LITTLE_ENDIAN__
50514#define vld3q_s64(__p0) __extension__ ({ \
50515  int64x2x3_t __ret; \
50516  __builtin_neon_vld3q_v(&__ret, __p0, 35); \
50517  __ret; \
50518})
50519#else
50520#define vld3q_s64(__p0) __extension__ ({ \
50521  int64x2x3_t __ret; \
50522  __builtin_neon_vld3q_v(&__ret, __p0, 35); \
50523 \
50524  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50525  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50526  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50527  __ret; \
50528})
50529#endif
50530
50531#ifdef __LITTLE_ENDIAN__
50532#define vld3_f64(__p0) __extension__ ({ \
50533  float64x1x3_t __ret; \
50534  __builtin_neon_vld3_v(&__ret, __p0, 10); \
50535  __ret; \
50536})
50537#else
50538#define vld3_f64(__p0) __extension__ ({ \
50539  float64x1x3_t __ret; \
50540  __builtin_neon_vld3_v(&__ret, __p0, 10); \
50541  __ret; \
50542})
50543#endif
50544
50545#ifdef __LITTLE_ENDIAN__
50546#define vld3_dup_p64(__p0) __extension__ ({ \
50547  poly64x1x3_t __ret; \
50548  __builtin_neon_vld3_dup_v(&__ret, __p0, 6); \
50549  __ret; \
50550})
50551#else
50552#define vld3_dup_p64(__p0) __extension__ ({ \
50553  poly64x1x3_t __ret; \
50554  __builtin_neon_vld3_dup_v(&__ret, __p0, 6); \
50555  __ret; \
50556})
50557#endif
50558
50559#ifdef __LITTLE_ENDIAN__
50560#define vld3q_dup_p8(__p0) __extension__ ({ \
50561  poly8x16x3_t __ret; \
50562  __builtin_neon_vld3q_dup_v(&__ret, __p0, 36); \
50563  __ret; \
50564})
50565#else
50566#define vld3q_dup_p8(__p0) __extension__ ({ \
50567  poly8x16x3_t __ret; \
50568  __builtin_neon_vld3q_dup_v(&__ret, __p0, 36); \
50569 \
50570  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50571  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50572  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50573  __ret; \
50574})
50575#endif
50576
50577#ifdef __LITTLE_ENDIAN__
50578#define vld3q_dup_p64(__p0) __extension__ ({ \
50579  poly64x2x3_t __ret; \
50580  __builtin_neon_vld3q_dup_v(&__ret, __p0, 38); \
50581  __ret; \
50582})
50583#else
50584#define vld3q_dup_p64(__p0) __extension__ ({ \
50585  poly64x2x3_t __ret; \
50586  __builtin_neon_vld3q_dup_v(&__ret, __p0, 38); \
50587 \
50588  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50589  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50590  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50591  __ret; \
50592})
50593#endif
50594
50595#ifdef __LITTLE_ENDIAN__
50596#define vld3q_dup_p16(__p0) __extension__ ({ \
50597  poly16x8x3_t __ret; \
50598  __builtin_neon_vld3q_dup_v(&__ret, __p0, 37); \
50599  __ret; \
50600})
50601#else
50602#define vld3q_dup_p16(__p0) __extension__ ({ \
50603  poly16x8x3_t __ret; \
50604  __builtin_neon_vld3q_dup_v(&__ret, __p0, 37); \
50605 \
50606  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50607  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50608  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50609  __ret; \
50610})
50611#endif
50612
50613#ifdef __LITTLE_ENDIAN__
50614#define vld3q_dup_u8(__p0) __extension__ ({ \
50615  uint8x16x3_t __ret; \
50616  __builtin_neon_vld3q_dup_v(&__ret, __p0, 48); \
50617  __ret; \
50618})
50619#else
50620#define vld3q_dup_u8(__p0) __extension__ ({ \
50621  uint8x16x3_t __ret; \
50622  __builtin_neon_vld3q_dup_v(&__ret, __p0, 48); \
50623 \
50624  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50625  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50626  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50627  __ret; \
50628})
50629#endif
50630
50631#ifdef __LITTLE_ENDIAN__
50632#define vld3q_dup_u32(__p0) __extension__ ({ \
50633  uint32x4x3_t __ret; \
50634  __builtin_neon_vld3q_dup_v(&__ret, __p0, 50); \
50635  __ret; \
50636})
50637#else
50638#define vld3q_dup_u32(__p0) __extension__ ({ \
50639  uint32x4x3_t __ret; \
50640  __builtin_neon_vld3q_dup_v(&__ret, __p0, 50); \
50641 \
50642  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50643  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50644  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
50645  __ret; \
50646})
50647#endif
50648
50649#ifdef __LITTLE_ENDIAN__
50650#define vld3q_dup_u64(__p0) __extension__ ({ \
50651  uint64x2x3_t __ret; \
50652  __builtin_neon_vld3q_dup_v(&__ret, __p0, 51); \
50653  __ret; \
50654})
50655#else
50656#define vld3q_dup_u64(__p0) __extension__ ({ \
50657  uint64x2x3_t __ret; \
50658  __builtin_neon_vld3q_dup_v(&__ret, __p0, 51); \
50659 \
50660  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50661  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50662  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50663  __ret; \
50664})
50665#endif
50666
50667#ifdef __LITTLE_ENDIAN__
50668#define vld3q_dup_u16(__p0) __extension__ ({ \
50669  uint16x8x3_t __ret; \
50670  __builtin_neon_vld3q_dup_v(&__ret, __p0, 49); \
50671  __ret; \
50672})
50673#else
50674#define vld3q_dup_u16(__p0) __extension__ ({ \
50675  uint16x8x3_t __ret; \
50676  __builtin_neon_vld3q_dup_v(&__ret, __p0, 49); \
50677 \
50678  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50679  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50680  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50681  __ret; \
50682})
50683#endif
50684
50685#ifdef __LITTLE_ENDIAN__
50686#define vld3q_dup_s8(__p0) __extension__ ({ \
50687  int8x16x3_t __ret; \
50688  __builtin_neon_vld3q_dup_v(&__ret, __p0, 32); \
50689  __ret; \
50690})
50691#else
50692#define vld3q_dup_s8(__p0) __extension__ ({ \
50693  int8x16x3_t __ret; \
50694  __builtin_neon_vld3q_dup_v(&__ret, __p0, 32); \
50695 \
50696  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50697  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50698  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50699  __ret; \
50700})
50701#endif
50702
50703#ifdef __LITTLE_ENDIAN__
50704#define vld3q_dup_f64(__p0) __extension__ ({ \
50705  float64x2x3_t __ret; \
50706  __builtin_neon_vld3q_dup_v(&__ret, __p0, 42); \
50707  __ret; \
50708})
50709#else
50710#define vld3q_dup_f64(__p0) __extension__ ({ \
50711  float64x2x3_t __ret; \
50712  __builtin_neon_vld3q_dup_v(&__ret, __p0, 42); \
50713 \
50714  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50715  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50716  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50717  __ret; \
50718})
50719#endif
50720
50721#ifdef __LITTLE_ENDIAN__
50722#define vld3q_dup_f32(__p0) __extension__ ({ \
50723  float32x4x3_t __ret; \
50724  __builtin_neon_vld3q_dup_v(&__ret, __p0, 41); \
50725  __ret; \
50726})
50727#else
50728#define vld3q_dup_f32(__p0) __extension__ ({ \
50729  float32x4x3_t __ret; \
50730  __builtin_neon_vld3q_dup_v(&__ret, __p0, 41); \
50731 \
50732  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50733  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50734  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
50735  __ret; \
50736})
50737#endif
50738
50739#ifdef __LITTLE_ENDIAN__
50740#define vld3q_dup_f16(__p0) __extension__ ({ \
50741  float16x8x3_t __ret; \
50742  __builtin_neon_vld3q_dup_v(&__ret, __p0, 40); \
50743  __ret; \
50744})
50745#else
50746#define vld3q_dup_f16(__p0) __extension__ ({ \
50747  float16x8x3_t __ret; \
50748  __builtin_neon_vld3q_dup_v(&__ret, __p0, 40); \
50749 \
50750  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50751  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50752  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50753  __ret; \
50754})
50755#endif
50756
50757#ifdef __LITTLE_ENDIAN__
50758#define vld3q_dup_s32(__p0) __extension__ ({ \
50759  int32x4x3_t __ret; \
50760  __builtin_neon_vld3q_dup_v(&__ret, __p0, 34); \
50761  __ret; \
50762})
50763#else
50764#define vld3q_dup_s32(__p0) __extension__ ({ \
50765  int32x4x3_t __ret; \
50766  __builtin_neon_vld3q_dup_v(&__ret, __p0, 34); \
50767 \
50768  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
50769  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
50770  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
50771  __ret; \
50772})
50773#endif
50774
50775#ifdef __LITTLE_ENDIAN__
50776#define vld3q_dup_s64(__p0) __extension__ ({ \
50777  int64x2x3_t __ret; \
50778  __builtin_neon_vld3q_dup_v(&__ret, __p0, 35); \
50779  __ret; \
50780})
50781#else
50782#define vld3q_dup_s64(__p0) __extension__ ({ \
50783  int64x2x3_t __ret; \
50784  __builtin_neon_vld3q_dup_v(&__ret, __p0, 35); \
50785 \
50786  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50787  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50788  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50789  __ret; \
50790})
50791#endif
50792
50793#ifdef __LITTLE_ENDIAN__
50794#define vld3q_dup_s16(__p0) __extension__ ({ \
50795  int16x8x3_t __ret; \
50796  __builtin_neon_vld3q_dup_v(&__ret, __p0, 33); \
50797  __ret; \
50798})
50799#else
50800#define vld3q_dup_s16(__p0) __extension__ ({ \
50801  int16x8x3_t __ret; \
50802  __builtin_neon_vld3q_dup_v(&__ret, __p0, 33); \
50803 \
50804  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
50805  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
50806  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
50807  __ret; \
50808})
50809#endif
50810
50811#ifdef __LITTLE_ENDIAN__
50812#define vld3_dup_f64(__p0) __extension__ ({ \
50813  float64x1x3_t __ret; \
50814  __builtin_neon_vld3_dup_v(&__ret, __p0, 10); \
50815  __ret; \
50816})
50817#else
50818#define vld3_dup_f64(__p0) __extension__ ({ \
50819  float64x1x3_t __ret; \
50820  __builtin_neon_vld3_dup_v(&__ret, __p0, 10); \
50821  __ret; \
50822})
50823#endif
50824
50825#ifdef __LITTLE_ENDIAN__
50826#define vld3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50827  poly64x1x3_t __s1 = __p1; \
50828  poly64x1x3_t __ret; \
50829  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
50830  __ret; \
50831})
50832#else
50833#define vld3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50834  poly64x1x3_t __s1 = __p1; \
50835  poly64x1x3_t __ret; \
50836  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
50837  __ret; \
50838})
50839#endif
50840
50841#ifdef __LITTLE_ENDIAN__
50842#define vld3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
50843  poly8x16x3_t __s1 = __p1; \
50844  poly8x16x3_t __ret; \
50845  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 36); \
50846  __ret; \
50847})
50848#else
50849#define vld3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
50850  poly8x16x3_t __s1 = __p1; \
50851  poly8x16x3_t __rev1; \
50852  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50853  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50854  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50855  poly8x16x3_t __ret; \
50856  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 36); \
50857 \
50858  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50859  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50860  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50861  __ret; \
50862})
50863#endif
50864
50865#ifdef __LITTLE_ENDIAN__
50866#define vld3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50867  poly64x2x3_t __s1 = __p1; \
50868  poly64x2x3_t __ret; \
50869  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 38); \
50870  __ret; \
50871})
50872#else
50873#define vld3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
50874  poly64x2x3_t __s1 = __p1; \
50875  poly64x2x3_t __rev1; \
50876  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50877  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50878  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50879  poly64x2x3_t __ret; \
50880  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 38); \
50881 \
50882  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50883  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50884  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50885  __ret; \
50886})
50887#endif
50888
50889#ifdef __LITTLE_ENDIAN__
50890#define vld3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
50891  uint8x16x3_t __s1 = __p1; \
50892  uint8x16x3_t __ret; \
50893  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 48); \
50894  __ret; \
50895})
50896#else
50897#define vld3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
50898  uint8x16x3_t __s1 = __p1; \
50899  uint8x16x3_t __rev1; \
50900  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50901  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50902  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50903  uint8x16x3_t __ret; \
50904  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 48); \
50905 \
50906  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50907  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50908  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50909  __ret; \
50910})
50911#endif
50912
50913#ifdef __LITTLE_ENDIAN__
50914#define vld3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50915  uint64x2x3_t __s1 = __p1; \
50916  uint64x2x3_t __ret; \
50917  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 51); \
50918  __ret; \
50919})
50920#else
50921#define vld3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
50922  uint64x2x3_t __s1 = __p1; \
50923  uint64x2x3_t __rev1; \
50924  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50925  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50926  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50927  uint64x2x3_t __ret; \
50928  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 51); \
50929 \
50930  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50931  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50932  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50933  __ret; \
50934})
50935#endif
50936
50937#ifdef __LITTLE_ENDIAN__
50938#define vld3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
50939  int8x16x3_t __s1 = __p1; \
50940  int8x16x3_t __ret; \
50941  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 32); \
50942  __ret; \
50943})
50944#else
50945#define vld3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
50946  int8x16x3_t __s1 = __p1; \
50947  int8x16x3_t __rev1; \
50948  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50949  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50950  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50951  int8x16x3_t __ret; \
50952  __builtin_neon_vld3q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 32); \
50953 \
50954  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50955  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50956  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
50957  __ret; \
50958})
50959#endif
50960
50961#ifdef __LITTLE_ENDIAN__
50962#define vld3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50963  float64x2x3_t __s1 = __p1; \
50964  float64x2x3_t __ret; \
50965  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 42); \
50966  __ret; \
50967})
50968#else
50969#define vld3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
50970  float64x2x3_t __s1 = __p1; \
50971  float64x2x3_t __rev1; \
50972  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50973  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50974  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50975  float64x2x3_t __ret; \
50976  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 42); \
50977 \
50978  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
50979  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
50980  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
50981  __ret; \
50982})
50983#endif
50984
50985#ifdef __LITTLE_ENDIAN__
50986#define vld3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50987  int64x2x3_t __s1 = __p1; \
50988  int64x2x3_t __ret; \
50989  __builtin_neon_vld3q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 35); \
50990  __ret; \
50991})
50992#else
50993#define vld3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
50994  int64x2x3_t __s1 = __p1; \
50995  int64x2x3_t __rev1; \
50996  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
50997  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
50998  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
50999  int64x2x3_t __ret; \
51000  __builtin_neon_vld3q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 35); \
51001 \
51002  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51003  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51004  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51005  __ret; \
51006})
51007#endif
51008
51009#ifdef __LITTLE_ENDIAN__
51010#define vld3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51011  uint64x1x3_t __s1 = __p1; \
51012  uint64x1x3_t __ret; \
51013  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
51014  __ret; \
51015})
51016#else
51017#define vld3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51018  uint64x1x3_t __s1 = __p1; \
51019  uint64x1x3_t __ret; \
51020  __builtin_neon_vld3_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
51021  __ret; \
51022})
51023#endif
51024
51025#ifdef __LITTLE_ENDIAN__
51026#define vld3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51027  float64x1x3_t __s1 = __p1; \
51028  float64x1x3_t __ret; \
51029  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
51030  __ret; \
51031})
51032#else
51033#define vld3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51034  float64x1x3_t __s1 = __p1; \
51035  float64x1x3_t __ret; \
51036  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
51037  __ret; \
51038})
51039#endif
51040
51041#ifdef __LITTLE_ENDIAN__
51042#define vld3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51043  int64x1x3_t __s1 = __p1; \
51044  int64x1x3_t __ret; \
51045  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
51046  __ret; \
51047})
51048#else
51049#define vld3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51050  int64x1x3_t __s1 = __p1; \
51051  int64x1x3_t __ret; \
51052  __builtin_neon_vld3_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
51053  __ret; \
51054})
51055#endif
51056
51057#ifdef __LITTLE_ENDIAN__
51058#define vld4_p64(__p0) __extension__ ({ \
51059  poly64x1x4_t __ret; \
51060  __builtin_neon_vld4_v(&__ret, __p0, 6); \
51061  __ret; \
51062})
51063#else
51064#define vld4_p64(__p0) __extension__ ({ \
51065  poly64x1x4_t __ret; \
51066  __builtin_neon_vld4_v(&__ret, __p0, 6); \
51067  __ret; \
51068})
51069#endif
51070
51071#ifdef __LITTLE_ENDIAN__
51072#define vld4q_p64(__p0) __extension__ ({ \
51073  poly64x2x4_t __ret; \
51074  __builtin_neon_vld4q_v(&__ret, __p0, 38); \
51075  __ret; \
51076})
51077#else
51078#define vld4q_p64(__p0) __extension__ ({ \
51079  poly64x2x4_t __ret; \
51080  __builtin_neon_vld4q_v(&__ret, __p0, 38); \
51081 \
51082  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51083  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51084  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51085  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51086  __ret; \
51087})
51088#endif
51089
51090#ifdef __LITTLE_ENDIAN__
51091#define vld4q_u64(__p0) __extension__ ({ \
51092  uint64x2x4_t __ret; \
51093  __builtin_neon_vld4q_v(&__ret, __p0, 51); \
51094  __ret; \
51095})
51096#else
51097#define vld4q_u64(__p0) __extension__ ({ \
51098  uint64x2x4_t __ret; \
51099  __builtin_neon_vld4q_v(&__ret, __p0, 51); \
51100 \
51101  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51102  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51103  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51104  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51105  __ret; \
51106})
51107#endif
51108
51109#ifdef __LITTLE_ENDIAN__
51110#define vld4q_f64(__p0) __extension__ ({ \
51111  float64x2x4_t __ret; \
51112  __builtin_neon_vld4q_v(&__ret, __p0, 42); \
51113  __ret; \
51114})
51115#else
51116#define vld4q_f64(__p0) __extension__ ({ \
51117  float64x2x4_t __ret; \
51118  __builtin_neon_vld4q_v(&__ret, __p0, 42); \
51119 \
51120  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51121  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51122  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51123  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51124  __ret; \
51125})
51126#endif
51127
51128#ifdef __LITTLE_ENDIAN__
51129#define vld4q_s64(__p0) __extension__ ({ \
51130  int64x2x4_t __ret; \
51131  __builtin_neon_vld4q_v(&__ret, __p0, 35); \
51132  __ret; \
51133})
51134#else
51135#define vld4q_s64(__p0) __extension__ ({ \
51136  int64x2x4_t __ret; \
51137  __builtin_neon_vld4q_v(&__ret, __p0, 35); \
51138 \
51139  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51140  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51141  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51142  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51143  __ret; \
51144})
51145#endif
51146
51147#ifdef __LITTLE_ENDIAN__
51148#define vld4_f64(__p0) __extension__ ({ \
51149  float64x1x4_t __ret; \
51150  __builtin_neon_vld4_v(&__ret, __p0, 10); \
51151  __ret; \
51152})
51153#else
51154#define vld4_f64(__p0) __extension__ ({ \
51155  float64x1x4_t __ret; \
51156  __builtin_neon_vld4_v(&__ret, __p0, 10); \
51157  __ret; \
51158})
51159#endif
51160
51161#ifdef __LITTLE_ENDIAN__
51162#define vld4_dup_p64(__p0) __extension__ ({ \
51163  poly64x1x4_t __ret; \
51164  __builtin_neon_vld4_dup_v(&__ret, __p0, 6); \
51165  __ret; \
51166})
51167#else
51168#define vld4_dup_p64(__p0) __extension__ ({ \
51169  poly64x1x4_t __ret; \
51170  __builtin_neon_vld4_dup_v(&__ret, __p0, 6); \
51171  __ret; \
51172})
51173#endif
51174
51175#ifdef __LITTLE_ENDIAN__
51176#define vld4q_dup_p8(__p0) __extension__ ({ \
51177  poly8x16x4_t __ret; \
51178  __builtin_neon_vld4q_dup_v(&__ret, __p0, 36); \
51179  __ret; \
51180})
51181#else
51182#define vld4q_dup_p8(__p0) __extension__ ({ \
51183  poly8x16x4_t __ret; \
51184  __builtin_neon_vld4q_dup_v(&__ret, __p0, 36); \
51185 \
51186  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51187  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51188  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51189  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51190  __ret; \
51191})
51192#endif
51193
51194#ifdef __LITTLE_ENDIAN__
51195#define vld4q_dup_p64(__p0) __extension__ ({ \
51196  poly64x2x4_t __ret; \
51197  __builtin_neon_vld4q_dup_v(&__ret, __p0, 38); \
51198  __ret; \
51199})
51200#else
51201#define vld4q_dup_p64(__p0) __extension__ ({ \
51202  poly64x2x4_t __ret; \
51203  __builtin_neon_vld4q_dup_v(&__ret, __p0, 38); \
51204 \
51205  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51206  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51207  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51208  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51209  __ret; \
51210})
51211#endif
51212
51213#ifdef __LITTLE_ENDIAN__
51214#define vld4q_dup_p16(__p0) __extension__ ({ \
51215  poly16x8x4_t __ret; \
51216  __builtin_neon_vld4q_dup_v(&__ret, __p0, 37); \
51217  __ret; \
51218})
51219#else
51220#define vld4q_dup_p16(__p0) __extension__ ({ \
51221  poly16x8x4_t __ret; \
51222  __builtin_neon_vld4q_dup_v(&__ret, __p0, 37); \
51223 \
51224  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
51225  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
51226  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
51227  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
51228  __ret; \
51229})
51230#endif
51231
51232#ifdef __LITTLE_ENDIAN__
51233#define vld4q_dup_u8(__p0) __extension__ ({ \
51234  uint8x16x4_t __ret; \
51235  __builtin_neon_vld4q_dup_v(&__ret, __p0, 48); \
51236  __ret; \
51237})
51238#else
51239#define vld4q_dup_u8(__p0) __extension__ ({ \
51240  uint8x16x4_t __ret; \
51241  __builtin_neon_vld4q_dup_v(&__ret, __p0, 48); \
51242 \
51243  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51244  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51245  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51246  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51247  __ret; \
51248})
51249#endif
51250
51251#ifdef __LITTLE_ENDIAN__
51252#define vld4q_dup_u32(__p0) __extension__ ({ \
51253  uint32x4x4_t __ret; \
51254  __builtin_neon_vld4q_dup_v(&__ret, __p0, 50); \
51255  __ret; \
51256})
51257#else
51258#define vld4q_dup_u32(__p0) __extension__ ({ \
51259  uint32x4x4_t __ret; \
51260  __builtin_neon_vld4q_dup_v(&__ret, __p0, 50); \
51261 \
51262  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
51263  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
51264  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
51265  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
51266  __ret; \
51267})
51268#endif
51269
51270#ifdef __LITTLE_ENDIAN__
51271#define vld4q_dup_u64(__p0) __extension__ ({ \
51272  uint64x2x4_t __ret; \
51273  __builtin_neon_vld4q_dup_v(&__ret, __p0, 51); \
51274  __ret; \
51275})
51276#else
51277#define vld4q_dup_u64(__p0) __extension__ ({ \
51278  uint64x2x4_t __ret; \
51279  __builtin_neon_vld4q_dup_v(&__ret, __p0, 51); \
51280 \
51281  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51282  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51283  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51284  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51285  __ret; \
51286})
51287#endif
51288
51289#ifdef __LITTLE_ENDIAN__
51290#define vld4q_dup_u16(__p0) __extension__ ({ \
51291  uint16x8x4_t __ret; \
51292  __builtin_neon_vld4q_dup_v(&__ret, __p0, 49); \
51293  __ret; \
51294})
51295#else
51296#define vld4q_dup_u16(__p0) __extension__ ({ \
51297  uint16x8x4_t __ret; \
51298  __builtin_neon_vld4q_dup_v(&__ret, __p0, 49); \
51299 \
51300  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
51301  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
51302  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
51303  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
51304  __ret; \
51305})
51306#endif
51307
51308#ifdef __LITTLE_ENDIAN__
51309#define vld4q_dup_s8(__p0) __extension__ ({ \
51310  int8x16x4_t __ret; \
51311  __builtin_neon_vld4q_dup_v(&__ret, __p0, 32); \
51312  __ret; \
51313})
51314#else
51315#define vld4q_dup_s8(__p0) __extension__ ({ \
51316  int8x16x4_t __ret; \
51317  __builtin_neon_vld4q_dup_v(&__ret, __p0, 32); \
51318 \
51319  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51320  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51321  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51322  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51323  __ret; \
51324})
51325#endif
51326
51327#ifdef __LITTLE_ENDIAN__
51328#define vld4q_dup_f64(__p0) __extension__ ({ \
51329  float64x2x4_t __ret; \
51330  __builtin_neon_vld4q_dup_v(&__ret, __p0, 42); \
51331  __ret; \
51332})
51333#else
51334#define vld4q_dup_f64(__p0) __extension__ ({ \
51335  float64x2x4_t __ret; \
51336  __builtin_neon_vld4q_dup_v(&__ret, __p0, 42); \
51337 \
51338  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51339  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51340  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51341  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51342  __ret; \
51343})
51344#endif
51345
51346#ifdef __LITTLE_ENDIAN__
51347#define vld4q_dup_f32(__p0) __extension__ ({ \
51348  float32x4x4_t __ret; \
51349  __builtin_neon_vld4q_dup_v(&__ret, __p0, 41); \
51350  __ret; \
51351})
51352#else
51353#define vld4q_dup_f32(__p0) __extension__ ({ \
51354  float32x4x4_t __ret; \
51355  __builtin_neon_vld4q_dup_v(&__ret, __p0, 41); \
51356 \
51357  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
51358  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
51359  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
51360  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
51361  __ret; \
51362})
51363#endif
51364
51365#ifdef __LITTLE_ENDIAN__
51366#define vld4q_dup_f16(__p0) __extension__ ({ \
51367  float16x8x4_t __ret; \
51368  __builtin_neon_vld4q_dup_v(&__ret, __p0, 40); \
51369  __ret; \
51370})
51371#else
51372#define vld4q_dup_f16(__p0) __extension__ ({ \
51373  float16x8x4_t __ret; \
51374  __builtin_neon_vld4q_dup_v(&__ret, __p0, 40); \
51375 \
51376  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
51377  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
51378  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
51379  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
51380  __ret; \
51381})
51382#endif
51383
51384#ifdef __LITTLE_ENDIAN__
51385#define vld4q_dup_s32(__p0) __extension__ ({ \
51386  int32x4x4_t __ret; \
51387  __builtin_neon_vld4q_dup_v(&__ret, __p0, 34); \
51388  __ret; \
51389})
51390#else
51391#define vld4q_dup_s32(__p0) __extension__ ({ \
51392  int32x4x4_t __ret; \
51393  __builtin_neon_vld4q_dup_v(&__ret, __p0, 34); \
51394 \
51395  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 3, 2, 1, 0); \
51396  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 3, 2, 1, 0); \
51397  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 3, 2, 1, 0); \
51398  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 3, 2, 1, 0); \
51399  __ret; \
51400})
51401#endif
51402
51403#ifdef __LITTLE_ENDIAN__
51404#define vld4q_dup_s64(__p0) __extension__ ({ \
51405  int64x2x4_t __ret; \
51406  __builtin_neon_vld4q_dup_v(&__ret, __p0, 35); \
51407  __ret; \
51408})
51409#else
51410#define vld4q_dup_s64(__p0) __extension__ ({ \
51411  int64x2x4_t __ret; \
51412  __builtin_neon_vld4q_dup_v(&__ret, __p0, 35); \
51413 \
51414  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51415  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51416  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51417  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51418  __ret; \
51419})
51420#endif
51421
51422#ifdef __LITTLE_ENDIAN__
51423#define vld4q_dup_s16(__p0) __extension__ ({ \
51424  int16x8x4_t __ret; \
51425  __builtin_neon_vld4q_dup_v(&__ret, __p0, 33); \
51426  __ret; \
51427})
51428#else
51429#define vld4q_dup_s16(__p0) __extension__ ({ \
51430  int16x8x4_t __ret; \
51431  __builtin_neon_vld4q_dup_v(&__ret, __p0, 33); \
51432 \
51433  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
51434  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
51435  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
51436  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
51437  __ret; \
51438})
51439#endif
51440
51441#ifdef __LITTLE_ENDIAN__
51442#define vld4_dup_f64(__p0) __extension__ ({ \
51443  float64x1x4_t __ret; \
51444  __builtin_neon_vld4_dup_v(&__ret, __p0, 10); \
51445  __ret; \
51446})
51447#else
51448#define vld4_dup_f64(__p0) __extension__ ({ \
51449  float64x1x4_t __ret; \
51450  __builtin_neon_vld4_dup_v(&__ret, __p0, 10); \
51451  __ret; \
51452})
51453#endif
51454
51455#ifdef __LITTLE_ENDIAN__
51456#define vld4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51457  poly64x1x4_t __s1 = __p1; \
51458  poly64x1x4_t __ret; \
51459  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
51460  __ret; \
51461})
51462#else
51463#define vld4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51464  poly64x1x4_t __s1 = __p1; \
51465  poly64x1x4_t __ret; \
51466  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
51467  __ret; \
51468})
51469#endif
51470
51471#ifdef __LITTLE_ENDIAN__
51472#define vld4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
51473  poly8x16x4_t __s1 = __p1; \
51474  poly8x16x4_t __ret; \
51475  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 36); \
51476  __ret; \
51477})
51478#else
51479#define vld4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
51480  poly8x16x4_t __s1 = __p1; \
51481  poly8x16x4_t __rev1; \
51482  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51483  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51484  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51485  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51486  poly8x16x4_t __ret; \
51487  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 36); \
51488 \
51489  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51490  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51491  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51492  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51493  __ret; \
51494})
51495#endif
51496
51497#ifdef __LITTLE_ENDIAN__
51498#define vld4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51499  poly64x2x4_t __s1 = __p1; \
51500  poly64x2x4_t __ret; \
51501  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 38); \
51502  __ret; \
51503})
51504#else
51505#define vld4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
51506  poly64x2x4_t __s1 = __p1; \
51507  poly64x2x4_t __rev1; \
51508  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51509  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51510  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51511  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51512  poly64x2x4_t __ret; \
51513  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 38); \
51514 \
51515  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51516  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51517  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51518  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51519  __ret; \
51520})
51521#endif
51522
51523#ifdef __LITTLE_ENDIAN__
51524#define vld4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
51525  uint8x16x4_t __s1 = __p1; \
51526  uint8x16x4_t __ret; \
51527  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 48); \
51528  __ret; \
51529})
51530#else
51531#define vld4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
51532  uint8x16x4_t __s1 = __p1; \
51533  uint8x16x4_t __rev1; \
51534  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51535  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51536  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51537  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51538  uint8x16x4_t __ret; \
51539  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 48); \
51540 \
51541  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51542  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51543  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51544  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51545  __ret; \
51546})
51547#endif
51548
51549#ifdef __LITTLE_ENDIAN__
51550#define vld4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51551  uint64x2x4_t __s1 = __p1; \
51552  uint64x2x4_t __ret; \
51553  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 51); \
51554  __ret; \
51555})
51556#else
51557#define vld4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51558  uint64x2x4_t __s1 = __p1; \
51559  uint64x2x4_t __rev1; \
51560  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51561  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51562  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51563  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51564  uint64x2x4_t __ret; \
51565  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 51); \
51566 \
51567  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51568  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51569  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51570  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51571  __ret; \
51572})
51573#endif
51574
51575#ifdef __LITTLE_ENDIAN__
51576#define vld4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
51577  int8x16x4_t __s1 = __p1; \
51578  int8x16x4_t __ret; \
51579  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 32); \
51580  __ret; \
51581})
51582#else
51583#define vld4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
51584  int8x16x4_t __s1 = __p1; \
51585  int8x16x4_t __rev1; \
51586  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51587  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51588  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51589  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51590  int8x16x4_t __ret; \
51591  __builtin_neon_vld4q_lane_v(&__ret, __p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 32); \
51592 \
51593  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51594  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51595  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51596  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
51597  __ret; \
51598})
51599#endif
51600
51601#ifdef __LITTLE_ENDIAN__
51602#define vld4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51603  float64x2x4_t __s1 = __p1; \
51604  float64x2x4_t __ret; \
51605  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 42); \
51606  __ret; \
51607})
51608#else
51609#define vld4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51610  float64x2x4_t __s1 = __p1; \
51611  float64x2x4_t __rev1; \
51612  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51613  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51614  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51615  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51616  float64x2x4_t __ret; \
51617  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 42); \
51618 \
51619  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51620  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51621  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51622  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51623  __ret; \
51624})
51625#endif
51626
51627#ifdef __LITTLE_ENDIAN__
51628#define vld4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51629  int64x2x4_t __s1 = __p1; \
51630  int64x2x4_t __ret; \
51631  __builtin_neon_vld4q_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 35); \
51632  __ret; \
51633})
51634#else
51635#define vld4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51636  int64x2x4_t __s1 = __p1; \
51637  int64x2x4_t __rev1; \
51638  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
51639  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
51640  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
51641  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
51642  int64x2x4_t __ret; \
51643  __builtin_neon_vld4q_lane_v(&__ret, __p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 35); \
51644 \
51645  __ret.val[0] = __builtin_shufflevector(__ret.val[0], __ret.val[0], 1, 0); \
51646  __ret.val[1] = __builtin_shufflevector(__ret.val[1], __ret.val[1], 1, 0); \
51647  __ret.val[2] = __builtin_shufflevector(__ret.val[2], __ret.val[2], 1, 0); \
51648  __ret.val[3] = __builtin_shufflevector(__ret.val[3], __ret.val[3], 1, 0); \
51649  __ret; \
51650})
51651#endif
51652
51653#ifdef __LITTLE_ENDIAN__
51654#define vld4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51655  uint64x1x4_t __s1 = __p1; \
51656  uint64x1x4_t __ret; \
51657  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
51658  __ret; \
51659})
51660#else
51661#define vld4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
51662  uint64x1x4_t __s1 = __p1; \
51663  uint64x1x4_t __ret; \
51664  __builtin_neon_vld4_lane_v(&__ret, __p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
51665  __ret; \
51666})
51667#endif
51668
51669#ifdef __LITTLE_ENDIAN__
51670#define vld4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51671  float64x1x4_t __s1 = __p1; \
51672  float64x1x4_t __ret; \
51673  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
51674  __ret; \
51675})
51676#else
51677#define vld4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
51678  float64x1x4_t __s1 = __p1; \
51679  float64x1x4_t __ret; \
51680  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
51681  __ret; \
51682})
51683#endif
51684
51685#ifdef __LITTLE_ENDIAN__
51686#define vld4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51687  int64x1x4_t __s1 = __p1; \
51688  int64x1x4_t __ret; \
51689  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
51690  __ret; \
51691})
51692#else
51693#define vld4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
51694  int64x1x4_t __s1 = __p1; \
51695  int64x1x4_t __ret; \
51696  __builtin_neon_vld4_lane_v(&__ret, __p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
51697  __ret; \
51698})
51699#endif
51700
51701#ifdef __LITTLE_ENDIAN__
51702#define vldrq_p128(__p0) __extension__ ({ \
51703  poly128_t __ret; \
51704  __ret = (poly128_t) __builtin_neon_vldrq_p128(__p0); \
51705  __ret; \
51706})
51707#else
51708#define vldrq_p128(__p0) __extension__ ({ \
51709  poly128_t __ret; \
51710  __ret = (poly128_t) __builtin_neon_vldrq_p128(__p0); \
51711  __ret; \
51712})
51713#endif
51714
51715#ifdef __LITTLE_ENDIAN__
51716__ai float64x2_t vmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
51717  float64x2_t __ret;
51718  __ret = (float64x2_t) __builtin_neon_vmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
51719  return __ret;
51720}
51721#else
51722__ai float64x2_t vmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
51723  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51724  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
51725  float64x2_t __ret;
51726  __ret = (float64x2_t) __builtin_neon_vmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
51727  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
51728  return __ret;
51729}
51730#endif
51731
51732#ifdef __LITTLE_ENDIAN__
51733__ai float64x1_t vmax_f64(float64x1_t __p0, float64x1_t __p1) {
51734  float64x1_t __ret;
51735  __ret = (float64x1_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
51736  return __ret;
51737}
51738#else
51739__ai float64x1_t vmax_f64(float64x1_t __p0, float64x1_t __p1) {
51740  float64x1_t __ret;
51741  __ret = (float64x1_t) __builtin_neon_vmax_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
51742  return __ret;
51743}
51744#endif
51745
51746#ifdef __LITTLE_ENDIAN__
51747__ai float64_t vmaxnmvq_f64(float64x2_t __p0) {
51748  float64_t __ret;
51749  __ret = (float64_t) __builtin_neon_vmaxnmvq_f64((int8x16_t)__p0);
51750  return __ret;
51751}
51752#else
51753__ai float64_t vmaxnmvq_f64(float64x2_t __p0) {
51754  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51755  float64_t __ret;
51756  __ret = (float64_t) __builtin_neon_vmaxnmvq_f64((int8x16_t)__rev0);
51757  return __ret;
51758}
51759#endif
51760
51761#ifdef __LITTLE_ENDIAN__
51762__ai float32_t vmaxnmvq_f32(float32x4_t __p0) {
51763  float32_t __ret;
51764  __ret = (float32_t) __builtin_neon_vmaxnmvq_f32((int8x16_t)__p0);
51765  return __ret;
51766}
51767#else
51768__ai float32_t vmaxnmvq_f32(float32x4_t __p0) {
51769  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51770  float32_t __ret;
51771  __ret = (float32_t) __builtin_neon_vmaxnmvq_f32((int8x16_t)__rev0);
51772  return __ret;
51773}
51774#endif
51775
51776#ifdef __LITTLE_ENDIAN__
51777__ai float32_t vmaxnmv_f32(float32x2_t __p0) {
51778  float32_t __ret;
51779  __ret = (float32_t) __builtin_neon_vmaxnmv_f32((int8x8_t)__p0);
51780  return __ret;
51781}
51782#else
51783__ai float32_t vmaxnmv_f32(float32x2_t __p0) {
51784  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51785  float32_t __ret;
51786  __ret = (float32_t) __builtin_neon_vmaxnmv_f32((int8x8_t)__rev0);
51787  return __ret;
51788}
51789#endif
51790
51791#ifdef __LITTLE_ENDIAN__
51792__ai uint8_t vmaxvq_u8(uint8x16_t __p0) {
51793  uint8_t __ret;
51794  __ret = (uint8_t) __builtin_neon_vmaxvq_u8((int8x16_t)__p0);
51795  return __ret;
51796}
51797#else
51798__ai uint8_t vmaxvq_u8(uint8x16_t __p0) {
51799  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51800  uint8_t __ret;
51801  __ret = (uint8_t) __builtin_neon_vmaxvq_u8((int8x16_t)__rev0);
51802  return __ret;
51803}
51804#endif
51805
51806#ifdef __LITTLE_ENDIAN__
51807__ai uint32_t vmaxvq_u32(uint32x4_t __p0) {
51808  uint32_t __ret;
51809  __ret = (uint32_t) __builtin_neon_vmaxvq_u32((int8x16_t)__p0);
51810  return __ret;
51811}
51812#else
51813__ai uint32_t vmaxvq_u32(uint32x4_t __p0) {
51814  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51815  uint32_t __ret;
51816  __ret = (uint32_t) __builtin_neon_vmaxvq_u32((int8x16_t)__rev0);
51817  return __ret;
51818}
51819#endif
51820
51821#ifdef __LITTLE_ENDIAN__
51822__ai uint16_t vmaxvq_u16(uint16x8_t __p0) {
51823  uint16_t __ret;
51824  __ret = (uint16_t) __builtin_neon_vmaxvq_u16((int8x16_t)__p0);
51825  return __ret;
51826}
51827#else
51828__ai uint16_t vmaxvq_u16(uint16x8_t __p0) {
51829  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51830  uint16_t __ret;
51831  __ret = (uint16_t) __builtin_neon_vmaxvq_u16((int8x16_t)__rev0);
51832  return __ret;
51833}
51834#endif
51835
51836#ifdef __LITTLE_ENDIAN__
51837__ai int8_t vmaxvq_s8(int8x16_t __p0) {
51838  int8_t __ret;
51839  __ret = (int8_t) __builtin_neon_vmaxvq_s8((int8x16_t)__p0);
51840  return __ret;
51841}
51842#else
51843__ai int8_t vmaxvq_s8(int8x16_t __p0) {
51844  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
51845  int8_t __ret;
51846  __ret = (int8_t) __builtin_neon_vmaxvq_s8((int8x16_t)__rev0);
51847  return __ret;
51848}
51849#endif
51850
51851#ifdef __LITTLE_ENDIAN__
51852__ai float64_t vmaxvq_f64(float64x2_t __p0) {
51853  float64_t __ret;
51854  __ret = (float64_t) __builtin_neon_vmaxvq_f64((int8x16_t)__p0);
51855  return __ret;
51856}
51857#else
51858__ai float64_t vmaxvq_f64(float64x2_t __p0) {
51859  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51860  float64_t __ret;
51861  __ret = (float64_t) __builtin_neon_vmaxvq_f64((int8x16_t)__rev0);
51862  return __ret;
51863}
51864#endif
51865
51866#ifdef __LITTLE_ENDIAN__
51867__ai float32_t vmaxvq_f32(float32x4_t __p0) {
51868  float32_t __ret;
51869  __ret = (float32_t) __builtin_neon_vmaxvq_f32((int8x16_t)__p0);
51870  return __ret;
51871}
51872#else
51873__ai float32_t vmaxvq_f32(float32x4_t __p0) {
51874  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51875  float32_t __ret;
51876  __ret = (float32_t) __builtin_neon_vmaxvq_f32((int8x16_t)__rev0);
51877  return __ret;
51878}
51879#endif
51880
51881#ifdef __LITTLE_ENDIAN__
51882__ai int32_t vmaxvq_s32(int32x4_t __p0) {
51883  int32_t __ret;
51884  __ret = (int32_t) __builtin_neon_vmaxvq_s32((int8x16_t)__p0);
51885  return __ret;
51886}
51887#else
51888__ai int32_t vmaxvq_s32(int32x4_t __p0) {
51889  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51890  int32_t __ret;
51891  __ret = (int32_t) __builtin_neon_vmaxvq_s32((int8x16_t)__rev0);
51892  return __ret;
51893}
51894#endif
51895
51896#ifdef __LITTLE_ENDIAN__
51897__ai int16_t vmaxvq_s16(int16x8_t __p0) {
51898  int16_t __ret;
51899  __ret = (int16_t) __builtin_neon_vmaxvq_s16((int8x16_t)__p0);
51900  return __ret;
51901}
51902#else
51903__ai int16_t vmaxvq_s16(int16x8_t __p0) {
51904  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51905  int16_t __ret;
51906  __ret = (int16_t) __builtin_neon_vmaxvq_s16((int8x16_t)__rev0);
51907  return __ret;
51908}
51909#endif
51910
51911#ifdef __LITTLE_ENDIAN__
51912__ai uint8_t vmaxv_u8(uint8x8_t __p0) {
51913  uint8_t __ret;
51914  __ret = (uint8_t) __builtin_neon_vmaxv_u8((int8x8_t)__p0);
51915  return __ret;
51916}
51917#else
51918__ai uint8_t vmaxv_u8(uint8x8_t __p0) {
51919  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51920  uint8_t __ret;
51921  __ret = (uint8_t) __builtin_neon_vmaxv_u8((int8x8_t)__rev0);
51922  return __ret;
51923}
51924#endif
51925
51926#ifdef __LITTLE_ENDIAN__
51927__ai uint32_t vmaxv_u32(uint32x2_t __p0) {
51928  uint32_t __ret;
51929  __ret = (uint32_t) __builtin_neon_vmaxv_u32((int8x8_t)__p0);
51930  return __ret;
51931}
51932#else
51933__ai uint32_t vmaxv_u32(uint32x2_t __p0) {
51934  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51935  uint32_t __ret;
51936  __ret = (uint32_t) __builtin_neon_vmaxv_u32((int8x8_t)__rev0);
51937  return __ret;
51938}
51939#endif
51940
51941#ifdef __LITTLE_ENDIAN__
51942__ai uint16_t vmaxv_u16(uint16x4_t __p0) {
51943  uint16_t __ret;
51944  __ret = (uint16_t) __builtin_neon_vmaxv_u16((int8x8_t)__p0);
51945  return __ret;
51946}
51947#else
51948__ai uint16_t vmaxv_u16(uint16x4_t __p0) {
51949  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
51950  uint16_t __ret;
51951  __ret = (uint16_t) __builtin_neon_vmaxv_u16((int8x8_t)__rev0);
51952  return __ret;
51953}
51954#endif
51955
51956#ifdef __LITTLE_ENDIAN__
51957__ai int8_t vmaxv_s8(int8x8_t __p0) {
51958  int8_t __ret;
51959  __ret = (int8_t) __builtin_neon_vmaxv_s8((int8x8_t)__p0);
51960  return __ret;
51961}
51962#else
51963__ai int8_t vmaxv_s8(int8x8_t __p0) {
51964  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
51965  int8_t __ret;
51966  __ret = (int8_t) __builtin_neon_vmaxv_s8((int8x8_t)__rev0);
51967  return __ret;
51968}
51969#endif
51970
51971#ifdef __LITTLE_ENDIAN__
51972__ai float32_t vmaxv_f32(float32x2_t __p0) {
51973  float32_t __ret;
51974  __ret = (float32_t) __builtin_neon_vmaxv_f32((int8x8_t)__p0);
51975  return __ret;
51976}
51977#else
51978__ai float32_t vmaxv_f32(float32x2_t __p0) {
51979  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51980  float32_t __ret;
51981  __ret = (float32_t) __builtin_neon_vmaxv_f32((int8x8_t)__rev0);
51982  return __ret;
51983}
51984#endif
51985
51986#ifdef __LITTLE_ENDIAN__
51987__ai int32_t vmaxv_s32(int32x2_t __p0) {
51988  int32_t __ret;
51989  __ret = (int32_t) __builtin_neon_vmaxv_s32((int8x8_t)__p0);
51990  return __ret;
51991}
51992#else
51993__ai int32_t vmaxv_s32(int32x2_t __p0) {
51994  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
51995  int32_t __ret;
51996  __ret = (int32_t) __builtin_neon_vmaxv_s32((int8x8_t)__rev0);
51997  return __ret;
51998}
51999#endif
52000
52001#ifdef __LITTLE_ENDIAN__
52002__ai int16_t vmaxv_s16(int16x4_t __p0) {
52003  int16_t __ret;
52004  __ret = (int16_t) __builtin_neon_vmaxv_s16((int8x8_t)__p0);
52005  return __ret;
52006}
52007#else
52008__ai int16_t vmaxv_s16(int16x4_t __p0) {
52009  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52010  int16_t __ret;
52011  __ret = (int16_t) __builtin_neon_vmaxv_s16((int8x8_t)__rev0);
52012  return __ret;
52013}
52014#endif
52015
52016#ifdef __LITTLE_ENDIAN__
52017__ai float64x2_t vminq_f64(float64x2_t __p0, float64x2_t __p1) {
52018  float64x2_t __ret;
52019  __ret = (float64x2_t) __builtin_neon_vminq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
52020  return __ret;
52021}
52022#else
52023__ai float64x2_t vminq_f64(float64x2_t __p0, float64x2_t __p1) {
52024  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52025  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52026  float64x2_t __ret;
52027  __ret = (float64x2_t) __builtin_neon_vminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
52028  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52029  return __ret;
52030}
52031#endif
52032
52033#ifdef __LITTLE_ENDIAN__
52034__ai float64x1_t vmin_f64(float64x1_t __p0, float64x1_t __p1) {
52035  float64x1_t __ret;
52036  __ret = (float64x1_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
52037  return __ret;
52038}
52039#else
52040__ai float64x1_t vmin_f64(float64x1_t __p0, float64x1_t __p1) {
52041  float64x1_t __ret;
52042  __ret = (float64x1_t) __builtin_neon_vmin_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
52043  return __ret;
52044}
52045#endif
52046
52047#ifdef __LITTLE_ENDIAN__
52048__ai float64_t vminnmvq_f64(float64x2_t __p0) {
52049  float64_t __ret;
52050  __ret = (float64_t) __builtin_neon_vminnmvq_f64((int8x16_t)__p0);
52051  return __ret;
52052}
52053#else
52054__ai float64_t vminnmvq_f64(float64x2_t __p0) {
52055  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52056  float64_t __ret;
52057  __ret = (float64_t) __builtin_neon_vminnmvq_f64((int8x16_t)__rev0);
52058  return __ret;
52059}
52060#endif
52061
52062#ifdef __LITTLE_ENDIAN__
52063__ai float32_t vminnmvq_f32(float32x4_t __p0) {
52064  float32_t __ret;
52065  __ret = (float32_t) __builtin_neon_vminnmvq_f32((int8x16_t)__p0);
52066  return __ret;
52067}
52068#else
52069__ai float32_t vminnmvq_f32(float32x4_t __p0) {
52070  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52071  float32_t __ret;
52072  __ret = (float32_t) __builtin_neon_vminnmvq_f32((int8x16_t)__rev0);
52073  return __ret;
52074}
52075#endif
52076
52077#ifdef __LITTLE_ENDIAN__
52078__ai float32_t vminnmv_f32(float32x2_t __p0) {
52079  float32_t __ret;
52080  __ret = (float32_t) __builtin_neon_vminnmv_f32((int8x8_t)__p0);
52081  return __ret;
52082}
52083#else
52084__ai float32_t vminnmv_f32(float32x2_t __p0) {
52085  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52086  float32_t __ret;
52087  __ret = (float32_t) __builtin_neon_vminnmv_f32((int8x8_t)__rev0);
52088  return __ret;
52089}
52090#endif
52091
52092#ifdef __LITTLE_ENDIAN__
52093__ai uint8_t vminvq_u8(uint8x16_t __p0) {
52094  uint8_t __ret;
52095  __ret = (uint8_t) __builtin_neon_vminvq_u8((int8x16_t)__p0);
52096  return __ret;
52097}
52098#else
52099__ai uint8_t vminvq_u8(uint8x16_t __p0) {
52100  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52101  uint8_t __ret;
52102  __ret = (uint8_t) __builtin_neon_vminvq_u8((int8x16_t)__rev0);
52103  return __ret;
52104}
52105#endif
52106
52107#ifdef __LITTLE_ENDIAN__
52108__ai uint32_t vminvq_u32(uint32x4_t __p0) {
52109  uint32_t __ret;
52110  __ret = (uint32_t) __builtin_neon_vminvq_u32((int8x16_t)__p0);
52111  return __ret;
52112}
52113#else
52114__ai uint32_t vminvq_u32(uint32x4_t __p0) {
52115  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52116  uint32_t __ret;
52117  __ret = (uint32_t) __builtin_neon_vminvq_u32((int8x16_t)__rev0);
52118  return __ret;
52119}
52120#endif
52121
52122#ifdef __LITTLE_ENDIAN__
52123__ai uint16_t vminvq_u16(uint16x8_t __p0) {
52124  uint16_t __ret;
52125  __ret = (uint16_t) __builtin_neon_vminvq_u16((int8x16_t)__p0);
52126  return __ret;
52127}
52128#else
52129__ai uint16_t vminvq_u16(uint16x8_t __p0) {
52130  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52131  uint16_t __ret;
52132  __ret = (uint16_t) __builtin_neon_vminvq_u16((int8x16_t)__rev0);
52133  return __ret;
52134}
52135#endif
52136
52137#ifdef __LITTLE_ENDIAN__
52138__ai int8_t vminvq_s8(int8x16_t __p0) {
52139  int8_t __ret;
52140  __ret = (int8_t) __builtin_neon_vminvq_s8((int8x16_t)__p0);
52141  return __ret;
52142}
52143#else
52144__ai int8_t vminvq_s8(int8x16_t __p0) {
52145  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
52146  int8_t __ret;
52147  __ret = (int8_t) __builtin_neon_vminvq_s8((int8x16_t)__rev0);
52148  return __ret;
52149}
52150#endif
52151
52152#ifdef __LITTLE_ENDIAN__
52153__ai float64_t vminvq_f64(float64x2_t __p0) {
52154  float64_t __ret;
52155  __ret = (float64_t) __builtin_neon_vminvq_f64((int8x16_t)__p0);
52156  return __ret;
52157}
52158#else
52159__ai float64_t vminvq_f64(float64x2_t __p0) {
52160  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52161  float64_t __ret;
52162  __ret = (float64_t) __builtin_neon_vminvq_f64((int8x16_t)__rev0);
52163  return __ret;
52164}
52165#endif
52166
52167#ifdef __LITTLE_ENDIAN__
52168__ai float32_t vminvq_f32(float32x4_t __p0) {
52169  float32_t __ret;
52170  __ret = (float32_t) __builtin_neon_vminvq_f32((int8x16_t)__p0);
52171  return __ret;
52172}
52173#else
52174__ai float32_t vminvq_f32(float32x4_t __p0) {
52175  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52176  float32_t __ret;
52177  __ret = (float32_t) __builtin_neon_vminvq_f32((int8x16_t)__rev0);
52178  return __ret;
52179}
52180#endif
52181
52182#ifdef __LITTLE_ENDIAN__
52183__ai int32_t vminvq_s32(int32x4_t __p0) {
52184  int32_t __ret;
52185  __ret = (int32_t) __builtin_neon_vminvq_s32((int8x16_t)__p0);
52186  return __ret;
52187}
52188#else
52189__ai int32_t vminvq_s32(int32x4_t __p0) {
52190  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52191  int32_t __ret;
52192  __ret = (int32_t) __builtin_neon_vminvq_s32((int8x16_t)__rev0);
52193  return __ret;
52194}
52195#endif
52196
52197#ifdef __LITTLE_ENDIAN__
52198__ai int16_t vminvq_s16(int16x8_t __p0) {
52199  int16_t __ret;
52200  __ret = (int16_t) __builtin_neon_vminvq_s16((int8x16_t)__p0);
52201  return __ret;
52202}
52203#else
52204__ai int16_t vminvq_s16(int16x8_t __p0) {
52205  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52206  int16_t __ret;
52207  __ret = (int16_t) __builtin_neon_vminvq_s16((int8x16_t)__rev0);
52208  return __ret;
52209}
52210#endif
52211
52212#ifdef __LITTLE_ENDIAN__
52213__ai uint8_t vminv_u8(uint8x8_t __p0) {
52214  uint8_t __ret;
52215  __ret = (uint8_t) __builtin_neon_vminv_u8((int8x8_t)__p0);
52216  return __ret;
52217}
52218#else
52219__ai uint8_t vminv_u8(uint8x8_t __p0) {
52220  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52221  uint8_t __ret;
52222  __ret = (uint8_t) __builtin_neon_vminv_u8((int8x8_t)__rev0);
52223  return __ret;
52224}
52225#endif
52226
52227#ifdef __LITTLE_ENDIAN__
52228__ai uint32_t vminv_u32(uint32x2_t __p0) {
52229  uint32_t __ret;
52230  __ret = (uint32_t) __builtin_neon_vminv_u32((int8x8_t)__p0);
52231  return __ret;
52232}
52233#else
52234__ai uint32_t vminv_u32(uint32x2_t __p0) {
52235  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52236  uint32_t __ret;
52237  __ret = (uint32_t) __builtin_neon_vminv_u32((int8x8_t)__rev0);
52238  return __ret;
52239}
52240#endif
52241
52242#ifdef __LITTLE_ENDIAN__
52243__ai uint16_t vminv_u16(uint16x4_t __p0) {
52244  uint16_t __ret;
52245  __ret = (uint16_t) __builtin_neon_vminv_u16((int8x8_t)__p0);
52246  return __ret;
52247}
52248#else
52249__ai uint16_t vminv_u16(uint16x4_t __p0) {
52250  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52251  uint16_t __ret;
52252  __ret = (uint16_t) __builtin_neon_vminv_u16((int8x8_t)__rev0);
52253  return __ret;
52254}
52255#endif
52256
52257#ifdef __LITTLE_ENDIAN__
52258__ai int8_t vminv_s8(int8x8_t __p0) {
52259  int8_t __ret;
52260  __ret = (int8_t) __builtin_neon_vminv_s8((int8x8_t)__p0);
52261  return __ret;
52262}
52263#else
52264__ai int8_t vminv_s8(int8x8_t __p0) {
52265  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
52266  int8_t __ret;
52267  __ret = (int8_t) __builtin_neon_vminv_s8((int8x8_t)__rev0);
52268  return __ret;
52269}
52270#endif
52271
52272#ifdef __LITTLE_ENDIAN__
52273__ai float32_t vminv_f32(float32x2_t __p0) {
52274  float32_t __ret;
52275  __ret = (float32_t) __builtin_neon_vminv_f32((int8x8_t)__p0);
52276  return __ret;
52277}
52278#else
52279__ai float32_t vminv_f32(float32x2_t __p0) {
52280  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52281  float32_t __ret;
52282  __ret = (float32_t) __builtin_neon_vminv_f32((int8x8_t)__rev0);
52283  return __ret;
52284}
52285#endif
52286
52287#ifdef __LITTLE_ENDIAN__
52288__ai int32_t vminv_s32(int32x2_t __p0) {
52289  int32_t __ret;
52290  __ret = (int32_t) __builtin_neon_vminv_s32((int8x8_t)__p0);
52291  return __ret;
52292}
52293#else
52294__ai int32_t vminv_s32(int32x2_t __p0) {
52295  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52296  int32_t __ret;
52297  __ret = (int32_t) __builtin_neon_vminv_s32((int8x8_t)__rev0);
52298  return __ret;
52299}
52300#endif
52301
52302#ifdef __LITTLE_ENDIAN__
52303__ai int16_t vminv_s16(int16x4_t __p0) {
52304  int16_t __ret;
52305  __ret = (int16_t) __builtin_neon_vminv_s16((int8x8_t)__p0);
52306  return __ret;
52307}
52308#else
52309__ai int16_t vminv_s16(int16x4_t __p0) {
52310  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
52311  int16_t __ret;
52312  __ret = (int16_t) __builtin_neon_vminv_s16((int8x8_t)__rev0);
52313  return __ret;
52314}
52315#endif
52316
52317#ifdef __LITTLE_ENDIAN__
52318__ai float64x2_t vmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
52319  float64x2_t __ret;
52320  __ret = __p0 + __p1 * __p2;
52321  return __ret;
52322}
52323#else
52324__ai float64x2_t vmlaq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
52325  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52326  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52327  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
52328  float64x2_t __ret;
52329  __ret = __rev0 + __rev1 * __rev2;
52330  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52331  return __ret;
52332}
52333#endif
52334
52335#ifdef __LITTLE_ENDIAN__
52336__ai float64x1_t vmla_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
52337  float64x1_t __ret;
52338  __ret = __p0 + __p1 * __p2;
52339  return __ret;
52340}
52341#else
52342__ai float64x1_t vmla_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
52343  float64x1_t __ret;
52344  __ret = __p0 + __p1 * __p2;
52345  return __ret;
52346}
52347#endif
52348
52349#ifdef __LITTLE_ENDIAN__
52350#define vmlaq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52351  uint32x4_t __s0 = __p0; \
52352  uint32x4_t __s1 = __p1; \
52353  uint32x4_t __s2 = __p2; \
52354  uint32x4_t __ret; \
52355  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52356  __ret; \
52357})
52358#else
52359#define vmlaq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52360  uint32x4_t __s0 = __p0; \
52361  uint32x4_t __s1 = __p1; \
52362  uint32x4_t __s2 = __p2; \
52363  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52364  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52365  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52366  uint32x4_t __ret; \
52367  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52368  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52369  __ret; \
52370})
52371#endif
52372
52373#ifdef __LITTLE_ENDIAN__
52374#define vmlaq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52375  uint16x8_t __s0 = __p0; \
52376  uint16x8_t __s1 = __p1; \
52377  uint16x8_t __s2 = __p2; \
52378  uint16x8_t __ret; \
52379  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52380  __ret; \
52381})
52382#else
52383#define vmlaq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52384  uint16x8_t __s0 = __p0; \
52385  uint16x8_t __s1 = __p1; \
52386  uint16x8_t __s2 = __p2; \
52387  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
52388  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52389  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52390  uint16x8_t __ret; \
52391  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52392  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
52393  __ret; \
52394})
52395#endif
52396
52397#ifdef __LITTLE_ENDIAN__
52398#define vmlaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52399  float32x4_t __s0 = __p0; \
52400  float32x4_t __s1 = __p1; \
52401  float32x4_t __s2 = __p2; \
52402  float32x4_t __ret; \
52403  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52404  __ret; \
52405})
52406#else
52407#define vmlaq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52408  float32x4_t __s0 = __p0; \
52409  float32x4_t __s1 = __p1; \
52410  float32x4_t __s2 = __p2; \
52411  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52412  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52413  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52414  float32x4_t __ret; \
52415  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52416  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52417  __ret; \
52418})
52419#endif
52420
52421#ifdef __LITTLE_ENDIAN__
52422#define vmlaq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52423  int32x4_t __s0 = __p0; \
52424  int32x4_t __s1 = __p1; \
52425  int32x4_t __s2 = __p2; \
52426  int32x4_t __ret; \
52427  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52428  __ret; \
52429})
52430#else
52431#define vmlaq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52432  int32x4_t __s0 = __p0; \
52433  int32x4_t __s1 = __p1; \
52434  int32x4_t __s2 = __p2; \
52435  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52436  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52437  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52438  int32x4_t __ret; \
52439  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52440  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52441  __ret; \
52442})
52443#endif
52444
52445#ifdef __LITTLE_ENDIAN__
52446#define vmlaq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52447  int16x8_t __s0 = __p0; \
52448  int16x8_t __s1 = __p1; \
52449  int16x8_t __s2 = __p2; \
52450  int16x8_t __ret; \
52451  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52452  __ret; \
52453})
52454#else
52455#define vmlaq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52456  int16x8_t __s0 = __p0; \
52457  int16x8_t __s1 = __p1; \
52458  int16x8_t __s2 = __p2; \
52459  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
52460  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52461  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52462  int16x8_t __ret; \
52463  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52464  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
52465  __ret; \
52466})
52467#endif
52468
52469#ifdef __LITTLE_ENDIAN__
52470#define vmla_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52471  uint32x2_t __s0 = __p0; \
52472  uint32x2_t __s1 = __p1; \
52473  uint32x4_t __s2 = __p2; \
52474  uint32x2_t __ret; \
52475  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
52476  __ret; \
52477})
52478#else
52479#define vmla_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52480  uint32x2_t __s0 = __p0; \
52481  uint32x2_t __s1 = __p1; \
52482  uint32x4_t __s2 = __p2; \
52483  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52484  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52485  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52486  uint32x2_t __ret; \
52487  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
52488  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52489  __ret; \
52490})
52491#endif
52492
52493#ifdef __LITTLE_ENDIAN__
52494#define vmla_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52495  uint16x4_t __s0 = __p0; \
52496  uint16x4_t __s1 = __p1; \
52497  uint16x8_t __s2 = __p2; \
52498  uint16x4_t __ret; \
52499  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52500  __ret; \
52501})
52502#else
52503#define vmla_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52504  uint16x4_t __s0 = __p0; \
52505  uint16x4_t __s1 = __p1; \
52506  uint16x8_t __s2 = __p2; \
52507  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52508  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52509  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52510  uint16x4_t __ret; \
52511  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52512  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52513  __ret; \
52514})
52515#endif
52516
52517#ifdef __LITTLE_ENDIAN__
52518#define vmla_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52519  float32x2_t __s0 = __p0; \
52520  float32x2_t __s1 = __p1; \
52521  float32x4_t __s2 = __p2; \
52522  float32x2_t __ret; \
52523  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
52524  __ret; \
52525})
52526#else
52527#define vmla_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52528  float32x2_t __s0 = __p0; \
52529  float32x2_t __s1 = __p1; \
52530  float32x4_t __s2 = __p2; \
52531  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52532  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52533  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52534  float32x2_t __ret; \
52535  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
52536  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52537  __ret; \
52538})
52539#endif
52540
52541#ifdef __LITTLE_ENDIAN__
52542#define vmla_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52543  int32x2_t __s0 = __p0; \
52544  int32x2_t __s1 = __p1; \
52545  int32x4_t __s2 = __p2; \
52546  int32x2_t __ret; \
52547  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
52548  __ret; \
52549})
52550#else
52551#define vmla_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52552  int32x2_t __s0 = __p0; \
52553  int32x2_t __s1 = __p1; \
52554  int32x4_t __s2 = __p2; \
52555  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52556  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52557  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52558  int32x2_t __ret; \
52559  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
52560  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52561  __ret; \
52562})
52563#endif
52564
52565#ifdef __LITTLE_ENDIAN__
52566#define vmla_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52567  int16x4_t __s0 = __p0; \
52568  int16x4_t __s1 = __p1; \
52569  int16x8_t __s2 = __p2; \
52570  int16x4_t __ret; \
52571  __ret = __s0 + __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52572  __ret; \
52573})
52574#else
52575#define vmla_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52576  int16x4_t __s0 = __p0; \
52577  int16x4_t __s1 = __p1; \
52578  int16x8_t __s2 = __p2; \
52579  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52580  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52581  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52582  int16x4_t __ret; \
52583  __ret = __rev0 + __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52584  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52585  __ret; \
52586})
52587#endif
52588
52589#ifdef __LITTLE_ENDIAN__
52590__ai float64x2_t vmlaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
52591  float64x2_t __ret;
52592  __ret = __p0 + __p1 * (float64x2_t) {__p2, __p2};
52593  return __ret;
52594}
52595#else
52596__ai float64x2_t vmlaq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
52597  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52598  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52599  float64x2_t __ret;
52600  __ret = __rev0 + __rev1 * (float64x2_t) {__p2, __p2};
52601  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52602  return __ret;
52603}
52604#endif
52605
52606#ifdef __LITTLE_ENDIAN__
52607#define vmlal_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52608  uint64x2_t __s0 = __p0; \
52609  uint32x4_t __s1 = __p1; \
52610  uint32x2_t __s2 = __p2; \
52611  uint64x2_t __ret; \
52612  __ret = __s0 + vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52613  __ret; \
52614})
52615#else
52616#define vmlal_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52617  uint64x2_t __s0 = __p0; \
52618  uint32x4_t __s1 = __p1; \
52619  uint32x2_t __s2 = __p2; \
52620  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52621  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52622  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
52623  uint64x2_t __ret; \
52624  __ret = __rev0 + __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52625  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52626  __ret; \
52627})
52628#endif
52629
52630#ifdef __LITTLE_ENDIAN__
52631#define vmlal_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52632  uint32x4_t __s0 = __p0; \
52633  uint16x8_t __s1 = __p1; \
52634  uint16x4_t __s2 = __p2; \
52635  uint32x4_t __ret; \
52636  __ret = __s0 + vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52637  __ret; \
52638})
52639#else
52640#define vmlal_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52641  uint32x4_t __s0 = __p0; \
52642  uint16x8_t __s1 = __p1; \
52643  uint16x4_t __s2 = __p2; \
52644  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52645  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52646  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52647  uint32x4_t __ret; \
52648  __ret = __rev0 + __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52649  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52650  __ret; \
52651})
52652#endif
52653
52654#ifdef __LITTLE_ENDIAN__
52655#define vmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52656  int64x2_t __s0 = __p0; \
52657  int32x4_t __s1 = __p1; \
52658  int32x2_t __s2 = __p2; \
52659  int64x2_t __ret; \
52660  __ret = __s0 + vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52661  __ret; \
52662})
52663#else
52664#define vmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52665  int64x2_t __s0 = __p0; \
52666  int32x4_t __s1 = __p1; \
52667  int32x2_t __s2 = __p2; \
52668  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52669  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52670  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
52671  int64x2_t __ret; \
52672  __ret = __rev0 + __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52673  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52674  __ret; \
52675})
52676#endif
52677
52678#ifdef __LITTLE_ENDIAN__
52679#define vmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52680  int32x4_t __s0 = __p0; \
52681  int16x8_t __s1 = __p1; \
52682  int16x4_t __s2 = __p2; \
52683  int32x4_t __ret; \
52684  __ret = __s0 + vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52685  __ret; \
52686})
52687#else
52688#define vmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52689  int32x4_t __s0 = __p0; \
52690  int16x8_t __s1 = __p1; \
52691  int16x4_t __s2 = __p2; \
52692  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52693  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52694  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52695  int32x4_t __ret; \
52696  __ret = __rev0 + __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52697  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52698  __ret; \
52699})
52700#endif
52701
52702#ifdef __LITTLE_ENDIAN__
52703#define vmlal_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52704  uint64x2_t __s0 = __p0; \
52705  uint32x4_t __s1 = __p1; \
52706  uint32x4_t __s2 = __p2; \
52707  uint64x2_t __ret; \
52708  __ret = __s0 + vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52709  __ret; \
52710})
52711#else
52712#define vmlal_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52713  uint64x2_t __s0 = __p0; \
52714  uint32x4_t __s1 = __p1; \
52715  uint32x4_t __s2 = __p2; \
52716  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52717  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52718  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52719  uint64x2_t __ret; \
52720  __ret = __rev0 + __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52721  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52722  __ret; \
52723})
52724#endif
52725
52726#ifdef __LITTLE_ENDIAN__
52727#define vmlal_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52728  uint32x4_t __s0 = __p0; \
52729  uint16x8_t __s1 = __p1; \
52730  uint16x8_t __s2 = __p2; \
52731  uint32x4_t __ret; \
52732  __ret = __s0 + vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52733  __ret; \
52734})
52735#else
52736#define vmlal_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52737  uint32x4_t __s0 = __p0; \
52738  uint16x8_t __s1 = __p1; \
52739  uint16x8_t __s2 = __p2; \
52740  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52741  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52742  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52743  uint32x4_t __ret; \
52744  __ret = __rev0 + __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52745  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52746  __ret; \
52747})
52748#endif
52749
52750#ifdef __LITTLE_ENDIAN__
52751#define vmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52752  int64x2_t __s0 = __p0; \
52753  int32x4_t __s1 = __p1; \
52754  int32x4_t __s2 = __p2; \
52755  int64x2_t __ret; \
52756  __ret = __s0 + vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52757  __ret; \
52758})
52759#else
52760#define vmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52761  int64x2_t __s0 = __p0; \
52762  int32x4_t __s1 = __p1; \
52763  int32x4_t __s2 = __p2; \
52764  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52765  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52766  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52767  int64x2_t __ret; \
52768  __ret = __rev0 + __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52769  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52770  __ret; \
52771})
52772#endif
52773
52774#ifdef __LITTLE_ENDIAN__
52775#define vmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52776  int32x4_t __s0 = __p0; \
52777  int16x8_t __s1 = __p1; \
52778  int16x8_t __s2 = __p2; \
52779  int32x4_t __ret; \
52780  __ret = __s0 + vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52781  __ret; \
52782})
52783#else
52784#define vmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52785  int32x4_t __s0 = __p0; \
52786  int16x8_t __s1 = __p1; \
52787  int16x8_t __s2 = __p2; \
52788  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52789  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52790  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52791  int32x4_t __ret; \
52792  __ret = __rev0 + __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52793  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52794  __ret; \
52795})
52796#endif
52797
52798#ifdef __LITTLE_ENDIAN__
52799#define vmlal_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52800  uint64x2_t __s0 = __p0; \
52801  uint32x2_t __s1 = __p1; \
52802  uint32x4_t __s2 = __p2; \
52803  uint64x2_t __ret; \
52804  __ret = __s0 + vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52805  __ret; \
52806})
52807#else
52808#define vmlal_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52809  uint64x2_t __s0 = __p0; \
52810  uint32x2_t __s1 = __p1; \
52811  uint32x4_t __s2 = __p2; \
52812  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52813  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52814  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52815  uint64x2_t __ret; \
52816  __ret = __rev0 + __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52818  __ret; \
52819})
52820#endif
52821
52822#ifdef __LITTLE_ENDIAN__
52823#define vmlal_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52824  uint32x4_t __s0 = __p0; \
52825  uint16x4_t __s1 = __p1; \
52826  uint16x8_t __s2 = __p2; \
52827  uint32x4_t __ret; \
52828  __ret = __s0 + vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52829  __ret; \
52830})
52831#else
52832#define vmlal_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52833  uint32x4_t __s0 = __p0; \
52834  uint16x4_t __s1 = __p1; \
52835  uint16x8_t __s2 = __p2; \
52836  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52837  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52838  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52839  uint32x4_t __ret; \
52840  __ret = __rev0 + __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52841  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52842  __ret; \
52843})
52844#endif
52845
52846#ifdef __LITTLE_ENDIAN__
52847#define vmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52848  int64x2_t __s0 = __p0; \
52849  int32x2_t __s1 = __p1; \
52850  int32x4_t __s2 = __p2; \
52851  int64x2_t __ret; \
52852  __ret = __s0 + vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
52853  __ret; \
52854})
52855#else
52856#define vmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
52857  int64x2_t __s0 = __p0; \
52858  int32x2_t __s1 = __p1; \
52859  int32x4_t __s2 = __p2; \
52860  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
52861  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
52862  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52863  int64x2_t __ret; \
52864  __ret = __rev0 + __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
52865  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
52866  __ret; \
52867})
52868#endif
52869
52870#ifdef __LITTLE_ENDIAN__
52871#define vmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52872  int32x4_t __s0 = __p0; \
52873  int16x4_t __s1 = __p1; \
52874  int16x8_t __s2 = __p2; \
52875  int32x4_t __ret; \
52876  __ret = __s0 + vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
52877  __ret; \
52878})
52879#else
52880#define vmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
52881  int32x4_t __s0 = __p0; \
52882  int16x4_t __s1 = __p1; \
52883  int16x8_t __s2 = __p2; \
52884  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52885  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52886  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52887  int32x4_t __ret; \
52888  __ret = __rev0 + __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
52889  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52890  __ret; \
52891})
52892#endif
52893
52894#ifdef __LITTLE_ENDIAN__
52895__ai float64x2_t vmlsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
52896  float64x2_t __ret;
52897  __ret = __p0 - __p1 * __p2;
52898  return __ret;
52899}
52900#else
52901__ai float64x2_t vmlsq_f64(float64x2_t __p0, float64x2_t __p1, float64x2_t __p2) {
52902  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
52903  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
52904  float64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
52905  float64x2_t __ret;
52906  __ret = __rev0 - __rev1 * __rev2;
52907  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
52908  return __ret;
52909}
52910#endif
52911
52912#ifdef __LITTLE_ENDIAN__
52913__ai float64x1_t vmls_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
52914  float64x1_t __ret;
52915  __ret = __p0 - __p1 * __p2;
52916  return __ret;
52917}
52918#else
52919__ai float64x1_t vmls_f64(float64x1_t __p0, float64x1_t __p1, float64x1_t __p2) {
52920  float64x1_t __ret;
52921  __ret = __p0 - __p1 * __p2;
52922  return __ret;
52923}
52924#endif
52925
52926#ifdef __LITTLE_ENDIAN__
52927#define vmlsq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52928  uint32x4_t __s0 = __p0; \
52929  uint32x4_t __s1 = __p1; \
52930  uint32x4_t __s2 = __p2; \
52931  uint32x4_t __ret; \
52932  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52933  __ret; \
52934})
52935#else
52936#define vmlsq_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
52937  uint32x4_t __s0 = __p0; \
52938  uint32x4_t __s1 = __p1; \
52939  uint32x4_t __s2 = __p2; \
52940  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52941  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52942  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52943  uint32x4_t __ret; \
52944  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52945  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52946  __ret; \
52947})
52948#endif
52949
52950#ifdef __LITTLE_ENDIAN__
52951#define vmlsq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52952  uint16x8_t __s0 = __p0; \
52953  uint16x8_t __s1 = __p1; \
52954  uint16x8_t __s2 = __p2; \
52955  uint16x8_t __ret; \
52956  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52957  __ret; \
52958})
52959#else
52960#define vmlsq_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
52961  uint16x8_t __s0 = __p0; \
52962  uint16x8_t __s1 = __p1; \
52963  uint16x8_t __s2 = __p2; \
52964  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
52965  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
52966  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
52967  uint16x8_t __ret; \
52968  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
52969  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
52970  __ret; \
52971})
52972#endif
52973
52974#ifdef __LITTLE_ENDIAN__
52975#define vmlsq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52976  float32x4_t __s0 = __p0; \
52977  float32x4_t __s1 = __p1; \
52978  float32x4_t __s2 = __p2; \
52979  float32x4_t __ret; \
52980  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
52981  __ret; \
52982})
52983#else
52984#define vmlsq_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
52985  float32x4_t __s0 = __p0; \
52986  float32x4_t __s1 = __p1; \
52987  float32x4_t __s2 = __p2; \
52988  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
52989  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
52990  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
52991  float32x4_t __ret; \
52992  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
52993  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
52994  __ret; \
52995})
52996#endif
52997
52998#ifdef __LITTLE_ENDIAN__
52999#define vmlsq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53000  int32x4_t __s0 = __p0; \
53001  int32x4_t __s1 = __p1; \
53002  int32x4_t __s2 = __p2; \
53003  int32x4_t __ret; \
53004  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
53005  __ret; \
53006})
53007#else
53008#define vmlsq_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53009  int32x4_t __s0 = __p0; \
53010  int32x4_t __s1 = __p1; \
53011  int32x4_t __s2 = __p2; \
53012  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53013  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53014  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53015  int32x4_t __ret; \
53016  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
53017  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53018  __ret; \
53019})
53020#endif
53021
53022#ifdef __LITTLE_ENDIAN__
53023#define vmlsq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53024  int16x8_t __s0 = __p0; \
53025  int16x8_t __s1 = __p1; \
53026  int16x8_t __s2 = __p2; \
53027  int16x8_t __ret; \
53028  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
53029  __ret; \
53030})
53031#else
53032#define vmlsq_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53033  int16x8_t __s0 = __p0; \
53034  int16x8_t __s1 = __p1; \
53035  int16x8_t __s2 = __p2; \
53036  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
53037  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53038  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53039  int16x8_t __ret; \
53040  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3, __p3, __p3, __p3, __p3); \
53041  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
53042  __ret; \
53043})
53044#endif
53045
53046#ifdef __LITTLE_ENDIAN__
53047#define vmls_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53048  uint32x2_t __s0 = __p0; \
53049  uint32x2_t __s1 = __p1; \
53050  uint32x4_t __s2 = __p2; \
53051  uint32x2_t __ret; \
53052  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
53053  __ret; \
53054})
53055#else
53056#define vmls_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53057  uint32x2_t __s0 = __p0; \
53058  uint32x2_t __s1 = __p1; \
53059  uint32x4_t __s2 = __p2; \
53060  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53061  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53062  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53063  uint32x2_t __ret; \
53064  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
53065  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53066  __ret; \
53067})
53068#endif
53069
53070#ifdef __LITTLE_ENDIAN__
53071#define vmls_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53072  uint16x4_t __s0 = __p0; \
53073  uint16x4_t __s1 = __p1; \
53074  uint16x8_t __s2 = __p2; \
53075  uint16x4_t __ret; \
53076  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
53077  __ret; \
53078})
53079#else
53080#define vmls_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53081  uint16x4_t __s0 = __p0; \
53082  uint16x4_t __s1 = __p1; \
53083  uint16x8_t __s2 = __p2; \
53084  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53085  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53086  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53087  uint16x4_t __ret; \
53088  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
53089  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53090  __ret; \
53091})
53092#endif
53093
53094#ifdef __LITTLE_ENDIAN__
53095#define vmls_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
53096  float32x2_t __s0 = __p0; \
53097  float32x2_t __s1 = __p1; \
53098  float32x4_t __s2 = __p2; \
53099  float32x2_t __ret; \
53100  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
53101  __ret; \
53102})
53103#else
53104#define vmls_laneq_f32(__p0, __p1, __p2, __p3) __extension__ ({ \
53105  float32x2_t __s0 = __p0; \
53106  float32x2_t __s1 = __p1; \
53107  float32x4_t __s2 = __p2; \
53108  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53109  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53110  float32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53111  float32x2_t __ret; \
53112  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
53113  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53114  __ret; \
53115})
53116#endif
53117
53118#ifdef __LITTLE_ENDIAN__
53119#define vmls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53120  int32x2_t __s0 = __p0; \
53121  int32x2_t __s1 = __p1; \
53122  int32x4_t __s2 = __p2; \
53123  int32x2_t __ret; \
53124  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3); \
53125  __ret; \
53126})
53127#else
53128#define vmls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53129  int32x2_t __s0 = __p0; \
53130  int32x2_t __s1 = __p1; \
53131  int32x4_t __s2 = __p2; \
53132  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53133  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53134  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53135  int32x2_t __ret; \
53136  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3); \
53137  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53138  __ret; \
53139})
53140#endif
53141
53142#ifdef __LITTLE_ENDIAN__
53143#define vmls_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53144  int16x4_t __s0 = __p0; \
53145  int16x4_t __s1 = __p1; \
53146  int16x8_t __s2 = __p2; \
53147  int16x4_t __ret; \
53148  __ret = __s0 - __s1 * __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3); \
53149  __ret; \
53150})
53151#else
53152#define vmls_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53153  int16x4_t __s0 = __p0; \
53154  int16x4_t __s1 = __p1; \
53155  int16x8_t __s2 = __p2; \
53156  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53157  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53158  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53159  int16x4_t __ret; \
53160  __ret = __rev0 - __rev1 * __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3); \
53161  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53162  __ret; \
53163})
53164#endif
53165
53166#ifdef __LITTLE_ENDIAN__
53167__ai float64x2_t vmlsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
53168  float64x2_t __ret;
53169  __ret = __p0 - __p1 * (float64x2_t) {__p2, __p2};
53170  return __ret;
53171}
53172#else
53173__ai float64x2_t vmlsq_n_f64(float64x2_t __p0, float64x2_t __p1, float64_t __p2) {
53174  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53175  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53176  float64x2_t __ret;
53177  __ret = __rev0 - __rev1 * (float64x2_t) {__p2, __p2};
53178  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53179  return __ret;
53180}
53181#endif
53182
53183#ifdef __LITTLE_ENDIAN__
53184#define vmlsl_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53185  uint64x2_t __s0 = __p0; \
53186  uint32x4_t __s1 = __p1; \
53187  uint32x2_t __s2 = __p2; \
53188  uint64x2_t __ret; \
53189  __ret = __s0 - vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53190  __ret; \
53191})
53192#else
53193#define vmlsl_high_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53194  uint64x2_t __s0 = __p0; \
53195  uint32x4_t __s1 = __p1; \
53196  uint32x2_t __s2 = __p2; \
53197  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53198  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53199  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
53200  uint64x2_t __ret; \
53201  __ret = __rev0 - __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53202  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53203  __ret; \
53204})
53205#endif
53206
53207#ifdef __LITTLE_ENDIAN__
53208#define vmlsl_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53209  uint32x4_t __s0 = __p0; \
53210  uint16x8_t __s1 = __p1; \
53211  uint16x4_t __s2 = __p2; \
53212  uint32x4_t __ret; \
53213  __ret = __s0 - vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53214  __ret; \
53215})
53216#else
53217#define vmlsl_high_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53218  uint32x4_t __s0 = __p0; \
53219  uint16x8_t __s1 = __p1; \
53220  uint16x4_t __s2 = __p2; \
53221  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53222  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53223  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53224  uint32x4_t __ret; \
53225  __ret = __rev0 - __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53226  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53227  __ret; \
53228})
53229#endif
53230
53231#ifdef __LITTLE_ENDIAN__
53232#define vmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53233  int64x2_t __s0 = __p0; \
53234  int32x4_t __s1 = __p1; \
53235  int32x2_t __s2 = __p2; \
53236  int64x2_t __ret; \
53237  __ret = __s0 - vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53238  __ret; \
53239})
53240#else
53241#define vmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53242  int64x2_t __s0 = __p0; \
53243  int32x4_t __s1 = __p1; \
53244  int32x2_t __s2 = __p2; \
53245  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53246  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53247  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
53248  int64x2_t __ret; \
53249  __ret = __rev0 - __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53250  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53251  __ret; \
53252})
53253#endif
53254
53255#ifdef __LITTLE_ENDIAN__
53256#define vmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53257  int32x4_t __s0 = __p0; \
53258  int16x8_t __s1 = __p1; \
53259  int16x4_t __s2 = __p2; \
53260  int32x4_t __ret; \
53261  __ret = __s0 - vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53262  __ret; \
53263})
53264#else
53265#define vmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53266  int32x4_t __s0 = __p0; \
53267  int16x8_t __s1 = __p1; \
53268  int16x4_t __s2 = __p2; \
53269  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53270  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53271  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53272  int32x4_t __ret; \
53273  __ret = __rev0 - __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53274  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53275  __ret; \
53276})
53277#endif
53278
53279#ifdef __LITTLE_ENDIAN__
53280#define vmlsl_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53281  uint64x2_t __s0 = __p0; \
53282  uint32x4_t __s1 = __p1; \
53283  uint32x4_t __s2 = __p2; \
53284  uint64x2_t __ret; \
53285  __ret = __s0 - vmull_u32(vget_high_u32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53286  __ret; \
53287})
53288#else
53289#define vmlsl_high_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53290  uint64x2_t __s0 = __p0; \
53291  uint32x4_t __s1 = __p1; \
53292  uint32x4_t __s2 = __p2; \
53293  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53294  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53295  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53296  uint64x2_t __ret; \
53297  __ret = __rev0 - __noswap_vmull_u32(__noswap_vget_high_u32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53298  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53299  __ret; \
53300})
53301#endif
53302
53303#ifdef __LITTLE_ENDIAN__
53304#define vmlsl_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53305  uint32x4_t __s0 = __p0; \
53306  uint16x8_t __s1 = __p1; \
53307  uint16x8_t __s2 = __p2; \
53308  uint32x4_t __ret; \
53309  __ret = __s0 - vmull_u16(vget_high_u16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53310  __ret; \
53311})
53312#else
53313#define vmlsl_high_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53314  uint32x4_t __s0 = __p0; \
53315  uint16x8_t __s1 = __p1; \
53316  uint16x8_t __s2 = __p2; \
53317  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53318  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53319  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53320  uint32x4_t __ret; \
53321  __ret = __rev0 - __noswap_vmull_u16(__noswap_vget_high_u16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53322  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53323  __ret; \
53324})
53325#endif
53326
53327#ifdef __LITTLE_ENDIAN__
53328#define vmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53329  int64x2_t __s0 = __p0; \
53330  int32x4_t __s1 = __p1; \
53331  int32x4_t __s2 = __p2; \
53332  int64x2_t __ret; \
53333  __ret = __s0 - vmull_s32(vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53334  __ret; \
53335})
53336#else
53337#define vmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53338  int64x2_t __s0 = __p0; \
53339  int32x4_t __s1 = __p1; \
53340  int32x4_t __s2 = __p2; \
53341  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53342  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53343  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53344  int64x2_t __ret; \
53345  __ret = __rev0 - __noswap_vmull_s32(__noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53346  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53347  __ret; \
53348})
53349#endif
53350
53351#ifdef __LITTLE_ENDIAN__
53352#define vmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53353  int32x4_t __s0 = __p0; \
53354  int16x8_t __s1 = __p1; \
53355  int16x8_t __s2 = __p2; \
53356  int32x4_t __ret; \
53357  __ret = __s0 - vmull_s16(vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53358  __ret; \
53359})
53360#else
53361#define vmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53362  int32x4_t __s0 = __p0; \
53363  int16x8_t __s1 = __p1; \
53364  int16x8_t __s2 = __p2; \
53365  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53366  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53367  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53368  int32x4_t __ret; \
53369  __ret = __rev0 - __noswap_vmull_s16(__noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53370  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53371  __ret; \
53372})
53373#endif
53374
53375#ifdef __LITTLE_ENDIAN__
53376#define vmlsl_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53377  uint64x2_t __s0 = __p0; \
53378  uint32x2_t __s1 = __p1; \
53379  uint32x4_t __s2 = __p2; \
53380  uint64x2_t __ret; \
53381  __ret = __s0 - vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53382  __ret; \
53383})
53384#else
53385#define vmlsl_laneq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
53386  uint64x2_t __s0 = __p0; \
53387  uint32x2_t __s1 = __p1; \
53388  uint32x4_t __s2 = __p2; \
53389  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53390  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53391  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53392  uint64x2_t __ret; \
53393  __ret = __rev0 - __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53394  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53395  __ret; \
53396})
53397#endif
53398
53399#ifdef __LITTLE_ENDIAN__
53400#define vmlsl_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53401  uint32x4_t __s0 = __p0; \
53402  uint16x4_t __s1 = __p1; \
53403  uint16x8_t __s2 = __p2; \
53404  uint32x4_t __ret; \
53405  __ret = __s0 - vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53406  __ret; \
53407})
53408#else
53409#define vmlsl_laneq_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
53410  uint32x4_t __s0 = __p0; \
53411  uint16x4_t __s1 = __p1; \
53412  uint16x8_t __s2 = __p2; \
53413  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53414  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53415  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53416  uint32x4_t __ret; \
53417  __ret = __rev0 - __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53418  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53419  __ret; \
53420})
53421#endif
53422
53423#ifdef __LITTLE_ENDIAN__
53424#define vmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53425  int64x2_t __s0 = __p0; \
53426  int32x2_t __s1 = __p1; \
53427  int32x4_t __s2 = __p2; \
53428  int64x2_t __ret; \
53429  __ret = __s0 - vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
53430  __ret; \
53431})
53432#else
53433#define vmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
53434  int64x2_t __s0 = __p0; \
53435  int32x2_t __s1 = __p1; \
53436  int32x4_t __s2 = __p2; \
53437  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53438  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53439  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
53440  int64x2_t __ret; \
53441  __ret = __rev0 - __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
53442  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53443  __ret; \
53444})
53445#endif
53446
53447#ifdef __LITTLE_ENDIAN__
53448#define vmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53449  int32x4_t __s0 = __p0; \
53450  int16x4_t __s1 = __p1; \
53451  int16x8_t __s2 = __p2; \
53452  int32x4_t __ret; \
53453  __ret = __s0 - vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
53454  __ret; \
53455})
53456#else
53457#define vmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
53458  int32x4_t __s0 = __p0; \
53459  int16x4_t __s1 = __p1; \
53460  int16x8_t __s2 = __p2; \
53461  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53462  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53463  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
53464  int32x4_t __ret; \
53465  __ret = __rev0 - __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
53466  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53467  __ret; \
53468})
53469#endif
53470
53471#ifdef __LITTLE_ENDIAN__
53472__ai poly64x1_t vmov_n_p64(poly64_t __p0) {
53473  poly64x1_t __ret;
53474  __ret = (poly64x1_t) {__p0};
53475  return __ret;
53476}
53477#else
53478__ai poly64x1_t vmov_n_p64(poly64_t __p0) {
53479  poly64x1_t __ret;
53480  __ret = (poly64x1_t) {__p0};
53481  return __ret;
53482}
53483#endif
53484
53485#ifdef __LITTLE_ENDIAN__
53486__ai poly64x2_t vmovq_n_p64(poly64_t __p0) {
53487  poly64x2_t __ret;
53488  __ret = (poly64x2_t) {__p0, __p0};
53489  return __ret;
53490}
53491#else
53492__ai poly64x2_t vmovq_n_p64(poly64_t __p0) {
53493  poly64x2_t __ret;
53494  __ret = (poly64x2_t) {__p0, __p0};
53495  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53496  return __ret;
53497}
53498#endif
53499
53500#ifdef __LITTLE_ENDIAN__
53501__ai float64x2_t vmovq_n_f64(float64_t __p0) {
53502  float64x2_t __ret;
53503  __ret = (float64x2_t) {__p0, __p0};
53504  return __ret;
53505}
53506#else
53507__ai float64x2_t vmovq_n_f64(float64_t __p0) {
53508  float64x2_t __ret;
53509  __ret = (float64x2_t) {__p0, __p0};
53510  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53511  return __ret;
53512}
53513#endif
53514
53515#ifdef __LITTLE_ENDIAN__
53516__ai float64x1_t vmov_n_f64(float64_t __p0) {
53517  float64x1_t __ret;
53518  __ret = (float64x1_t) {__p0};
53519  return __ret;
53520}
53521#else
53522__ai float64x1_t vmov_n_f64(float64_t __p0) {
53523  float64x1_t __ret;
53524  __ret = (float64x1_t) {__p0};
53525  return __ret;
53526}
53527#endif
53528
53529#ifdef __LITTLE_ENDIAN__
53530__ai uint16x8_t vmovl_high_u8(uint8x16_t __p0_112) {
53531  uint16x8_t __ret_112;
53532  uint8x8_t __a1_112 = vget_high_u8(__p0_112);
53533  __ret_112 = (uint16x8_t)(vshll_n_u8(__a1_112, 0));
53534  return __ret_112;
53535}
53536#else
53537__ai uint16x8_t vmovl_high_u8(uint8x16_t __p0_113) {
53538  uint8x16_t __rev0_113;  __rev0_113 = __builtin_shufflevector(__p0_113, __p0_113, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53539  uint16x8_t __ret_113;
53540  uint8x8_t __a1_113 = __noswap_vget_high_u8(__rev0_113);
53541  __ret_113 = (uint16x8_t)(__noswap_vshll_n_u8(__a1_113, 0));
53542  __ret_113 = __builtin_shufflevector(__ret_113, __ret_113, 7, 6, 5, 4, 3, 2, 1, 0);
53543  return __ret_113;
53544}
53545__ai uint16x8_t __noswap_vmovl_high_u8(uint8x16_t __p0_114) {
53546  uint16x8_t __ret_114;
53547  uint8x8_t __a1_114 = __noswap_vget_high_u8(__p0_114);
53548  __ret_114 = (uint16x8_t)(__noswap_vshll_n_u8(__a1_114, 0));
53549  return __ret_114;
53550}
53551#endif
53552
53553#ifdef __LITTLE_ENDIAN__
53554__ai uint64x2_t vmovl_high_u32(uint32x4_t __p0_115) {
53555  uint64x2_t __ret_115;
53556  uint32x2_t __a1_115 = vget_high_u32(__p0_115);
53557  __ret_115 = (uint64x2_t)(vshll_n_u32(__a1_115, 0));
53558  return __ret_115;
53559}
53560#else
53561__ai uint64x2_t vmovl_high_u32(uint32x4_t __p0_116) {
53562  uint32x4_t __rev0_116;  __rev0_116 = __builtin_shufflevector(__p0_116, __p0_116, 3, 2, 1, 0);
53563  uint64x2_t __ret_116;
53564  uint32x2_t __a1_116 = __noswap_vget_high_u32(__rev0_116);
53565  __ret_116 = (uint64x2_t)(__noswap_vshll_n_u32(__a1_116, 0));
53566  __ret_116 = __builtin_shufflevector(__ret_116, __ret_116, 1, 0);
53567  return __ret_116;
53568}
53569__ai uint64x2_t __noswap_vmovl_high_u32(uint32x4_t __p0_117) {
53570  uint64x2_t __ret_117;
53571  uint32x2_t __a1_117 = __noswap_vget_high_u32(__p0_117);
53572  __ret_117 = (uint64x2_t)(__noswap_vshll_n_u32(__a1_117, 0));
53573  return __ret_117;
53574}
53575#endif
53576
53577#ifdef __LITTLE_ENDIAN__
53578__ai uint32x4_t vmovl_high_u16(uint16x8_t __p0_118) {
53579  uint32x4_t __ret_118;
53580  uint16x4_t __a1_118 = vget_high_u16(__p0_118);
53581  __ret_118 = (uint32x4_t)(vshll_n_u16(__a1_118, 0));
53582  return __ret_118;
53583}
53584#else
53585__ai uint32x4_t vmovl_high_u16(uint16x8_t __p0_119) {
53586  uint16x8_t __rev0_119;  __rev0_119 = __builtin_shufflevector(__p0_119, __p0_119, 7, 6, 5, 4, 3, 2, 1, 0);
53587  uint32x4_t __ret_119;
53588  uint16x4_t __a1_119 = __noswap_vget_high_u16(__rev0_119);
53589  __ret_119 = (uint32x4_t)(__noswap_vshll_n_u16(__a1_119, 0));
53590  __ret_119 = __builtin_shufflevector(__ret_119, __ret_119, 3, 2, 1, 0);
53591  return __ret_119;
53592}
53593__ai uint32x4_t __noswap_vmovl_high_u16(uint16x8_t __p0_120) {
53594  uint32x4_t __ret_120;
53595  uint16x4_t __a1_120 = __noswap_vget_high_u16(__p0_120);
53596  __ret_120 = (uint32x4_t)(__noswap_vshll_n_u16(__a1_120, 0));
53597  return __ret_120;
53598}
53599#endif
53600
53601#ifdef __LITTLE_ENDIAN__
53602__ai int16x8_t vmovl_high_s8(int8x16_t __p0_121) {
53603  int16x8_t __ret_121;
53604  int8x8_t __a1_121 = vget_high_s8(__p0_121);
53605  __ret_121 = (int16x8_t)(vshll_n_s8(__a1_121, 0));
53606  return __ret_121;
53607}
53608#else
53609__ai int16x8_t vmovl_high_s8(int8x16_t __p0_122) {
53610  int8x16_t __rev0_122;  __rev0_122 = __builtin_shufflevector(__p0_122, __p0_122, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53611  int16x8_t __ret_122;
53612  int8x8_t __a1_122 = __noswap_vget_high_s8(__rev0_122);
53613  __ret_122 = (int16x8_t)(__noswap_vshll_n_s8(__a1_122, 0));
53614  __ret_122 = __builtin_shufflevector(__ret_122, __ret_122, 7, 6, 5, 4, 3, 2, 1, 0);
53615  return __ret_122;
53616}
53617__ai int16x8_t __noswap_vmovl_high_s8(int8x16_t __p0_123) {
53618  int16x8_t __ret_123;
53619  int8x8_t __a1_123 = __noswap_vget_high_s8(__p0_123);
53620  __ret_123 = (int16x8_t)(__noswap_vshll_n_s8(__a1_123, 0));
53621  return __ret_123;
53622}
53623#endif
53624
53625#ifdef __LITTLE_ENDIAN__
53626__ai int64x2_t vmovl_high_s32(int32x4_t __p0_124) {
53627  int64x2_t __ret_124;
53628  int32x2_t __a1_124 = vget_high_s32(__p0_124);
53629  __ret_124 = (int64x2_t)(vshll_n_s32(__a1_124, 0));
53630  return __ret_124;
53631}
53632#else
53633__ai int64x2_t vmovl_high_s32(int32x4_t __p0_125) {
53634  int32x4_t __rev0_125;  __rev0_125 = __builtin_shufflevector(__p0_125, __p0_125, 3, 2, 1, 0);
53635  int64x2_t __ret_125;
53636  int32x2_t __a1_125 = __noswap_vget_high_s32(__rev0_125);
53637  __ret_125 = (int64x2_t)(__noswap_vshll_n_s32(__a1_125, 0));
53638  __ret_125 = __builtin_shufflevector(__ret_125, __ret_125, 1, 0);
53639  return __ret_125;
53640}
53641__ai int64x2_t __noswap_vmovl_high_s32(int32x4_t __p0_126) {
53642  int64x2_t __ret_126;
53643  int32x2_t __a1_126 = __noswap_vget_high_s32(__p0_126);
53644  __ret_126 = (int64x2_t)(__noswap_vshll_n_s32(__a1_126, 0));
53645  return __ret_126;
53646}
53647#endif
53648
53649#ifdef __LITTLE_ENDIAN__
53650__ai int32x4_t vmovl_high_s16(int16x8_t __p0_127) {
53651  int32x4_t __ret_127;
53652  int16x4_t __a1_127 = vget_high_s16(__p0_127);
53653  __ret_127 = (int32x4_t)(vshll_n_s16(__a1_127, 0));
53654  return __ret_127;
53655}
53656#else
53657__ai int32x4_t vmovl_high_s16(int16x8_t __p0_128) {
53658  int16x8_t __rev0_128;  __rev0_128 = __builtin_shufflevector(__p0_128, __p0_128, 7, 6, 5, 4, 3, 2, 1, 0);
53659  int32x4_t __ret_128;
53660  int16x4_t __a1_128 = __noswap_vget_high_s16(__rev0_128);
53661  __ret_128 = (int32x4_t)(__noswap_vshll_n_s16(__a1_128, 0));
53662  __ret_128 = __builtin_shufflevector(__ret_128, __ret_128, 3, 2, 1, 0);
53663  return __ret_128;
53664}
53665__ai int32x4_t __noswap_vmovl_high_s16(int16x8_t __p0_129) {
53666  int32x4_t __ret_129;
53667  int16x4_t __a1_129 = __noswap_vget_high_s16(__p0_129);
53668  __ret_129 = (int32x4_t)(__noswap_vshll_n_s16(__a1_129, 0));
53669  return __ret_129;
53670}
53671#endif
53672
53673#ifdef __LITTLE_ENDIAN__
53674__ai uint16x8_t vmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
53675  uint16x8_t __ret;
53676  __ret = vcombine_u16(__p0, vmovn_u32(__p1));
53677  return __ret;
53678}
53679#else
53680__ai uint16x8_t vmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
53681  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53682  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53683  uint16x8_t __ret;
53684  __ret = __noswap_vcombine_u16(__rev0, __noswap_vmovn_u32(__rev1));
53685  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
53686  return __ret;
53687}
53688#endif
53689
53690#ifdef __LITTLE_ENDIAN__
53691__ai uint32x4_t vmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
53692  uint32x4_t __ret;
53693  __ret = vcombine_u32(__p0, vmovn_u64(__p1));
53694  return __ret;
53695}
53696#else
53697__ai uint32x4_t vmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
53698  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53699  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53700  uint32x4_t __ret;
53701  __ret = __noswap_vcombine_u32(__rev0, __noswap_vmovn_u64(__rev1));
53702  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53703  return __ret;
53704}
53705#endif
53706
53707#ifdef __LITTLE_ENDIAN__
53708__ai uint8x16_t vmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
53709  uint8x16_t __ret;
53710  __ret = vcombine_u8(__p0, vmovn_u16(__p1));
53711  return __ret;
53712}
53713#else
53714__ai uint8x16_t vmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
53715  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53716  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
53717  uint8x16_t __ret;
53718  __ret = __noswap_vcombine_u8(__rev0, __noswap_vmovn_u16(__rev1));
53719  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53720  return __ret;
53721}
53722#endif
53723
53724#ifdef __LITTLE_ENDIAN__
53725__ai int16x8_t vmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
53726  int16x8_t __ret;
53727  __ret = vcombine_s16(__p0, vmovn_s32(__p1));
53728  return __ret;
53729}
53730#else
53731__ai int16x8_t vmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
53732  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
53733  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
53734  int16x8_t __ret;
53735  __ret = __noswap_vcombine_s16(__rev0, __noswap_vmovn_s32(__rev1));
53736  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
53737  return __ret;
53738}
53739#endif
53740
53741#ifdef __LITTLE_ENDIAN__
53742__ai int32x4_t vmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
53743  int32x4_t __ret;
53744  __ret = vcombine_s32(__p0, vmovn_s64(__p1));
53745  return __ret;
53746}
53747#else
53748__ai int32x4_t vmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
53749  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53750  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53751  int32x4_t __ret;
53752  __ret = __noswap_vcombine_s32(__rev0, __noswap_vmovn_s64(__rev1));
53753  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
53754  return __ret;
53755}
53756#endif
53757
53758#ifdef __LITTLE_ENDIAN__
53759__ai int8x16_t vmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
53760  int8x16_t __ret;
53761  __ret = vcombine_s8(__p0, vmovn_s16(__p1));
53762  return __ret;
53763}
53764#else
53765__ai int8x16_t vmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
53766  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
53767  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
53768  int8x16_t __ret;
53769  __ret = __noswap_vcombine_s8(__rev0, __noswap_vmovn_s16(__rev1));
53770  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
53771  return __ret;
53772}
53773#endif
53774
53775#ifdef __LITTLE_ENDIAN__
53776__ai float64x2_t vmulq_f64(float64x2_t __p0, float64x2_t __p1) {
53777  float64x2_t __ret;
53778  __ret = __p0 * __p1;
53779  return __ret;
53780}
53781#else
53782__ai float64x2_t vmulq_f64(float64x2_t __p0, float64x2_t __p1) {
53783  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
53784  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
53785  float64x2_t __ret;
53786  __ret = __rev0 * __rev1;
53787  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
53788  return __ret;
53789}
53790#endif
53791
53792#ifdef __LITTLE_ENDIAN__
53793__ai float64x1_t vmul_f64(float64x1_t __p0, float64x1_t __p1) {
53794  float64x1_t __ret;
53795  __ret = __p0 * __p1;
53796  return __ret;
53797}
53798#else
53799__ai float64x1_t vmul_f64(float64x1_t __p0, float64x1_t __p1) {
53800  float64x1_t __ret;
53801  __ret = __p0 * __p1;
53802  return __ret;
53803}
53804#endif
53805
53806#ifdef __LITTLE_ENDIAN__
53807#define vmuld_lane_f64(__p0_130, __p1_130, __p2_130) __extension__ ({ \
53808  float64_t __s0_130 = __p0_130; \
53809  float64x1_t __s1_130 = __p1_130; \
53810  float64_t __ret_130; \
53811  __ret_130 = __s0_130 * vget_lane_f64(__s1_130, __p2_130); \
53812  __ret_130; \
53813})
53814#else
53815#define vmuld_lane_f64(__p0_131, __p1_131, __p2_131) __extension__ ({ \
53816  float64_t __s0_131 = __p0_131; \
53817  float64x1_t __s1_131 = __p1_131; \
53818  float64_t __ret_131; \
53819  __ret_131 = __s0_131 * __noswap_vget_lane_f64(__s1_131, __p2_131); \
53820  __ret_131; \
53821})
53822#endif
53823
53824#ifdef __LITTLE_ENDIAN__
53825#define vmuls_lane_f32(__p0_132, __p1_132, __p2_132) __extension__ ({ \
53826  float32_t __s0_132 = __p0_132; \
53827  float32x2_t __s1_132 = __p1_132; \
53828  float32_t __ret_132; \
53829  __ret_132 = __s0_132 * vget_lane_f32(__s1_132, __p2_132); \
53830  __ret_132; \
53831})
53832#else
53833#define vmuls_lane_f32(__p0_133, __p1_133, __p2_133) __extension__ ({ \
53834  float32_t __s0_133 = __p0_133; \
53835  float32x2_t __s1_133 = __p1_133; \
53836  float32x2_t __rev1_133;  __rev1_133 = __builtin_shufflevector(__s1_133, __s1_133, 1, 0); \
53837  float32_t __ret_133; \
53838  __ret_133 = __s0_133 * __noswap_vget_lane_f32(__rev1_133, __p2_133); \
53839  __ret_133; \
53840})
53841#endif
53842
53843#ifdef __LITTLE_ENDIAN__
53844#define vmul_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53845  float64x1_t __s0 = __p0; \
53846  float64x1_t __s1 = __p1; \
53847  float64x1_t __ret; \
53848  __ret = (float64x1_t) __builtin_neon_vmul_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
53849  __ret; \
53850})
53851#else
53852#define vmul_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53853  float64x1_t __s0 = __p0; \
53854  float64x1_t __s1 = __p1; \
53855  float64x1_t __ret; \
53856  __ret = (float64x1_t) __builtin_neon_vmul_lane_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 10); \
53857  __ret; \
53858})
53859#endif
53860
53861#ifdef __LITTLE_ENDIAN__
53862#define vmulq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53863  float64x2_t __s0 = __p0; \
53864  float64x1_t __s1 = __p1; \
53865  float64x2_t __ret; \
53866  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
53867  __ret; \
53868})
53869#else
53870#define vmulq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
53871  float64x2_t __s0 = __p0; \
53872  float64x1_t __s1 = __p1; \
53873  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53874  float64x2_t __ret; \
53875  __ret = __rev0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
53876  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53877  __ret; \
53878})
53879#endif
53880
53881#ifdef __LITTLE_ENDIAN__
53882#define vmuld_laneq_f64(__p0_134, __p1_134, __p2_134) __extension__ ({ \
53883  float64_t __s0_134 = __p0_134; \
53884  float64x2_t __s1_134 = __p1_134; \
53885  float64_t __ret_134; \
53886  __ret_134 = __s0_134 * vgetq_lane_f64(__s1_134, __p2_134); \
53887  __ret_134; \
53888})
53889#else
53890#define vmuld_laneq_f64(__p0_135, __p1_135, __p2_135) __extension__ ({ \
53891  float64_t __s0_135 = __p0_135; \
53892  float64x2_t __s1_135 = __p1_135; \
53893  float64x2_t __rev1_135;  __rev1_135 = __builtin_shufflevector(__s1_135, __s1_135, 1, 0); \
53894  float64_t __ret_135; \
53895  __ret_135 = __s0_135 * __noswap_vgetq_lane_f64(__rev1_135, __p2_135); \
53896  __ret_135; \
53897})
53898#endif
53899
53900#ifdef __LITTLE_ENDIAN__
53901#define vmuls_laneq_f32(__p0_136, __p1_136, __p2_136) __extension__ ({ \
53902  float32_t __s0_136 = __p0_136; \
53903  float32x4_t __s1_136 = __p1_136; \
53904  float32_t __ret_136; \
53905  __ret_136 = __s0_136 * vgetq_lane_f32(__s1_136, __p2_136); \
53906  __ret_136; \
53907})
53908#else
53909#define vmuls_laneq_f32(__p0_137, __p1_137, __p2_137) __extension__ ({ \
53910  float32_t __s0_137 = __p0_137; \
53911  float32x4_t __s1_137 = __p1_137; \
53912  float32x4_t __rev1_137;  __rev1_137 = __builtin_shufflevector(__s1_137, __s1_137, 3, 2, 1, 0); \
53913  float32_t __ret_137; \
53914  __ret_137 = __s0_137 * __noswap_vgetq_lane_f32(__rev1_137, __p2_137); \
53915  __ret_137; \
53916})
53917#endif
53918
53919#ifdef __LITTLE_ENDIAN__
53920#define vmul_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
53921  float64x1_t __s0 = __p0; \
53922  float64x2_t __s1 = __p1; \
53923  float64x1_t __ret; \
53924  __ret = (float64x1_t) __builtin_neon_vmul_laneq_v((int8x8_t)__s0, (int8x16_t)__s1, __p2, 10); \
53925  __ret; \
53926})
53927#else
53928#define vmul_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
53929  float64x1_t __s0 = __p0; \
53930  float64x2_t __s1 = __p1; \
53931  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53932  float64x1_t __ret; \
53933  __ret = (float64x1_t) __builtin_neon_vmul_laneq_v((int8x8_t)__s0, (int8x16_t)__rev1, __p2, 10); \
53934  __ret; \
53935})
53936#endif
53937
53938#ifdef __LITTLE_ENDIAN__
53939#define vmulq_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
53940  uint32x4_t __s0 = __p0; \
53941  uint32x4_t __s1 = __p1; \
53942  uint32x4_t __ret; \
53943  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
53944  __ret; \
53945})
53946#else
53947#define vmulq_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
53948  uint32x4_t __s0 = __p0; \
53949  uint32x4_t __s1 = __p1; \
53950  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
53951  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
53952  uint32x4_t __ret; \
53953  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
53954  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
53955  __ret; \
53956})
53957#endif
53958
53959#ifdef __LITTLE_ENDIAN__
53960#define vmulq_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
53961  uint16x8_t __s0 = __p0; \
53962  uint16x8_t __s1 = __p1; \
53963  uint16x8_t __ret; \
53964  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
53965  __ret; \
53966})
53967#else
53968#define vmulq_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
53969  uint16x8_t __s0 = __p0; \
53970  uint16x8_t __s1 = __p1; \
53971  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
53972  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
53973  uint16x8_t __ret; \
53974  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
53975  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
53976  __ret; \
53977})
53978#endif
53979
53980#ifdef __LITTLE_ENDIAN__
53981#define vmulq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
53982  float64x2_t __s0 = __p0; \
53983  float64x2_t __s1 = __p1; \
53984  float64x2_t __ret; \
53985  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
53986  __ret; \
53987})
53988#else
53989#define vmulq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
53990  float64x2_t __s0 = __p0; \
53991  float64x2_t __s1 = __p1; \
53992  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
53993  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
53994  float64x2_t __ret; \
53995  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
53996  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
53997  __ret; \
53998})
53999#endif
54000
54001#ifdef __LITTLE_ENDIAN__
54002#define vmulq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54003  float32x4_t __s0 = __p0; \
54004  float32x4_t __s1 = __p1; \
54005  float32x4_t __ret; \
54006  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
54007  __ret; \
54008})
54009#else
54010#define vmulq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54011  float32x4_t __s0 = __p0; \
54012  float32x4_t __s1 = __p1; \
54013  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54014  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54015  float32x4_t __ret; \
54016  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
54017  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54018  __ret; \
54019})
54020#endif
54021
54022#ifdef __LITTLE_ENDIAN__
54023#define vmulq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54024  int32x4_t __s0 = __p0; \
54025  int32x4_t __s1 = __p1; \
54026  int32x4_t __ret; \
54027  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
54028  __ret; \
54029})
54030#else
54031#define vmulq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54032  int32x4_t __s0 = __p0; \
54033  int32x4_t __s1 = __p1; \
54034  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54035  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54036  int32x4_t __ret; \
54037  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
54038  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54039  __ret; \
54040})
54041#endif
54042
54043#ifdef __LITTLE_ENDIAN__
54044#define vmulq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54045  int16x8_t __s0 = __p0; \
54046  int16x8_t __s1 = __p1; \
54047  int16x8_t __ret; \
54048  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
54049  __ret; \
54050})
54051#else
54052#define vmulq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54053  int16x8_t __s0 = __p0; \
54054  int16x8_t __s1 = __p1; \
54055  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54056  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54057  int16x8_t __ret; \
54058  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2); \
54059  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
54060  __ret; \
54061})
54062#endif
54063
54064#ifdef __LITTLE_ENDIAN__
54065#define vmul_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54066  uint32x2_t __s0 = __p0; \
54067  uint32x4_t __s1 = __p1; \
54068  uint32x2_t __ret; \
54069  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
54070  __ret; \
54071})
54072#else
54073#define vmul_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54074  uint32x2_t __s0 = __p0; \
54075  uint32x4_t __s1 = __p1; \
54076  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54077  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54078  uint32x2_t __ret; \
54079  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
54080  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54081  __ret; \
54082})
54083#endif
54084
54085#ifdef __LITTLE_ENDIAN__
54086#define vmul_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54087  uint16x4_t __s0 = __p0; \
54088  uint16x8_t __s1 = __p1; \
54089  uint16x4_t __ret; \
54090  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
54091  __ret; \
54092})
54093#else
54094#define vmul_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54095  uint16x4_t __s0 = __p0; \
54096  uint16x8_t __s1 = __p1; \
54097  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54098  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54099  uint16x4_t __ret; \
54100  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
54101  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54102  __ret; \
54103})
54104#endif
54105
54106#ifdef __LITTLE_ENDIAN__
54107#define vmul_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54108  float32x2_t __s0 = __p0; \
54109  float32x4_t __s1 = __p1; \
54110  float32x2_t __ret; \
54111  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
54112  __ret; \
54113})
54114#else
54115#define vmul_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54116  float32x2_t __s0 = __p0; \
54117  float32x4_t __s1 = __p1; \
54118  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54119  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54120  float32x2_t __ret; \
54121  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
54122  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54123  __ret; \
54124})
54125#endif
54126
54127#ifdef __LITTLE_ENDIAN__
54128#define vmul_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54129  int32x2_t __s0 = __p0; \
54130  int32x4_t __s1 = __p1; \
54131  int32x2_t __ret; \
54132  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2); \
54133  __ret; \
54134})
54135#else
54136#define vmul_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54137  int32x2_t __s0 = __p0; \
54138  int32x4_t __s1 = __p1; \
54139  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54140  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54141  int32x2_t __ret; \
54142  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2); \
54143  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54144  __ret; \
54145})
54146#endif
54147
54148#ifdef __LITTLE_ENDIAN__
54149#define vmul_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54150  int16x4_t __s0 = __p0; \
54151  int16x8_t __s1 = __p1; \
54152  int16x4_t __ret; \
54153  __ret = __s0 * __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2); \
54154  __ret; \
54155})
54156#else
54157#define vmul_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54158  int16x4_t __s0 = __p0; \
54159  int16x8_t __s1 = __p1; \
54160  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54161  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54162  int16x4_t __ret; \
54163  __ret = __rev0 * __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2); \
54164  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54165  __ret; \
54166})
54167#endif
54168
54169#ifdef __LITTLE_ENDIAN__
54170__ai float64x1_t vmul_n_f64(float64x1_t __p0, float64_t __p1) {
54171  float64x1_t __ret;
54172  __ret = (float64x1_t) __builtin_neon_vmul_n_f64((int8x8_t)__p0, __p1);
54173  return __ret;
54174}
54175#else
54176__ai float64x1_t vmul_n_f64(float64x1_t __p0, float64_t __p1) {
54177  float64x1_t __ret;
54178  __ret = (float64x1_t) __builtin_neon_vmul_n_f64((int8x8_t)__p0, __p1);
54179  return __ret;
54180}
54181#endif
54182
54183#ifdef __LITTLE_ENDIAN__
54184__ai float64x2_t vmulq_n_f64(float64x2_t __p0, float64_t __p1) {
54185  float64x2_t __ret;
54186  __ret = __p0 * (float64x2_t) {__p1, __p1};
54187  return __ret;
54188}
54189#else
54190__ai float64x2_t vmulq_n_f64(float64x2_t __p0, float64_t __p1) {
54191  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54192  float64x2_t __ret;
54193  __ret = __rev0 * (float64x2_t) {__p1, __p1};
54194  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54195  return __ret;
54196}
54197#endif
54198
54199#ifdef __LITTLE_ENDIAN__
54200__ai poly128_t vmull_p64(poly64_t __p0, poly64_t __p1) {
54201  poly128_t __ret;
54202  __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
54203  return __ret;
54204}
54205#else
54206__ai poly128_t vmull_p64(poly64_t __p0, poly64_t __p1) {
54207  poly128_t __ret;
54208  __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
54209  return __ret;
54210}
54211__ai poly128_t __noswap_vmull_p64(poly64_t __p0, poly64_t __p1) {
54212  poly128_t __ret;
54213  __ret = (poly128_t) __builtin_neon_vmull_p64(__p0, __p1);
54214  return __ret;
54215}
54216#endif
54217
54218#ifdef __LITTLE_ENDIAN__
54219__ai poly16x8_t vmull_high_p8(poly8x16_t __p0, poly8x16_t __p1) {
54220  poly16x8_t __ret;
54221  __ret = vmull_p8(vget_high_p8(__p0), vget_high_p8(__p1));
54222  return __ret;
54223}
54224#else
54225__ai poly16x8_t vmull_high_p8(poly8x16_t __p0, poly8x16_t __p1) {
54226  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54227  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54228  poly16x8_t __ret;
54229  __ret = __noswap_vmull_p8(__noswap_vget_high_p8(__rev0), __noswap_vget_high_p8(__rev1));
54230  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54231  return __ret;
54232}
54233#endif
54234
54235#ifdef __LITTLE_ENDIAN__
54236__ai uint16x8_t vmull_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
54237  uint16x8_t __ret;
54238  __ret = vmull_u8(vget_high_u8(__p0), vget_high_u8(__p1));
54239  return __ret;
54240}
54241#else
54242__ai uint16x8_t vmull_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
54243  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54244  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54245  uint16x8_t __ret;
54246  __ret = __noswap_vmull_u8(__noswap_vget_high_u8(__rev0), __noswap_vget_high_u8(__rev1));
54247  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54248  return __ret;
54249}
54250#endif
54251
54252#ifdef __LITTLE_ENDIAN__
54253__ai uint64x2_t vmull_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
54254  uint64x2_t __ret;
54255  __ret = vmull_u32(vget_high_u32(__p0), vget_high_u32(__p1));
54256  return __ret;
54257}
54258#else
54259__ai uint64x2_t vmull_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
54260  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54261  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54262  uint64x2_t __ret;
54263  __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __noswap_vget_high_u32(__rev1));
54264  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54265  return __ret;
54266}
54267#endif
54268
54269#ifdef __LITTLE_ENDIAN__
54270__ai uint32x4_t vmull_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
54271  uint32x4_t __ret;
54272  __ret = vmull_u16(vget_high_u16(__p0), vget_high_u16(__p1));
54273  return __ret;
54274}
54275#else
54276__ai uint32x4_t vmull_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
54277  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54278  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54279  uint32x4_t __ret;
54280  __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __noswap_vget_high_u16(__rev1));
54281  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54282  return __ret;
54283}
54284#endif
54285
54286#ifdef __LITTLE_ENDIAN__
54287__ai int16x8_t vmull_high_s8(int8x16_t __p0, int8x16_t __p1) {
54288  int16x8_t __ret;
54289  __ret = vmull_s8(vget_high_s8(__p0), vget_high_s8(__p1));
54290  return __ret;
54291}
54292#else
54293__ai int16x8_t vmull_high_s8(int8x16_t __p0, int8x16_t __p1) {
54294  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54295  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
54296  int16x8_t __ret;
54297  __ret = __noswap_vmull_s8(__noswap_vget_high_s8(__rev0), __noswap_vget_high_s8(__rev1));
54298  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
54299  return __ret;
54300}
54301#endif
54302
54303#ifdef __LITTLE_ENDIAN__
54304__ai int64x2_t vmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
54305  int64x2_t __ret;
54306  __ret = vmull_s32(vget_high_s32(__p0), vget_high_s32(__p1));
54307  return __ret;
54308}
54309#else
54310__ai int64x2_t vmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
54311  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54312  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54313  int64x2_t __ret;
54314  __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
54315  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54316  return __ret;
54317}
54318#endif
54319
54320#ifdef __LITTLE_ENDIAN__
54321__ai int32x4_t vmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
54322  int32x4_t __ret;
54323  __ret = vmull_s16(vget_high_s16(__p0), vget_high_s16(__p1));
54324  return __ret;
54325}
54326#else
54327__ai int32x4_t vmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
54328  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54329  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
54330  int32x4_t __ret;
54331  __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
54332  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54333  return __ret;
54334}
54335#endif
54336
54337#ifdef __LITTLE_ENDIAN__
54338__ai poly128_t vmull_high_p64(poly64x2_t __p0, poly64x2_t __p1) {
54339  poly128_t __ret;
54340  __ret = vmull_p64((poly64_t)(vget_high_p64(__p0)), (poly64_t)(vget_high_p64(__p1)));
54341  return __ret;
54342}
54343#else
54344__ai poly128_t vmull_high_p64(poly64x2_t __p0, poly64x2_t __p1) {
54345  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54346  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54347  poly128_t __ret;
54348  __ret = __noswap_vmull_p64((poly64_t)(__noswap_vget_high_p64(__rev0)), (poly64_t)(__noswap_vget_high_p64(__rev1)));
54349  return __ret;
54350}
54351#endif
54352
54353#ifdef __LITTLE_ENDIAN__
54354#define vmull_high_lane_u32(__p0, __p1, __p2) __extension__ ({ \
54355  uint32x4_t __s0 = __p0; \
54356  uint32x2_t __s1 = __p1; \
54357  uint64x2_t __ret; \
54358  __ret = vmull_u32(vget_high_u32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54359  __ret; \
54360})
54361#else
54362#define vmull_high_lane_u32(__p0, __p1, __p2) __extension__ ({ \
54363  uint32x4_t __s0 = __p0; \
54364  uint32x2_t __s1 = __p1; \
54365  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54366  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54367  uint64x2_t __ret; \
54368  __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54369  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54370  __ret; \
54371})
54372#endif
54373
54374#ifdef __LITTLE_ENDIAN__
54375#define vmull_high_lane_u16(__p0, __p1, __p2) __extension__ ({ \
54376  uint16x8_t __s0 = __p0; \
54377  uint16x4_t __s1 = __p1; \
54378  uint32x4_t __ret; \
54379  __ret = vmull_u16(vget_high_u16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54380  __ret; \
54381})
54382#else
54383#define vmull_high_lane_u16(__p0, __p1, __p2) __extension__ ({ \
54384  uint16x8_t __s0 = __p0; \
54385  uint16x4_t __s1 = __p1; \
54386  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54387  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54388  uint32x4_t __ret; \
54389  __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54390  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54391  __ret; \
54392})
54393#endif
54394
54395#ifdef __LITTLE_ENDIAN__
54396#define vmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
54397  int32x4_t __s0 = __p0; \
54398  int32x2_t __s1 = __p1; \
54399  int64x2_t __ret; \
54400  __ret = vmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54401  __ret; \
54402})
54403#else
54404#define vmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
54405  int32x4_t __s0 = __p0; \
54406  int32x2_t __s1 = __p1; \
54407  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54408  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54409  int64x2_t __ret; \
54410  __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54411  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54412  __ret; \
54413})
54414#endif
54415
54416#ifdef __LITTLE_ENDIAN__
54417#define vmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
54418  int16x8_t __s0 = __p0; \
54419  int16x4_t __s1 = __p1; \
54420  int32x4_t __ret; \
54421  __ret = vmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54422  __ret; \
54423})
54424#else
54425#define vmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
54426  int16x8_t __s0 = __p0; \
54427  int16x4_t __s1 = __p1; \
54428  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54429  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54430  int32x4_t __ret; \
54431  __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54432  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54433  __ret; \
54434})
54435#endif
54436
54437#ifdef __LITTLE_ENDIAN__
54438#define vmull_high_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54439  uint32x4_t __s0 = __p0; \
54440  uint32x4_t __s1 = __p1; \
54441  uint64x2_t __ret; \
54442  __ret = vmull_u32(vget_high_u32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54443  __ret; \
54444})
54445#else
54446#define vmull_high_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54447  uint32x4_t __s0 = __p0; \
54448  uint32x4_t __s1 = __p1; \
54449  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54450  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54451  uint64x2_t __ret; \
54452  __ret = __noswap_vmull_u32(__noswap_vget_high_u32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54453  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54454  __ret; \
54455})
54456#endif
54457
54458#ifdef __LITTLE_ENDIAN__
54459#define vmull_high_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54460  uint16x8_t __s0 = __p0; \
54461  uint16x8_t __s1 = __p1; \
54462  uint32x4_t __ret; \
54463  __ret = vmull_u16(vget_high_u16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54464  __ret; \
54465})
54466#else
54467#define vmull_high_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54468  uint16x8_t __s0 = __p0; \
54469  uint16x8_t __s1 = __p1; \
54470  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54471  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54472  uint32x4_t __ret; \
54473  __ret = __noswap_vmull_u16(__noswap_vget_high_u16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54474  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54475  __ret; \
54476})
54477#endif
54478
54479#ifdef __LITTLE_ENDIAN__
54480#define vmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54481  int32x4_t __s0 = __p0; \
54482  int32x4_t __s1 = __p1; \
54483  int64x2_t __ret; \
54484  __ret = vmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54485  __ret; \
54486})
54487#else
54488#define vmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54489  int32x4_t __s0 = __p0; \
54490  int32x4_t __s1 = __p1; \
54491  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54492  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54493  int64x2_t __ret; \
54494  __ret = __noswap_vmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54495  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54496  __ret; \
54497})
54498#endif
54499
54500#ifdef __LITTLE_ENDIAN__
54501#define vmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54502  int16x8_t __s0 = __p0; \
54503  int16x8_t __s1 = __p1; \
54504  int32x4_t __ret; \
54505  __ret = vmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54506  __ret; \
54507})
54508#else
54509#define vmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54510  int16x8_t __s0 = __p0; \
54511  int16x8_t __s1 = __p1; \
54512  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
54513  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54514  int32x4_t __ret; \
54515  __ret = __noswap_vmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54516  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54517  __ret; \
54518})
54519#endif
54520
54521#ifdef __LITTLE_ENDIAN__
54522__ai uint64x2_t vmull_high_n_u32(uint32x4_t __p0, uint32_t __p1) {
54523  uint64x2_t __ret;
54524  __ret = vmull_n_u32(vget_high_u32(__p0), __p1);
54525  return __ret;
54526}
54527#else
54528__ai uint64x2_t vmull_high_n_u32(uint32x4_t __p0, uint32_t __p1) {
54529  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54530  uint64x2_t __ret;
54531  __ret = __noswap_vmull_n_u32(__noswap_vget_high_u32(__rev0), __p1);
54532  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54533  return __ret;
54534}
54535#endif
54536
54537#ifdef __LITTLE_ENDIAN__
54538__ai uint32x4_t vmull_high_n_u16(uint16x8_t __p0, uint16_t __p1) {
54539  uint32x4_t __ret;
54540  __ret = vmull_n_u16(vget_high_u16(__p0), __p1);
54541  return __ret;
54542}
54543#else
54544__ai uint32x4_t vmull_high_n_u16(uint16x8_t __p0, uint16_t __p1) {
54545  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54546  uint32x4_t __ret;
54547  __ret = __noswap_vmull_n_u16(__noswap_vget_high_u16(__rev0), __p1);
54548  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54549  return __ret;
54550}
54551#endif
54552
54553#ifdef __LITTLE_ENDIAN__
54554__ai int64x2_t vmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
54555  int64x2_t __ret;
54556  __ret = vmull_n_s32(vget_high_s32(__p0), __p1);
54557  return __ret;
54558}
54559#else
54560__ai int64x2_t vmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
54561  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54562  int64x2_t __ret;
54563  __ret = __noswap_vmull_n_s32(__noswap_vget_high_s32(__rev0), __p1);
54564  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54565  return __ret;
54566}
54567#endif
54568
54569#ifdef __LITTLE_ENDIAN__
54570__ai int32x4_t vmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
54571  int32x4_t __ret;
54572  __ret = vmull_n_s16(vget_high_s16(__p0), __p1);
54573  return __ret;
54574}
54575#else
54576__ai int32x4_t vmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
54577  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
54578  int32x4_t __ret;
54579  __ret = __noswap_vmull_n_s16(__noswap_vget_high_s16(__rev0), __p1);
54580  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54581  return __ret;
54582}
54583#endif
54584
54585#ifdef __LITTLE_ENDIAN__
54586#define vmull_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54587  uint32x2_t __s0 = __p0; \
54588  uint32x4_t __s1 = __p1; \
54589  uint64x2_t __ret; \
54590  __ret = vmull_u32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54591  __ret; \
54592})
54593#else
54594#define vmull_laneq_u32(__p0, __p1, __p2) __extension__ ({ \
54595  uint32x2_t __s0 = __p0; \
54596  uint32x4_t __s1 = __p1; \
54597  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54598  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54599  uint64x2_t __ret; \
54600  __ret = __noswap_vmull_u32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54601  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54602  __ret; \
54603})
54604#endif
54605
54606#ifdef __LITTLE_ENDIAN__
54607#define vmull_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54608  uint16x4_t __s0 = __p0; \
54609  uint16x8_t __s1 = __p1; \
54610  uint32x4_t __ret; \
54611  __ret = vmull_u16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54612  __ret; \
54613})
54614#else
54615#define vmull_laneq_u16(__p0, __p1, __p2) __extension__ ({ \
54616  uint16x4_t __s0 = __p0; \
54617  uint16x8_t __s1 = __p1; \
54618  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54619  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54620  uint32x4_t __ret; \
54621  __ret = __noswap_vmull_u16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54622  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54623  __ret; \
54624})
54625#endif
54626
54627#ifdef __LITTLE_ENDIAN__
54628#define vmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54629  int32x2_t __s0 = __p0; \
54630  int32x4_t __s1 = __p1; \
54631  int64x2_t __ret; \
54632  __ret = vmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54633  __ret; \
54634})
54635#else
54636#define vmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
54637  int32x2_t __s0 = __p0; \
54638  int32x4_t __s1 = __p1; \
54639  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54640  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54641  int64x2_t __ret; \
54642  __ret = __noswap_vmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54643  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54644  __ret; \
54645})
54646#endif
54647
54648#ifdef __LITTLE_ENDIAN__
54649#define vmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54650  int16x4_t __s0 = __p0; \
54651  int16x8_t __s1 = __p1; \
54652  int32x4_t __ret; \
54653  __ret = vmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54654  __ret; \
54655})
54656#else
54657#define vmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
54658  int16x4_t __s0 = __p0; \
54659  int16x8_t __s1 = __p1; \
54660  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54661  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
54662  int32x4_t __ret; \
54663  __ret = __noswap_vmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54664  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54665  __ret; \
54666})
54667#endif
54668
54669#ifdef __LITTLE_ENDIAN__
54670__ai float64x2_t vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
54671  float64x2_t __ret;
54672  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
54673  return __ret;
54674}
54675#else
54676__ai float64x2_t vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
54677  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54678  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54679  float64x2_t __ret;
54680  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
54681  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54682  return __ret;
54683}
54684__ai float64x2_t __noswap_vmulxq_f64(float64x2_t __p0, float64x2_t __p1) {
54685  float64x2_t __ret;
54686  __ret = (float64x2_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
54687  return __ret;
54688}
54689#endif
54690
54691#ifdef __LITTLE_ENDIAN__
54692__ai float32x4_t vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
54693  float32x4_t __ret;
54694  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
54695  return __ret;
54696}
54697#else
54698__ai float32x4_t vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
54699  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
54700  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
54701  float32x4_t __ret;
54702  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
54703  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
54704  return __ret;
54705}
54706__ai float32x4_t __noswap_vmulxq_f32(float32x4_t __p0, float32x4_t __p1) {
54707  float32x4_t __ret;
54708  __ret = (float32x4_t) __builtin_neon_vmulxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
54709  return __ret;
54710}
54711#endif
54712
54713#ifdef __LITTLE_ENDIAN__
54714__ai float64x1_t vmulx_f64(float64x1_t __p0, float64x1_t __p1) {
54715  float64x1_t __ret;
54716  __ret = (float64x1_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
54717  return __ret;
54718}
54719#else
54720__ai float64x1_t vmulx_f64(float64x1_t __p0, float64x1_t __p1) {
54721  float64x1_t __ret;
54722  __ret = (float64x1_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
54723  return __ret;
54724}
54725#endif
54726
54727#ifdef __LITTLE_ENDIAN__
54728__ai float32x2_t vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
54729  float32x2_t __ret;
54730  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
54731  return __ret;
54732}
54733#else
54734__ai float32x2_t vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
54735  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54736  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
54737  float32x2_t __ret;
54738  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
54739  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54740  return __ret;
54741}
54742__ai float32x2_t __noswap_vmulx_f32(float32x2_t __p0, float32x2_t __p1) {
54743  float32x2_t __ret;
54744  __ret = (float32x2_t) __builtin_neon_vmulx_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
54745  return __ret;
54746}
54747#endif
54748
54749#ifdef __LITTLE_ENDIAN__
54750__ai float64_t vmulxd_f64(float64_t __p0, float64_t __p1) {
54751  float64_t __ret;
54752  __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
54753  return __ret;
54754}
54755#else
54756__ai float64_t vmulxd_f64(float64_t __p0, float64_t __p1) {
54757  float64_t __ret;
54758  __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
54759  return __ret;
54760}
54761__ai float64_t __noswap_vmulxd_f64(float64_t __p0, float64_t __p1) {
54762  float64_t __ret;
54763  __ret = (float64_t) __builtin_neon_vmulxd_f64(__p0, __p1);
54764  return __ret;
54765}
54766#endif
54767
54768#ifdef __LITTLE_ENDIAN__
54769__ai float32_t vmulxs_f32(float32_t __p0, float32_t __p1) {
54770  float32_t __ret;
54771  __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
54772  return __ret;
54773}
54774#else
54775__ai float32_t vmulxs_f32(float32_t __p0, float32_t __p1) {
54776  float32_t __ret;
54777  __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
54778  return __ret;
54779}
54780__ai float32_t __noswap_vmulxs_f32(float32_t __p0, float32_t __p1) {
54781  float32_t __ret;
54782  __ret = (float32_t) __builtin_neon_vmulxs_f32(__p0, __p1);
54783  return __ret;
54784}
54785#endif
54786
54787#ifdef __LITTLE_ENDIAN__
54788#define vmulxd_lane_f64(__p0_138, __p1_138, __p2_138) __extension__ ({ \
54789  float64_t __s0_138 = __p0_138; \
54790  float64x1_t __s1_138 = __p1_138; \
54791  float64_t __ret_138; \
54792  __ret_138 = vmulxd_f64(__s0_138, vget_lane_f64(__s1_138, __p2_138)); \
54793  __ret_138; \
54794})
54795#else
54796#define vmulxd_lane_f64(__p0_139, __p1_139, __p2_139) __extension__ ({ \
54797  float64_t __s0_139 = __p0_139; \
54798  float64x1_t __s1_139 = __p1_139; \
54799  float64_t __ret_139; \
54800  __ret_139 = __noswap_vmulxd_f64(__s0_139, __noswap_vget_lane_f64(__s1_139, __p2_139)); \
54801  __ret_139; \
54802})
54803#endif
54804
54805#ifdef __LITTLE_ENDIAN__
54806#define vmulxs_lane_f32(__p0_140, __p1_140, __p2_140) __extension__ ({ \
54807  float32_t __s0_140 = __p0_140; \
54808  float32x2_t __s1_140 = __p1_140; \
54809  float32_t __ret_140; \
54810  __ret_140 = vmulxs_f32(__s0_140, vget_lane_f32(__s1_140, __p2_140)); \
54811  __ret_140; \
54812})
54813#else
54814#define vmulxs_lane_f32(__p0_141, __p1_141, __p2_141) __extension__ ({ \
54815  float32_t __s0_141 = __p0_141; \
54816  float32x2_t __s1_141 = __p1_141; \
54817  float32x2_t __rev1_141;  __rev1_141 = __builtin_shufflevector(__s1_141, __s1_141, 1, 0); \
54818  float32_t __ret_141; \
54819  __ret_141 = __noswap_vmulxs_f32(__s0_141, __noswap_vget_lane_f32(__rev1_141, __p2_141)); \
54820  __ret_141; \
54821})
54822#endif
54823
54824#ifdef __LITTLE_ENDIAN__
54825#define vmulxq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
54826  float64x2_t __s0 = __p0; \
54827  float64x1_t __s1 = __p1; \
54828  float64x2_t __ret; \
54829  __ret = vmulxq_f64(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54830  __ret; \
54831})
54832#else
54833#define vmulxq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
54834  float64x2_t __s0 = __p0; \
54835  float64x1_t __s1 = __p1; \
54836  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54837  float64x2_t __ret; \
54838  __ret = __noswap_vmulxq_f64(__rev0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54839  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54840  __ret; \
54841})
54842#endif
54843
54844#ifdef __LITTLE_ENDIAN__
54845#define vmulxq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
54846  float32x4_t __s0 = __p0; \
54847  float32x2_t __s1 = __p1; \
54848  float32x4_t __ret; \
54849  __ret = vmulxq_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54850  __ret; \
54851})
54852#else
54853#define vmulxq_lane_f32(__p0, __p1, __p2) __extension__ ({ \
54854  float32x4_t __s0 = __p0; \
54855  float32x2_t __s1 = __p1; \
54856  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54857  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54858  float32x4_t __ret; \
54859  __ret = __noswap_vmulxq_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54860  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54861  __ret; \
54862})
54863#endif
54864
54865#ifdef __LITTLE_ENDIAN__
54866#define vmulx_lane_f32(__p0, __p1, __p2) __extension__ ({ \
54867  float32x2_t __s0 = __p0; \
54868  float32x2_t __s1 = __p1; \
54869  float32x2_t __ret; \
54870  __ret = vmulx_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54871  __ret; \
54872})
54873#else
54874#define vmulx_lane_f32(__p0, __p1, __p2) __extension__ ({ \
54875  float32x2_t __s0 = __p0; \
54876  float32x2_t __s1 = __p1; \
54877  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54878  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54879  float32x2_t __ret; \
54880  __ret = __noswap_vmulx_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54881  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54882  __ret; \
54883})
54884#endif
54885
54886#ifdef __LITTLE_ENDIAN__
54887#define vmulxd_laneq_f64(__p0_142, __p1_142, __p2_142) __extension__ ({ \
54888  float64_t __s0_142 = __p0_142; \
54889  float64x2_t __s1_142 = __p1_142; \
54890  float64_t __ret_142; \
54891  __ret_142 = vmulxd_f64(__s0_142, vgetq_lane_f64(__s1_142, __p2_142)); \
54892  __ret_142; \
54893})
54894#else
54895#define vmulxd_laneq_f64(__p0_143, __p1_143, __p2_143) __extension__ ({ \
54896  float64_t __s0_143 = __p0_143; \
54897  float64x2_t __s1_143 = __p1_143; \
54898  float64x2_t __rev1_143;  __rev1_143 = __builtin_shufflevector(__s1_143, __s1_143, 1, 0); \
54899  float64_t __ret_143; \
54900  __ret_143 = __noswap_vmulxd_f64(__s0_143, __noswap_vgetq_lane_f64(__rev1_143, __p2_143)); \
54901  __ret_143; \
54902})
54903#endif
54904
54905#ifdef __LITTLE_ENDIAN__
54906#define vmulxs_laneq_f32(__p0_144, __p1_144, __p2_144) __extension__ ({ \
54907  float32_t __s0_144 = __p0_144; \
54908  float32x4_t __s1_144 = __p1_144; \
54909  float32_t __ret_144; \
54910  __ret_144 = vmulxs_f32(__s0_144, vgetq_lane_f32(__s1_144, __p2_144)); \
54911  __ret_144; \
54912})
54913#else
54914#define vmulxs_laneq_f32(__p0_145, __p1_145, __p2_145) __extension__ ({ \
54915  float32_t __s0_145 = __p0_145; \
54916  float32x4_t __s1_145 = __p1_145; \
54917  float32x4_t __rev1_145;  __rev1_145 = __builtin_shufflevector(__s1_145, __s1_145, 3, 2, 1, 0); \
54918  float32_t __ret_145; \
54919  __ret_145 = __noswap_vmulxs_f32(__s0_145, __noswap_vgetq_lane_f32(__rev1_145, __p2_145)); \
54920  __ret_145; \
54921})
54922#endif
54923
54924#ifdef __LITTLE_ENDIAN__
54925#define vmulxq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
54926  float64x2_t __s0 = __p0; \
54927  float64x2_t __s1 = __p1; \
54928  float64x2_t __ret; \
54929  __ret = vmulxq_f64(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54930  __ret; \
54931})
54932#else
54933#define vmulxq_laneq_f64(__p0, __p1, __p2) __extension__ ({ \
54934  float64x2_t __s0 = __p0; \
54935  float64x2_t __s1 = __p1; \
54936  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54937  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
54938  float64x2_t __ret; \
54939  __ret = __noswap_vmulxq_f64(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54940  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54941  __ret; \
54942})
54943#endif
54944
54945#ifdef __LITTLE_ENDIAN__
54946#define vmulxq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54947  float32x4_t __s0 = __p0; \
54948  float32x4_t __s1 = __p1; \
54949  float32x4_t __ret; \
54950  __ret = vmulxq_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
54951  __ret; \
54952})
54953#else
54954#define vmulxq_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54955  float32x4_t __s0 = __p0; \
54956  float32x4_t __s1 = __p1; \
54957  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
54958  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54959  float32x4_t __ret; \
54960  __ret = __noswap_vmulxq_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
54961  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
54962  __ret; \
54963})
54964#endif
54965
54966#ifdef __LITTLE_ENDIAN__
54967#define vmulx_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54968  float32x2_t __s0 = __p0; \
54969  float32x4_t __s1 = __p1; \
54970  float32x2_t __ret; \
54971  __ret = vmulx_f32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
54972  __ret; \
54973})
54974#else
54975#define vmulx_laneq_f32(__p0, __p1, __p2) __extension__ ({ \
54976  float32x2_t __s0 = __p0; \
54977  float32x4_t __s1 = __p1; \
54978  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
54979  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
54980  float32x2_t __ret; \
54981  __ret = __noswap_vmulx_f32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
54982  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
54983  __ret; \
54984})
54985#endif
54986
54987#ifdef __LITTLE_ENDIAN__
54988__ai float64x2_t vnegq_f64(float64x2_t __p0) {
54989  float64x2_t __ret;
54990  __ret = -__p0;
54991  return __ret;
54992}
54993#else
54994__ai float64x2_t vnegq_f64(float64x2_t __p0) {
54995  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
54996  float64x2_t __ret;
54997  __ret = -__rev0;
54998  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
54999  return __ret;
55000}
55001#endif
55002
55003#ifdef __LITTLE_ENDIAN__
55004__ai int64x2_t vnegq_s64(int64x2_t __p0) {
55005  int64x2_t __ret;
55006  __ret = -__p0;
55007  return __ret;
55008}
55009#else
55010__ai int64x2_t vnegq_s64(int64x2_t __p0) {
55011  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55012  int64x2_t __ret;
55013  __ret = -__rev0;
55014  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55015  return __ret;
55016}
55017#endif
55018
55019#ifdef __LITTLE_ENDIAN__
55020__ai float64x1_t vneg_f64(float64x1_t __p0) {
55021  float64x1_t __ret;
55022  __ret = -__p0;
55023  return __ret;
55024}
55025#else
55026__ai float64x1_t vneg_f64(float64x1_t __p0) {
55027  float64x1_t __ret;
55028  __ret = -__p0;
55029  return __ret;
55030}
55031#endif
55032
55033#ifdef __LITTLE_ENDIAN__
55034__ai int64x1_t vneg_s64(int64x1_t __p0) {
55035  int64x1_t __ret;
55036  __ret = -__p0;
55037  return __ret;
55038}
55039#else
55040__ai int64x1_t vneg_s64(int64x1_t __p0) {
55041  int64x1_t __ret;
55042  __ret = -__p0;
55043  return __ret;
55044}
55045#endif
55046
55047#ifdef __LITTLE_ENDIAN__
55048__ai int64_t vnegd_s64(int64_t __p0) {
55049  int64_t __ret;
55050  __ret = (int64_t) __builtin_neon_vnegd_s64(__p0);
55051  return __ret;
55052}
55053#else
55054__ai int64_t vnegd_s64(int64_t __p0) {
55055  int64_t __ret;
55056  __ret = (int64_t) __builtin_neon_vnegd_s64(__p0);
55057  return __ret;
55058}
55059#endif
55060
55061#ifdef __LITTLE_ENDIAN__
55062__ai uint8x16_t vpaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55063  uint8x16_t __ret;
55064  __ret = (uint8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
55065  return __ret;
55066}
55067#else
55068__ai uint8x16_t vpaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55069  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55070  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55071  uint8x16_t __ret;
55072  __ret = (uint8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
55073  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55074  return __ret;
55075}
55076#endif
55077
55078#ifdef __LITTLE_ENDIAN__
55079__ai uint32x4_t vpaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55080  uint32x4_t __ret;
55081  __ret = (uint32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
55082  return __ret;
55083}
55084#else
55085__ai uint32x4_t vpaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55086  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55087  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55088  uint32x4_t __ret;
55089  __ret = (uint32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
55090  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55091  return __ret;
55092}
55093#endif
55094
55095#ifdef __LITTLE_ENDIAN__
55096__ai uint64x2_t vpaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
55097  uint64x2_t __ret;
55098  __ret = (uint64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
55099  return __ret;
55100}
55101#else
55102__ai uint64x2_t vpaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
55103  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55104  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55105  uint64x2_t __ret;
55106  __ret = (uint64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
55107  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55108  return __ret;
55109}
55110#endif
55111
55112#ifdef __LITTLE_ENDIAN__
55113__ai uint16x8_t vpaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55114  uint16x8_t __ret;
55115  __ret = (uint16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
55116  return __ret;
55117}
55118#else
55119__ai uint16x8_t vpaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55120  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55121  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55122  uint16x8_t __ret;
55123  __ret = (uint16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
55124  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55125  return __ret;
55126}
55127#endif
55128
55129#ifdef __LITTLE_ENDIAN__
55130__ai int8x16_t vpaddq_s8(int8x16_t __p0, int8x16_t __p1) {
55131  int8x16_t __ret;
55132  __ret = (int8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
55133  return __ret;
55134}
55135#else
55136__ai int8x16_t vpaddq_s8(int8x16_t __p0, int8x16_t __p1) {
55137  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55138  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55139  int8x16_t __ret;
55140  __ret = (int8x16_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
55141  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55142  return __ret;
55143}
55144#endif
55145
55146#ifdef __LITTLE_ENDIAN__
55147__ai float64x2_t vpaddq_f64(float64x2_t __p0, float64x2_t __p1) {
55148  float64x2_t __ret;
55149  __ret = (float64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55150  return __ret;
55151}
55152#else
55153__ai float64x2_t vpaddq_f64(float64x2_t __p0, float64x2_t __p1) {
55154  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55155  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55156  float64x2_t __ret;
55157  __ret = (float64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55158  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55159  return __ret;
55160}
55161#endif
55162
55163#ifdef __LITTLE_ENDIAN__
55164__ai float32x4_t vpaddq_f32(float32x4_t __p0, float32x4_t __p1) {
55165  float32x4_t __ret;
55166  __ret = (float32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55167  return __ret;
55168}
55169#else
55170__ai float32x4_t vpaddq_f32(float32x4_t __p0, float32x4_t __p1) {
55171  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55172  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55173  float32x4_t __ret;
55174  __ret = (float32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55175  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55176  return __ret;
55177}
55178#endif
55179
55180#ifdef __LITTLE_ENDIAN__
55181__ai int32x4_t vpaddq_s32(int32x4_t __p0, int32x4_t __p1) {
55182  int32x4_t __ret;
55183  __ret = (int32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
55184  return __ret;
55185}
55186#else
55187__ai int32x4_t vpaddq_s32(int32x4_t __p0, int32x4_t __p1) {
55188  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55189  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55190  int32x4_t __ret;
55191  __ret = (int32x4_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
55192  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55193  return __ret;
55194}
55195#endif
55196
55197#ifdef __LITTLE_ENDIAN__
55198__ai int64x2_t vpaddq_s64(int64x2_t __p0, int64x2_t __p1) {
55199  int64x2_t __ret;
55200  __ret = (int64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
55201  return __ret;
55202}
55203#else
55204__ai int64x2_t vpaddq_s64(int64x2_t __p0, int64x2_t __p1) {
55205  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55206  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55207  int64x2_t __ret;
55208  __ret = (int64x2_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
55209  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55210  return __ret;
55211}
55212#endif
55213
55214#ifdef __LITTLE_ENDIAN__
55215__ai int16x8_t vpaddq_s16(int16x8_t __p0, int16x8_t __p1) {
55216  int16x8_t __ret;
55217  __ret = (int16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
55218  return __ret;
55219}
55220#else
55221__ai int16x8_t vpaddq_s16(int16x8_t __p0, int16x8_t __p1) {
55222  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55223  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55224  int16x8_t __ret;
55225  __ret = (int16x8_t) __builtin_neon_vpaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
55226  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55227  return __ret;
55228}
55229#endif
55230
55231#ifdef __LITTLE_ENDIAN__
55232__ai uint64_t vpaddd_u64(uint64x2_t __p0) {
55233  uint64_t __ret;
55234  __ret = (uint64_t) __builtin_neon_vpaddd_u64((int8x16_t)__p0);
55235  return __ret;
55236}
55237#else
55238__ai uint64_t vpaddd_u64(uint64x2_t __p0) {
55239  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55240  uint64_t __ret;
55241  __ret = (uint64_t) __builtin_neon_vpaddd_u64((int8x16_t)__rev0);
55242  return __ret;
55243}
55244#endif
55245
55246#ifdef __LITTLE_ENDIAN__
55247__ai float64_t vpaddd_f64(float64x2_t __p0) {
55248  float64_t __ret;
55249  __ret = (float64_t) __builtin_neon_vpaddd_f64((int8x16_t)__p0);
55250  return __ret;
55251}
55252#else
55253__ai float64_t vpaddd_f64(float64x2_t __p0) {
55254  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55255  float64_t __ret;
55256  __ret = (float64_t) __builtin_neon_vpaddd_f64((int8x16_t)__rev0);
55257  return __ret;
55258}
55259#endif
55260
55261#ifdef __LITTLE_ENDIAN__
55262__ai int64_t vpaddd_s64(int64x2_t __p0) {
55263  int64_t __ret;
55264  __ret = (int64_t) __builtin_neon_vpaddd_s64((int8x16_t)__p0);
55265  return __ret;
55266}
55267#else
55268__ai int64_t vpaddd_s64(int64x2_t __p0) {
55269  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55270  int64_t __ret;
55271  __ret = (int64_t) __builtin_neon_vpaddd_s64((int8x16_t)__rev0);
55272  return __ret;
55273}
55274#endif
55275
55276#ifdef __LITTLE_ENDIAN__
55277__ai float32_t vpadds_f32(float32x2_t __p0) {
55278  float32_t __ret;
55279  __ret = (float32_t) __builtin_neon_vpadds_f32((int8x8_t)__p0);
55280  return __ret;
55281}
55282#else
55283__ai float32_t vpadds_f32(float32x2_t __p0) {
55284  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55285  float32_t __ret;
55286  __ret = (float32_t) __builtin_neon_vpadds_f32((int8x8_t)__rev0);
55287  return __ret;
55288}
55289#endif
55290
55291#ifdef __LITTLE_ENDIAN__
55292__ai uint8x16_t vpmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55293  uint8x16_t __ret;
55294  __ret = (uint8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
55295  return __ret;
55296}
55297#else
55298__ai uint8x16_t vpmaxq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55299  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55300  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55301  uint8x16_t __ret;
55302  __ret = (uint8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
55303  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55304  return __ret;
55305}
55306#endif
55307
55308#ifdef __LITTLE_ENDIAN__
55309__ai uint32x4_t vpmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55310  uint32x4_t __ret;
55311  __ret = (uint32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
55312  return __ret;
55313}
55314#else
55315__ai uint32x4_t vpmaxq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55316  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55317  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55318  uint32x4_t __ret;
55319  __ret = (uint32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
55320  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55321  return __ret;
55322}
55323#endif
55324
55325#ifdef __LITTLE_ENDIAN__
55326__ai uint16x8_t vpmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55327  uint16x8_t __ret;
55328  __ret = (uint16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
55329  return __ret;
55330}
55331#else
55332__ai uint16x8_t vpmaxq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55333  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55334  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55335  uint16x8_t __ret;
55336  __ret = (uint16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
55337  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55338  return __ret;
55339}
55340#endif
55341
55342#ifdef __LITTLE_ENDIAN__
55343__ai int8x16_t vpmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
55344  int8x16_t __ret;
55345  __ret = (int8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
55346  return __ret;
55347}
55348#else
55349__ai int8x16_t vpmaxq_s8(int8x16_t __p0, int8x16_t __p1) {
55350  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55351  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55352  int8x16_t __ret;
55353  __ret = (int8x16_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
55354  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55355  return __ret;
55356}
55357#endif
55358
55359#ifdef __LITTLE_ENDIAN__
55360__ai float64x2_t vpmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
55361  float64x2_t __ret;
55362  __ret = (float64x2_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55363  return __ret;
55364}
55365#else
55366__ai float64x2_t vpmaxq_f64(float64x2_t __p0, float64x2_t __p1) {
55367  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55368  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55369  float64x2_t __ret;
55370  __ret = (float64x2_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55371  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55372  return __ret;
55373}
55374#endif
55375
55376#ifdef __LITTLE_ENDIAN__
55377__ai float32x4_t vpmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
55378  float32x4_t __ret;
55379  __ret = (float32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55380  return __ret;
55381}
55382#else
55383__ai float32x4_t vpmaxq_f32(float32x4_t __p0, float32x4_t __p1) {
55384  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55385  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55386  float32x4_t __ret;
55387  __ret = (float32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55388  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55389  return __ret;
55390}
55391#endif
55392
55393#ifdef __LITTLE_ENDIAN__
55394__ai int32x4_t vpmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
55395  int32x4_t __ret;
55396  __ret = (int32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
55397  return __ret;
55398}
55399#else
55400__ai int32x4_t vpmaxq_s32(int32x4_t __p0, int32x4_t __p1) {
55401  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55402  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55403  int32x4_t __ret;
55404  __ret = (int32x4_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
55405  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55406  return __ret;
55407}
55408#endif
55409
55410#ifdef __LITTLE_ENDIAN__
55411__ai int16x8_t vpmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
55412  int16x8_t __ret;
55413  __ret = (int16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
55414  return __ret;
55415}
55416#else
55417__ai int16x8_t vpmaxq_s16(int16x8_t __p0, int16x8_t __p1) {
55418  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55419  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55420  int16x8_t __ret;
55421  __ret = (int16x8_t) __builtin_neon_vpmaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
55422  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55423  return __ret;
55424}
55425#endif
55426
55427#ifdef __LITTLE_ENDIAN__
55428__ai float64_t vpmaxqd_f64(float64x2_t __p0) {
55429  float64_t __ret;
55430  __ret = (float64_t) __builtin_neon_vpmaxqd_f64((int8x16_t)__p0);
55431  return __ret;
55432}
55433#else
55434__ai float64_t vpmaxqd_f64(float64x2_t __p0) {
55435  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55436  float64_t __ret;
55437  __ret = (float64_t) __builtin_neon_vpmaxqd_f64((int8x16_t)__rev0);
55438  return __ret;
55439}
55440#endif
55441
55442#ifdef __LITTLE_ENDIAN__
55443__ai float32_t vpmaxs_f32(float32x2_t __p0) {
55444  float32_t __ret;
55445  __ret = (float32_t) __builtin_neon_vpmaxs_f32((int8x8_t)__p0);
55446  return __ret;
55447}
55448#else
55449__ai float32_t vpmaxs_f32(float32x2_t __p0) {
55450  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55451  float32_t __ret;
55452  __ret = (float32_t) __builtin_neon_vpmaxs_f32((int8x8_t)__rev0);
55453  return __ret;
55454}
55455#endif
55456
55457#ifdef __LITTLE_ENDIAN__
55458__ai float64x2_t vpmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55459  float64x2_t __ret;
55460  __ret = (float64x2_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55461  return __ret;
55462}
55463#else
55464__ai float64x2_t vpmaxnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55465  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55466  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55467  float64x2_t __ret;
55468  __ret = (float64x2_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55469  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55470  return __ret;
55471}
55472#endif
55473
55474#ifdef __LITTLE_ENDIAN__
55475__ai float32x4_t vpmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55476  float32x4_t __ret;
55477  __ret = (float32x4_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55478  return __ret;
55479}
55480#else
55481__ai float32x4_t vpmaxnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55482  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55483  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55484  float32x4_t __ret;
55485  __ret = (float32x4_t) __builtin_neon_vpmaxnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55486  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55487  return __ret;
55488}
55489#endif
55490
55491#ifdef __LITTLE_ENDIAN__
55492__ai float32x2_t vpmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
55493  float32x2_t __ret;
55494  __ret = (float32x2_t) __builtin_neon_vpmaxnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
55495  return __ret;
55496}
55497#else
55498__ai float32x2_t vpmaxnm_f32(float32x2_t __p0, float32x2_t __p1) {
55499  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55500  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55501  float32x2_t __ret;
55502  __ret = (float32x2_t) __builtin_neon_vpmaxnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
55503  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55504  return __ret;
55505}
55506#endif
55507
55508#ifdef __LITTLE_ENDIAN__
55509__ai float64_t vpmaxnmqd_f64(float64x2_t __p0) {
55510  float64_t __ret;
55511  __ret = (float64_t) __builtin_neon_vpmaxnmqd_f64((int8x16_t)__p0);
55512  return __ret;
55513}
55514#else
55515__ai float64_t vpmaxnmqd_f64(float64x2_t __p0) {
55516  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55517  float64_t __ret;
55518  __ret = (float64_t) __builtin_neon_vpmaxnmqd_f64((int8x16_t)__rev0);
55519  return __ret;
55520}
55521#endif
55522
55523#ifdef __LITTLE_ENDIAN__
55524__ai float32_t vpmaxnms_f32(float32x2_t __p0) {
55525  float32_t __ret;
55526  __ret = (float32_t) __builtin_neon_vpmaxnms_f32((int8x8_t)__p0);
55527  return __ret;
55528}
55529#else
55530__ai float32_t vpmaxnms_f32(float32x2_t __p0) {
55531  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55532  float32_t __ret;
55533  __ret = (float32_t) __builtin_neon_vpmaxnms_f32((int8x8_t)__rev0);
55534  return __ret;
55535}
55536#endif
55537
55538#ifdef __LITTLE_ENDIAN__
55539__ai uint8x16_t vpminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55540  uint8x16_t __ret;
55541  __ret = (uint8x16_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
55542  return __ret;
55543}
55544#else
55545__ai uint8x16_t vpminq_u8(uint8x16_t __p0, uint8x16_t __p1) {
55546  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55547  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55548  uint8x16_t __ret;
55549  __ret = (uint8x16_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
55550  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55551  return __ret;
55552}
55553#endif
55554
55555#ifdef __LITTLE_ENDIAN__
55556__ai uint32x4_t vpminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55557  uint32x4_t __ret;
55558  __ret = (uint32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
55559  return __ret;
55560}
55561#else
55562__ai uint32x4_t vpminq_u32(uint32x4_t __p0, uint32x4_t __p1) {
55563  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55564  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55565  uint32x4_t __ret;
55566  __ret = (uint32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
55567  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55568  return __ret;
55569}
55570#endif
55571
55572#ifdef __LITTLE_ENDIAN__
55573__ai uint16x8_t vpminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55574  uint16x8_t __ret;
55575  __ret = (uint16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
55576  return __ret;
55577}
55578#else
55579__ai uint16x8_t vpminq_u16(uint16x8_t __p0, uint16x8_t __p1) {
55580  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55581  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55582  uint16x8_t __ret;
55583  __ret = (uint16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
55584  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55585  return __ret;
55586}
55587#endif
55588
55589#ifdef __LITTLE_ENDIAN__
55590__ai int8x16_t vpminq_s8(int8x16_t __p0, int8x16_t __p1) {
55591  int8x16_t __ret;
55592  __ret = (int8x16_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
55593  return __ret;
55594}
55595#else
55596__ai int8x16_t vpminq_s8(int8x16_t __p0, int8x16_t __p1) {
55597  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55598  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55599  int8x16_t __ret;
55600  __ret = (int8x16_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
55601  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
55602  return __ret;
55603}
55604#endif
55605
55606#ifdef __LITTLE_ENDIAN__
55607__ai float64x2_t vpminq_f64(float64x2_t __p0, float64x2_t __p1) {
55608  float64x2_t __ret;
55609  __ret = (float64x2_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55610  return __ret;
55611}
55612#else
55613__ai float64x2_t vpminq_f64(float64x2_t __p0, float64x2_t __p1) {
55614  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55615  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55616  float64x2_t __ret;
55617  __ret = (float64x2_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55618  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55619  return __ret;
55620}
55621#endif
55622
55623#ifdef __LITTLE_ENDIAN__
55624__ai float32x4_t vpminq_f32(float32x4_t __p0, float32x4_t __p1) {
55625  float32x4_t __ret;
55626  __ret = (float32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55627  return __ret;
55628}
55629#else
55630__ai float32x4_t vpminq_f32(float32x4_t __p0, float32x4_t __p1) {
55631  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55632  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55633  float32x4_t __ret;
55634  __ret = (float32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55635  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55636  return __ret;
55637}
55638#endif
55639
55640#ifdef __LITTLE_ENDIAN__
55641__ai int32x4_t vpminq_s32(int32x4_t __p0, int32x4_t __p1) {
55642  int32x4_t __ret;
55643  __ret = (int32x4_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
55644  return __ret;
55645}
55646#else
55647__ai int32x4_t vpminq_s32(int32x4_t __p0, int32x4_t __p1) {
55648  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55649  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55650  int32x4_t __ret;
55651  __ret = (int32x4_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
55652  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55653  return __ret;
55654}
55655#endif
55656
55657#ifdef __LITTLE_ENDIAN__
55658__ai int16x8_t vpminq_s16(int16x8_t __p0, int16x8_t __p1) {
55659  int16x8_t __ret;
55660  __ret = (int16x8_t) __builtin_neon_vpminq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
55661  return __ret;
55662}
55663#else
55664__ai int16x8_t vpminq_s16(int16x8_t __p0, int16x8_t __p1) {
55665  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
55666  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
55667  int16x8_t __ret;
55668  __ret = (int16x8_t) __builtin_neon_vpminq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
55669  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
55670  return __ret;
55671}
55672#endif
55673
55674#ifdef __LITTLE_ENDIAN__
55675__ai float64_t vpminqd_f64(float64x2_t __p0) {
55676  float64_t __ret;
55677  __ret = (float64_t) __builtin_neon_vpminqd_f64((int8x16_t)__p0);
55678  return __ret;
55679}
55680#else
55681__ai float64_t vpminqd_f64(float64x2_t __p0) {
55682  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55683  float64_t __ret;
55684  __ret = (float64_t) __builtin_neon_vpminqd_f64((int8x16_t)__rev0);
55685  return __ret;
55686}
55687#endif
55688
55689#ifdef __LITTLE_ENDIAN__
55690__ai float32_t vpmins_f32(float32x2_t __p0) {
55691  float32_t __ret;
55692  __ret = (float32_t) __builtin_neon_vpmins_f32((int8x8_t)__p0);
55693  return __ret;
55694}
55695#else
55696__ai float32_t vpmins_f32(float32x2_t __p0) {
55697  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55698  float32_t __ret;
55699  __ret = (float32_t) __builtin_neon_vpmins_f32((int8x8_t)__rev0);
55700  return __ret;
55701}
55702#endif
55703
55704#ifdef __LITTLE_ENDIAN__
55705__ai float64x2_t vpminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55706  float64x2_t __ret;
55707  __ret = (float64x2_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
55708  return __ret;
55709}
55710#else
55711__ai float64x2_t vpminnmq_f64(float64x2_t __p0, float64x2_t __p1) {
55712  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55713  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55714  float64x2_t __ret;
55715  __ret = (float64x2_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
55716  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55717  return __ret;
55718}
55719#endif
55720
55721#ifdef __LITTLE_ENDIAN__
55722__ai float32x4_t vpminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55723  float32x4_t __ret;
55724  __ret = (float32x4_t) __builtin_neon_vpminnmq_v((int8x16_t)__p0, (int8x16_t)__p1, 41);
55725  return __ret;
55726}
55727#else
55728__ai float32x4_t vpminnmq_f32(float32x4_t __p0, float32x4_t __p1) {
55729  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
55730  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
55731  float32x4_t __ret;
55732  __ret = (float32x4_t) __builtin_neon_vpminnmq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 41);
55733  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
55734  return __ret;
55735}
55736#endif
55737
55738#ifdef __LITTLE_ENDIAN__
55739__ai float32x2_t vpminnm_f32(float32x2_t __p0, float32x2_t __p1) {
55740  float32x2_t __ret;
55741  __ret = (float32x2_t) __builtin_neon_vpminnm_v((int8x8_t)__p0, (int8x8_t)__p1, 9);
55742  return __ret;
55743}
55744#else
55745__ai float32x2_t vpminnm_f32(float32x2_t __p0, float32x2_t __p1) {
55746  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55747  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
55748  float32x2_t __ret;
55749  __ret = (float32x2_t) __builtin_neon_vpminnm_v((int8x8_t)__rev0, (int8x8_t)__rev1, 9);
55750  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55751  return __ret;
55752}
55753#endif
55754
55755#ifdef __LITTLE_ENDIAN__
55756__ai float64_t vpminnmqd_f64(float64x2_t __p0) {
55757  float64_t __ret;
55758  __ret = (float64_t) __builtin_neon_vpminnmqd_f64((int8x16_t)__p0);
55759  return __ret;
55760}
55761#else
55762__ai float64_t vpminnmqd_f64(float64x2_t __p0) {
55763  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55764  float64_t __ret;
55765  __ret = (float64_t) __builtin_neon_vpminnmqd_f64((int8x16_t)__rev0);
55766  return __ret;
55767}
55768#endif
55769
55770#ifdef __LITTLE_ENDIAN__
55771__ai float32_t vpminnms_f32(float32x2_t __p0) {
55772  float32_t __ret;
55773  __ret = (float32_t) __builtin_neon_vpminnms_f32((int8x8_t)__p0);
55774  return __ret;
55775}
55776#else
55777__ai float32_t vpminnms_f32(float32x2_t __p0) {
55778  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55779  float32_t __ret;
55780  __ret = (float32_t) __builtin_neon_vpminnms_f32((int8x8_t)__rev0);
55781  return __ret;
55782}
55783#endif
55784
55785#ifdef __LITTLE_ENDIAN__
55786__ai int64x2_t vqabsq_s64(int64x2_t __p0) {
55787  int64x2_t __ret;
55788  __ret = (int64x2_t) __builtin_neon_vqabsq_v((int8x16_t)__p0, 35);
55789  return __ret;
55790}
55791#else
55792__ai int64x2_t vqabsq_s64(int64x2_t __p0) {
55793  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
55794  int64x2_t __ret;
55795  __ret = (int64x2_t) __builtin_neon_vqabsq_v((int8x16_t)__rev0, 35);
55796  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
55797  return __ret;
55798}
55799#endif
55800
55801#ifdef __LITTLE_ENDIAN__
55802__ai int64x1_t vqabs_s64(int64x1_t __p0) {
55803  int64x1_t __ret;
55804  __ret = (int64x1_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 3);
55805  return __ret;
55806}
55807#else
55808__ai int64x1_t vqabs_s64(int64x1_t __p0) {
55809  int64x1_t __ret;
55810  __ret = (int64x1_t) __builtin_neon_vqabs_v((int8x8_t)__p0, 3);
55811  return __ret;
55812}
55813#endif
55814
55815#ifdef __LITTLE_ENDIAN__
55816__ai int8_t vqabsb_s8(int8_t __p0) {
55817  int8_t __ret;
55818  __ret = (int8_t) __builtin_neon_vqabsb_s8(__p0);
55819  return __ret;
55820}
55821#else
55822__ai int8_t vqabsb_s8(int8_t __p0) {
55823  int8_t __ret;
55824  __ret = (int8_t) __builtin_neon_vqabsb_s8(__p0);
55825  return __ret;
55826}
55827#endif
55828
55829#ifdef __LITTLE_ENDIAN__
55830__ai int32_t vqabss_s32(int32_t __p0) {
55831  int32_t __ret;
55832  __ret = (int32_t) __builtin_neon_vqabss_s32(__p0);
55833  return __ret;
55834}
55835#else
55836__ai int32_t vqabss_s32(int32_t __p0) {
55837  int32_t __ret;
55838  __ret = (int32_t) __builtin_neon_vqabss_s32(__p0);
55839  return __ret;
55840}
55841#endif
55842
55843#ifdef __LITTLE_ENDIAN__
55844__ai int64_t vqabsd_s64(int64_t __p0) {
55845  int64_t __ret;
55846  __ret = (int64_t) __builtin_neon_vqabsd_s64(__p0);
55847  return __ret;
55848}
55849#else
55850__ai int64_t vqabsd_s64(int64_t __p0) {
55851  int64_t __ret;
55852  __ret = (int64_t) __builtin_neon_vqabsd_s64(__p0);
55853  return __ret;
55854}
55855#endif
55856
55857#ifdef __LITTLE_ENDIAN__
55858__ai int16_t vqabsh_s16(int16_t __p0) {
55859  int16_t __ret;
55860  __ret = (int16_t) __builtin_neon_vqabsh_s16(__p0);
55861  return __ret;
55862}
55863#else
55864__ai int16_t vqabsh_s16(int16_t __p0) {
55865  int16_t __ret;
55866  __ret = (int16_t) __builtin_neon_vqabsh_s16(__p0);
55867  return __ret;
55868}
55869#endif
55870
55871#ifdef __LITTLE_ENDIAN__
55872__ai uint8_t vqaddb_u8(uint8_t __p0, uint8_t __p1) {
55873  uint8_t __ret;
55874  __ret = (uint8_t) __builtin_neon_vqaddb_u8(__p0, __p1);
55875  return __ret;
55876}
55877#else
55878__ai uint8_t vqaddb_u8(uint8_t __p0, uint8_t __p1) {
55879  uint8_t __ret;
55880  __ret = (uint8_t) __builtin_neon_vqaddb_u8(__p0, __p1);
55881  return __ret;
55882}
55883#endif
55884
55885#ifdef __LITTLE_ENDIAN__
55886__ai uint32_t vqadds_u32(uint32_t __p0, uint32_t __p1) {
55887  uint32_t __ret;
55888  __ret = (uint32_t) __builtin_neon_vqadds_u32(__p0, __p1);
55889  return __ret;
55890}
55891#else
55892__ai uint32_t vqadds_u32(uint32_t __p0, uint32_t __p1) {
55893  uint32_t __ret;
55894  __ret = (uint32_t) __builtin_neon_vqadds_u32(__p0, __p1);
55895  return __ret;
55896}
55897#endif
55898
55899#ifdef __LITTLE_ENDIAN__
55900__ai uint64_t vqaddd_u64(uint64_t __p0, uint64_t __p1) {
55901  uint64_t __ret;
55902  __ret = (uint64_t) __builtin_neon_vqaddd_u64(__p0, __p1);
55903  return __ret;
55904}
55905#else
55906__ai uint64_t vqaddd_u64(uint64_t __p0, uint64_t __p1) {
55907  uint64_t __ret;
55908  __ret = (uint64_t) __builtin_neon_vqaddd_u64(__p0, __p1);
55909  return __ret;
55910}
55911#endif
55912
55913#ifdef __LITTLE_ENDIAN__
55914__ai uint16_t vqaddh_u16(uint16_t __p0, uint16_t __p1) {
55915  uint16_t __ret;
55916  __ret = (uint16_t) __builtin_neon_vqaddh_u16(__p0, __p1);
55917  return __ret;
55918}
55919#else
55920__ai uint16_t vqaddh_u16(uint16_t __p0, uint16_t __p1) {
55921  uint16_t __ret;
55922  __ret = (uint16_t) __builtin_neon_vqaddh_u16(__p0, __p1);
55923  return __ret;
55924}
55925#endif
55926
55927#ifdef __LITTLE_ENDIAN__
55928__ai int8_t vqaddb_s8(int8_t __p0, int8_t __p1) {
55929  int8_t __ret;
55930  __ret = (int8_t) __builtin_neon_vqaddb_s8(__p0, __p1);
55931  return __ret;
55932}
55933#else
55934__ai int8_t vqaddb_s8(int8_t __p0, int8_t __p1) {
55935  int8_t __ret;
55936  __ret = (int8_t) __builtin_neon_vqaddb_s8(__p0, __p1);
55937  return __ret;
55938}
55939#endif
55940
55941#ifdef __LITTLE_ENDIAN__
55942__ai int32_t vqadds_s32(int32_t __p0, int32_t __p1) {
55943  int32_t __ret;
55944  __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
55945  return __ret;
55946}
55947#else
55948__ai int32_t vqadds_s32(int32_t __p0, int32_t __p1) {
55949  int32_t __ret;
55950  __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
55951  return __ret;
55952}
55953__ai int32_t __noswap_vqadds_s32(int32_t __p0, int32_t __p1) {
55954  int32_t __ret;
55955  __ret = (int32_t) __builtin_neon_vqadds_s32(__p0, __p1);
55956  return __ret;
55957}
55958#endif
55959
55960#ifdef __LITTLE_ENDIAN__
55961__ai int64_t vqaddd_s64(int64_t __p0, int64_t __p1) {
55962  int64_t __ret;
55963  __ret = (int64_t) __builtin_neon_vqaddd_s64(__p0, __p1);
55964  return __ret;
55965}
55966#else
55967__ai int64_t vqaddd_s64(int64_t __p0, int64_t __p1) {
55968  int64_t __ret;
55969  __ret = (int64_t) __builtin_neon_vqaddd_s64(__p0, __p1);
55970  return __ret;
55971}
55972#endif
55973
55974#ifdef __LITTLE_ENDIAN__
55975__ai int16_t vqaddh_s16(int16_t __p0, int16_t __p1) {
55976  int16_t __ret;
55977  __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
55978  return __ret;
55979}
55980#else
55981__ai int16_t vqaddh_s16(int16_t __p0, int16_t __p1) {
55982  int16_t __ret;
55983  __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
55984  return __ret;
55985}
55986__ai int16_t __noswap_vqaddh_s16(int16_t __p0, int16_t __p1) {
55987  int16_t __ret;
55988  __ret = (int16_t) __builtin_neon_vqaddh_s16(__p0, __p1);
55989  return __ret;
55990}
55991#endif
55992
55993#ifdef __LITTLE_ENDIAN__
55994__ai int64_t vqdmlals_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
55995  int64_t __ret;
55996  __ret = (int64_t) __builtin_neon_vqdmlals_s32(__p0, __p1, __p2);
55997  return __ret;
55998}
55999#else
56000__ai int64_t vqdmlals_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
56001  int64_t __ret;
56002  __ret = (int64_t) __builtin_neon_vqdmlals_s32(__p0, __p1, __p2);
56003  return __ret;
56004}
56005#endif
56006
56007#ifdef __LITTLE_ENDIAN__
56008__ai int32_t vqdmlalh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
56009  int32_t __ret;
56010  __ret = (int32_t) __builtin_neon_vqdmlalh_s16(__p0, __p1, __p2);
56011  return __ret;
56012}
56013#else
56014__ai int32_t vqdmlalh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
56015  int32_t __ret;
56016  __ret = (int32_t) __builtin_neon_vqdmlalh_s16(__p0, __p1, __p2);
56017  return __ret;
56018}
56019#endif
56020
56021#ifdef __LITTLE_ENDIAN__
56022__ai int64x2_t vqdmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
56023  int64x2_t __ret;
56024  __ret = vqdmlal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
56025  return __ret;
56026}
56027#else
56028__ai int64x2_t vqdmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
56029  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56030  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56031  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
56032  int64x2_t __ret;
56033  __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
56034  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56035  return __ret;
56036}
56037#endif
56038
56039#ifdef __LITTLE_ENDIAN__
56040__ai int32x4_t vqdmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
56041  int32x4_t __ret;
56042  __ret = vqdmlal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
56043  return __ret;
56044}
56045#else
56046__ai int32x4_t vqdmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
56047  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56048  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56049  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
56050  int32x4_t __ret;
56051  __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
56052  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56053  return __ret;
56054}
56055#endif
56056
56057#ifdef __LITTLE_ENDIAN__
56058#define vqdmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56059  int64x2_t __s0 = __p0; \
56060  int32x4_t __s1 = __p1; \
56061  int32x2_t __s2 = __p2; \
56062  int64x2_t __ret; \
56063  __ret = vqdmlal_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56064  __ret; \
56065})
56066#else
56067#define vqdmlal_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56068  int64x2_t __s0 = __p0; \
56069  int32x4_t __s1 = __p1; \
56070  int32x2_t __s2 = __p2; \
56071  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56072  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56073  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
56074  int64x2_t __ret; \
56075  __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56076  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56077  __ret; \
56078})
56079#endif
56080
56081#ifdef __LITTLE_ENDIAN__
56082#define vqdmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56083  int32x4_t __s0 = __p0; \
56084  int16x8_t __s1 = __p1; \
56085  int16x4_t __s2 = __p2; \
56086  int32x4_t __ret; \
56087  __ret = vqdmlal_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56088  __ret; \
56089})
56090#else
56091#define vqdmlal_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56092  int32x4_t __s0 = __p0; \
56093  int16x8_t __s1 = __p1; \
56094  int16x4_t __s2 = __p2; \
56095  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56096  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56097  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56098  int32x4_t __ret; \
56099  __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56100  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56101  __ret; \
56102})
56103#endif
56104
56105#ifdef __LITTLE_ENDIAN__
56106#define vqdmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56107  int64x2_t __s0 = __p0; \
56108  int32x4_t __s1 = __p1; \
56109  int32x4_t __s2 = __p2; \
56110  int64x2_t __ret; \
56111  __ret = vqdmlal_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56112  __ret; \
56113})
56114#else
56115#define vqdmlal_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56116  int64x2_t __s0 = __p0; \
56117  int32x4_t __s1 = __p1; \
56118  int32x4_t __s2 = __p2; \
56119  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56120  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56121  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56122  int64x2_t __ret; \
56123  __ret = __noswap_vqdmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56124  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56125  __ret; \
56126})
56127#endif
56128
56129#ifdef __LITTLE_ENDIAN__
56130#define vqdmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56131  int32x4_t __s0 = __p0; \
56132  int16x8_t __s1 = __p1; \
56133  int16x8_t __s2 = __p2; \
56134  int32x4_t __ret; \
56135  __ret = vqdmlal_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56136  __ret; \
56137})
56138#else
56139#define vqdmlal_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56140  int32x4_t __s0 = __p0; \
56141  int16x8_t __s1 = __p1; \
56142  int16x8_t __s2 = __p2; \
56143  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56144  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56145  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56146  int32x4_t __ret; \
56147  __ret = __noswap_vqdmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56148  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56149  __ret; \
56150})
56151#endif
56152
56153#ifdef __LITTLE_ENDIAN__
56154__ai int64x2_t vqdmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56155  int64x2_t __ret;
56156  __ret = vqdmlal_n_s32(__p0, vget_high_s32(__p1), __p2);
56157  return __ret;
56158}
56159#else
56160__ai int64x2_t vqdmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56161  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56162  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56163  int64x2_t __ret;
56164  __ret = __noswap_vqdmlal_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
56165  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56166  return __ret;
56167}
56168#endif
56169
56170#ifdef __LITTLE_ENDIAN__
56171__ai int32x4_t vqdmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56172  int32x4_t __ret;
56173  __ret = vqdmlal_n_s16(__p0, vget_high_s16(__p1), __p2);
56174  return __ret;
56175}
56176#else
56177__ai int32x4_t vqdmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56178  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56179  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56180  int32x4_t __ret;
56181  __ret = __noswap_vqdmlal_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
56182  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56183  return __ret;
56184}
56185#endif
56186
56187#ifdef __LITTLE_ENDIAN__
56188#define vqdmlals_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56189  int64_t __s0 = __p0; \
56190  int32_t __s1 = __p1; \
56191  int32x2_t __s2 = __p2; \
56192  int64_t __ret; \
56193  __ret = (int64_t) __builtin_neon_vqdmlals_lane_s32(__s0, __s1, (int8x8_t)__s2, __p3); \
56194  __ret; \
56195})
56196#else
56197#define vqdmlals_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56198  int64_t __s0 = __p0; \
56199  int32_t __s1 = __p1; \
56200  int32x2_t __s2 = __p2; \
56201  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
56202  int64_t __ret; \
56203  __ret = (int64_t) __builtin_neon_vqdmlals_lane_s32(__s0, __s1, (int8x8_t)__rev2, __p3); \
56204  __ret; \
56205})
56206#endif
56207
56208#ifdef __LITTLE_ENDIAN__
56209#define vqdmlalh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56210  int32_t __s0 = __p0; \
56211  int16_t __s1 = __p1; \
56212  int16x4_t __s2 = __p2; \
56213  int32_t __ret; \
56214  __ret = (int32_t) __builtin_neon_vqdmlalh_lane_s16(__s0, __s1, (int8x8_t)__s2, __p3); \
56215  __ret; \
56216})
56217#else
56218#define vqdmlalh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56219  int32_t __s0 = __p0; \
56220  int16_t __s1 = __p1; \
56221  int16x4_t __s2 = __p2; \
56222  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56223  int32_t __ret; \
56224  __ret = (int32_t) __builtin_neon_vqdmlalh_lane_s16(__s0, __s1, (int8x8_t)__rev2, __p3); \
56225  __ret; \
56226})
56227#endif
56228
56229#ifdef __LITTLE_ENDIAN__
56230#define vqdmlals_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56231  int64_t __s0 = __p0; \
56232  int32_t __s1 = __p1; \
56233  int32x4_t __s2 = __p2; \
56234  int64_t __ret; \
56235  __ret = (int64_t) __builtin_neon_vqdmlals_laneq_s32(__s0, __s1, (int8x16_t)__s2, __p3); \
56236  __ret; \
56237})
56238#else
56239#define vqdmlals_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56240  int64_t __s0 = __p0; \
56241  int32_t __s1 = __p1; \
56242  int32x4_t __s2 = __p2; \
56243  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56244  int64_t __ret; \
56245  __ret = (int64_t) __builtin_neon_vqdmlals_laneq_s32(__s0, __s1, (int8x16_t)__rev2, __p3); \
56246  __ret; \
56247})
56248#endif
56249
56250#ifdef __LITTLE_ENDIAN__
56251#define vqdmlalh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56252  int32_t __s0 = __p0; \
56253  int16_t __s1 = __p1; \
56254  int16x8_t __s2 = __p2; \
56255  int32_t __ret; \
56256  __ret = (int32_t) __builtin_neon_vqdmlalh_laneq_s16(__s0, __s1, (int8x16_t)__s2, __p3); \
56257  __ret; \
56258})
56259#else
56260#define vqdmlalh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56261  int32_t __s0 = __p0; \
56262  int16_t __s1 = __p1; \
56263  int16x8_t __s2 = __p2; \
56264  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56265  int32_t __ret; \
56266  __ret = (int32_t) __builtin_neon_vqdmlalh_laneq_s16(__s0, __s1, (int8x16_t)__rev2, __p3); \
56267  __ret; \
56268})
56269#endif
56270
56271#ifdef __LITTLE_ENDIAN__
56272#define vqdmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56273  int64x2_t __s0 = __p0; \
56274  int32x2_t __s1 = __p1; \
56275  int32x4_t __s2 = __p2; \
56276  int64x2_t __ret; \
56277  __ret = vqdmlal_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56278  __ret; \
56279})
56280#else
56281#define vqdmlal_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56282  int64x2_t __s0 = __p0; \
56283  int32x2_t __s1 = __p1; \
56284  int32x4_t __s2 = __p2; \
56285  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56286  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
56287  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56288  int64x2_t __ret; \
56289  __ret = __noswap_vqdmlal_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56290  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56291  __ret; \
56292})
56293#endif
56294
56295#ifdef __LITTLE_ENDIAN__
56296#define vqdmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56297  int32x4_t __s0 = __p0; \
56298  int16x4_t __s1 = __p1; \
56299  int16x8_t __s2 = __p2; \
56300  int32x4_t __ret; \
56301  __ret = vqdmlal_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56302  __ret; \
56303})
56304#else
56305#define vqdmlal_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56306  int32x4_t __s0 = __p0; \
56307  int16x4_t __s1 = __p1; \
56308  int16x8_t __s2 = __p2; \
56309  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56310  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56311  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56312  int32x4_t __ret; \
56313  __ret = __noswap_vqdmlal_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56314  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56315  __ret; \
56316})
56317#endif
56318
56319#ifdef __LITTLE_ENDIAN__
56320__ai int64_t vqdmlsls_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
56321  int64_t __ret;
56322  __ret = (int64_t) __builtin_neon_vqdmlsls_s32(__p0, __p1, __p2);
56323  return __ret;
56324}
56325#else
56326__ai int64_t vqdmlsls_s32(int64_t __p0, int32_t __p1, int32_t __p2) {
56327  int64_t __ret;
56328  __ret = (int64_t) __builtin_neon_vqdmlsls_s32(__p0, __p1, __p2);
56329  return __ret;
56330}
56331#endif
56332
56333#ifdef __LITTLE_ENDIAN__
56334__ai int32_t vqdmlslh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
56335  int32_t __ret;
56336  __ret = (int32_t) __builtin_neon_vqdmlslh_s16(__p0, __p1, __p2);
56337  return __ret;
56338}
56339#else
56340__ai int32_t vqdmlslh_s16(int32_t __p0, int16_t __p1, int16_t __p2) {
56341  int32_t __ret;
56342  __ret = (int32_t) __builtin_neon_vqdmlslh_s16(__p0, __p1, __p2);
56343  return __ret;
56344}
56345#endif
56346
56347#ifdef __LITTLE_ENDIAN__
56348__ai int64x2_t vqdmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
56349  int64x2_t __ret;
56350  __ret = vqdmlsl_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
56351  return __ret;
56352}
56353#else
56354__ai int64x2_t vqdmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
56355  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56356  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56357  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
56358  int64x2_t __ret;
56359  __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
56360  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56361  return __ret;
56362}
56363#endif
56364
56365#ifdef __LITTLE_ENDIAN__
56366__ai int32x4_t vqdmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
56367  int32x4_t __ret;
56368  __ret = vqdmlsl_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
56369  return __ret;
56370}
56371#else
56372__ai int32x4_t vqdmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
56373  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56374  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56375  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
56376  int32x4_t __ret;
56377  __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
56378  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56379  return __ret;
56380}
56381#endif
56382
56383#ifdef __LITTLE_ENDIAN__
56384#define vqdmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56385  int64x2_t __s0 = __p0; \
56386  int32x4_t __s1 = __p1; \
56387  int32x2_t __s2 = __p2; \
56388  int64x2_t __ret; \
56389  __ret = vqdmlsl_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56390  __ret; \
56391})
56392#else
56393#define vqdmlsl_high_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56394  int64x2_t __s0 = __p0; \
56395  int32x4_t __s1 = __p1; \
56396  int32x2_t __s2 = __p2; \
56397  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56398  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56399  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
56400  int64x2_t __ret; \
56401  __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56402  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56403  __ret; \
56404})
56405#endif
56406
56407#ifdef __LITTLE_ENDIAN__
56408#define vqdmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56409  int32x4_t __s0 = __p0; \
56410  int16x8_t __s1 = __p1; \
56411  int16x4_t __s2 = __p2; \
56412  int32x4_t __ret; \
56413  __ret = vqdmlsl_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56414  __ret; \
56415})
56416#else
56417#define vqdmlsl_high_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56418  int32x4_t __s0 = __p0; \
56419  int16x8_t __s1 = __p1; \
56420  int16x4_t __s2 = __p2; \
56421  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56422  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56423  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56424  int32x4_t __ret; \
56425  __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56426  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56427  __ret; \
56428})
56429#endif
56430
56431#ifdef __LITTLE_ENDIAN__
56432#define vqdmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56433  int64x2_t __s0 = __p0; \
56434  int32x4_t __s1 = __p1; \
56435  int32x4_t __s2 = __p2; \
56436  int64x2_t __ret; \
56437  __ret = vqdmlsl_s32(__s0, vget_high_s32(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56438  __ret; \
56439})
56440#else
56441#define vqdmlsl_high_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56442  int64x2_t __s0 = __p0; \
56443  int32x4_t __s1 = __p1; \
56444  int32x4_t __s2 = __p2; \
56445  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56446  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56447  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56448  int64x2_t __ret; \
56449  __ret = __noswap_vqdmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56450  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56451  __ret; \
56452})
56453#endif
56454
56455#ifdef __LITTLE_ENDIAN__
56456#define vqdmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56457  int32x4_t __s0 = __p0; \
56458  int16x8_t __s1 = __p1; \
56459  int16x8_t __s2 = __p2; \
56460  int32x4_t __ret; \
56461  __ret = vqdmlsl_s16(__s0, vget_high_s16(__s1), __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56462  __ret; \
56463})
56464#else
56465#define vqdmlsl_high_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56466  int32x4_t __s0 = __p0; \
56467  int16x8_t __s1 = __p1; \
56468  int16x8_t __s2 = __p2; \
56469  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56470  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56471  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56472  int32x4_t __ret; \
56473  __ret = __noswap_vqdmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56474  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56475  __ret; \
56476})
56477#endif
56478
56479#ifdef __LITTLE_ENDIAN__
56480__ai int64x2_t vqdmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56481  int64x2_t __ret;
56482  __ret = vqdmlsl_n_s32(__p0, vget_high_s32(__p1), __p2);
56483  return __ret;
56484}
56485#else
56486__ai int64x2_t vqdmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
56487  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
56488  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56489  int64x2_t __ret;
56490  __ret = __noswap_vqdmlsl_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
56491  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56492  return __ret;
56493}
56494#endif
56495
56496#ifdef __LITTLE_ENDIAN__
56497__ai int32x4_t vqdmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56498  int32x4_t __ret;
56499  __ret = vqdmlsl_n_s16(__p0, vget_high_s16(__p1), __p2);
56500  return __ret;
56501}
56502#else
56503__ai int32x4_t vqdmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
56504  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56505  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56506  int32x4_t __ret;
56507  __ret = __noswap_vqdmlsl_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
56508  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56509  return __ret;
56510}
56511#endif
56512
56513#ifdef __LITTLE_ENDIAN__
56514#define vqdmlsls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56515  int64_t __s0 = __p0; \
56516  int32_t __s1 = __p1; \
56517  int32x2_t __s2 = __p2; \
56518  int64_t __ret; \
56519  __ret = (int64_t) __builtin_neon_vqdmlsls_lane_s32(__s0, __s1, (int8x8_t)__s2, __p3); \
56520  __ret; \
56521})
56522#else
56523#define vqdmlsls_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56524  int64_t __s0 = __p0; \
56525  int32_t __s1 = __p1; \
56526  int32x2_t __s2 = __p2; \
56527  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
56528  int64_t __ret; \
56529  __ret = (int64_t) __builtin_neon_vqdmlsls_lane_s32(__s0, __s1, (int8x8_t)__rev2, __p3); \
56530  __ret; \
56531})
56532#endif
56533
56534#ifdef __LITTLE_ENDIAN__
56535#define vqdmlslh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56536  int32_t __s0 = __p0; \
56537  int16_t __s1 = __p1; \
56538  int16x4_t __s2 = __p2; \
56539  int32_t __ret; \
56540  __ret = (int32_t) __builtin_neon_vqdmlslh_lane_s16(__s0, __s1, (int8x8_t)__s2, __p3); \
56541  __ret; \
56542})
56543#else
56544#define vqdmlslh_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56545  int32_t __s0 = __p0; \
56546  int16_t __s1 = __p1; \
56547  int16x4_t __s2 = __p2; \
56548  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56549  int32_t __ret; \
56550  __ret = (int32_t) __builtin_neon_vqdmlslh_lane_s16(__s0, __s1, (int8x8_t)__rev2, __p3); \
56551  __ret; \
56552})
56553#endif
56554
56555#ifdef __LITTLE_ENDIAN__
56556#define vqdmlsls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56557  int64_t __s0 = __p0; \
56558  int32_t __s1 = __p1; \
56559  int32x4_t __s2 = __p2; \
56560  int64_t __ret; \
56561  __ret = (int64_t) __builtin_neon_vqdmlsls_laneq_s32(__s0, __s1, (int8x16_t)__s2, __p3); \
56562  __ret; \
56563})
56564#else
56565#define vqdmlsls_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56566  int64_t __s0 = __p0; \
56567  int32_t __s1 = __p1; \
56568  int32x4_t __s2 = __p2; \
56569  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56570  int64_t __ret; \
56571  __ret = (int64_t) __builtin_neon_vqdmlsls_laneq_s32(__s0, __s1, (int8x16_t)__rev2, __p3); \
56572  __ret; \
56573})
56574#endif
56575
56576#ifdef __LITTLE_ENDIAN__
56577#define vqdmlslh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56578  int32_t __s0 = __p0; \
56579  int16_t __s1 = __p1; \
56580  int16x8_t __s2 = __p2; \
56581  int32_t __ret; \
56582  __ret = (int32_t) __builtin_neon_vqdmlslh_laneq_s16(__s0, __s1, (int8x16_t)__s2, __p3); \
56583  __ret; \
56584})
56585#else
56586#define vqdmlslh_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56587  int32_t __s0 = __p0; \
56588  int16_t __s1 = __p1; \
56589  int16x8_t __s2 = __p2; \
56590  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56591  int32_t __ret; \
56592  __ret = (int32_t) __builtin_neon_vqdmlslh_laneq_s16(__s0, __s1, (int8x16_t)__rev2, __p3); \
56593  __ret; \
56594})
56595#endif
56596
56597#ifdef __LITTLE_ENDIAN__
56598#define vqdmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56599  int64x2_t __s0 = __p0; \
56600  int32x2_t __s1 = __p1; \
56601  int32x4_t __s2 = __p2; \
56602  int64x2_t __ret; \
56603  __ret = vqdmlsl_s32(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
56604  __ret; \
56605})
56606#else
56607#define vqdmlsl_laneq_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
56608  int64x2_t __s0 = __p0; \
56609  int32x2_t __s1 = __p1; \
56610  int32x4_t __s2 = __p2; \
56611  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56612  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
56613  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
56614  int64x2_t __ret; \
56615  __ret = __noswap_vqdmlsl_s32(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
56616  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56617  __ret; \
56618})
56619#endif
56620
56621#ifdef __LITTLE_ENDIAN__
56622#define vqdmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56623  int32x4_t __s0 = __p0; \
56624  int16x4_t __s1 = __p1; \
56625  int16x8_t __s2 = __p2; \
56626  int32x4_t __ret; \
56627  __ret = vqdmlsl_s16(__s0, __s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
56628  __ret; \
56629})
56630#else
56631#define vqdmlsl_laneq_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
56632  int32x4_t __s0 = __p0; \
56633  int16x4_t __s1 = __p1; \
56634  int16x8_t __s2 = __p2; \
56635  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56636  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56637  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 7, 6, 5, 4, 3, 2, 1, 0); \
56638  int32x4_t __ret; \
56639  __ret = __noswap_vqdmlsl_s16(__rev0, __rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
56640  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56641  __ret; \
56642})
56643#endif
56644
56645#ifdef __LITTLE_ENDIAN__
56646__ai int32_t vqdmulhs_s32(int32_t __p0, int32_t __p1) {
56647  int32_t __ret;
56648  __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
56649  return __ret;
56650}
56651#else
56652__ai int32_t vqdmulhs_s32(int32_t __p0, int32_t __p1) {
56653  int32_t __ret;
56654  __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
56655  return __ret;
56656}
56657__ai int32_t __noswap_vqdmulhs_s32(int32_t __p0, int32_t __p1) {
56658  int32_t __ret;
56659  __ret = (int32_t) __builtin_neon_vqdmulhs_s32(__p0, __p1);
56660  return __ret;
56661}
56662#endif
56663
56664#ifdef __LITTLE_ENDIAN__
56665__ai int16_t vqdmulhh_s16(int16_t __p0, int16_t __p1) {
56666  int16_t __ret;
56667  __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
56668  return __ret;
56669}
56670#else
56671__ai int16_t vqdmulhh_s16(int16_t __p0, int16_t __p1) {
56672  int16_t __ret;
56673  __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
56674  return __ret;
56675}
56676__ai int16_t __noswap_vqdmulhh_s16(int16_t __p0, int16_t __p1) {
56677  int16_t __ret;
56678  __ret = (int16_t) __builtin_neon_vqdmulhh_s16(__p0, __p1);
56679  return __ret;
56680}
56681#endif
56682
56683#ifdef __LITTLE_ENDIAN__
56684#define vqdmulhs_lane_s32(__p0_146, __p1_146, __p2_146) __extension__ ({ \
56685  int32_t __s0_146 = __p0_146; \
56686  int32x2_t __s1_146 = __p1_146; \
56687  int32_t __ret_146; \
56688  __ret_146 = vqdmulhs_s32(__s0_146, vget_lane_s32(__s1_146, __p2_146)); \
56689  __ret_146; \
56690})
56691#else
56692#define vqdmulhs_lane_s32(__p0_147, __p1_147, __p2_147) __extension__ ({ \
56693  int32_t __s0_147 = __p0_147; \
56694  int32x2_t __s1_147 = __p1_147; \
56695  int32x2_t __rev1_147;  __rev1_147 = __builtin_shufflevector(__s1_147, __s1_147, 1, 0); \
56696  int32_t __ret_147; \
56697  __ret_147 = __noswap_vqdmulhs_s32(__s0_147, __noswap_vget_lane_s32(__rev1_147, __p2_147)); \
56698  __ret_147; \
56699})
56700#endif
56701
56702#ifdef __LITTLE_ENDIAN__
56703#define vqdmulhh_lane_s16(__p0_148, __p1_148, __p2_148) __extension__ ({ \
56704  int16_t __s0_148 = __p0_148; \
56705  int16x4_t __s1_148 = __p1_148; \
56706  int16_t __ret_148; \
56707  __ret_148 = vqdmulhh_s16(__s0_148, vget_lane_s16(__s1_148, __p2_148)); \
56708  __ret_148; \
56709})
56710#else
56711#define vqdmulhh_lane_s16(__p0_149, __p1_149, __p2_149) __extension__ ({ \
56712  int16_t __s0_149 = __p0_149; \
56713  int16x4_t __s1_149 = __p1_149; \
56714  int16x4_t __rev1_149;  __rev1_149 = __builtin_shufflevector(__s1_149, __s1_149, 3, 2, 1, 0); \
56715  int16_t __ret_149; \
56716  __ret_149 = __noswap_vqdmulhh_s16(__s0_149, __noswap_vget_lane_s16(__rev1_149, __p2_149)); \
56717  __ret_149; \
56718})
56719#endif
56720
56721#ifdef __LITTLE_ENDIAN__
56722#define vqdmulhs_laneq_s32(__p0_150, __p1_150, __p2_150) __extension__ ({ \
56723  int32_t __s0_150 = __p0_150; \
56724  int32x4_t __s1_150 = __p1_150; \
56725  int32_t __ret_150; \
56726  __ret_150 = vqdmulhs_s32(__s0_150, vgetq_lane_s32(__s1_150, __p2_150)); \
56727  __ret_150; \
56728})
56729#else
56730#define vqdmulhs_laneq_s32(__p0_151, __p1_151, __p2_151) __extension__ ({ \
56731  int32_t __s0_151 = __p0_151; \
56732  int32x4_t __s1_151 = __p1_151; \
56733  int32x4_t __rev1_151;  __rev1_151 = __builtin_shufflevector(__s1_151, __s1_151, 3, 2, 1, 0); \
56734  int32_t __ret_151; \
56735  __ret_151 = __noswap_vqdmulhs_s32(__s0_151, __noswap_vgetq_lane_s32(__rev1_151, __p2_151)); \
56736  __ret_151; \
56737})
56738#endif
56739
56740#ifdef __LITTLE_ENDIAN__
56741#define vqdmulhh_laneq_s16(__p0_152, __p1_152, __p2_152) __extension__ ({ \
56742  int16_t __s0_152 = __p0_152; \
56743  int16x8_t __s1_152 = __p1_152; \
56744  int16_t __ret_152; \
56745  __ret_152 = vqdmulhh_s16(__s0_152, vgetq_lane_s16(__s1_152, __p2_152)); \
56746  __ret_152; \
56747})
56748#else
56749#define vqdmulhh_laneq_s16(__p0_153, __p1_153, __p2_153) __extension__ ({ \
56750  int16_t __s0_153 = __p0_153; \
56751  int16x8_t __s1_153 = __p1_153; \
56752  int16x8_t __rev1_153;  __rev1_153 = __builtin_shufflevector(__s1_153, __s1_153, 7, 6, 5, 4, 3, 2, 1, 0); \
56753  int16_t __ret_153; \
56754  __ret_153 = __noswap_vqdmulhh_s16(__s0_153, __noswap_vgetq_lane_s16(__rev1_153, __p2_153)); \
56755  __ret_153; \
56756})
56757#endif
56758
56759#ifdef __LITTLE_ENDIAN__
56760#define vqdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56761  int32x4_t __s0 = __p0; \
56762  int32x4_t __s1 = __p1; \
56763  int32x4_t __ret; \
56764  __ret = vqdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56765  __ret; \
56766})
56767#else
56768#define vqdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56769  int32x4_t __s0 = __p0; \
56770  int32x4_t __s1 = __p1; \
56771  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56772  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56773  int32x4_t __ret; \
56774  __ret = __noswap_vqdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56775  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56776  __ret; \
56777})
56778#endif
56779
56780#ifdef __LITTLE_ENDIAN__
56781#define vqdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56782  int16x8_t __s0 = __p0; \
56783  int16x8_t __s1 = __p1; \
56784  int16x8_t __ret; \
56785  __ret = vqdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
56786  __ret; \
56787})
56788#else
56789#define vqdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56790  int16x8_t __s0 = __p0; \
56791  int16x8_t __s1 = __p1; \
56792  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
56793  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56794  int16x8_t __ret; \
56795  __ret = __noswap_vqdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
56796  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
56797  __ret; \
56798})
56799#endif
56800
56801#ifdef __LITTLE_ENDIAN__
56802#define vqdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56803  int32x2_t __s0 = __p0; \
56804  int32x4_t __s1 = __p1; \
56805  int32x2_t __ret; \
56806  __ret = vqdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
56807  __ret; \
56808})
56809#else
56810#define vqdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56811  int32x2_t __s0 = __p0; \
56812  int32x4_t __s1 = __p1; \
56813  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
56814  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56815  int32x2_t __ret; \
56816  __ret = __noswap_vqdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
56817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56818  __ret; \
56819})
56820#endif
56821
56822#ifdef __LITTLE_ENDIAN__
56823#define vqdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56824  int16x4_t __s0 = __p0; \
56825  int16x8_t __s1 = __p1; \
56826  int16x4_t __ret; \
56827  __ret = vqdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56828  __ret; \
56829})
56830#else
56831#define vqdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56832  int16x4_t __s0 = __p0; \
56833  int16x8_t __s1 = __p1; \
56834  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56835  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56836  int16x4_t __ret; \
56837  __ret = __noswap_vqdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56838  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56839  __ret; \
56840})
56841#endif
56842
56843#ifdef __LITTLE_ENDIAN__
56844__ai int64_t vqdmulls_s32(int32_t __p0, int32_t __p1) {
56845  int64_t __ret;
56846  __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
56847  return __ret;
56848}
56849#else
56850__ai int64_t vqdmulls_s32(int32_t __p0, int32_t __p1) {
56851  int64_t __ret;
56852  __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
56853  return __ret;
56854}
56855__ai int64_t __noswap_vqdmulls_s32(int32_t __p0, int32_t __p1) {
56856  int64_t __ret;
56857  __ret = (int64_t) __builtin_neon_vqdmulls_s32(__p0, __p1);
56858  return __ret;
56859}
56860#endif
56861
56862#ifdef __LITTLE_ENDIAN__
56863__ai int32_t vqdmullh_s16(int16_t __p0, int16_t __p1) {
56864  int32_t __ret;
56865  __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
56866  return __ret;
56867}
56868#else
56869__ai int32_t vqdmullh_s16(int16_t __p0, int16_t __p1) {
56870  int32_t __ret;
56871  __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
56872  return __ret;
56873}
56874__ai int32_t __noswap_vqdmullh_s16(int16_t __p0, int16_t __p1) {
56875  int32_t __ret;
56876  __ret = (int32_t) __builtin_neon_vqdmullh_s16(__p0, __p1);
56877  return __ret;
56878}
56879#endif
56880
56881#ifdef __LITTLE_ENDIAN__
56882__ai int64x2_t vqdmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
56883  int64x2_t __ret;
56884  __ret = vqdmull_s32(vget_high_s32(__p0), vget_high_s32(__p1));
56885  return __ret;
56886}
56887#else
56888__ai int64x2_t vqdmull_high_s32(int32x4_t __p0, int32x4_t __p1) {
56889  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
56890  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
56891  int64x2_t __ret;
56892  __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
56893  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
56894  return __ret;
56895}
56896#endif
56897
56898#ifdef __LITTLE_ENDIAN__
56899__ai int32x4_t vqdmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
56900  int32x4_t __ret;
56901  __ret = vqdmull_s16(vget_high_s16(__p0), vget_high_s16(__p1));
56902  return __ret;
56903}
56904#else
56905__ai int32x4_t vqdmull_high_s16(int16x8_t __p0, int16x8_t __p1) {
56906  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
56907  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
56908  int32x4_t __ret;
56909  __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
56910  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
56911  return __ret;
56912}
56913#endif
56914
56915#ifdef __LITTLE_ENDIAN__
56916#define vqdmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
56917  int32x4_t __s0 = __p0; \
56918  int32x2_t __s1 = __p1; \
56919  int64x2_t __ret; \
56920  __ret = vqdmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
56921  __ret; \
56922})
56923#else
56924#define vqdmull_high_lane_s32(__p0, __p1, __p2) __extension__ ({ \
56925  int32x4_t __s0 = __p0; \
56926  int32x2_t __s1 = __p1; \
56927  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56928  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
56929  int64x2_t __ret; \
56930  __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
56931  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56932  __ret; \
56933})
56934#endif
56935
56936#ifdef __LITTLE_ENDIAN__
56937#define vqdmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
56938  int16x8_t __s0 = __p0; \
56939  int16x4_t __s1 = __p1; \
56940  int32x4_t __ret; \
56941  __ret = vqdmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56942  __ret; \
56943})
56944#else
56945#define vqdmull_high_lane_s16(__p0, __p1, __p2) __extension__ ({ \
56946  int16x8_t __s0 = __p0; \
56947  int16x4_t __s1 = __p1; \
56948  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
56949  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56950  int32x4_t __ret; \
56951  __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56952  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56953  __ret; \
56954})
56955#endif
56956
56957#ifdef __LITTLE_ENDIAN__
56958#define vqdmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56959  int32x4_t __s0 = __p0; \
56960  int32x4_t __s1 = __p1; \
56961  int64x2_t __ret; \
56962  __ret = vqdmull_s32(vget_high_s32(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
56963  __ret; \
56964})
56965#else
56966#define vqdmull_high_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
56967  int32x4_t __s0 = __p0; \
56968  int32x4_t __s1 = __p1; \
56969  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
56970  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
56971  int64x2_t __ret; \
56972  __ret = __noswap_vqdmull_s32(__noswap_vget_high_s32(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
56973  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
56974  __ret; \
56975})
56976#endif
56977
56978#ifdef __LITTLE_ENDIAN__
56979#define vqdmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56980  int16x8_t __s0 = __p0; \
56981  int16x8_t __s1 = __p1; \
56982  int32x4_t __ret; \
56983  __ret = vqdmull_s16(vget_high_s16(__s0), __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
56984  __ret; \
56985})
56986#else
56987#define vqdmull_high_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
56988  int16x8_t __s0 = __p0; \
56989  int16x8_t __s1 = __p1; \
56990  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
56991  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
56992  int32x4_t __ret; \
56993  __ret = __noswap_vqdmull_s16(__noswap_vget_high_s16(__rev0), __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
56994  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
56995  __ret; \
56996})
56997#endif
56998
56999#ifdef __LITTLE_ENDIAN__
57000__ai int64x2_t vqdmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
57001  int64x2_t __ret;
57002  __ret = vqdmull_n_s32(vget_high_s32(__p0), __p1);
57003  return __ret;
57004}
57005#else
57006__ai int64x2_t vqdmull_high_n_s32(int32x4_t __p0, int32_t __p1) {
57007  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57008  int64x2_t __ret;
57009  __ret = __noswap_vqdmull_n_s32(__noswap_vget_high_s32(__rev0), __p1);
57010  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
57011  return __ret;
57012}
57013#endif
57014
57015#ifdef __LITTLE_ENDIAN__
57016__ai int32x4_t vqdmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
57017  int32x4_t __ret;
57018  __ret = vqdmull_n_s16(vget_high_s16(__p0), __p1);
57019  return __ret;
57020}
57021#else
57022__ai int32x4_t vqdmull_high_n_s16(int16x8_t __p0, int16_t __p1) {
57023  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57024  int32x4_t __ret;
57025  __ret = __noswap_vqdmull_n_s16(__noswap_vget_high_s16(__rev0), __p1);
57026  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57027  return __ret;
57028}
57029#endif
57030
57031#ifdef __LITTLE_ENDIAN__
57032#define vqdmulls_lane_s32(__p0_154, __p1_154, __p2_154) __extension__ ({ \
57033  int32_t __s0_154 = __p0_154; \
57034  int32x2_t __s1_154 = __p1_154; \
57035  int64_t __ret_154; \
57036  __ret_154 = vqdmulls_s32(__s0_154, vget_lane_s32(__s1_154, __p2_154)); \
57037  __ret_154; \
57038})
57039#else
57040#define vqdmulls_lane_s32(__p0_155, __p1_155, __p2_155) __extension__ ({ \
57041  int32_t __s0_155 = __p0_155; \
57042  int32x2_t __s1_155 = __p1_155; \
57043  int32x2_t __rev1_155;  __rev1_155 = __builtin_shufflevector(__s1_155, __s1_155, 1, 0); \
57044  int64_t __ret_155; \
57045  __ret_155 = __noswap_vqdmulls_s32(__s0_155, __noswap_vget_lane_s32(__rev1_155, __p2_155)); \
57046  __ret_155; \
57047})
57048#endif
57049
57050#ifdef __LITTLE_ENDIAN__
57051#define vqdmullh_lane_s16(__p0_156, __p1_156, __p2_156) __extension__ ({ \
57052  int16_t __s0_156 = __p0_156; \
57053  int16x4_t __s1_156 = __p1_156; \
57054  int32_t __ret_156; \
57055  __ret_156 = vqdmullh_s16(__s0_156, vget_lane_s16(__s1_156, __p2_156)); \
57056  __ret_156; \
57057})
57058#else
57059#define vqdmullh_lane_s16(__p0_157, __p1_157, __p2_157) __extension__ ({ \
57060  int16_t __s0_157 = __p0_157; \
57061  int16x4_t __s1_157 = __p1_157; \
57062  int16x4_t __rev1_157;  __rev1_157 = __builtin_shufflevector(__s1_157, __s1_157, 3, 2, 1, 0); \
57063  int32_t __ret_157; \
57064  __ret_157 = __noswap_vqdmullh_s16(__s0_157, __noswap_vget_lane_s16(__rev1_157, __p2_157)); \
57065  __ret_157; \
57066})
57067#endif
57068
57069#ifdef __LITTLE_ENDIAN__
57070#define vqdmulls_laneq_s32(__p0_158, __p1_158, __p2_158) __extension__ ({ \
57071  int32_t __s0_158 = __p0_158; \
57072  int32x4_t __s1_158 = __p1_158; \
57073  int64_t __ret_158; \
57074  __ret_158 = vqdmulls_s32(__s0_158, vgetq_lane_s32(__s1_158, __p2_158)); \
57075  __ret_158; \
57076})
57077#else
57078#define vqdmulls_laneq_s32(__p0_159, __p1_159, __p2_159) __extension__ ({ \
57079  int32_t __s0_159 = __p0_159; \
57080  int32x4_t __s1_159 = __p1_159; \
57081  int32x4_t __rev1_159;  __rev1_159 = __builtin_shufflevector(__s1_159, __s1_159, 3, 2, 1, 0); \
57082  int64_t __ret_159; \
57083  __ret_159 = __noswap_vqdmulls_s32(__s0_159, __noswap_vgetq_lane_s32(__rev1_159, __p2_159)); \
57084  __ret_159; \
57085})
57086#endif
57087
57088#ifdef __LITTLE_ENDIAN__
57089#define vqdmullh_laneq_s16(__p0_160, __p1_160, __p2_160) __extension__ ({ \
57090  int16_t __s0_160 = __p0_160; \
57091  int16x8_t __s1_160 = __p1_160; \
57092  int32_t __ret_160; \
57093  __ret_160 = vqdmullh_s16(__s0_160, vgetq_lane_s16(__s1_160, __p2_160)); \
57094  __ret_160; \
57095})
57096#else
57097#define vqdmullh_laneq_s16(__p0_161, __p1_161, __p2_161) __extension__ ({ \
57098  int16_t __s0_161 = __p0_161; \
57099  int16x8_t __s1_161 = __p1_161; \
57100  int16x8_t __rev1_161;  __rev1_161 = __builtin_shufflevector(__s1_161, __s1_161, 7, 6, 5, 4, 3, 2, 1, 0); \
57101  int32_t __ret_161; \
57102  __ret_161 = __noswap_vqdmullh_s16(__s0_161, __noswap_vgetq_lane_s16(__rev1_161, __p2_161)); \
57103  __ret_161; \
57104})
57105#endif
57106
57107#ifdef __LITTLE_ENDIAN__
57108#define vqdmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57109  int32x2_t __s0 = __p0; \
57110  int32x4_t __s1 = __p1; \
57111  int64x2_t __ret; \
57112  __ret = vqdmull_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
57113  __ret; \
57114})
57115#else
57116#define vqdmull_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57117  int32x2_t __s0 = __p0; \
57118  int32x4_t __s1 = __p1; \
57119  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
57120  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
57121  int64x2_t __ret; \
57122  __ret = __noswap_vqdmull_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
57123  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
57124  __ret; \
57125})
57126#endif
57127
57128#ifdef __LITTLE_ENDIAN__
57129#define vqdmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57130  int16x4_t __s0 = __p0; \
57131  int16x8_t __s1 = __p1; \
57132  int32x4_t __ret; \
57133  __ret = vqdmull_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
57134  __ret; \
57135})
57136#else
57137#define vqdmull_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57138  int16x4_t __s0 = __p0; \
57139  int16x8_t __s1 = __p1; \
57140  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
57141  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
57142  int32x4_t __ret; \
57143  __ret = __noswap_vqdmull_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
57144  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
57145  __ret; \
57146})
57147#endif
57148
57149#ifdef __LITTLE_ENDIAN__
57150__ai int16_t vqmovns_s32(int32_t __p0) {
57151  int16_t __ret;
57152  __ret = (int16_t) __builtin_neon_vqmovns_s32(__p0);
57153  return __ret;
57154}
57155#else
57156__ai int16_t vqmovns_s32(int32_t __p0) {
57157  int16_t __ret;
57158  __ret = (int16_t) __builtin_neon_vqmovns_s32(__p0);
57159  return __ret;
57160}
57161#endif
57162
57163#ifdef __LITTLE_ENDIAN__
57164__ai int32_t vqmovnd_s64(int64_t __p0) {
57165  int32_t __ret;
57166  __ret = (int32_t) __builtin_neon_vqmovnd_s64(__p0);
57167  return __ret;
57168}
57169#else
57170__ai int32_t vqmovnd_s64(int64_t __p0) {
57171  int32_t __ret;
57172  __ret = (int32_t) __builtin_neon_vqmovnd_s64(__p0);
57173  return __ret;
57174}
57175#endif
57176
57177#ifdef __LITTLE_ENDIAN__
57178__ai int8_t vqmovnh_s16(int16_t __p0) {
57179  int8_t __ret;
57180  __ret = (int8_t) __builtin_neon_vqmovnh_s16(__p0);
57181  return __ret;
57182}
57183#else
57184__ai int8_t vqmovnh_s16(int16_t __p0) {
57185  int8_t __ret;
57186  __ret = (int8_t) __builtin_neon_vqmovnh_s16(__p0);
57187  return __ret;
57188}
57189#endif
57190
57191#ifdef __LITTLE_ENDIAN__
57192__ai uint16_t vqmovns_u32(uint32_t __p0) {
57193  uint16_t __ret;
57194  __ret = (uint16_t) __builtin_neon_vqmovns_u32(__p0);
57195  return __ret;
57196}
57197#else
57198__ai uint16_t vqmovns_u32(uint32_t __p0) {
57199  uint16_t __ret;
57200  __ret = (uint16_t) __builtin_neon_vqmovns_u32(__p0);
57201  return __ret;
57202}
57203#endif
57204
57205#ifdef __LITTLE_ENDIAN__
57206__ai uint32_t vqmovnd_u64(uint64_t __p0) {
57207  uint32_t __ret;
57208  __ret = (uint32_t) __builtin_neon_vqmovnd_u64(__p0);
57209  return __ret;
57210}
57211#else
57212__ai uint32_t vqmovnd_u64(uint64_t __p0) {
57213  uint32_t __ret;
57214  __ret = (uint32_t) __builtin_neon_vqmovnd_u64(__p0);
57215  return __ret;
57216}
57217#endif
57218
57219#ifdef __LITTLE_ENDIAN__
57220__ai uint8_t vqmovnh_u16(uint16_t __p0) {
57221  uint8_t __ret;
57222  __ret = (uint8_t) __builtin_neon_vqmovnh_u16(__p0);
57223  return __ret;
57224}
57225#else
57226__ai uint8_t vqmovnh_u16(uint16_t __p0) {
57227  uint8_t __ret;
57228  __ret = (uint8_t) __builtin_neon_vqmovnh_u16(__p0);
57229  return __ret;
57230}
57231#endif
57232
57233#ifdef __LITTLE_ENDIAN__
57234__ai uint16x8_t vqmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
57235  uint16x8_t __ret;
57236  __ret = vcombine_u16(__p0, vqmovn_u32(__p1));
57237  return __ret;
57238}
57239#else
57240__ai uint16x8_t vqmovn_high_u32(uint16x4_t __p0, uint32x4_t __p1) {
57241  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57242  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
57243  uint16x8_t __ret;
57244  __ret = __noswap_vcombine_u16(__rev0, __noswap_vqmovn_u32(__rev1));
57245  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57246  return __ret;
57247}
57248#endif
57249
57250#ifdef __LITTLE_ENDIAN__
57251__ai uint32x4_t vqmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
57252  uint32x4_t __ret;
57253  __ret = vcombine_u32(__p0, vqmovn_u64(__p1));
57254  return __ret;
57255}
57256#else
57257__ai uint32x4_t vqmovn_high_u64(uint32x2_t __p0, uint64x2_t __p1) {
57258  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57259  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
57260  uint32x4_t __ret;
57261  __ret = __noswap_vcombine_u32(__rev0, __noswap_vqmovn_u64(__rev1));
57262  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57263  return __ret;
57264}
57265#endif
57266
57267#ifdef __LITTLE_ENDIAN__
57268__ai uint8x16_t vqmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
57269  uint8x16_t __ret;
57270  __ret = vcombine_u8(__p0, vqmovn_u16(__p1));
57271  return __ret;
57272}
57273#else
57274__ai uint8x16_t vqmovn_high_u16(uint8x8_t __p0, uint16x8_t __p1) {
57275  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57276  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57277  uint8x16_t __ret;
57278  __ret = __noswap_vcombine_u8(__rev0, __noswap_vqmovn_u16(__rev1));
57279  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57280  return __ret;
57281}
57282#endif
57283
57284#ifdef __LITTLE_ENDIAN__
57285__ai int16x8_t vqmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
57286  int16x8_t __ret;
57287  __ret = vcombine_s16(__p0, vqmovn_s32(__p1));
57288  return __ret;
57289}
57290#else
57291__ai int16x8_t vqmovn_high_s32(int16x4_t __p0, int32x4_t __p1) {
57292  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57293  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
57294  int16x8_t __ret;
57295  __ret = __noswap_vcombine_s16(__rev0, __noswap_vqmovn_s32(__rev1));
57296  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57297  return __ret;
57298}
57299#endif
57300
57301#ifdef __LITTLE_ENDIAN__
57302__ai int32x4_t vqmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
57303  int32x4_t __ret;
57304  __ret = vcombine_s32(__p0, vqmovn_s64(__p1));
57305  return __ret;
57306}
57307#else
57308__ai int32x4_t vqmovn_high_s64(int32x2_t __p0, int64x2_t __p1) {
57309  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57310  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
57311  int32x4_t __ret;
57312  __ret = __noswap_vcombine_s32(__rev0, __noswap_vqmovn_s64(__rev1));
57313  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57314  return __ret;
57315}
57316#endif
57317
57318#ifdef __LITTLE_ENDIAN__
57319__ai int8x16_t vqmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
57320  int8x16_t __ret;
57321  __ret = vcombine_s8(__p0, vqmovn_s16(__p1));
57322  return __ret;
57323}
57324#else
57325__ai int8x16_t vqmovn_high_s16(int8x8_t __p0, int16x8_t __p1) {
57326  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57327  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57328  int8x16_t __ret;
57329  __ret = __noswap_vcombine_s8(__rev0, __noswap_vqmovn_s16(__rev1));
57330  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57331  return __ret;
57332}
57333#endif
57334
57335#ifdef __LITTLE_ENDIAN__
57336__ai int16_t vqmovuns_s32(int32_t __p0) {
57337  int16_t __ret;
57338  __ret = (int16_t) __builtin_neon_vqmovuns_s32(__p0);
57339  return __ret;
57340}
57341#else
57342__ai int16_t vqmovuns_s32(int32_t __p0) {
57343  int16_t __ret;
57344  __ret = (int16_t) __builtin_neon_vqmovuns_s32(__p0);
57345  return __ret;
57346}
57347#endif
57348
57349#ifdef __LITTLE_ENDIAN__
57350__ai int32_t vqmovund_s64(int64_t __p0) {
57351  int32_t __ret;
57352  __ret = (int32_t) __builtin_neon_vqmovund_s64(__p0);
57353  return __ret;
57354}
57355#else
57356__ai int32_t vqmovund_s64(int64_t __p0) {
57357  int32_t __ret;
57358  __ret = (int32_t) __builtin_neon_vqmovund_s64(__p0);
57359  return __ret;
57360}
57361#endif
57362
57363#ifdef __LITTLE_ENDIAN__
57364__ai int8_t vqmovunh_s16(int16_t __p0) {
57365  int8_t __ret;
57366  __ret = (int8_t) __builtin_neon_vqmovunh_s16(__p0);
57367  return __ret;
57368}
57369#else
57370__ai int8_t vqmovunh_s16(int16_t __p0) {
57371  int8_t __ret;
57372  __ret = (int8_t) __builtin_neon_vqmovunh_s16(__p0);
57373  return __ret;
57374}
57375#endif
57376
57377#ifdef __LITTLE_ENDIAN__
57378__ai int16x8_t vqmovun_high_s32(int16x4_t __p0, int32x4_t __p1) {
57379  int16x8_t __ret;
57380  __ret = vcombine_u16((uint16x4_t)(__p0), vqmovun_s32(__p1));
57381  return __ret;
57382}
57383#else
57384__ai int16x8_t vqmovun_high_s32(int16x4_t __p0, int32x4_t __p1) {
57385  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
57386  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
57387  int16x8_t __ret;
57388  __ret = __noswap_vcombine_u16((uint16x4_t)(__rev0), __noswap_vqmovun_s32(__rev1));
57389  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
57390  return __ret;
57391}
57392#endif
57393
57394#ifdef __LITTLE_ENDIAN__
57395__ai int32x4_t vqmovun_high_s64(int32x2_t __p0, int64x2_t __p1) {
57396  int32x4_t __ret;
57397  __ret = vcombine_u32((uint32x2_t)(__p0), vqmovun_s64(__p1));
57398  return __ret;
57399}
57400#else
57401__ai int32x4_t vqmovun_high_s64(int32x2_t __p0, int64x2_t __p1) {
57402  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57403  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
57404  int32x4_t __ret;
57405  __ret = __noswap_vcombine_u32((uint32x2_t)(__rev0), __noswap_vqmovun_s64(__rev1));
57406  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
57407  return __ret;
57408}
57409#endif
57410
57411#ifdef __LITTLE_ENDIAN__
57412__ai int8x16_t vqmovun_high_s16(int8x8_t __p0, int16x8_t __p1) {
57413  int8x16_t __ret;
57414  __ret = vcombine_u8((uint8x8_t)(__p0), vqmovun_s16(__p1));
57415  return __ret;
57416}
57417#else
57418__ai int8x16_t vqmovun_high_s16(int8x8_t __p0, int16x8_t __p1) {
57419  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
57420  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
57421  int8x16_t __ret;
57422  __ret = __noswap_vcombine_u8((uint8x8_t)(__rev0), __noswap_vqmovun_s16(__rev1));
57423  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
57424  return __ret;
57425}
57426#endif
57427
57428#ifdef __LITTLE_ENDIAN__
57429__ai int64x2_t vqnegq_s64(int64x2_t __p0) {
57430  int64x2_t __ret;
57431  __ret = (int64x2_t) __builtin_neon_vqnegq_v((int8x16_t)__p0, 35);
57432  return __ret;
57433}
57434#else
57435__ai int64x2_t vqnegq_s64(int64x2_t __p0) {
57436  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
57437  int64x2_t __ret;
57438  __ret = (int64x2_t) __builtin_neon_vqnegq_v((int8x16_t)__rev0, 35);
57439  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
57440  return __ret;
57441}
57442#endif
57443
57444#ifdef __LITTLE_ENDIAN__
57445__ai int64x1_t vqneg_s64(int64x1_t __p0) {
57446  int64x1_t __ret;
57447  __ret = (int64x1_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 3);
57448  return __ret;
57449}
57450#else
57451__ai int64x1_t vqneg_s64(int64x1_t __p0) {
57452  int64x1_t __ret;
57453  __ret = (int64x1_t) __builtin_neon_vqneg_v((int8x8_t)__p0, 3);
57454  return __ret;
57455}
57456#endif
57457
57458#ifdef __LITTLE_ENDIAN__
57459__ai int8_t vqnegb_s8(int8_t __p0) {
57460  int8_t __ret;
57461  __ret = (int8_t) __builtin_neon_vqnegb_s8(__p0);
57462  return __ret;
57463}
57464#else
57465__ai int8_t vqnegb_s8(int8_t __p0) {
57466  int8_t __ret;
57467  __ret = (int8_t) __builtin_neon_vqnegb_s8(__p0);
57468  return __ret;
57469}
57470#endif
57471
57472#ifdef __LITTLE_ENDIAN__
57473__ai int32_t vqnegs_s32(int32_t __p0) {
57474  int32_t __ret;
57475  __ret = (int32_t) __builtin_neon_vqnegs_s32(__p0);
57476  return __ret;
57477}
57478#else
57479__ai int32_t vqnegs_s32(int32_t __p0) {
57480  int32_t __ret;
57481  __ret = (int32_t) __builtin_neon_vqnegs_s32(__p0);
57482  return __ret;
57483}
57484#endif
57485
57486#ifdef __LITTLE_ENDIAN__
57487__ai int64_t vqnegd_s64(int64_t __p0) {
57488  int64_t __ret;
57489  __ret = (int64_t) __builtin_neon_vqnegd_s64(__p0);
57490  return __ret;
57491}
57492#else
57493__ai int64_t vqnegd_s64(int64_t __p0) {
57494  int64_t __ret;
57495  __ret = (int64_t) __builtin_neon_vqnegd_s64(__p0);
57496  return __ret;
57497}
57498#endif
57499
57500#ifdef __LITTLE_ENDIAN__
57501__ai int16_t vqnegh_s16(int16_t __p0) {
57502  int16_t __ret;
57503  __ret = (int16_t) __builtin_neon_vqnegh_s16(__p0);
57504  return __ret;
57505}
57506#else
57507__ai int16_t vqnegh_s16(int16_t __p0) {
57508  int16_t __ret;
57509  __ret = (int16_t) __builtin_neon_vqnegh_s16(__p0);
57510  return __ret;
57511}
57512#endif
57513
57514#ifdef __LITTLE_ENDIAN__
57515__ai int32_t vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
57516  int32_t __ret;
57517  __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
57518  return __ret;
57519}
57520#else
57521__ai int32_t vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
57522  int32_t __ret;
57523  __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
57524  return __ret;
57525}
57526__ai int32_t __noswap_vqrdmulhs_s32(int32_t __p0, int32_t __p1) {
57527  int32_t __ret;
57528  __ret = (int32_t) __builtin_neon_vqrdmulhs_s32(__p0, __p1);
57529  return __ret;
57530}
57531#endif
57532
57533#ifdef __LITTLE_ENDIAN__
57534__ai int16_t vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
57535  int16_t __ret;
57536  __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
57537  return __ret;
57538}
57539#else
57540__ai int16_t vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
57541  int16_t __ret;
57542  __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
57543  return __ret;
57544}
57545__ai int16_t __noswap_vqrdmulhh_s16(int16_t __p0, int16_t __p1) {
57546  int16_t __ret;
57547  __ret = (int16_t) __builtin_neon_vqrdmulhh_s16(__p0, __p1);
57548  return __ret;
57549}
57550#endif
57551
57552#ifdef __LITTLE_ENDIAN__
57553#define vqrdmulhs_lane_s32(__p0_162, __p1_162, __p2_162) __extension__ ({ \
57554  int32_t __s0_162 = __p0_162; \
57555  int32x2_t __s1_162 = __p1_162; \
57556  int32_t __ret_162; \
57557  __ret_162 = vqrdmulhs_s32(__s0_162, vget_lane_s32(__s1_162, __p2_162)); \
57558  __ret_162; \
57559})
57560#else
57561#define vqrdmulhs_lane_s32(__p0_163, __p1_163, __p2_163) __extension__ ({ \
57562  int32_t __s0_163 = __p0_163; \
57563  int32x2_t __s1_163 = __p1_163; \
57564  int32x2_t __rev1_163;  __rev1_163 = __builtin_shufflevector(__s1_163, __s1_163, 1, 0); \
57565  int32_t __ret_163; \
57566  __ret_163 = __noswap_vqrdmulhs_s32(__s0_163, __noswap_vget_lane_s32(__rev1_163, __p2_163)); \
57567  __ret_163; \
57568})
57569#endif
57570
57571#ifdef __LITTLE_ENDIAN__
57572#define vqrdmulhh_lane_s16(__p0_164, __p1_164, __p2_164) __extension__ ({ \
57573  int16_t __s0_164 = __p0_164; \
57574  int16x4_t __s1_164 = __p1_164; \
57575  int16_t __ret_164; \
57576  __ret_164 = vqrdmulhh_s16(__s0_164, vget_lane_s16(__s1_164, __p2_164)); \
57577  __ret_164; \
57578})
57579#else
57580#define vqrdmulhh_lane_s16(__p0_165, __p1_165, __p2_165) __extension__ ({ \
57581  int16_t __s0_165 = __p0_165; \
57582  int16x4_t __s1_165 = __p1_165; \
57583  int16x4_t __rev1_165;  __rev1_165 = __builtin_shufflevector(__s1_165, __s1_165, 3, 2, 1, 0); \
57584  int16_t __ret_165; \
57585  __ret_165 = __noswap_vqrdmulhh_s16(__s0_165, __noswap_vget_lane_s16(__rev1_165, __p2_165)); \
57586  __ret_165; \
57587})
57588#endif
57589
57590#ifdef __LITTLE_ENDIAN__
57591#define vqrdmulhs_laneq_s32(__p0_166, __p1_166, __p2_166) __extension__ ({ \
57592  int32_t __s0_166 = __p0_166; \
57593  int32x4_t __s1_166 = __p1_166; \
57594  int32_t __ret_166; \
57595  __ret_166 = vqrdmulhs_s32(__s0_166, vgetq_lane_s32(__s1_166, __p2_166)); \
57596  __ret_166; \
57597})
57598#else
57599#define vqrdmulhs_laneq_s32(__p0_167, __p1_167, __p2_167) __extension__ ({ \
57600  int32_t __s0_167 = __p0_167; \
57601  int32x4_t __s1_167 = __p1_167; \
57602  int32x4_t __rev1_167;  __rev1_167 = __builtin_shufflevector(__s1_167, __s1_167, 3, 2, 1, 0); \
57603  int32_t __ret_167; \
57604  __ret_167 = __noswap_vqrdmulhs_s32(__s0_167, __noswap_vgetq_lane_s32(__rev1_167, __p2_167)); \
57605  __ret_167; \
57606})
57607#endif
57608
57609#ifdef __LITTLE_ENDIAN__
57610#define vqrdmulhh_laneq_s16(__p0_168, __p1_168, __p2_168) __extension__ ({ \
57611  int16_t __s0_168 = __p0_168; \
57612  int16x8_t __s1_168 = __p1_168; \
57613  int16_t __ret_168; \
57614  __ret_168 = vqrdmulhh_s16(__s0_168, vgetq_lane_s16(__s1_168, __p2_168)); \
57615  __ret_168; \
57616})
57617#else
57618#define vqrdmulhh_laneq_s16(__p0_169, __p1_169, __p2_169) __extension__ ({ \
57619  int16_t __s0_169 = __p0_169; \
57620  int16x8_t __s1_169 = __p1_169; \
57621  int16x8_t __rev1_169;  __rev1_169 = __builtin_shufflevector(__s1_169, __s1_169, 7, 6, 5, 4, 3, 2, 1, 0); \
57622  int16_t __ret_169; \
57623  __ret_169 = __noswap_vqrdmulhh_s16(__s0_169, __noswap_vgetq_lane_s16(__rev1_169, __p2_169)); \
57624  __ret_169; \
57625})
57626#endif
57627
57628#ifdef __LITTLE_ENDIAN__
57629#define vqrdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57630  int32x4_t __s0 = __p0; \
57631  int32x4_t __s1 = __p1; \
57632  int32x4_t __ret; \
57633  __ret = vqrdmulhq_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
57634  __ret; \
57635})
57636#else
57637#define vqrdmulhq_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57638  int32x4_t __s0 = __p0; \
57639  int32x4_t __s1 = __p1; \
57640  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
57641  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
57642  int32x4_t __ret; \
57643  __ret = __noswap_vqrdmulhq_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
57644  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
57645  __ret; \
57646})
57647#endif
57648
57649#ifdef __LITTLE_ENDIAN__
57650#define vqrdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57651  int16x8_t __s0 = __p0; \
57652  int16x8_t __s1 = __p1; \
57653  int16x8_t __ret; \
57654  __ret = vqrdmulhq_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
57655  __ret; \
57656})
57657#else
57658#define vqrdmulhq_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57659  int16x8_t __s0 = __p0; \
57660  int16x8_t __s1 = __p1; \
57661  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 7, 6, 5, 4, 3, 2, 1, 0); \
57662  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
57663  int16x8_t __ret; \
57664  __ret = __noswap_vqrdmulhq_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2, __p2, __p2, __p2, __p2)); \
57665  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); \
57666  __ret; \
57667})
57668#endif
57669
57670#ifdef __LITTLE_ENDIAN__
57671#define vqrdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57672  int32x2_t __s0 = __p0; \
57673  int32x4_t __s1 = __p1; \
57674  int32x2_t __ret; \
57675  __ret = vqrdmulh_s32(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2)); \
57676  __ret; \
57677})
57678#else
57679#define vqrdmulh_laneq_s32(__p0, __p1, __p2) __extension__ ({ \
57680  int32x2_t __s0 = __p0; \
57681  int32x4_t __s1 = __p1; \
57682  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
57683  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
57684  int32x2_t __ret; \
57685  __ret = __noswap_vqrdmulh_s32(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2)); \
57686  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
57687  __ret; \
57688})
57689#endif
57690
57691#ifdef __LITTLE_ENDIAN__
57692#define vqrdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57693  int16x4_t __s0 = __p0; \
57694  int16x8_t __s1 = __p1; \
57695  int16x4_t __ret; \
57696  __ret = vqrdmulh_s16(__s0, __builtin_shufflevector(__s1, __s1, __p2, __p2, __p2, __p2)); \
57697  __ret; \
57698})
57699#else
57700#define vqrdmulh_laneq_s16(__p0, __p1, __p2) __extension__ ({ \
57701  int16x4_t __s0 = __p0; \
57702  int16x8_t __s1 = __p1; \
57703  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
57704  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 7, 6, 5, 4, 3, 2, 1, 0); \
57705  int16x4_t __ret; \
57706  __ret = __noswap_vqrdmulh_s16(__rev0, __builtin_shufflevector(__rev1, __rev1, __p2, __p2, __p2, __p2)); \
57707  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
57708  __ret; \
57709})
57710#endif
57711
57712#ifdef __LITTLE_ENDIAN__
57713__ai uint8_t vqrshlb_u8(uint8_t __p0, uint8_t __p1) {
57714  uint8_t __ret;
57715  __ret = (uint8_t) __builtin_neon_vqrshlb_u8(__p0, __p1);
57716  return __ret;
57717}
57718#else
57719__ai uint8_t vqrshlb_u8(uint8_t __p0, uint8_t __p1) {
57720  uint8_t __ret;
57721  __ret = (uint8_t) __builtin_neon_vqrshlb_u8(__p0, __p1);
57722  return __ret;
57723}
57724#endif
57725
57726#ifdef __LITTLE_ENDIAN__
57727__ai uint32_t vqrshls_u32(uint32_t __p0, uint32_t __p1) {
57728  uint32_t __ret;
57729  __ret = (uint32_t) __builtin_neon_vqrshls_u32(__p0, __p1);
57730  return __ret;
57731}
57732#else
57733__ai uint32_t vqrshls_u32(uint32_t __p0, uint32_t __p1) {
57734  uint32_t __ret;
57735  __ret = (uint32_t) __builtin_neon_vqrshls_u32(__p0, __p1);
57736  return __ret;
57737}
57738#endif
57739
57740#ifdef __LITTLE_ENDIAN__
57741__ai uint64_t vqrshld_u64(uint64_t __p0, uint64_t __p1) {
57742  uint64_t __ret;
57743  __ret = (uint64_t) __builtin_neon_vqrshld_u64(__p0, __p1);
57744  return __ret;
57745}
57746#else
57747__ai uint64_t vqrshld_u64(uint64_t __p0, uint64_t __p1) {
57748  uint64_t __ret;
57749  __ret = (uint64_t) __builtin_neon_vqrshld_u64(__p0, __p1);
57750  return __ret;
57751}
57752#endif
57753
57754#ifdef __LITTLE_ENDIAN__
57755__ai uint16_t vqrshlh_u16(uint16_t __p0, uint16_t __p1) {
57756  uint16_t __ret;
57757  __ret = (uint16_t) __builtin_neon_vqrshlh_u16(__p0, __p1);
57758  return __ret;
57759}
57760#else
57761__ai uint16_t vqrshlh_u16(uint16_t __p0, uint16_t __p1) {
57762  uint16_t __ret;
57763  __ret = (uint16_t) __builtin_neon_vqrshlh_u16(__p0, __p1);
57764  return __ret;
57765}
57766#endif
57767
57768#ifdef __LITTLE_ENDIAN__
57769__ai int8_t vqrshlb_s8(int8_t __p0, int8_t __p1) {
57770  int8_t __ret;
57771  __ret = (int8_t) __builtin_neon_vqrshlb_s8(__p0, __p1);
57772  return __ret;
57773}
57774#else
57775__ai int8_t vqrshlb_s8(int8_t __p0, int8_t __p1) {
57776  int8_t __ret;
57777  __ret = (int8_t) __builtin_neon_vqrshlb_s8(__p0, __p1);
57778  return __ret;
57779}
57780#endif
57781
57782#ifdef __LITTLE_ENDIAN__
57783__ai int32_t vqrshls_s32(int32_t __p0, int32_t __p1) {
57784  int32_t __ret;
57785  __ret = (int32_t) __builtin_neon_vqrshls_s32(__p0, __p1);
57786  return __ret;
57787}
57788#else
57789__ai int32_t vqrshls_s32(int32_t __p0, int32_t __p1) {
57790  int32_t __ret;
57791  __ret = (int32_t) __builtin_neon_vqrshls_s32(__p0, __p1);
57792  return __ret;
57793}
57794#endif
57795
57796#ifdef __LITTLE_ENDIAN__
57797__ai int64_t vqrshld_s64(int64_t __p0, int64_t __p1) {
57798  int64_t __ret;
57799  __ret = (int64_t) __builtin_neon_vqrshld_s64(__p0, __p1);
57800  return __ret;
57801}
57802#else
57803__ai int64_t vqrshld_s64(int64_t __p0, int64_t __p1) {
57804  int64_t __ret;
57805  __ret = (int64_t) __builtin_neon_vqrshld_s64(__p0, __p1);
57806  return __ret;
57807}
57808#endif
57809
57810#ifdef __LITTLE_ENDIAN__
57811__ai int16_t vqrshlh_s16(int16_t __p0, int16_t __p1) {
57812  int16_t __ret;
57813  __ret = (int16_t) __builtin_neon_vqrshlh_s16(__p0, __p1);
57814  return __ret;
57815}
57816#else
57817__ai int16_t vqrshlh_s16(int16_t __p0, int16_t __p1) {
57818  int16_t __ret;
57819  __ret = (int16_t) __builtin_neon_vqrshlh_s16(__p0, __p1);
57820  return __ret;
57821}
57822#endif
57823
57824#ifdef __LITTLE_ENDIAN__
57825#define vqrshrn_high_n_u32(__p0_170, __p1_170, __p2_170) __extension__ ({ \
57826  uint16x4_t __s0_170 = __p0_170; \
57827  uint32x4_t __s1_170 = __p1_170; \
57828  uint16x8_t __ret_170; \
57829  __ret_170 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_170), (uint16x4_t)(vqrshrn_n_u32(__s1_170, __p2_170)))); \
57830  __ret_170; \
57831})
57832#else
57833#define vqrshrn_high_n_u32(__p0_171, __p1_171, __p2_171) __extension__ ({ \
57834  uint16x4_t __s0_171 = __p0_171; \
57835  uint32x4_t __s1_171 = __p1_171; \
57836  uint16x4_t __rev0_171;  __rev0_171 = __builtin_shufflevector(__s0_171, __s0_171, 3, 2, 1, 0); \
57837  uint32x4_t __rev1_171;  __rev1_171 = __builtin_shufflevector(__s1_171, __s1_171, 3, 2, 1, 0); \
57838  uint16x8_t __ret_171; \
57839  __ret_171 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_171), (uint16x4_t)(__noswap_vqrshrn_n_u32(__rev1_171, __p2_171)))); \
57840  __ret_171 = __builtin_shufflevector(__ret_171, __ret_171, 7, 6, 5, 4, 3, 2, 1, 0); \
57841  __ret_171; \
57842})
57843#endif
57844
57845#ifdef __LITTLE_ENDIAN__
57846#define vqrshrn_high_n_u64(__p0_172, __p1_172, __p2_172) __extension__ ({ \
57847  uint32x2_t __s0_172 = __p0_172; \
57848  uint64x2_t __s1_172 = __p1_172; \
57849  uint32x4_t __ret_172; \
57850  __ret_172 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_172), (uint32x2_t)(vqrshrn_n_u64(__s1_172, __p2_172)))); \
57851  __ret_172; \
57852})
57853#else
57854#define vqrshrn_high_n_u64(__p0_173, __p1_173, __p2_173) __extension__ ({ \
57855  uint32x2_t __s0_173 = __p0_173; \
57856  uint64x2_t __s1_173 = __p1_173; \
57857  uint32x2_t __rev0_173;  __rev0_173 = __builtin_shufflevector(__s0_173, __s0_173, 1, 0); \
57858  uint64x2_t __rev1_173;  __rev1_173 = __builtin_shufflevector(__s1_173, __s1_173, 1, 0); \
57859  uint32x4_t __ret_173; \
57860  __ret_173 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_173), (uint32x2_t)(__noswap_vqrshrn_n_u64(__rev1_173, __p2_173)))); \
57861  __ret_173 = __builtin_shufflevector(__ret_173, __ret_173, 3, 2, 1, 0); \
57862  __ret_173; \
57863})
57864#endif
57865
57866#ifdef __LITTLE_ENDIAN__
57867#define vqrshrn_high_n_u16(__p0_174, __p1_174, __p2_174) __extension__ ({ \
57868  uint8x8_t __s0_174 = __p0_174; \
57869  uint16x8_t __s1_174 = __p1_174; \
57870  uint8x16_t __ret_174; \
57871  __ret_174 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_174), (uint8x8_t)(vqrshrn_n_u16(__s1_174, __p2_174)))); \
57872  __ret_174; \
57873})
57874#else
57875#define vqrshrn_high_n_u16(__p0_175, __p1_175, __p2_175) __extension__ ({ \
57876  uint8x8_t __s0_175 = __p0_175; \
57877  uint16x8_t __s1_175 = __p1_175; \
57878  uint8x8_t __rev0_175;  __rev0_175 = __builtin_shufflevector(__s0_175, __s0_175, 7, 6, 5, 4, 3, 2, 1, 0); \
57879  uint16x8_t __rev1_175;  __rev1_175 = __builtin_shufflevector(__s1_175, __s1_175, 7, 6, 5, 4, 3, 2, 1, 0); \
57880  uint8x16_t __ret_175; \
57881  __ret_175 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_175), (uint8x8_t)(__noswap_vqrshrn_n_u16(__rev1_175, __p2_175)))); \
57882  __ret_175 = __builtin_shufflevector(__ret_175, __ret_175, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
57883  __ret_175; \
57884})
57885#endif
57886
57887#ifdef __LITTLE_ENDIAN__
57888#define vqrshrn_high_n_s32(__p0_176, __p1_176, __p2_176) __extension__ ({ \
57889  int16x4_t __s0_176 = __p0_176; \
57890  int32x4_t __s1_176 = __p1_176; \
57891  int16x8_t __ret_176; \
57892  __ret_176 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_176), (int16x4_t)(vqrshrn_n_s32(__s1_176, __p2_176)))); \
57893  __ret_176; \
57894})
57895#else
57896#define vqrshrn_high_n_s32(__p0_177, __p1_177, __p2_177) __extension__ ({ \
57897  int16x4_t __s0_177 = __p0_177; \
57898  int32x4_t __s1_177 = __p1_177; \
57899  int16x4_t __rev0_177;  __rev0_177 = __builtin_shufflevector(__s0_177, __s0_177, 3, 2, 1, 0); \
57900  int32x4_t __rev1_177;  __rev1_177 = __builtin_shufflevector(__s1_177, __s1_177, 3, 2, 1, 0); \
57901  int16x8_t __ret_177; \
57902  __ret_177 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_177), (int16x4_t)(__noswap_vqrshrn_n_s32(__rev1_177, __p2_177)))); \
57903  __ret_177 = __builtin_shufflevector(__ret_177, __ret_177, 7, 6, 5, 4, 3, 2, 1, 0); \
57904  __ret_177; \
57905})
57906#endif
57907
57908#ifdef __LITTLE_ENDIAN__
57909#define vqrshrn_high_n_s64(__p0_178, __p1_178, __p2_178) __extension__ ({ \
57910  int32x2_t __s0_178 = __p0_178; \
57911  int64x2_t __s1_178 = __p1_178; \
57912  int32x4_t __ret_178; \
57913  __ret_178 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_178), (int32x2_t)(vqrshrn_n_s64(__s1_178, __p2_178)))); \
57914  __ret_178; \
57915})
57916#else
57917#define vqrshrn_high_n_s64(__p0_179, __p1_179, __p2_179) __extension__ ({ \
57918  int32x2_t __s0_179 = __p0_179; \
57919  int64x2_t __s1_179 = __p1_179; \
57920  int32x2_t __rev0_179;  __rev0_179 = __builtin_shufflevector(__s0_179, __s0_179, 1, 0); \
57921  int64x2_t __rev1_179;  __rev1_179 = __builtin_shufflevector(__s1_179, __s1_179, 1, 0); \
57922  int32x4_t __ret_179; \
57923  __ret_179 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_179), (int32x2_t)(__noswap_vqrshrn_n_s64(__rev1_179, __p2_179)))); \
57924  __ret_179 = __builtin_shufflevector(__ret_179, __ret_179, 3, 2, 1, 0); \
57925  __ret_179; \
57926})
57927#endif
57928
57929#ifdef __LITTLE_ENDIAN__
57930#define vqrshrn_high_n_s16(__p0_180, __p1_180, __p2_180) __extension__ ({ \
57931  int8x8_t __s0_180 = __p0_180; \
57932  int16x8_t __s1_180 = __p1_180; \
57933  int8x16_t __ret_180; \
57934  __ret_180 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_180), (int8x8_t)(vqrshrn_n_s16(__s1_180, __p2_180)))); \
57935  __ret_180; \
57936})
57937#else
57938#define vqrshrn_high_n_s16(__p0_181, __p1_181, __p2_181) __extension__ ({ \
57939  int8x8_t __s0_181 = __p0_181; \
57940  int16x8_t __s1_181 = __p1_181; \
57941  int8x8_t __rev0_181;  __rev0_181 = __builtin_shufflevector(__s0_181, __s0_181, 7, 6, 5, 4, 3, 2, 1, 0); \
57942  int16x8_t __rev1_181;  __rev1_181 = __builtin_shufflevector(__s1_181, __s1_181, 7, 6, 5, 4, 3, 2, 1, 0); \
57943  int8x16_t __ret_181; \
57944  __ret_181 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_181), (int8x8_t)(__noswap_vqrshrn_n_s16(__rev1_181, __p2_181)))); \
57945  __ret_181 = __builtin_shufflevector(__ret_181, __ret_181, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
57946  __ret_181; \
57947})
57948#endif
57949
57950#ifdef __LITTLE_ENDIAN__
57951#define vqrshrns_n_u32(__p0, __p1) __extension__ ({ \
57952  uint32_t __s0 = __p0; \
57953  uint16_t __ret; \
57954  __ret = (uint16_t) __builtin_neon_vqrshrns_n_u32(__s0, __p1); \
57955  __ret; \
57956})
57957#else
57958#define vqrshrns_n_u32(__p0, __p1) __extension__ ({ \
57959  uint32_t __s0 = __p0; \
57960  uint16_t __ret; \
57961  __ret = (uint16_t) __builtin_neon_vqrshrns_n_u32(__s0, __p1); \
57962  __ret; \
57963})
57964#endif
57965
57966#ifdef __LITTLE_ENDIAN__
57967#define vqrshrnd_n_u64(__p0, __p1) __extension__ ({ \
57968  uint64_t __s0 = __p0; \
57969  uint32_t __ret; \
57970  __ret = (uint32_t) __builtin_neon_vqrshrnd_n_u64(__s0, __p1); \
57971  __ret; \
57972})
57973#else
57974#define vqrshrnd_n_u64(__p0, __p1) __extension__ ({ \
57975  uint64_t __s0 = __p0; \
57976  uint32_t __ret; \
57977  __ret = (uint32_t) __builtin_neon_vqrshrnd_n_u64(__s0, __p1); \
57978  __ret; \
57979})
57980#endif
57981
57982#ifdef __LITTLE_ENDIAN__
57983#define vqrshrnh_n_u16(__p0, __p1) __extension__ ({ \
57984  uint16_t __s0 = __p0; \
57985  uint8_t __ret; \
57986  __ret = (uint8_t) __builtin_neon_vqrshrnh_n_u16(__s0, __p1); \
57987  __ret; \
57988})
57989#else
57990#define vqrshrnh_n_u16(__p0, __p1) __extension__ ({ \
57991  uint16_t __s0 = __p0; \
57992  uint8_t __ret; \
57993  __ret = (uint8_t) __builtin_neon_vqrshrnh_n_u16(__s0, __p1); \
57994  __ret; \
57995})
57996#endif
57997
57998#ifdef __LITTLE_ENDIAN__
57999#define vqrshrns_n_s32(__p0, __p1) __extension__ ({ \
58000  int32_t __s0 = __p0; \
58001  int16_t __ret; \
58002  __ret = (int16_t) __builtin_neon_vqrshrns_n_s32(__s0, __p1); \
58003  __ret; \
58004})
58005#else
58006#define vqrshrns_n_s32(__p0, __p1) __extension__ ({ \
58007  int32_t __s0 = __p0; \
58008  int16_t __ret; \
58009  __ret = (int16_t) __builtin_neon_vqrshrns_n_s32(__s0, __p1); \
58010  __ret; \
58011})
58012#endif
58013
58014#ifdef __LITTLE_ENDIAN__
58015#define vqrshrnd_n_s64(__p0, __p1) __extension__ ({ \
58016  int64_t __s0 = __p0; \
58017  int32_t __ret; \
58018  __ret = (int32_t) __builtin_neon_vqrshrnd_n_s64(__s0, __p1); \
58019  __ret; \
58020})
58021#else
58022#define vqrshrnd_n_s64(__p0, __p1) __extension__ ({ \
58023  int64_t __s0 = __p0; \
58024  int32_t __ret; \
58025  __ret = (int32_t) __builtin_neon_vqrshrnd_n_s64(__s0, __p1); \
58026  __ret; \
58027})
58028#endif
58029
58030#ifdef __LITTLE_ENDIAN__
58031#define vqrshrnh_n_s16(__p0, __p1) __extension__ ({ \
58032  int16_t __s0 = __p0; \
58033  int8_t __ret; \
58034  __ret = (int8_t) __builtin_neon_vqrshrnh_n_s16(__s0, __p1); \
58035  __ret; \
58036})
58037#else
58038#define vqrshrnh_n_s16(__p0, __p1) __extension__ ({ \
58039  int16_t __s0 = __p0; \
58040  int8_t __ret; \
58041  __ret = (int8_t) __builtin_neon_vqrshrnh_n_s16(__s0, __p1); \
58042  __ret; \
58043})
58044#endif
58045
58046#ifdef __LITTLE_ENDIAN__
58047#define vqrshrun_high_n_s32(__p0_182, __p1_182, __p2_182) __extension__ ({ \
58048  int16x4_t __s0_182 = __p0_182; \
58049  int32x4_t __s1_182 = __p1_182; \
58050  int16x8_t __ret_182; \
58051  __ret_182 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_182), (int16x4_t)(vqrshrun_n_s32(__s1_182, __p2_182)))); \
58052  __ret_182; \
58053})
58054#else
58055#define vqrshrun_high_n_s32(__p0_183, __p1_183, __p2_183) __extension__ ({ \
58056  int16x4_t __s0_183 = __p0_183; \
58057  int32x4_t __s1_183 = __p1_183; \
58058  int16x4_t __rev0_183;  __rev0_183 = __builtin_shufflevector(__s0_183, __s0_183, 3, 2, 1, 0); \
58059  int32x4_t __rev1_183;  __rev1_183 = __builtin_shufflevector(__s1_183, __s1_183, 3, 2, 1, 0); \
58060  int16x8_t __ret_183; \
58061  __ret_183 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_183), (int16x4_t)(__noswap_vqrshrun_n_s32(__rev1_183, __p2_183)))); \
58062  __ret_183 = __builtin_shufflevector(__ret_183, __ret_183, 7, 6, 5, 4, 3, 2, 1, 0); \
58063  __ret_183; \
58064})
58065#endif
58066
58067#ifdef __LITTLE_ENDIAN__
58068#define vqrshrun_high_n_s64(__p0_184, __p1_184, __p2_184) __extension__ ({ \
58069  int32x2_t __s0_184 = __p0_184; \
58070  int64x2_t __s1_184 = __p1_184; \
58071  int32x4_t __ret_184; \
58072  __ret_184 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_184), (int32x2_t)(vqrshrun_n_s64(__s1_184, __p2_184)))); \
58073  __ret_184; \
58074})
58075#else
58076#define vqrshrun_high_n_s64(__p0_185, __p1_185, __p2_185) __extension__ ({ \
58077  int32x2_t __s0_185 = __p0_185; \
58078  int64x2_t __s1_185 = __p1_185; \
58079  int32x2_t __rev0_185;  __rev0_185 = __builtin_shufflevector(__s0_185, __s0_185, 1, 0); \
58080  int64x2_t __rev1_185;  __rev1_185 = __builtin_shufflevector(__s1_185, __s1_185, 1, 0); \
58081  int32x4_t __ret_185; \
58082  __ret_185 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_185), (int32x2_t)(__noswap_vqrshrun_n_s64(__rev1_185, __p2_185)))); \
58083  __ret_185 = __builtin_shufflevector(__ret_185, __ret_185, 3, 2, 1, 0); \
58084  __ret_185; \
58085})
58086#endif
58087
58088#ifdef __LITTLE_ENDIAN__
58089#define vqrshrun_high_n_s16(__p0_186, __p1_186, __p2_186) __extension__ ({ \
58090  int8x8_t __s0_186 = __p0_186; \
58091  int16x8_t __s1_186 = __p1_186; \
58092  int8x16_t __ret_186; \
58093  __ret_186 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_186), (int8x8_t)(vqrshrun_n_s16(__s1_186, __p2_186)))); \
58094  __ret_186; \
58095})
58096#else
58097#define vqrshrun_high_n_s16(__p0_187, __p1_187, __p2_187) __extension__ ({ \
58098  int8x8_t __s0_187 = __p0_187; \
58099  int16x8_t __s1_187 = __p1_187; \
58100  int8x8_t __rev0_187;  __rev0_187 = __builtin_shufflevector(__s0_187, __s0_187, 7, 6, 5, 4, 3, 2, 1, 0); \
58101  int16x8_t __rev1_187;  __rev1_187 = __builtin_shufflevector(__s1_187, __s1_187, 7, 6, 5, 4, 3, 2, 1, 0); \
58102  int8x16_t __ret_187; \
58103  __ret_187 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_187), (int8x8_t)(__noswap_vqrshrun_n_s16(__rev1_187, __p2_187)))); \
58104  __ret_187 = __builtin_shufflevector(__ret_187, __ret_187, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58105  __ret_187; \
58106})
58107#endif
58108
58109#ifdef __LITTLE_ENDIAN__
58110#define vqrshruns_n_s32(__p0, __p1) __extension__ ({ \
58111  int32_t __s0 = __p0; \
58112  int16_t __ret; \
58113  __ret = (int16_t) __builtin_neon_vqrshruns_n_s32(__s0, __p1); \
58114  __ret; \
58115})
58116#else
58117#define vqrshruns_n_s32(__p0, __p1) __extension__ ({ \
58118  int32_t __s0 = __p0; \
58119  int16_t __ret; \
58120  __ret = (int16_t) __builtin_neon_vqrshruns_n_s32(__s0, __p1); \
58121  __ret; \
58122})
58123#endif
58124
58125#ifdef __LITTLE_ENDIAN__
58126#define vqrshrund_n_s64(__p0, __p1) __extension__ ({ \
58127  int64_t __s0 = __p0; \
58128  int32_t __ret; \
58129  __ret = (int32_t) __builtin_neon_vqrshrund_n_s64(__s0, __p1); \
58130  __ret; \
58131})
58132#else
58133#define vqrshrund_n_s64(__p0, __p1) __extension__ ({ \
58134  int64_t __s0 = __p0; \
58135  int32_t __ret; \
58136  __ret = (int32_t) __builtin_neon_vqrshrund_n_s64(__s0, __p1); \
58137  __ret; \
58138})
58139#endif
58140
58141#ifdef __LITTLE_ENDIAN__
58142#define vqrshrunh_n_s16(__p0, __p1) __extension__ ({ \
58143  int16_t __s0 = __p0; \
58144  int8_t __ret; \
58145  __ret = (int8_t) __builtin_neon_vqrshrunh_n_s16(__s0, __p1); \
58146  __ret; \
58147})
58148#else
58149#define vqrshrunh_n_s16(__p0, __p1) __extension__ ({ \
58150  int16_t __s0 = __p0; \
58151  int8_t __ret; \
58152  __ret = (int8_t) __builtin_neon_vqrshrunh_n_s16(__s0, __p1); \
58153  __ret; \
58154})
58155#endif
58156
58157#ifdef __LITTLE_ENDIAN__
58158__ai uint8_t vqshlb_u8(uint8_t __p0, uint8_t __p1) {
58159  uint8_t __ret;
58160  __ret = (uint8_t) __builtin_neon_vqshlb_u8(__p0, __p1);
58161  return __ret;
58162}
58163#else
58164__ai uint8_t vqshlb_u8(uint8_t __p0, uint8_t __p1) {
58165  uint8_t __ret;
58166  __ret = (uint8_t) __builtin_neon_vqshlb_u8(__p0, __p1);
58167  return __ret;
58168}
58169#endif
58170
58171#ifdef __LITTLE_ENDIAN__
58172__ai uint32_t vqshls_u32(uint32_t __p0, uint32_t __p1) {
58173  uint32_t __ret;
58174  __ret = (uint32_t) __builtin_neon_vqshls_u32(__p0, __p1);
58175  return __ret;
58176}
58177#else
58178__ai uint32_t vqshls_u32(uint32_t __p0, uint32_t __p1) {
58179  uint32_t __ret;
58180  __ret = (uint32_t) __builtin_neon_vqshls_u32(__p0, __p1);
58181  return __ret;
58182}
58183#endif
58184
58185#ifdef __LITTLE_ENDIAN__
58186__ai uint64_t vqshld_u64(uint64_t __p0, uint64_t __p1) {
58187  uint64_t __ret;
58188  __ret = (uint64_t) __builtin_neon_vqshld_u64(__p0, __p1);
58189  return __ret;
58190}
58191#else
58192__ai uint64_t vqshld_u64(uint64_t __p0, uint64_t __p1) {
58193  uint64_t __ret;
58194  __ret = (uint64_t) __builtin_neon_vqshld_u64(__p0, __p1);
58195  return __ret;
58196}
58197#endif
58198
58199#ifdef __LITTLE_ENDIAN__
58200__ai uint16_t vqshlh_u16(uint16_t __p0, uint16_t __p1) {
58201  uint16_t __ret;
58202  __ret = (uint16_t) __builtin_neon_vqshlh_u16(__p0, __p1);
58203  return __ret;
58204}
58205#else
58206__ai uint16_t vqshlh_u16(uint16_t __p0, uint16_t __p1) {
58207  uint16_t __ret;
58208  __ret = (uint16_t) __builtin_neon_vqshlh_u16(__p0, __p1);
58209  return __ret;
58210}
58211#endif
58212
58213#ifdef __LITTLE_ENDIAN__
58214__ai int8_t vqshlb_s8(int8_t __p0, int8_t __p1) {
58215  int8_t __ret;
58216  __ret = (int8_t) __builtin_neon_vqshlb_s8(__p0, __p1);
58217  return __ret;
58218}
58219#else
58220__ai int8_t vqshlb_s8(int8_t __p0, int8_t __p1) {
58221  int8_t __ret;
58222  __ret = (int8_t) __builtin_neon_vqshlb_s8(__p0, __p1);
58223  return __ret;
58224}
58225#endif
58226
58227#ifdef __LITTLE_ENDIAN__
58228__ai int32_t vqshls_s32(int32_t __p0, int32_t __p1) {
58229  int32_t __ret;
58230  __ret = (int32_t) __builtin_neon_vqshls_s32(__p0, __p1);
58231  return __ret;
58232}
58233#else
58234__ai int32_t vqshls_s32(int32_t __p0, int32_t __p1) {
58235  int32_t __ret;
58236  __ret = (int32_t) __builtin_neon_vqshls_s32(__p0, __p1);
58237  return __ret;
58238}
58239#endif
58240
58241#ifdef __LITTLE_ENDIAN__
58242__ai int64_t vqshld_s64(int64_t __p0, int64_t __p1) {
58243  int64_t __ret;
58244  __ret = (int64_t) __builtin_neon_vqshld_s64(__p0, __p1);
58245  return __ret;
58246}
58247#else
58248__ai int64_t vqshld_s64(int64_t __p0, int64_t __p1) {
58249  int64_t __ret;
58250  __ret = (int64_t) __builtin_neon_vqshld_s64(__p0, __p1);
58251  return __ret;
58252}
58253#endif
58254
58255#ifdef __LITTLE_ENDIAN__
58256__ai int16_t vqshlh_s16(int16_t __p0, int16_t __p1) {
58257  int16_t __ret;
58258  __ret = (int16_t) __builtin_neon_vqshlh_s16(__p0, __p1);
58259  return __ret;
58260}
58261#else
58262__ai int16_t vqshlh_s16(int16_t __p0, int16_t __p1) {
58263  int16_t __ret;
58264  __ret = (int16_t) __builtin_neon_vqshlh_s16(__p0, __p1);
58265  return __ret;
58266}
58267#endif
58268
58269#ifdef __LITTLE_ENDIAN__
58270#define vqshlb_n_u8(__p0, __p1) __extension__ ({ \
58271  uint8_t __s0 = __p0; \
58272  uint8_t __ret; \
58273  __ret = (uint8_t) __builtin_neon_vqshlb_n_u8(__s0, __p1); \
58274  __ret; \
58275})
58276#else
58277#define vqshlb_n_u8(__p0, __p1) __extension__ ({ \
58278  uint8_t __s0 = __p0; \
58279  uint8_t __ret; \
58280  __ret = (uint8_t) __builtin_neon_vqshlb_n_u8(__s0, __p1); \
58281  __ret; \
58282})
58283#endif
58284
58285#ifdef __LITTLE_ENDIAN__
58286#define vqshls_n_u32(__p0, __p1) __extension__ ({ \
58287  uint32_t __s0 = __p0; \
58288  uint32_t __ret; \
58289  __ret = (uint32_t) __builtin_neon_vqshls_n_u32(__s0, __p1); \
58290  __ret; \
58291})
58292#else
58293#define vqshls_n_u32(__p0, __p1) __extension__ ({ \
58294  uint32_t __s0 = __p0; \
58295  uint32_t __ret; \
58296  __ret = (uint32_t) __builtin_neon_vqshls_n_u32(__s0, __p1); \
58297  __ret; \
58298})
58299#endif
58300
58301#ifdef __LITTLE_ENDIAN__
58302#define vqshld_n_u64(__p0, __p1) __extension__ ({ \
58303  uint64_t __s0 = __p0; \
58304  uint64_t __ret; \
58305  __ret = (uint64_t) __builtin_neon_vqshld_n_u64(__s0, __p1); \
58306  __ret; \
58307})
58308#else
58309#define vqshld_n_u64(__p0, __p1) __extension__ ({ \
58310  uint64_t __s0 = __p0; \
58311  uint64_t __ret; \
58312  __ret = (uint64_t) __builtin_neon_vqshld_n_u64(__s0, __p1); \
58313  __ret; \
58314})
58315#endif
58316
58317#ifdef __LITTLE_ENDIAN__
58318#define vqshlh_n_u16(__p0, __p1) __extension__ ({ \
58319  uint16_t __s0 = __p0; \
58320  uint16_t __ret; \
58321  __ret = (uint16_t) __builtin_neon_vqshlh_n_u16(__s0, __p1); \
58322  __ret; \
58323})
58324#else
58325#define vqshlh_n_u16(__p0, __p1) __extension__ ({ \
58326  uint16_t __s0 = __p0; \
58327  uint16_t __ret; \
58328  __ret = (uint16_t) __builtin_neon_vqshlh_n_u16(__s0, __p1); \
58329  __ret; \
58330})
58331#endif
58332
58333#ifdef __LITTLE_ENDIAN__
58334#define vqshlb_n_s8(__p0, __p1) __extension__ ({ \
58335  int8_t __s0 = __p0; \
58336  int8_t __ret; \
58337  __ret = (int8_t) __builtin_neon_vqshlb_n_s8(__s0, __p1); \
58338  __ret; \
58339})
58340#else
58341#define vqshlb_n_s8(__p0, __p1) __extension__ ({ \
58342  int8_t __s0 = __p0; \
58343  int8_t __ret; \
58344  __ret = (int8_t) __builtin_neon_vqshlb_n_s8(__s0, __p1); \
58345  __ret; \
58346})
58347#endif
58348
58349#ifdef __LITTLE_ENDIAN__
58350#define vqshls_n_s32(__p0, __p1) __extension__ ({ \
58351  int32_t __s0 = __p0; \
58352  int32_t __ret; \
58353  __ret = (int32_t) __builtin_neon_vqshls_n_s32(__s0, __p1); \
58354  __ret; \
58355})
58356#else
58357#define vqshls_n_s32(__p0, __p1) __extension__ ({ \
58358  int32_t __s0 = __p0; \
58359  int32_t __ret; \
58360  __ret = (int32_t) __builtin_neon_vqshls_n_s32(__s0, __p1); \
58361  __ret; \
58362})
58363#endif
58364
58365#ifdef __LITTLE_ENDIAN__
58366#define vqshld_n_s64(__p0, __p1) __extension__ ({ \
58367  int64_t __s0 = __p0; \
58368  int64_t __ret; \
58369  __ret = (int64_t) __builtin_neon_vqshld_n_s64(__s0, __p1); \
58370  __ret; \
58371})
58372#else
58373#define vqshld_n_s64(__p0, __p1) __extension__ ({ \
58374  int64_t __s0 = __p0; \
58375  int64_t __ret; \
58376  __ret = (int64_t) __builtin_neon_vqshld_n_s64(__s0, __p1); \
58377  __ret; \
58378})
58379#endif
58380
58381#ifdef __LITTLE_ENDIAN__
58382#define vqshlh_n_s16(__p0, __p1) __extension__ ({ \
58383  int16_t __s0 = __p0; \
58384  int16_t __ret; \
58385  __ret = (int16_t) __builtin_neon_vqshlh_n_s16(__s0, __p1); \
58386  __ret; \
58387})
58388#else
58389#define vqshlh_n_s16(__p0, __p1) __extension__ ({ \
58390  int16_t __s0 = __p0; \
58391  int16_t __ret; \
58392  __ret = (int16_t) __builtin_neon_vqshlh_n_s16(__s0, __p1); \
58393  __ret; \
58394})
58395#endif
58396
58397#ifdef __LITTLE_ENDIAN__
58398#define vqshlub_n_s8(__p0, __p1) __extension__ ({ \
58399  int8_t __s0 = __p0; \
58400  int8_t __ret; \
58401  __ret = (int8_t) __builtin_neon_vqshlub_n_s8(__s0, __p1); \
58402  __ret; \
58403})
58404#else
58405#define vqshlub_n_s8(__p0, __p1) __extension__ ({ \
58406  int8_t __s0 = __p0; \
58407  int8_t __ret; \
58408  __ret = (int8_t) __builtin_neon_vqshlub_n_s8(__s0, __p1); \
58409  __ret; \
58410})
58411#endif
58412
58413#ifdef __LITTLE_ENDIAN__
58414#define vqshlus_n_s32(__p0, __p1) __extension__ ({ \
58415  int32_t __s0 = __p0; \
58416  int32_t __ret; \
58417  __ret = (int32_t) __builtin_neon_vqshlus_n_s32(__s0, __p1); \
58418  __ret; \
58419})
58420#else
58421#define vqshlus_n_s32(__p0, __p1) __extension__ ({ \
58422  int32_t __s0 = __p0; \
58423  int32_t __ret; \
58424  __ret = (int32_t) __builtin_neon_vqshlus_n_s32(__s0, __p1); \
58425  __ret; \
58426})
58427#endif
58428
58429#ifdef __LITTLE_ENDIAN__
58430#define vqshlud_n_s64(__p0, __p1) __extension__ ({ \
58431  int64_t __s0 = __p0; \
58432  int64_t __ret; \
58433  __ret = (int64_t) __builtin_neon_vqshlud_n_s64(__s0, __p1); \
58434  __ret; \
58435})
58436#else
58437#define vqshlud_n_s64(__p0, __p1) __extension__ ({ \
58438  int64_t __s0 = __p0; \
58439  int64_t __ret; \
58440  __ret = (int64_t) __builtin_neon_vqshlud_n_s64(__s0, __p1); \
58441  __ret; \
58442})
58443#endif
58444
58445#ifdef __LITTLE_ENDIAN__
58446#define vqshluh_n_s16(__p0, __p1) __extension__ ({ \
58447  int16_t __s0 = __p0; \
58448  int16_t __ret; \
58449  __ret = (int16_t) __builtin_neon_vqshluh_n_s16(__s0, __p1); \
58450  __ret; \
58451})
58452#else
58453#define vqshluh_n_s16(__p0, __p1) __extension__ ({ \
58454  int16_t __s0 = __p0; \
58455  int16_t __ret; \
58456  __ret = (int16_t) __builtin_neon_vqshluh_n_s16(__s0, __p1); \
58457  __ret; \
58458})
58459#endif
58460
58461#ifdef __LITTLE_ENDIAN__
58462#define vqshrn_high_n_u32(__p0_188, __p1_188, __p2_188) __extension__ ({ \
58463  uint16x4_t __s0_188 = __p0_188; \
58464  uint32x4_t __s1_188 = __p1_188; \
58465  uint16x8_t __ret_188; \
58466  __ret_188 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_188), (uint16x4_t)(vqshrn_n_u32(__s1_188, __p2_188)))); \
58467  __ret_188; \
58468})
58469#else
58470#define vqshrn_high_n_u32(__p0_189, __p1_189, __p2_189) __extension__ ({ \
58471  uint16x4_t __s0_189 = __p0_189; \
58472  uint32x4_t __s1_189 = __p1_189; \
58473  uint16x4_t __rev0_189;  __rev0_189 = __builtin_shufflevector(__s0_189, __s0_189, 3, 2, 1, 0); \
58474  uint32x4_t __rev1_189;  __rev1_189 = __builtin_shufflevector(__s1_189, __s1_189, 3, 2, 1, 0); \
58475  uint16x8_t __ret_189; \
58476  __ret_189 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_189), (uint16x4_t)(__noswap_vqshrn_n_u32(__rev1_189, __p2_189)))); \
58477  __ret_189 = __builtin_shufflevector(__ret_189, __ret_189, 7, 6, 5, 4, 3, 2, 1, 0); \
58478  __ret_189; \
58479})
58480#endif
58481
58482#ifdef __LITTLE_ENDIAN__
58483#define vqshrn_high_n_u64(__p0_190, __p1_190, __p2_190) __extension__ ({ \
58484  uint32x2_t __s0_190 = __p0_190; \
58485  uint64x2_t __s1_190 = __p1_190; \
58486  uint32x4_t __ret_190; \
58487  __ret_190 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_190), (uint32x2_t)(vqshrn_n_u64(__s1_190, __p2_190)))); \
58488  __ret_190; \
58489})
58490#else
58491#define vqshrn_high_n_u64(__p0_191, __p1_191, __p2_191) __extension__ ({ \
58492  uint32x2_t __s0_191 = __p0_191; \
58493  uint64x2_t __s1_191 = __p1_191; \
58494  uint32x2_t __rev0_191;  __rev0_191 = __builtin_shufflevector(__s0_191, __s0_191, 1, 0); \
58495  uint64x2_t __rev1_191;  __rev1_191 = __builtin_shufflevector(__s1_191, __s1_191, 1, 0); \
58496  uint32x4_t __ret_191; \
58497  __ret_191 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_191), (uint32x2_t)(__noswap_vqshrn_n_u64(__rev1_191, __p2_191)))); \
58498  __ret_191 = __builtin_shufflevector(__ret_191, __ret_191, 3, 2, 1, 0); \
58499  __ret_191; \
58500})
58501#endif
58502
58503#ifdef __LITTLE_ENDIAN__
58504#define vqshrn_high_n_u16(__p0_192, __p1_192, __p2_192) __extension__ ({ \
58505  uint8x8_t __s0_192 = __p0_192; \
58506  uint16x8_t __s1_192 = __p1_192; \
58507  uint8x16_t __ret_192; \
58508  __ret_192 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_192), (uint8x8_t)(vqshrn_n_u16(__s1_192, __p2_192)))); \
58509  __ret_192; \
58510})
58511#else
58512#define vqshrn_high_n_u16(__p0_193, __p1_193, __p2_193) __extension__ ({ \
58513  uint8x8_t __s0_193 = __p0_193; \
58514  uint16x8_t __s1_193 = __p1_193; \
58515  uint8x8_t __rev0_193;  __rev0_193 = __builtin_shufflevector(__s0_193, __s0_193, 7, 6, 5, 4, 3, 2, 1, 0); \
58516  uint16x8_t __rev1_193;  __rev1_193 = __builtin_shufflevector(__s1_193, __s1_193, 7, 6, 5, 4, 3, 2, 1, 0); \
58517  uint8x16_t __ret_193; \
58518  __ret_193 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_193), (uint8x8_t)(__noswap_vqshrn_n_u16(__rev1_193, __p2_193)))); \
58519  __ret_193 = __builtin_shufflevector(__ret_193, __ret_193, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58520  __ret_193; \
58521})
58522#endif
58523
58524#ifdef __LITTLE_ENDIAN__
58525#define vqshrn_high_n_s32(__p0_194, __p1_194, __p2_194) __extension__ ({ \
58526  int16x4_t __s0_194 = __p0_194; \
58527  int32x4_t __s1_194 = __p1_194; \
58528  int16x8_t __ret_194; \
58529  __ret_194 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_194), (int16x4_t)(vqshrn_n_s32(__s1_194, __p2_194)))); \
58530  __ret_194; \
58531})
58532#else
58533#define vqshrn_high_n_s32(__p0_195, __p1_195, __p2_195) __extension__ ({ \
58534  int16x4_t __s0_195 = __p0_195; \
58535  int32x4_t __s1_195 = __p1_195; \
58536  int16x4_t __rev0_195;  __rev0_195 = __builtin_shufflevector(__s0_195, __s0_195, 3, 2, 1, 0); \
58537  int32x4_t __rev1_195;  __rev1_195 = __builtin_shufflevector(__s1_195, __s1_195, 3, 2, 1, 0); \
58538  int16x8_t __ret_195; \
58539  __ret_195 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_195), (int16x4_t)(__noswap_vqshrn_n_s32(__rev1_195, __p2_195)))); \
58540  __ret_195 = __builtin_shufflevector(__ret_195, __ret_195, 7, 6, 5, 4, 3, 2, 1, 0); \
58541  __ret_195; \
58542})
58543#endif
58544
58545#ifdef __LITTLE_ENDIAN__
58546#define vqshrn_high_n_s64(__p0_196, __p1_196, __p2_196) __extension__ ({ \
58547  int32x2_t __s0_196 = __p0_196; \
58548  int64x2_t __s1_196 = __p1_196; \
58549  int32x4_t __ret_196; \
58550  __ret_196 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_196), (int32x2_t)(vqshrn_n_s64(__s1_196, __p2_196)))); \
58551  __ret_196; \
58552})
58553#else
58554#define vqshrn_high_n_s64(__p0_197, __p1_197, __p2_197) __extension__ ({ \
58555  int32x2_t __s0_197 = __p0_197; \
58556  int64x2_t __s1_197 = __p1_197; \
58557  int32x2_t __rev0_197;  __rev0_197 = __builtin_shufflevector(__s0_197, __s0_197, 1, 0); \
58558  int64x2_t __rev1_197;  __rev1_197 = __builtin_shufflevector(__s1_197, __s1_197, 1, 0); \
58559  int32x4_t __ret_197; \
58560  __ret_197 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_197), (int32x2_t)(__noswap_vqshrn_n_s64(__rev1_197, __p2_197)))); \
58561  __ret_197 = __builtin_shufflevector(__ret_197, __ret_197, 3, 2, 1, 0); \
58562  __ret_197; \
58563})
58564#endif
58565
58566#ifdef __LITTLE_ENDIAN__
58567#define vqshrn_high_n_s16(__p0_198, __p1_198, __p2_198) __extension__ ({ \
58568  int8x8_t __s0_198 = __p0_198; \
58569  int16x8_t __s1_198 = __p1_198; \
58570  int8x16_t __ret_198; \
58571  __ret_198 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_198), (int8x8_t)(vqshrn_n_s16(__s1_198, __p2_198)))); \
58572  __ret_198; \
58573})
58574#else
58575#define vqshrn_high_n_s16(__p0_199, __p1_199, __p2_199) __extension__ ({ \
58576  int8x8_t __s0_199 = __p0_199; \
58577  int16x8_t __s1_199 = __p1_199; \
58578  int8x8_t __rev0_199;  __rev0_199 = __builtin_shufflevector(__s0_199, __s0_199, 7, 6, 5, 4, 3, 2, 1, 0); \
58579  int16x8_t __rev1_199;  __rev1_199 = __builtin_shufflevector(__s1_199, __s1_199, 7, 6, 5, 4, 3, 2, 1, 0); \
58580  int8x16_t __ret_199; \
58581  __ret_199 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_199), (int8x8_t)(__noswap_vqshrn_n_s16(__rev1_199, __p2_199)))); \
58582  __ret_199 = __builtin_shufflevector(__ret_199, __ret_199, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58583  __ret_199; \
58584})
58585#endif
58586
58587#ifdef __LITTLE_ENDIAN__
58588#define vqshrns_n_u32(__p0, __p1) __extension__ ({ \
58589  uint32_t __s0 = __p0; \
58590  uint16_t __ret; \
58591  __ret = (uint16_t) __builtin_neon_vqshrns_n_u32(__s0, __p1); \
58592  __ret; \
58593})
58594#else
58595#define vqshrns_n_u32(__p0, __p1) __extension__ ({ \
58596  uint32_t __s0 = __p0; \
58597  uint16_t __ret; \
58598  __ret = (uint16_t) __builtin_neon_vqshrns_n_u32(__s0, __p1); \
58599  __ret; \
58600})
58601#endif
58602
58603#ifdef __LITTLE_ENDIAN__
58604#define vqshrnd_n_u64(__p0, __p1) __extension__ ({ \
58605  uint64_t __s0 = __p0; \
58606  uint32_t __ret; \
58607  __ret = (uint32_t) __builtin_neon_vqshrnd_n_u64(__s0, __p1); \
58608  __ret; \
58609})
58610#else
58611#define vqshrnd_n_u64(__p0, __p1) __extension__ ({ \
58612  uint64_t __s0 = __p0; \
58613  uint32_t __ret; \
58614  __ret = (uint32_t) __builtin_neon_vqshrnd_n_u64(__s0, __p1); \
58615  __ret; \
58616})
58617#endif
58618
58619#ifdef __LITTLE_ENDIAN__
58620#define vqshrnh_n_u16(__p0, __p1) __extension__ ({ \
58621  uint16_t __s0 = __p0; \
58622  uint8_t __ret; \
58623  __ret = (uint8_t) __builtin_neon_vqshrnh_n_u16(__s0, __p1); \
58624  __ret; \
58625})
58626#else
58627#define vqshrnh_n_u16(__p0, __p1) __extension__ ({ \
58628  uint16_t __s0 = __p0; \
58629  uint8_t __ret; \
58630  __ret = (uint8_t) __builtin_neon_vqshrnh_n_u16(__s0, __p1); \
58631  __ret; \
58632})
58633#endif
58634
58635#ifdef __LITTLE_ENDIAN__
58636#define vqshrns_n_s32(__p0, __p1) __extension__ ({ \
58637  int32_t __s0 = __p0; \
58638  int16_t __ret; \
58639  __ret = (int16_t) __builtin_neon_vqshrns_n_s32(__s0, __p1); \
58640  __ret; \
58641})
58642#else
58643#define vqshrns_n_s32(__p0, __p1) __extension__ ({ \
58644  int32_t __s0 = __p0; \
58645  int16_t __ret; \
58646  __ret = (int16_t) __builtin_neon_vqshrns_n_s32(__s0, __p1); \
58647  __ret; \
58648})
58649#endif
58650
58651#ifdef __LITTLE_ENDIAN__
58652#define vqshrnd_n_s64(__p0, __p1) __extension__ ({ \
58653  int64_t __s0 = __p0; \
58654  int32_t __ret; \
58655  __ret = (int32_t) __builtin_neon_vqshrnd_n_s64(__s0, __p1); \
58656  __ret; \
58657})
58658#else
58659#define vqshrnd_n_s64(__p0, __p1) __extension__ ({ \
58660  int64_t __s0 = __p0; \
58661  int32_t __ret; \
58662  __ret = (int32_t) __builtin_neon_vqshrnd_n_s64(__s0, __p1); \
58663  __ret; \
58664})
58665#endif
58666
58667#ifdef __LITTLE_ENDIAN__
58668#define vqshrnh_n_s16(__p0, __p1) __extension__ ({ \
58669  int16_t __s0 = __p0; \
58670  int8_t __ret; \
58671  __ret = (int8_t) __builtin_neon_vqshrnh_n_s16(__s0, __p1); \
58672  __ret; \
58673})
58674#else
58675#define vqshrnh_n_s16(__p0, __p1) __extension__ ({ \
58676  int16_t __s0 = __p0; \
58677  int8_t __ret; \
58678  __ret = (int8_t) __builtin_neon_vqshrnh_n_s16(__s0, __p1); \
58679  __ret; \
58680})
58681#endif
58682
58683#ifdef __LITTLE_ENDIAN__
58684#define vqshrun_high_n_s32(__p0_200, __p1_200, __p2_200) __extension__ ({ \
58685  int16x4_t __s0_200 = __p0_200; \
58686  int32x4_t __s1_200 = __p1_200; \
58687  int16x8_t __ret_200; \
58688  __ret_200 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_200), (int16x4_t)(vqshrun_n_s32(__s1_200, __p2_200)))); \
58689  __ret_200; \
58690})
58691#else
58692#define vqshrun_high_n_s32(__p0_201, __p1_201, __p2_201) __extension__ ({ \
58693  int16x4_t __s0_201 = __p0_201; \
58694  int32x4_t __s1_201 = __p1_201; \
58695  int16x4_t __rev0_201;  __rev0_201 = __builtin_shufflevector(__s0_201, __s0_201, 3, 2, 1, 0); \
58696  int32x4_t __rev1_201;  __rev1_201 = __builtin_shufflevector(__s1_201, __s1_201, 3, 2, 1, 0); \
58697  int16x8_t __ret_201; \
58698  __ret_201 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_201), (int16x4_t)(__noswap_vqshrun_n_s32(__rev1_201, __p2_201)))); \
58699  __ret_201 = __builtin_shufflevector(__ret_201, __ret_201, 7, 6, 5, 4, 3, 2, 1, 0); \
58700  __ret_201; \
58701})
58702#endif
58703
58704#ifdef __LITTLE_ENDIAN__
58705#define vqshrun_high_n_s64(__p0_202, __p1_202, __p2_202) __extension__ ({ \
58706  int32x2_t __s0_202 = __p0_202; \
58707  int64x2_t __s1_202 = __p1_202; \
58708  int32x4_t __ret_202; \
58709  __ret_202 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_202), (int32x2_t)(vqshrun_n_s64(__s1_202, __p2_202)))); \
58710  __ret_202; \
58711})
58712#else
58713#define vqshrun_high_n_s64(__p0_203, __p1_203, __p2_203) __extension__ ({ \
58714  int32x2_t __s0_203 = __p0_203; \
58715  int64x2_t __s1_203 = __p1_203; \
58716  int32x2_t __rev0_203;  __rev0_203 = __builtin_shufflevector(__s0_203, __s0_203, 1, 0); \
58717  int64x2_t __rev1_203;  __rev1_203 = __builtin_shufflevector(__s1_203, __s1_203, 1, 0); \
58718  int32x4_t __ret_203; \
58719  __ret_203 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_203), (int32x2_t)(__noswap_vqshrun_n_s64(__rev1_203, __p2_203)))); \
58720  __ret_203 = __builtin_shufflevector(__ret_203, __ret_203, 3, 2, 1, 0); \
58721  __ret_203; \
58722})
58723#endif
58724
58725#ifdef __LITTLE_ENDIAN__
58726#define vqshrun_high_n_s16(__p0_204, __p1_204, __p2_204) __extension__ ({ \
58727  int8x8_t __s0_204 = __p0_204; \
58728  int16x8_t __s1_204 = __p1_204; \
58729  int8x16_t __ret_204; \
58730  __ret_204 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_204), (int8x8_t)(vqshrun_n_s16(__s1_204, __p2_204)))); \
58731  __ret_204; \
58732})
58733#else
58734#define vqshrun_high_n_s16(__p0_205, __p1_205, __p2_205) __extension__ ({ \
58735  int8x8_t __s0_205 = __p0_205; \
58736  int16x8_t __s1_205 = __p1_205; \
58737  int8x8_t __rev0_205;  __rev0_205 = __builtin_shufflevector(__s0_205, __s0_205, 7, 6, 5, 4, 3, 2, 1, 0); \
58738  int16x8_t __rev1_205;  __rev1_205 = __builtin_shufflevector(__s1_205, __s1_205, 7, 6, 5, 4, 3, 2, 1, 0); \
58739  int8x16_t __ret_205; \
58740  __ret_205 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_205), (int8x8_t)(__noswap_vqshrun_n_s16(__rev1_205, __p2_205)))); \
58741  __ret_205 = __builtin_shufflevector(__ret_205, __ret_205, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
58742  __ret_205; \
58743})
58744#endif
58745
58746#ifdef __LITTLE_ENDIAN__
58747#define vqshruns_n_s32(__p0, __p1) __extension__ ({ \
58748  int32_t __s0 = __p0; \
58749  int16_t __ret; \
58750  __ret = (int16_t) __builtin_neon_vqshruns_n_s32(__s0, __p1); \
58751  __ret; \
58752})
58753#else
58754#define vqshruns_n_s32(__p0, __p1) __extension__ ({ \
58755  int32_t __s0 = __p0; \
58756  int16_t __ret; \
58757  __ret = (int16_t) __builtin_neon_vqshruns_n_s32(__s0, __p1); \
58758  __ret; \
58759})
58760#endif
58761
58762#ifdef __LITTLE_ENDIAN__
58763#define vqshrund_n_s64(__p0, __p1) __extension__ ({ \
58764  int64_t __s0 = __p0; \
58765  int32_t __ret; \
58766  __ret = (int32_t) __builtin_neon_vqshrund_n_s64(__s0, __p1); \
58767  __ret; \
58768})
58769#else
58770#define vqshrund_n_s64(__p0, __p1) __extension__ ({ \
58771  int64_t __s0 = __p0; \
58772  int32_t __ret; \
58773  __ret = (int32_t) __builtin_neon_vqshrund_n_s64(__s0, __p1); \
58774  __ret; \
58775})
58776#endif
58777
58778#ifdef __LITTLE_ENDIAN__
58779#define vqshrunh_n_s16(__p0, __p1) __extension__ ({ \
58780  int16_t __s0 = __p0; \
58781  int8_t __ret; \
58782  __ret = (int8_t) __builtin_neon_vqshrunh_n_s16(__s0, __p1); \
58783  __ret; \
58784})
58785#else
58786#define vqshrunh_n_s16(__p0, __p1) __extension__ ({ \
58787  int16_t __s0 = __p0; \
58788  int8_t __ret; \
58789  __ret = (int8_t) __builtin_neon_vqshrunh_n_s16(__s0, __p1); \
58790  __ret; \
58791})
58792#endif
58793
58794#ifdef __LITTLE_ENDIAN__
58795__ai uint8_t vqsubb_u8(uint8_t __p0, uint8_t __p1) {
58796  uint8_t __ret;
58797  __ret = (uint8_t) __builtin_neon_vqsubb_u8(__p0, __p1);
58798  return __ret;
58799}
58800#else
58801__ai uint8_t vqsubb_u8(uint8_t __p0, uint8_t __p1) {
58802  uint8_t __ret;
58803  __ret = (uint8_t) __builtin_neon_vqsubb_u8(__p0, __p1);
58804  return __ret;
58805}
58806#endif
58807
58808#ifdef __LITTLE_ENDIAN__
58809__ai uint32_t vqsubs_u32(uint32_t __p0, uint32_t __p1) {
58810  uint32_t __ret;
58811  __ret = (uint32_t) __builtin_neon_vqsubs_u32(__p0, __p1);
58812  return __ret;
58813}
58814#else
58815__ai uint32_t vqsubs_u32(uint32_t __p0, uint32_t __p1) {
58816  uint32_t __ret;
58817  __ret = (uint32_t) __builtin_neon_vqsubs_u32(__p0, __p1);
58818  return __ret;
58819}
58820#endif
58821
58822#ifdef __LITTLE_ENDIAN__
58823__ai uint64_t vqsubd_u64(uint64_t __p0, uint64_t __p1) {
58824  uint64_t __ret;
58825  __ret = (uint64_t) __builtin_neon_vqsubd_u64(__p0, __p1);
58826  return __ret;
58827}
58828#else
58829__ai uint64_t vqsubd_u64(uint64_t __p0, uint64_t __p1) {
58830  uint64_t __ret;
58831  __ret = (uint64_t) __builtin_neon_vqsubd_u64(__p0, __p1);
58832  return __ret;
58833}
58834#endif
58835
58836#ifdef __LITTLE_ENDIAN__
58837__ai uint16_t vqsubh_u16(uint16_t __p0, uint16_t __p1) {
58838  uint16_t __ret;
58839  __ret = (uint16_t) __builtin_neon_vqsubh_u16(__p0, __p1);
58840  return __ret;
58841}
58842#else
58843__ai uint16_t vqsubh_u16(uint16_t __p0, uint16_t __p1) {
58844  uint16_t __ret;
58845  __ret = (uint16_t) __builtin_neon_vqsubh_u16(__p0, __p1);
58846  return __ret;
58847}
58848#endif
58849
58850#ifdef __LITTLE_ENDIAN__
58851__ai int8_t vqsubb_s8(int8_t __p0, int8_t __p1) {
58852  int8_t __ret;
58853  __ret = (int8_t) __builtin_neon_vqsubb_s8(__p0, __p1);
58854  return __ret;
58855}
58856#else
58857__ai int8_t vqsubb_s8(int8_t __p0, int8_t __p1) {
58858  int8_t __ret;
58859  __ret = (int8_t) __builtin_neon_vqsubb_s8(__p0, __p1);
58860  return __ret;
58861}
58862#endif
58863
58864#ifdef __LITTLE_ENDIAN__
58865__ai int32_t vqsubs_s32(int32_t __p0, int32_t __p1) {
58866  int32_t __ret;
58867  __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
58868  return __ret;
58869}
58870#else
58871__ai int32_t vqsubs_s32(int32_t __p0, int32_t __p1) {
58872  int32_t __ret;
58873  __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
58874  return __ret;
58875}
58876__ai int32_t __noswap_vqsubs_s32(int32_t __p0, int32_t __p1) {
58877  int32_t __ret;
58878  __ret = (int32_t) __builtin_neon_vqsubs_s32(__p0, __p1);
58879  return __ret;
58880}
58881#endif
58882
58883#ifdef __LITTLE_ENDIAN__
58884__ai int64_t vqsubd_s64(int64_t __p0, int64_t __p1) {
58885  int64_t __ret;
58886  __ret = (int64_t) __builtin_neon_vqsubd_s64(__p0, __p1);
58887  return __ret;
58888}
58889#else
58890__ai int64_t vqsubd_s64(int64_t __p0, int64_t __p1) {
58891  int64_t __ret;
58892  __ret = (int64_t) __builtin_neon_vqsubd_s64(__p0, __p1);
58893  return __ret;
58894}
58895#endif
58896
58897#ifdef __LITTLE_ENDIAN__
58898__ai int16_t vqsubh_s16(int16_t __p0, int16_t __p1) {
58899  int16_t __ret;
58900  __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
58901  return __ret;
58902}
58903#else
58904__ai int16_t vqsubh_s16(int16_t __p0, int16_t __p1) {
58905  int16_t __ret;
58906  __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
58907  return __ret;
58908}
58909__ai int16_t __noswap_vqsubh_s16(int16_t __p0, int16_t __p1) {
58910  int16_t __ret;
58911  __ret = (int16_t) __builtin_neon_vqsubh_s16(__p0, __p1);
58912  return __ret;
58913}
58914#endif
58915
58916#ifdef __LITTLE_ENDIAN__
58917__ai poly8x8_t vqtbl1_p8(poly8x16_t __p0, uint8x8_t __p1) {
58918  poly8x8_t __ret;
58919  __ret = (poly8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 4);
58920  return __ret;
58921}
58922#else
58923__ai poly8x8_t vqtbl1_p8(poly8x16_t __p0, uint8x8_t __p1) {
58924  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58925  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58926  poly8x8_t __ret;
58927  __ret = (poly8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 4);
58928  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58929  return __ret;
58930}
58931#endif
58932
58933#ifdef __LITTLE_ENDIAN__
58934__ai poly8x16_t vqtbl1q_p8(poly8x16_t __p0, uint8x16_t __p1) {
58935  poly8x16_t __ret;
58936  __ret = (poly8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 36);
58937  return __ret;
58938}
58939#else
58940__ai poly8x16_t vqtbl1q_p8(poly8x16_t __p0, uint8x16_t __p1) {
58941  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58942  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58943  poly8x16_t __ret;
58944  __ret = (poly8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36);
58945  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58946  return __ret;
58947}
58948#endif
58949
58950#ifdef __LITTLE_ENDIAN__
58951__ai uint8x16_t vqtbl1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
58952  uint8x16_t __ret;
58953  __ret = (uint8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
58954  return __ret;
58955}
58956#else
58957__ai uint8x16_t vqtbl1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
58958  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58959  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58960  uint8x16_t __ret;
58961  __ret = (uint8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
58962  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58963  return __ret;
58964}
58965#endif
58966
58967#ifdef __LITTLE_ENDIAN__
58968__ai int8x16_t vqtbl1q_s8(int8x16_t __p0, int8x16_t __p1) {
58969  int8x16_t __ret;
58970  __ret = (int8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
58971  return __ret;
58972}
58973#else
58974__ai int8x16_t vqtbl1q_s8(int8x16_t __p0, int8x16_t __p1) {
58975  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58976  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58977  int8x16_t __ret;
58978  __ret = (int8x16_t) __builtin_neon_vqtbl1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
58979  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58980  return __ret;
58981}
58982#endif
58983
58984#ifdef __LITTLE_ENDIAN__
58985__ai uint8x8_t vqtbl1_u8(uint8x16_t __p0, uint8x8_t __p1) {
58986  uint8x8_t __ret;
58987  __ret = (uint8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 16);
58988  return __ret;
58989}
58990#else
58991__ai uint8x8_t vqtbl1_u8(uint8x16_t __p0, uint8x8_t __p1) {
58992  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
58993  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
58994  uint8x8_t __ret;
58995  __ret = (uint8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 16);
58996  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
58997  return __ret;
58998}
58999#endif
59000
59001#ifdef __LITTLE_ENDIAN__
59002__ai int8x8_t vqtbl1_s8(int8x16_t __p0, int8x8_t __p1) {
59003  int8x8_t __ret;
59004  __ret = (int8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__p0, (int8x8_t)__p1, 0);
59005  return __ret;
59006}
59007#else
59008__ai int8x8_t vqtbl1_s8(int8x16_t __p0, int8x8_t __p1) {
59009  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59010  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59011  int8x8_t __ret;
59012  __ret = (int8x8_t) __builtin_neon_vqtbl1_v((int8x16_t)__rev0, (int8x8_t)__rev1, 0);
59013  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59014  return __ret;
59015}
59016#endif
59017
59018#ifdef __LITTLE_ENDIAN__
59019__ai poly8x8_t vqtbl2_p8(poly8x16x2_t __p0, uint8x8_t __p1) {
59020  poly8x8_t __ret;
59021  __ret = (poly8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 4);
59022  return __ret;
59023}
59024#else
59025__ai poly8x8_t vqtbl2_p8(poly8x16x2_t __p0, uint8x8_t __p1) {
59026  poly8x16x2_t __rev0;
59027  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59028  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59029  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59030  poly8x8_t __ret;
59031  __ret = (poly8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 4);
59032  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59033  return __ret;
59034}
59035#endif
59036
59037#ifdef __LITTLE_ENDIAN__
59038__ai poly8x16_t vqtbl2q_p8(poly8x16x2_t __p0, uint8x16_t __p1) {
59039  poly8x16_t __ret;
59040  __ret = (poly8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 36);
59041  return __ret;
59042}
59043#else
59044__ai poly8x16_t vqtbl2q_p8(poly8x16x2_t __p0, uint8x16_t __p1) {
59045  poly8x16x2_t __rev0;
59046  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59047  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59048  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59049  poly8x16_t __ret;
59050  __ret = (poly8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 36);
59051  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59052  return __ret;
59053}
59054#endif
59055
59056#ifdef __LITTLE_ENDIAN__
59057__ai uint8x16_t vqtbl2q_u8(uint8x16x2_t __p0, uint8x16_t __p1) {
59058  uint8x16_t __ret;
59059  __ret = (uint8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 48);
59060  return __ret;
59061}
59062#else
59063__ai uint8x16_t vqtbl2q_u8(uint8x16x2_t __p0, uint8x16_t __p1) {
59064  uint8x16x2_t __rev0;
59065  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59066  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59067  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59068  uint8x16_t __ret;
59069  __ret = (uint8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 48);
59070  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59071  return __ret;
59072}
59073#endif
59074
59075#ifdef __LITTLE_ENDIAN__
59076__ai int8x16_t vqtbl2q_s8(int8x16x2_t __p0, int8x16_t __p1) {
59077  int8x16_t __ret;
59078  __ret = (int8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p1, 32);
59079  return __ret;
59080}
59081#else
59082__ai int8x16_t vqtbl2q_s8(int8x16x2_t __p0, int8x16_t __p1) {
59083  int8x16x2_t __rev0;
59084  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59085  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59086  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59087  int8x16_t __ret;
59088  __ret = (int8x16_t) __builtin_neon_vqtbl2q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev1, 32);
59089  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59090  return __ret;
59091}
59092#endif
59093
59094#ifdef __LITTLE_ENDIAN__
59095__ai uint8x8_t vqtbl2_u8(uint8x16x2_t __p0, uint8x8_t __p1) {
59096  uint8x8_t __ret;
59097  __ret = (uint8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 16);
59098  return __ret;
59099}
59100#else
59101__ai uint8x8_t vqtbl2_u8(uint8x16x2_t __p0, uint8x8_t __p1) {
59102  uint8x16x2_t __rev0;
59103  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59104  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59105  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59106  uint8x8_t __ret;
59107  __ret = (uint8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 16);
59108  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59109  return __ret;
59110}
59111#endif
59112
59113#ifdef __LITTLE_ENDIAN__
59114__ai int8x8_t vqtbl2_s8(int8x16x2_t __p0, int8x8_t __p1) {
59115  int8x8_t __ret;
59116  __ret = (int8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x8_t)__p1, 0);
59117  return __ret;
59118}
59119#else
59120__ai int8x8_t vqtbl2_s8(int8x16x2_t __p0, int8x8_t __p1) {
59121  int8x16x2_t __rev0;
59122  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59123  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59124  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59125  int8x8_t __ret;
59126  __ret = (int8x8_t) __builtin_neon_vqtbl2_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x8_t)__rev1, 0);
59127  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59128  return __ret;
59129}
59130#endif
59131
59132#ifdef __LITTLE_ENDIAN__
59133__ai poly8x8_t vqtbl3_p8(poly8x16x3_t __p0, uint8x8_t __p1) {
59134  poly8x8_t __ret;
59135  __ret = (poly8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 4);
59136  return __ret;
59137}
59138#else
59139__ai poly8x8_t vqtbl3_p8(poly8x16x3_t __p0, uint8x8_t __p1) {
59140  poly8x16x3_t __rev0;
59141  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59142  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59143  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59144  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59145  poly8x8_t __ret;
59146  __ret = (poly8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 4);
59147  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59148  return __ret;
59149}
59150#endif
59151
59152#ifdef __LITTLE_ENDIAN__
59153__ai poly8x16_t vqtbl3q_p8(poly8x16x3_t __p0, uint8x16_t __p1) {
59154  poly8x16_t __ret;
59155  __ret = (poly8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 36);
59156  return __ret;
59157}
59158#else
59159__ai poly8x16_t vqtbl3q_p8(poly8x16x3_t __p0, uint8x16_t __p1) {
59160  poly8x16x3_t __rev0;
59161  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59162  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59163  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59164  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59165  poly8x16_t __ret;
59166  __ret = (poly8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 36);
59167  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59168  return __ret;
59169}
59170#endif
59171
59172#ifdef __LITTLE_ENDIAN__
59173__ai uint8x16_t vqtbl3q_u8(uint8x16x3_t __p0, uint8x16_t __p1) {
59174  uint8x16_t __ret;
59175  __ret = (uint8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 48);
59176  return __ret;
59177}
59178#else
59179__ai uint8x16_t vqtbl3q_u8(uint8x16x3_t __p0, uint8x16_t __p1) {
59180  uint8x16x3_t __rev0;
59181  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59182  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59183  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59184  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59185  uint8x16_t __ret;
59186  __ret = (uint8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 48);
59187  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59188  return __ret;
59189}
59190#endif
59191
59192#ifdef __LITTLE_ENDIAN__
59193__ai int8x16_t vqtbl3q_s8(int8x16x3_t __p0, int8x16_t __p1) {
59194  int8x16_t __ret;
59195  __ret = (int8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p1, 32);
59196  return __ret;
59197}
59198#else
59199__ai int8x16_t vqtbl3q_s8(int8x16x3_t __p0, int8x16_t __p1) {
59200  int8x16x3_t __rev0;
59201  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59202  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59203  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59204  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59205  int8x16_t __ret;
59206  __ret = (int8x16_t) __builtin_neon_vqtbl3q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev1, 32);
59207  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59208  return __ret;
59209}
59210#endif
59211
59212#ifdef __LITTLE_ENDIAN__
59213__ai uint8x8_t vqtbl3_u8(uint8x16x3_t __p0, uint8x8_t __p1) {
59214  uint8x8_t __ret;
59215  __ret = (uint8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 16);
59216  return __ret;
59217}
59218#else
59219__ai uint8x8_t vqtbl3_u8(uint8x16x3_t __p0, uint8x8_t __p1) {
59220  uint8x16x3_t __rev0;
59221  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59222  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59223  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59224  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59225  uint8x8_t __ret;
59226  __ret = (uint8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 16);
59227  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59228  return __ret;
59229}
59230#endif
59231
59232#ifdef __LITTLE_ENDIAN__
59233__ai int8x8_t vqtbl3_s8(int8x16x3_t __p0, int8x8_t __p1) {
59234  int8x8_t __ret;
59235  __ret = (int8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x8_t)__p1, 0);
59236  return __ret;
59237}
59238#else
59239__ai int8x8_t vqtbl3_s8(int8x16x3_t __p0, int8x8_t __p1) {
59240  int8x16x3_t __rev0;
59241  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59242  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59243  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59244  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59245  int8x8_t __ret;
59246  __ret = (int8x8_t) __builtin_neon_vqtbl3_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x8_t)__rev1, 0);
59247  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59248  return __ret;
59249}
59250#endif
59251
59252#ifdef __LITTLE_ENDIAN__
59253__ai poly8x8_t vqtbl4_p8(poly8x16x4_t __p0, uint8x8_t __p1) {
59254  poly8x8_t __ret;
59255  __ret = (poly8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 4);
59256  return __ret;
59257}
59258#else
59259__ai poly8x8_t vqtbl4_p8(poly8x16x4_t __p0, uint8x8_t __p1) {
59260  poly8x16x4_t __rev0;
59261  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59262  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59263  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59264  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59265  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59266  poly8x8_t __ret;
59267  __ret = (poly8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 4);
59268  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59269  return __ret;
59270}
59271#endif
59272
59273#ifdef __LITTLE_ENDIAN__
59274__ai poly8x16_t vqtbl4q_p8(poly8x16x4_t __p0, uint8x16_t __p1) {
59275  poly8x16_t __ret;
59276  __ret = (poly8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 36);
59277  return __ret;
59278}
59279#else
59280__ai poly8x16_t vqtbl4q_p8(poly8x16x4_t __p0, uint8x16_t __p1) {
59281  poly8x16x4_t __rev0;
59282  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59283  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59284  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59285  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59286  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59287  poly8x16_t __ret;
59288  __ret = (poly8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 36);
59289  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59290  return __ret;
59291}
59292#endif
59293
59294#ifdef __LITTLE_ENDIAN__
59295__ai uint8x16_t vqtbl4q_u8(uint8x16x4_t __p0, uint8x16_t __p1) {
59296  uint8x16_t __ret;
59297  __ret = (uint8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 48);
59298  return __ret;
59299}
59300#else
59301__ai uint8x16_t vqtbl4q_u8(uint8x16x4_t __p0, uint8x16_t __p1) {
59302  uint8x16x4_t __rev0;
59303  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59304  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59305  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59306  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59307  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59308  uint8x16_t __ret;
59309  __ret = (uint8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 48);
59310  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59311  return __ret;
59312}
59313#endif
59314
59315#ifdef __LITTLE_ENDIAN__
59316__ai int8x16_t vqtbl4q_s8(int8x16x4_t __p0, int8x16_t __p1) {
59317  int8x16_t __ret;
59318  __ret = (int8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x16_t)__p1, 32);
59319  return __ret;
59320}
59321#else
59322__ai int8x16_t vqtbl4q_s8(int8x16x4_t __p0, int8x16_t __p1) {
59323  int8x16x4_t __rev0;
59324  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59325  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59326  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59327  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59328  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59329  int8x16_t __ret;
59330  __ret = (int8x16_t) __builtin_neon_vqtbl4q_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x16_t)__rev1, 32);
59331  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59332  return __ret;
59333}
59334#endif
59335
59336#ifdef __LITTLE_ENDIAN__
59337__ai uint8x8_t vqtbl4_u8(uint8x16x4_t __p0, uint8x8_t __p1) {
59338  uint8x8_t __ret;
59339  __ret = (uint8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 16);
59340  return __ret;
59341}
59342#else
59343__ai uint8x8_t vqtbl4_u8(uint8x16x4_t __p0, uint8x8_t __p1) {
59344  uint8x16x4_t __rev0;
59345  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59346  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59347  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59348  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59349  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59350  uint8x8_t __ret;
59351  __ret = (uint8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 16);
59352  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59353  return __ret;
59354}
59355#endif
59356
59357#ifdef __LITTLE_ENDIAN__
59358__ai int8x8_t vqtbl4_s8(int8x16x4_t __p0, int8x8_t __p1) {
59359  int8x8_t __ret;
59360  __ret = (int8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__p0.val[0], (int8x16_t)__p0.val[1], (int8x16_t)__p0.val[2], (int8x16_t)__p0.val[3], (int8x8_t)__p1, 0);
59361  return __ret;
59362}
59363#else
59364__ai int8x8_t vqtbl4_s8(int8x16x4_t __p0, int8x8_t __p1) {
59365  int8x16x4_t __rev0;
59366  __rev0.val[0] = __builtin_shufflevector(__p0.val[0], __p0.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59367  __rev0.val[1] = __builtin_shufflevector(__p0.val[1], __p0.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59368  __rev0.val[2] = __builtin_shufflevector(__p0.val[2], __p0.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59369  __rev0.val[3] = __builtin_shufflevector(__p0.val[3], __p0.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59370  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59371  int8x8_t __ret;
59372  __ret = (int8x8_t) __builtin_neon_vqtbl4_v((int8x16_t)__rev0.val[0], (int8x16_t)__rev0.val[1], (int8x16_t)__rev0.val[2], (int8x16_t)__rev0.val[3], (int8x8_t)__rev1, 0);
59373  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59374  return __ret;
59375}
59376#endif
59377
59378#ifdef __LITTLE_ENDIAN__
59379__ai poly8x8_t vqtbx1_p8(poly8x8_t __p0, poly8x16_t __p1, uint8x8_t __p2) {
59380  poly8x8_t __ret;
59381  __ret = (poly8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 4);
59382  return __ret;
59383}
59384#else
59385__ai poly8x8_t vqtbx1_p8(poly8x8_t __p0, poly8x16_t __p1, uint8x8_t __p2) {
59386  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59387  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59388  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59389  poly8x8_t __ret;
59390  __ret = (poly8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 4);
59391  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59392  return __ret;
59393}
59394#endif
59395
59396#ifdef __LITTLE_ENDIAN__
59397__ai poly8x16_t vqtbx1q_p8(poly8x16_t __p0, poly8x16_t __p1, uint8x16_t __p2) {
59398  poly8x16_t __ret;
59399  __ret = (poly8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 36);
59400  return __ret;
59401}
59402#else
59403__ai poly8x16_t vqtbx1q_p8(poly8x16_t __p0, poly8x16_t __p1, uint8x16_t __p2) {
59404  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59405  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59406  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59407  poly8x16_t __ret;
59408  __ret = (poly8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 36);
59409  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59410  return __ret;
59411}
59412#endif
59413
59414#ifdef __LITTLE_ENDIAN__
59415__ai uint8x16_t vqtbx1q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
59416  uint8x16_t __ret;
59417  __ret = (uint8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48);
59418  return __ret;
59419}
59420#else
59421__ai uint8x16_t vqtbx1q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
59422  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59423  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59424  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59425  uint8x16_t __ret;
59426  __ret = (uint8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48);
59427  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59428  return __ret;
59429}
59430#endif
59431
59432#ifdef __LITTLE_ENDIAN__
59433__ai int8x16_t vqtbx1q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
59434  int8x16_t __ret;
59435  __ret = (int8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32);
59436  return __ret;
59437}
59438#else
59439__ai int8x16_t vqtbx1q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
59440  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59441  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59442  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59443  int8x16_t __ret;
59444  __ret = (int8x16_t) __builtin_neon_vqtbx1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32);
59445  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59446  return __ret;
59447}
59448#endif
59449
59450#ifdef __LITTLE_ENDIAN__
59451__ai uint8x8_t vqtbx1_u8(uint8x8_t __p0, uint8x16_t __p1, uint8x8_t __p2) {
59452  uint8x8_t __ret;
59453  __ret = (uint8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 16);
59454  return __ret;
59455}
59456#else
59457__ai uint8x8_t vqtbx1_u8(uint8x8_t __p0, uint8x16_t __p1, uint8x8_t __p2) {
59458  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59459  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59460  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59461  uint8x8_t __ret;
59462  __ret = (uint8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 16);
59463  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59464  return __ret;
59465}
59466#endif
59467
59468#ifdef __LITTLE_ENDIAN__
59469__ai int8x8_t vqtbx1_s8(int8x8_t __p0, int8x16_t __p1, int8x8_t __p2) {
59470  int8x8_t __ret;
59471  __ret = (int8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__p0, (int8x16_t)__p1, (int8x8_t)__p2, 0);
59472  return __ret;
59473}
59474#else
59475__ai int8x8_t vqtbx1_s8(int8x8_t __p0, int8x16_t __p1, int8x8_t __p2) {
59476  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59477  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59478  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59479  int8x8_t __ret;
59480  __ret = (int8x8_t) __builtin_neon_vqtbx1_v((int8x8_t)__rev0, (int8x16_t)__rev1, (int8x8_t)__rev2, 0);
59481  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59482  return __ret;
59483}
59484#endif
59485
59486#ifdef __LITTLE_ENDIAN__
59487__ai poly8x8_t vqtbx2_p8(poly8x8_t __p0, poly8x16x2_t __p1, uint8x8_t __p2) {
59488  poly8x8_t __ret;
59489  __ret = (poly8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 4);
59490  return __ret;
59491}
59492#else
59493__ai poly8x8_t vqtbx2_p8(poly8x8_t __p0, poly8x16x2_t __p1, uint8x8_t __p2) {
59494  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59495  poly8x16x2_t __rev1;
59496  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59497  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59498  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59499  poly8x8_t __ret;
59500  __ret = (poly8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 4);
59501  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59502  return __ret;
59503}
59504#endif
59505
59506#ifdef __LITTLE_ENDIAN__
59507__ai poly8x16_t vqtbx2q_p8(poly8x16_t __p0, poly8x16x2_t __p1, uint8x16_t __p2) {
59508  poly8x16_t __ret;
59509  __ret = (poly8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 36);
59510  return __ret;
59511}
59512#else
59513__ai poly8x16_t vqtbx2q_p8(poly8x16_t __p0, poly8x16x2_t __p1, uint8x16_t __p2) {
59514  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59515  poly8x16x2_t __rev1;
59516  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59517  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59518  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59519  poly8x16_t __ret;
59520  __ret = (poly8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 36);
59521  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59522  return __ret;
59523}
59524#endif
59525
59526#ifdef __LITTLE_ENDIAN__
59527__ai uint8x16_t vqtbx2q_u8(uint8x16_t __p0, uint8x16x2_t __p1, uint8x16_t __p2) {
59528  uint8x16_t __ret;
59529  __ret = (uint8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 48);
59530  return __ret;
59531}
59532#else
59533__ai uint8x16_t vqtbx2q_u8(uint8x16_t __p0, uint8x16x2_t __p1, uint8x16_t __p2) {
59534  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59535  uint8x16x2_t __rev1;
59536  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59537  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59538  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59539  uint8x16_t __ret;
59540  __ret = (uint8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 48);
59541  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59542  return __ret;
59543}
59544#endif
59545
59546#ifdef __LITTLE_ENDIAN__
59547__ai int8x16_t vqtbx2q_s8(int8x16_t __p0, int8x16x2_t __p1, int8x16_t __p2) {
59548  int8x16_t __ret;
59549  __ret = (int8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p2, 32);
59550  return __ret;
59551}
59552#else
59553__ai int8x16_t vqtbx2q_s8(int8x16_t __p0, int8x16x2_t __p1, int8x16_t __p2) {
59554  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59555  int8x16x2_t __rev1;
59556  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59557  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59558  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59559  int8x16_t __ret;
59560  __ret = (int8x16_t) __builtin_neon_vqtbx2q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev2, 32);
59561  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59562  return __ret;
59563}
59564#endif
59565
59566#ifdef __LITTLE_ENDIAN__
59567__ai uint8x8_t vqtbx2_u8(uint8x8_t __p0, uint8x16x2_t __p1, uint8x8_t __p2) {
59568  uint8x8_t __ret;
59569  __ret = (uint8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 16);
59570  return __ret;
59571}
59572#else
59573__ai uint8x8_t vqtbx2_u8(uint8x8_t __p0, uint8x16x2_t __p1, uint8x8_t __p2) {
59574  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59575  uint8x16x2_t __rev1;
59576  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59577  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59578  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59579  uint8x8_t __ret;
59580  __ret = (uint8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 16);
59581  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59582  return __ret;
59583}
59584#endif
59585
59586#ifdef __LITTLE_ENDIAN__
59587__ai int8x8_t vqtbx2_s8(int8x8_t __p0, int8x16x2_t __p1, int8x8_t __p2) {
59588  int8x8_t __ret;
59589  __ret = (int8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x8_t)__p2, 0);
59590  return __ret;
59591}
59592#else
59593__ai int8x8_t vqtbx2_s8(int8x8_t __p0, int8x16x2_t __p1, int8x8_t __p2) {
59594  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59595  int8x16x2_t __rev1;
59596  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59597  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59598  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59599  int8x8_t __ret;
59600  __ret = (int8x8_t) __builtin_neon_vqtbx2_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x8_t)__rev2, 0);
59601  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59602  return __ret;
59603}
59604#endif
59605
59606#ifdef __LITTLE_ENDIAN__
59607__ai poly8x8_t vqtbx3_p8(poly8x8_t __p0, poly8x16x3_t __p1, uint8x8_t __p2) {
59608  poly8x8_t __ret;
59609  __ret = (poly8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 4);
59610  return __ret;
59611}
59612#else
59613__ai poly8x8_t vqtbx3_p8(poly8x8_t __p0, poly8x16x3_t __p1, uint8x8_t __p2) {
59614  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59615  poly8x16x3_t __rev1;
59616  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59617  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59618  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59619  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59620  poly8x8_t __ret;
59621  __ret = (poly8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 4);
59622  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59623  return __ret;
59624}
59625#endif
59626
59627#ifdef __LITTLE_ENDIAN__
59628__ai poly8x16_t vqtbx3q_p8(poly8x16_t __p0, poly8x16x3_t __p1, uint8x16_t __p2) {
59629  poly8x16_t __ret;
59630  __ret = (poly8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 36);
59631  return __ret;
59632}
59633#else
59634__ai poly8x16_t vqtbx3q_p8(poly8x16_t __p0, poly8x16x3_t __p1, uint8x16_t __p2) {
59635  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59636  poly8x16x3_t __rev1;
59637  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59638  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59639  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59640  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59641  poly8x16_t __ret;
59642  __ret = (poly8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 36);
59643  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59644  return __ret;
59645}
59646#endif
59647
59648#ifdef __LITTLE_ENDIAN__
59649__ai uint8x16_t vqtbx3q_u8(uint8x16_t __p0, uint8x16x3_t __p1, uint8x16_t __p2) {
59650  uint8x16_t __ret;
59651  __ret = (uint8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 48);
59652  return __ret;
59653}
59654#else
59655__ai uint8x16_t vqtbx3q_u8(uint8x16_t __p0, uint8x16x3_t __p1, uint8x16_t __p2) {
59656  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59657  uint8x16x3_t __rev1;
59658  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59659  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59660  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59661  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59662  uint8x16_t __ret;
59663  __ret = (uint8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 48);
59664  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59665  return __ret;
59666}
59667#endif
59668
59669#ifdef __LITTLE_ENDIAN__
59670__ai int8x16_t vqtbx3q_s8(int8x16_t __p0, int8x16x3_t __p1, int8x16_t __p2) {
59671  int8x16_t __ret;
59672  __ret = (int8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p2, 32);
59673  return __ret;
59674}
59675#else
59676__ai int8x16_t vqtbx3q_s8(int8x16_t __p0, int8x16x3_t __p1, int8x16_t __p2) {
59677  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59678  int8x16x3_t __rev1;
59679  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59680  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59681  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59682  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59683  int8x16_t __ret;
59684  __ret = (int8x16_t) __builtin_neon_vqtbx3q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev2, 32);
59685  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59686  return __ret;
59687}
59688#endif
59689
59690#ifdef __LITTLE_ENDIAN__
59691__ai uint8x8_t vqtbx3_u8(uint8x8_t __p0, uint8x16x3_t __p1, uint8x8_t __p2) {
59692  uint8x8_t __ret;
59693  __ret = (uint8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 16);
59694  return __ret;
59695}
59696#else
59697__ai uint8x8_t vqtbx3_u8(uint8x8_t __p0, uint8x16x3_t __p1, uint8x8_t __p2) {
59698  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59699  uint8x16x3_t __rev1;
59700  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59701  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59702  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59703  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59704  uint8x8_t __ret;
59705  __ret = (uint8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 16);
59706  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59707  return __ret;
59708}
59709#endif
59710
59711#ifdef __LITTLE_ENDIAN__
59712__ai int8x8_t vqtbx3_s8(int8x8_t __p0, int8x16x3_t __p1, int8x8_t __p2) {
59713  int8x8_t __ret;
59714  __ret = (int8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x8_t)__p2, 0);
59715  return __ret;
59716}
59717#else
59718__ai int8x8_t vqtbx3_s8(int8x8_t __p0, int8x16x3_t __p1, int8x8_t __p2) {
59719  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59720  int8x16x3_t __rev1;
59721  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59722  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59723  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59724  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59725  int8x8_t __ret;
59726  __ret = (int8x8_t) __builtin_neon_vqtbx3_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x8_t)__rev2, 0);
59727  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59728  return __ret;
59729}
59730#endif
59731
59732#ifdef __LITTLE_ENDIAN__
59733__ai poly8x8_t vqtbx4_p8(poly8x8_t __p0, poly8x16x4_t __p1, uint8x8_t __p2) {
59734  poly8x8_t __ret;
59735  __ret = (poly8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 4);
59736  return __ret;
59737}
59738#else
59739__ai poly8x8_t vqtbx4_p8(poly8x8_t __p0, poly8x16x4_t __p1, uint8x8_t __p2) {
59740  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59741  poly8x16x4_t __rev1;
59742  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59743  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59744  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59745  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59746  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59747  poly8x8_t __ret;
59748  __ret = (poly8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 4);
59749  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59750  return __ret;
59751}
59752#endif
59753
59754#ifdef __LITTLE_ENDIAN__
59755__ai poly8x16_t vqtbx4q_p8(poly8x16_t __p0, poly8x16x4_t __p1, uint8x16_t __p2) {
59756  poly8x16_t __ret;
59757  __ret = (poly8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 36);
59758  return __ret;
59759}
59760#else
59761__ai poly8x16_t vqtbx4q_p8(poly8x16_t __p0, poly8x16x4_t __p1, uint8x16_t __p2) {
59762  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59763  poly8x16x4_t __rev1;
59764  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59765  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59766  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59767  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59768  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59769  poly8x16_t __ret;
59770  __ret = (poly8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 36);
59771  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59772  return __ret;
59773}
59774#endif
59775
59776#ifdef __LITTLE_ENDIAN__
59777__ai uint8x16_t vqtbx4q_u8(uint8x16_t __p0, uint8x16x4_t __p1, uint8x16_t __p2) {
59778  uint8x16_t __ret;
59779  __ret = (uint8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 48);
59780  return __ret;
59781}
59782#else
59783__ai uint8x16_t vqtbx4q_u8(uint8x16_t __p0, uint8x16x4_t __p1, uint8x16_t __p2) {
59784  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59785  uint8x16x4_t __rev1;
59786  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59787  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59788  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59789  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59790  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59791  uint8x16_t __ret;
59792  __ret = (uint8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 48);
59793  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59794  return __ret;
59795}
59796#endif
59797
59798#ifdef __LITTLE_ENDIAN__
59799__ai int8x16_t vqtbx4q_s8(int8x16_t __p0, int8x16x4_t __p1, int8x16_t __p2) {
59800  int8x16_t __ret;
59801  __ret = (int8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x16_t)__p2, 32);
59802  return __ret;
59803}
59804#else
59805__ai int8x16_t vqtbx4q_s8(int8x16_t __p0, int8x16x4_t __p1, int8x16_t __p2) {
59806  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59807  int8x16x4_t __rev1;
59808  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59809  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59810  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59811  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59812  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59813  int8x16_t __ret;
59814  __ret = (int8x16_t) __builtin_neon_vqtbx4q_v((int8x16_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x16_t)__rev2, 32);
59815  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59816  return __ret;
59817}
59818#endif
59819
59820#ifdef __LITTLE_ENDIAN__
59821__ai uint8x8_t vqtbx4_u8(uint8x8_t __p0, uint8x16x4_t __p1, uint8x8_t __p2) {
59822  uint8x8_t __ret;
59823  __ret = (uint8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 16);
59824  return __ret;
59825}
59826#else
59827__ai uint8x8_t vqtbx4_u8(uint8x8_t __p0, uint8x16x4_t __p1, uint8x8_t __p2) {
59828  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59829  uint8x16x4_t __rev1;
59830  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59831  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59832  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59833  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59834  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59835  uint8x8_t __ret;
59836  __ret = (uint8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 16);
59837  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59838  return __ret;
59839}
59840#endif
59841
59842#ifdef __LITTLE_ENDIAN__
59843__ai int8x8_t vqtbx4_s8(int8x8_t __p0, int8x16x4_t __p1, int8x8_t __p2) {
59844  int8x8_t __ret;
59845  __ret = (int8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__p0, (int8x16_t)__p1.val[0], (int8x16_t)__p1.val[1], (int8x16_t)__p1.val[2], (int8x16_t)__p1.val[3], (int8x8_t)__p2, 0);
59846  return __ret;
59847}
59848#else
59849__ai int8x8_t vqtbx4_s8(int8x8_t __p0, int8x16x4_t __p1, int8x8_t __p2) {
59850  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59851  int8x16x4_t __rev1;
59852  __rev1.val[0] = __builtin_shufflevector(__p1.val[0], __p1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59853  __rev1.val[1] = __builtin_shufflevector(__p1.val[1], __p1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59854  __rev1.val[2] = __builtin_shufflevector(__p1.val[2], __p1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59855  __rev1.val[3] = __builtin_shufflevector(__p1.val[3], __p1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59856  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59857  int8x8_t __ret;
59858  __ret = (int8x8_t) __builtin_neon_vqtbx4_v((int8x8_t)__rev0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], (int8x8_t)__rev2, 0);
59859  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59860  return __ret;
59861}
59862#endif
59863
59864#ifdef __LITTLE_ENDIAN__
59865__ai uint16x8_t vraddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
59866  uint16x8_t __ret;
59867  __ret = vcombine_u16(__p0, vraddhn_u32(__p1, __p2));
59868  return __ret;
59869}
59870#else
59871__ai uint16x8_t vraddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
59872  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59873  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59874  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
59875  uint16x8_t __ret;
59876  __ret = __noswap_vcombine_u16(__rev0, __noswap_vraddhn_u32(__rev1, __rev2));
59877  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59878  return __ret;
59879}
59880#endif
59881
59882#ifdef __LITTLE_ENDIAN__
59883__ai uint32x4_t vraddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
59884  uint32x4_t __ret;
59885  __ret = vcombine_u32(__p0, vraddhn_u64(__p1, __p2));
59886  return __ret;
59887}
59888#else
59889__ai uint32x4_t vraddhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
59890  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59891  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59892  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
59893  uint32x4_t __ret;
59894  __ret = __noswap_vcombine_u32(__rev0, __noswap_vraddhn_u64(__rev1, __rev2));
59895  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59896  return __ret;
59897}
59898#endif
59899
59900#ifdef __LITTLE_ENDIAN__
59901__ai uint8x16_t vraddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
59902  uint8x16_t __ret;
59903  __ret = vcombine_u8(__p0, vraddhn_u16(__p1, __p2));
59904  return __ret;
59905}
59906#else
59907__ai uint8x16_t vraddhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
59908  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59909  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59910  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59911  uint8x16_t __ret;
59912  __ret = __noswap_vcombine_u8(__rev0, __noswap_vraddhn_u16(__rev1, __rev2));
59913  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59914  return __ret;
59915}
59916#endif
59917
59918#ifdef __LITTLE_ENDIAN__
59919__ai int16x8_t vraddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
59920  int16x8_t __ret;
59921  __ret = vcombine_s16(__p0, vraddhn_s32(__p1, __p2));
59922  return __ret;
59923}
59924#else
59925__ai int16x8_t vraddhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
59926  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
59927  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
59928  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
59929  int16x8_t __ret;
59930  __ret = __noswap_vcombine_s16(__rev0, __noswap_vraddhn_s32(__rev1, __rev2));
59931  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59932  return __ret;
59933}
59934#endif
59935
59936#ifdef __LITTLE_ENDIAN__
59937__ai int32x4_t vraddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
59938  int32x4_t __ret;
59939  __ret = vcombine_s32(__p0, vraddhn_s64(__p1, __p2));
59940  return __ret;
59941}
59942#else
59943__ai int32x4_t vraddhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
59944  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
59945  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
59946  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
59947  int32x4_t __ret;
59948  __ret = __noswap_vcombine_s32(__rev0, __noswap_vraddhn_s64(__rev1, __rev2));
59949  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
59950  return __ret;
59951}
59952#endif
59953
59954#ifdef __LITTLE_ENDIAN__
59955__ai int8x16_t vraddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
59956  int8x16_t __ret;
59957  __ret = vcombine_s8(__p0, vraddhn_s16(__p1, __p2));
59958  return __ret;
59959}
59960#else
59961__ai int8x16_t vraddhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
59962  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59963  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
59964  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
59965  int8x16_t __ret;
59966  __ret = __noswap_vcombine_s8(__rev0, __noswap_vraddhn_s16(__rev1, __rev2));
59967  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59968  return __ret;
59969}
59970#endif
59971
59972#ifdef __LITTLE_ENDIAN__
59973__ai poly8x8_t vrbit_p8(poly8x8_t __p0) {
59974  poly8x8_t __ret;
59975  __ret = (poly8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 4);
59976  return __ret;
59977}
59978#else
59979__ai poly8x8_t vrbit_p8(poly8x8_t __p0) {
59980  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
59981  poly8x8_t __ret;
59982  __ret = (poly8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 4);
59983  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
59984  return __ret;
59985}
59986#endif
59987
59988#ifdef __LITTLE_ENDIAN__
59989__ai poly8x16_t vrbitq_p8(poly8x16_t __p0) {
59990  poly8x16_t __ret;
59991  __ret = (poly8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 36);
59992  return __ret;
59993}
59994#else
59995__ai poly8x16_t vrbitq_p8(poly8x16_t __p0) {
59996  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
59997  poly8x16_t __ret;
59998  __ret = (poly8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 36);
59999  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60000  return __ret;
60001}
60002#endif
60003
60004#ifdef __LITTLE_ENDIAN__
60005__ai uint8x16_t vrbitq_u8(uint8x16_t __p0) {
60006  uint8x16_t __ret;
60007  __ret = (uint8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 48);
60008  return __ret;
60009}
60010#else
60011__ai uint8x16_t vrbitq_u8(uint8x16_t __p0) {
60012  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60013  uint8x16_t __ret;
60014  __ret = (uint8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 48);
60015  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60016  return __ret;
60017}
60018#endif
60019
60020#ifdef __LITTLE_ENDIAN__
60021__ai int8x16_t vrbitq_s8(int8x16_t __p0) {
60022  int8x16_t __ret;
60023  __ret = (int8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__p0, 32);
60024  return __ret;
60025}
60026#else
60027__ai int8x16_t vrbitq_s8(int8x16_t __p0) {
60028  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60029  int8x16_t __ret;
60030  __ret = (int8x16_t) __builtin_neon_vrbitq_v((int8x16_t)__rev0, 32);
60031  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60032  return __ret;
60033}
60034#endif
60035
60036#ifdef __LITTLE_ENDIAN__
60037__ai uint8x8_t vrbit_u8(uint8x8_t __p0) {
60038  uint8x8_t __ret;
60039  __ret = (uint8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 16);
60040  return __ret;
60041}
60042#else
60043__ai uint8x8_t vrbit_u8(uint8x8_t __p0) {
60044  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60045  uint8x8_t __ret;
60046  __ret = (uint8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 16);
60047  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60048  return __ret;
60049}
60050#endif
60051
60052#ifdef __LITTLE_ENDIAN__
60053__ai int8x8_t vrbit_s8(int8x8_t __p0) {
60054  int8x8_t __ret;
60055  __ret = (int8x8_t) __builtin_neon_vrbit_v((int8x8_t)__p0, 0);
60056  return __ret;
60057}
60058#else
60059__ai int8x8_t vrbit_s8(int8x8_t __p0) {
60060  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60061  int8x8_t __ret;
60062  __ret = (int8x8_t) __builtin_neon_vrbit_v((int8x8_t)__rev0, 0);
60063  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60064  return __ret;
60065}
60066#endif
60067
60068#ifdef __LITTLE_ENDIAN__
60069__ai float64x2_t vrecpeq_f64(float64x2_t __p0) {
60070  float64x2_t __ret;
60071  __ret = (float64x2_t) __builtin_neon_vrecpeq_v((int8x16_t)__p0, 42);
60072  return __ret;
60073}
60074#else
60075__ai float64x2_t vrecpeq_f64(float64x2_t __p0) {
60076  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60077  float64x2_t __ret;
60078  __ret = (float64x2_t) __builtin_neon_vrecpeq_v((int8x16_t)__rev0, 42);
60079  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60080  return __ret;
60081}
60082#endif
60083
60084#ifdef __LITTLE_ENDIAN__
60085__ai float64x1_t vrecpe_f64(float64x1_t __p0) {
60086  float64x1_t __ret;
60087  __ret = (float64x1_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 10);
60088  return __ret;
60089}
60090#else
60091__ai float64x1_t vrecpe_f64(float64x1_t __p0) {
60092  float64x1_t __ret;
60093  __ret = (float64x1_t) __builtin_neon_vrecpe_v((int8x8_t)__p0, 10);
60094  return __ret;
60095}
60096#endif
60097
60098#ifdef __LITTLE_ENDIAN__
60099__ai float64_t vrecped_f64(float64_t __p0) {
60100  float64_t __ret;
60101  __ret = (float64_t) __builtin_neon_vrecped_f64(__p0);
60102  return __ret;
60103}
60104#else
60105__ai float64_t vrecped_f64(float64_t __p0) {
60106  float64_t __ret;
60107  __ret = (float64_t) __builtin_neon_vrecped_f64(__p0);
60108  return __ret;
60109}
60110#endif
60111
60112#ifdef __LITTLE_ENDIAN__
60113__ai float32_t vrecpes_f32(float32_t __p0) {
60114  float32_t __ret;
60115  __ret = (float32_t) __builtin_neon_vrecpes_f32(__p0);
60116  return __ret;
60117}
60118#else
60119__ai float32_t vrecpes_f32(float32_t __p0) {
60120  float32_t __ret;
60121  __ret = (float32_t) __builtin_neon_vrecpes_f32(__p0);
60122  return __ret;
60123}
60124#endif
60125
60126#ifdef __LITTLE_ENDIAN__
60127__ai float64x2_t vrecpsq_f64(float64x2_t __p0, float64x2_t __p1) {
60128  float64x2_t __ret;
60129  __ret = (float64x2_t) __builtin_neon_vrecpsq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
60130  return __ret;
60131}
60132#else
60133__ai float64x2_t vrecpsq_f64(float64x2_t __p0, float64x2_t __p1) {
60134  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60135  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60136  float64x2_t __ret;
60137  __ret = (float64x2_t) __builtin_neon_vrecpsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
60138  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60139  return __ret;
60140}
60141#endif
60142
60143#ifdef __LITTLE_ENDIAN__
60144__ai float64x1_t vrecps_f64(float64x1_t __p0, float64x1_t __p1) {
60145  float64x1_t __ret;
60146  __ret = (float64x1_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
60147  return __ret;
60148}
60149#else
60150__ai float64x1_t vrecps_f64(float64x1_t __p0, float64x1_t __p1) {
60151  float64x1_t __ret;
60152  __ret = (float64x1_t) __builtin_neon_vrecps_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
60153  return __ret;
60154}
60155#endif
60156
60157#ifdef __LITTLE_ENDIAN__
60158__ai float64_t vrecpsd_f64(float64_t __p0, float64_t __p1) {
60159  float64_t __ret;
60160  __ret = (float64_t) __builtin_neon_vrecpsd_f64(__p0, __p1);
60161  return __ret;
60162}
60163#else
60164__ai float64_t vrecpsd_f64(float64_t __p0, float64_t __p1) {
60165  float64_t __ret;
60166  __ret = (float64_t) __builtin_neon_vrecpsd_f64(__p0, __p1);
60167  return __ret;
60168}
60169#endif
60170
60171#ifdef __LITTLE_ENDIAN__
60172__ai float32_t vrecpss_f32(float32_t __p0, float32_t __p1) {
60173  float32_t __ret;
60174  __ret = (float32_t) __builtin_neon_vrecpss_f32(__p0, __p1);
60175  return __ret;
60176}
60177#else
60178__ai float32_t vrecpss_f32(float32_t __p0, float32_t __p1) {
60179  float32_t __ret;
60180  __ret = (float32_t) __builtin_neon_vrecpss_f32(__p0, __p1);
60181  return __ret;
60182}
60183#endif
60184
60185#ifdef __LITTLE_ENDIAN__
60186__ai float64_t vrecpxd_f64(float64_t __p0) {
60187  float64_t __ret;
60188  __ret = (float64_t) __builtin_neon_vrecpxd_f64(__p0);
60189  return __ret;
60190}
60191#else
60192__ai float64_t vrecpxd_f64(float64_t __p0) {
60193  float64_t __ret;
60194  __ret = (float64_t) __builtin_neon_vrecpxd_f64(__p0);
60195  return __ret;
60196}
60197#endif
60198
60199#ifdef __LITTLE_ENDIAN__
60200__ai float32_t vrecpxs_f32(float32_t __p0) {
60201  float32_t __ret;
60202  __ret = (float32_t) __builtin_neon_vrecpxs_f32(__p0);
60203  return __ret;
60204}
60205#else
60206__ai float32_t vrecpxs_f32(float32_t __p0) {
60207  float32_t __ret;
60208  __ret = (float32_t) __builtin_neon_vrecpxs_f32(__p0);
60209  return __ret;
60210}
60211#endif
60212
60213#ifdef __LITTLE_ENDIAN__
60214__ai uint64_t vrshld_u64(uint64_t __p0, uint64_t __p1) {
60215  uint64_t __ret;
60216  __ret = (uint64_t) __builtin_neon_vrshld_u64(__p0, __p1);
60217  return __ret;
60218}
60219#else
60220__ai uint64_t vrshld_u64(uint64_t __p0, uint64_t __p1) {
60221  uint64_t __ret;
60222  __ret = (uint64_t) __builtin_neon_vrshld_u64(__p0, __p1);
60223  return __ret;
60224}
60225#endif
60226
60227#ifdef __LITTLE_ENDIAN__
60228__ai int64_t vrshld_s64(int64_t __p0, int64_t __p1) {
60229  int64_t __ret;
60230  __ret = (int64_t) __builtin_neon_vrshld_s64(__p0, __p1);
60231  return __ret;
60232}
60233#else
60234__ai int64_t vrshld_s64(int64_t __p0, int64_t __p1) {
60235  int64_t __ret;
60236  __ret = (int64_t) __builtin_neon_vrshld_s64(__p0, __p1);
60237  return __ret;
60238}
60239#endif
60240
60241#ifdef __LITTLE_ENDIAN__
60242#define vrshrd_n_u64(__p0, __p1) __extension__ ({ \
60243  uint64_t __s0 = __p0; \
60244  uint64_t __ret; \
60245  __ret = (uint64_t) __builtin_neon_vrshrd_n_u64(__s0, __p1); \
60246  __ret; \
60247})
60248#else
60249#define vrshrd_n_u64(__p0, __p1) __extension__ ({ \
60250  uint64_t __s0 = __p0; \
60251  uint64_t __ret; \
60252  __ret = (uint64_t) __builtin_neon_vrshrd_n_u64(__s0, __p1); \
60253  __ret; \
60254})
60255#endif
60256
60257#ifdef __LITTLE_ENDIAN__
60258#define vrshrd_n_s64(__p0, __p1) __extension__ ({ \
60259  int64_t __s0 = __p0; \
60260  int64_t __ret; \
60261  __ret = (int64_t) __builtin_neon_vrshrd_n_s64(__s0, __p1); \
60262  __ret; \
60263})
60264#else
60265#define vrshrd_n_s64(__p0, __p1) __extension__ ({ \
60266  int64_t __s0 = __p0; \
60267  int64_t __ret; \
60268  __ret = (int64_t) __builtin_neon_vrshrd_n_s64(__s0, __p1); \
60269  __ret; \
60270})
60271#endif
60272
60273#ifdef __LITTLE_ENDIAN__
60274#define vrshrn_high_n_u32(__p0_206, __p1_206, __p2_206) __extension__ ({ \
60275  uint16x4_t __s0_206 = __p0_206; \
60276  uint32x4_t __s1_206 = __p1_206; \
60277  uint16x8_t __ret_206; \
60278  __ret_206 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_206), (uint16x4_t)(vrshrn_n_u32(__s1_206, __p2_206)))); \
60279  __ret_206; \
60280})
60281#else
60282#define vrshrn_high_n_u32(__p0_207, __p1_207, __p2_207) __extension__ ({ \
60283  uint16x4_t __s0_207 = __p0_207; \
60284  uint32x4_t __s1_207 = __p1_207; \
60285  uint16x4_t __rev0_207;  __rev0_207 = __builtin_shufflevector(__s0_207, __s0_207, 3, 2, 1, 0); \
60286  uint32x4_t __rev1_207;  __rev1_207 = __builtin_shufflevector(__s1_207, __s1_207, 3, 2, 1, 0); \
60287  uint16x8_t __ret_207; \
60288  __ret_207 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_207), (uint16x4_t)(__noswap_vrshrn_n_u32(__rev1_207, __p2_207)))); \
60289  __ret_207 = __builtin_shufflevector(__ret_207, __ret_207, 7, 6, 5, 4, 3, 2, 1, 0); \
60290  __ret_207; \
60291})
60292#endif
60293
60294#ifdef __LITTLE_ENDIAN__
60295#define vrshrn_high_n_u64(__p0_208, __p1_208, __p2_208) __extension__ ({ \
60296  uint32x2_t __s0_208 = __p0_208; \
60297  uint64x2_t __s1_208 = __p1_208; \
60298  uint32x4_t __ret_208; \
60299  __ret_208 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_208), (uint32x2_t)(vrshrn_n_u64(__s1_208, __p2_208)))); \
60300  __ret_208; \
60301})
60302#else
60303#define vrshrn_high_n_u64(__p0_209, __p1_209, __p2_209) __extension__ ({ \
60304  uint32x2_t __s0_209 = __p0_209; \
60305  uint64x2_t __s1_209 = __p1_209; \
60306  uint32x2_t __rev0_209;  __rev0_209 = __builtin_shufflevector(__s0_209, __s0_209, 1, 0); \
60307  uint64x2_t __rev1_209;  __rev1_209 = __builtin_shufflevector(__s1_209, __s1_209, 1, 0); \
60308  uint32x4_t __ret_209; \
60309  __ret_209 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_209), (uint32x2_t)(__noswap_vrshrn_n_u64(__rev1_209, __p2_209)))); \
60310  __ret_209 = __builtin_shufflevector(__ret_209, __ret_209, 3, 2, 1, 0); \
60311  __ret_209; \
60312})
60313#endif
60314
60315#ifdef __LITTLE_ENDIAN__
60316#define vrshrn_high_n_u16(__p0_210, __p1_210, __p2_210) __extension__ ({ \
60317  uint8x8_t __s0_210 = __p0_210; \
60318  uint16x8_t __s1_210 = __p1_210; \
60319  uint8x16_t __ret_210; \
60320  __ret_210 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_210), (uint8x8_t)(vrshrn_n_u16(__s1_210, __p2_210)))); \
60321  __ret_210; \
60322})
60323#else
60324#define vrshrn_high_n_u16(__p0_211, __p1_211, __p2_211) __extension__ ({ \
60325  uint8x8_t __s0_211 = __p0_211; \
60326  uint16x8_t __s1_211 = __p1_211; \
60327  uint8x8_t __rev0_211;  __rev0_211 = __builtin_shufflevector(__s0_211, __s0_211, 7, 6, 5, 4, 3, 2, 1, 0); \
60328  uint16x8_t __rev1_211;  __rev1_211 = __builtin_shufflevector(__s1_211, __s1_211, 7, 6, 5, 4, 3, 2, 1, 0); \
60329  uint8x16_t __ret_211; \
60330  __ret_211 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_211), (uint8x8_t)(__noswap_vrshrn_n_u16(__rev1_211, __p2_211)))); \
60331  __ret_211 = __builtin_shufflevector(__ret_211, __ret_211, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60332  __ret_211; \
60333})
60334#endif
60335
60336#ifdef __LITTLE_ENDIAN__
60337#define vrshrn_high_n_s32(__p0_212, __p1_212, __p2_212) __extension__ ({ \
60338  int16x4_t __s0_212 = __p0_212; \
60339  int32x4_t __s1_212 = __p1_212; \
60340  int16x8_t __ret_212; \
60341  __ret_212 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_212), (int16x4_t)(vrshrn_n_s32(__s1_212, __p2_212)))); \
60342  __ret_212; \
60343})
60344#else
60345#define vrshrn_high_n_s32(__p0_213, __p1_213, __p2_213) __extension__ ({ \
60346  int16x4_t __s0_213 = __p0_213; \
60347  int32x4_t __s1_213 = __p1_213; \
60348  int16x4_t __rev0_213;  __rev0_213 = __builtin_shufflevector(__s0_213, __s0_213, 3, 2, 1, 0); \
60349  int32x4_t __rev1_213;  __rev1_213 = __builtin_shufflevector(__s1_213, __s1_213, 3, 2, 1, 0); \
60350  int16x8_t __ret_213; \
60351  __ret_213 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_213), (int16x4_t)(__noswap_vrshrn_n_s32(__rev1_213, __p2_213)))); \
60352  __ret_213 = __builtin_shufflevector(__ret_213, __ret_213, 7, 6, 5, 4, 3, 2, 1, 0); \
60353  __ret_213; \
60354})
60355#endif
60356
60357#ifdef __LITTLE_ENDIAN__
60358#define vrshrn_high_n_s64(__p0_214, __p1_214, __p2_214) __extension__ ({ \
60359  int32x2_t __s0_214 = __p0_214; \
60360  int64x2_t __s1_214 = __p1_214; \
60361  int32x4_t __ret_214; \
60362  __ret_214 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_214), (int32x2_t)(vrshrn_n_s64(__s1_214, __p2_214)))); \
60363  __ret_214; \
60364})
60365#else
60366#define vrshrn_high_n_s64(__p0_215, __p1_215, __p2_215) __extension__ ({ \
60367  int32x2_t __s0_215 = __p0_215; \
60368  int64x2_t __s1_215 = __p1_215; \
60369  int32x2_t __rev0_215;  __rev0_215 = __builtin_shufflevector(__s0_215, __s0_215, 1, 0); \
60370  int64x2_t __rev1_215;  __rev1_215 = __builtin_shufflevector(__s1_215, __s1_215, 1, 0); \
60371  int32x4_t __ret_215; \
60372  __ret_215 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_215), (int32x2_t)(__noswap_vrshrn_n_s64(__rev1_215, __p2_215)))); \
60373  __ret_215 = __builtin_shufflevector(__ret_215, __ret_215, 3, 2, 1, 0); \
60374  __ret_215; \
60375})
60376#endif
60377
60378#ifdef __LITTLE_ENDIAN__
60379#define vrshrn_high_n_s16(__p0_216, __p1_216, __p2_216) __extension__ ({ \
60380  int8x8_t __s0_216 = __p0_216; \
60381  int16x8_t __s1_216 = __p1_216; \
60382  int8x16_t __ret_216; \
60383  __ret_216 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_216), (int8x8_t)(vrshrn_n_s16(__s1_216, __p2_216)))); \
60384  __ret_216; \
60385})
60386#else
60387#define vrshrn_high_n_s16(__p0_217, __p1_217, __p2_217) __extension__ ({ \
60388  int8x8_t __s0_217 = __p0_217; \
60389  int16x8_t __s1_217 = __p1_217; \
60390  int8x8_t __rev0_217;  __rev0_217 = __builtin_shufflevector(__s0_217, __s0_217, 7, 6, 5, 4, 3, 2, 1, 0); \
60391  int16x8_t __rev1_217;  __rev1_217 = __builtin_shufflevector(__s1_217, __s1_217, 7, 6, 5, 4, 3, 2, 1, 0); \
60392  int8x16_t __ret_217; \
60393  __ret_217 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_217), (int8x8_t)(__noswap_vrshrn_n_s16(__rev1_217, __p2_217)))); \
60394  __ret_217 = __builtin_shufflevector(__ret_217, __ret_217, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60395  __ret_217; \
60396})
60397#endif
60398
60399#ifdef __LITTLE_ENDIAN__
60400__ai float64x2_t vrsqrteq_f64(float64x2_t __p0) {
60401  float64x2_t __ret;
60402  __ret = (float64x2_t) __builtin_neon_vrsqrteq_v((int8x16_t)__p0, 42);
60403  return __ret;
60404}
60405#else
60406__ai float64x2_t vrsqrteq_f64(float64x2_t __p0) {
60407  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60408  float64x2_t __ret;
60409  __ret = (float64x2_t) __builtin_neon_vrsqrteq_v((int8x16_t)__rev0, 42);
60410  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60411  return __ret;
60412}
60413#endif
60414
60415#ifdef __LITTLE_ENDIAN__
60416__ai float64x1_t vrsqrte_f64(float64x1_t __p0) {
60417  float64x1_t __ret;
60418  __ret = (float64x1_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 10);
60419  return __ret;
60420}
60421#else
60422__ai float64x1_t vrsqrte_f64(float64x1_t __p0) {
60423  float64x1_t __ret;
60424  __ret = (float64x1_t) __builtin_neon_vrsqrte_v((int8x8_t)__p0, 10);
60425  return __ret;
60426}
60427#endif
60428
60429#ifdef __LITTLE_ENDIAN__
60430__ai float64_t vrsqrted_f64(float64_t __p0) {
60431  float64_t __ret;
60432  __ret = (float64_t) __builtin_neon_vrsqrted_f64(__p0);
60433  return __ret;
60434}
60435#else
60436__ai float64_t vrsqrted_f64(float64_t __p0) {
60437  float64_t __ret;
60438  __ret = (float64_t) __builtin_neon_vrsqrted_f64(__p0);
60439  return __ret;
60440}
60441#endif
60442
60443#ifdef __LITTLE_ENDIAN__
60444__ai float32_t vrsqrtes_f32(float32_t __p0) {
60445  float32_t __ret;
60446  __ret = (float32_t) __builtin_neon_vrsqrtes_f32(__p0);
60447  return __ret;
60448}
60449#else
60450__ai float32_t vrsqrtes_f32(float32_t __p0) {
60451  float32_t __ret;
60452  __ret = (float32_t) __builtin_neon_vrsqrtes_f32(__p0);
60453  return __ret;
60454}
60455#endif
60456
60457#ifdef __LITTLE_ENDIAN__
60458__ai float64x2_t vrsqrtsq_f64(float64x2_t __p0, float64x2_t __p1) {
60459  float64x2_t __ret;
60460  __ret = (float64x2_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__p0, (int8x16_t)__p1, 42);
60461  return __ret;
60462}
60463#else
60464__ai float64x2_t vrsqrtsq_f64(float64x2_t __p0, float64x2_t __p1) {
60465  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60466  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60467  float64x2_t __ret;
60468  __ret = (float64x2_t) __builtin_neon_vrsqrtsq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 42);
60469  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
60470  return __ret;
60471}
60472#endif
60473
60474#ifdef __LITTLE_ENDIAN__
60475__ai float64x1_t vrsqrts_f64(float64x1_t __p0, float64x1_t __p1) {
60476  float64x1_t __ret;
60477  __ret = (float64x1_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
60478  return __ret;
60479}
60480#else
60481__ai float64x1_t vrsqrts_f64(float64x1_t __p0, float64x1_t __p1) {
60482  float64x1_t __ret;
60483  __ret = (float64x1_t) __builtin_neon_vrsqrts_v((int8x8_t)__p0, (int8x8_t)__p1, 10);
60484  return __ret;
60485}
60486#endif
60487
60488#ifdef __LITTLE_ENDIAN__
60489__ai float64_t vrsqrtsd_f64(float64_t __p0, float64_t __p1) {
60490  float64_t __ret;
60491  __ret = (float64_t) __builtin_neon_vrsqrtsd_f64(__p0, __p1);
60492  return __ret;
60493}
60494#else
60495__ai float64_t vrsqrtsd_f64(float64_t __p0, float64_t __p1) {
60496  float64_t __ret;
60497  __ret = (float64_t) __builtin_neon_vrsqrtsd_f64(__p0, __p1);
60498  return __ret;
60499}
60500#endif
60501
60502#ifdef __LITTLE_ENDIAN__
60503__ai float32_t vrsqrtss_f32(float32_t __p0, float32_t __p1) {
60504  float32_t __ret;
60505  __ret = (float32_t) __builtin_neon_vrsqrtss_f32(__p0, __p1);
60506  return __ret;
60507}
60508#else
60509__ai float32_t vrsqrtss_f32(float32_t __p0, float32_t __p1) {
60510  float32_t __ret;
60511  __ret = (float32_t) __builtin_neon_vrsqrtss_f32(__p0, __p1);
60512  return __ret;
60513}
60514#endif
60515
60516#ifdef __LITTLE_ENDIAN__
60517#define vrsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
60518  uint64_t __s0 = __p0; \
60519  uint64_t __s1 = __p1; \
60520  uint64_t __ret; \
60521  __ret = (uint64_t) __builtin_neon_vrsrad_n_u64(__s0, __s1, __p2); \
60522  __ret; \
60523})
60524#else
60525#define vrsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
60526  uint64_t __s0 = __p0; \
60527  uint64_t __s1 = __p1; \
60528  uint64_t __ret; \
60529  __ret = (uint64_t) __builtin_neon_vrsrad_n_u64(__s0, __s1, __p2); \
60530  __ret; \
60531})
60532#endif
60533
60534#ifdef __LITTLE_ENDIAN__
60535#define vrsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
60536  int64_t __s0 = __p0; \
60537  int64_t __s1 = __p1; \
60538  int64_t __ret; \
60539  __ret = (int64_t) __builtin_neon_vrsrad_n_s64(__s0, __s1, __p2); \
60540  __ret; \
60541})
60542#else
60543#define vrsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
60544  int64_t __s0 = __p0; \
60545  int64_t __s1 = __p1; \
60546  int64_t __ret; \
60547  __ret = (int64_t) __builtin_neon_vrsrad_n_s64(__s0, __s1, __p2); \
60548  __ret; \
60549})
60550#endif
60551
60552#ifdef __LITTLE_ENDIAN__
60553__ai uint16x8_t vrsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
60554  uint16x8_t __ret;
60555  __ret = vcombine_u16(__p0, vrsubhn_u32(__p1, __p2));
60556  return __ret;
60557}
60558#else
60559__ai uint16x8_t vrsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
60560  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60561  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60562  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
60563  uint16x8_t __ret;
60564  __ret = __noswap_vcombine_u16(__rev0, __noswap_vrsubhn_u32(__rev1, __rev2));
60565  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60566  return __ret;
60567}
60568#endif
60569
60570#ifdef __LITTLE_ENDIAN__
60571__ai uint32x4_t vrsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
60572  uint32x4_t __ret;
60573  __ret = vcombine_u32(__p0, vrsubhn_u64(__p1, __p2));
60574  return __ret;
60575}
60576#else
60577__ai uint32x4_t vrsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
60578  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60579  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60580  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
60581  uint32x4_t __ret;
60582  __ret = __noswap_vcombine_u32(__rev0, __noswap_vrsubhn_u64(__rev1, __rev2));
60583  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60584  return __ret;
60585}
60586#endif
60587
60588#ifdef __LITTLE_ENDIAN__
60589__ai uint8x16_t vrsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
60590  uint8x16_t __ret;
60591  __ret = vcombine_u8(__p0, vrsubhn_u16(__p1, __p2));
60592  return __ret;
60593}
60594#else
60595__ai uint8x16_t vrsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
60596  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60597  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60598  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
60599  uint8x16_t __ret;
60600  __ret = __noswap_vcombine_u8(__rev0, __noswap_vrsubhn_u16(__rev1, __rev2));
60601  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60602  return __ret;
60603}
60604#endif
60605
60606#ifdef __LITTLE_ENDIAN__
60607__ai int16x8_t vrsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
60608  int16x8_t __ret;
60609  __ret = vcombine_s16(__p0, vrsubhn_s32(__p1, __p2));
60610  return __ret;
60611}
60612#else
60613__ai int16x8_t vrsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
60614  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
60615  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
60616  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
60617  int16x8_t __ret;
60618  __ret = __noswap_vcombine_s16(__rev0, __noswap_vrsubhn_s32(__rev1, __rev2));
60619  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
60620  return __ret;
60621}
60622#endif
60623
60624#ifdef __LITTLE_ENDIAN__
60625__ai int32x4_t vrsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
60626  int32x4_t __ret;
60627  __ret = vcombine_s32(__p0, vrsubhn_s64(__p1, __p2));
60628  return __ret;
60629}
60630#else
60631__ai int32x4_t vrsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
60632  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
60633  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
60634  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
60635  int32x4_t __ret;
60636  __ret = __noswap_vcombine_s32(__rev0, __noswap_vrsubhn_s64(__rev1, __rev2));
60637  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
60638  return __ret;
60639}
60640#endif
60641
60642#ifdef __LITTLE_ENDIAN__
60643__ai int8x16_t vrsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
60644  int8x16_t __ret;
60645  __ret = vcombine_s8(__p0, vrsubhn_s16(__p1, __p2));
60646  return __ret;
60647}
60648#else
60649__ai int8x16_t vrsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
60650  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
60651  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
60652  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
60653  int8x16_t __ret;
60654  __ret = __noswap_vcombine_s8(__rev0, __noswap_vrsubhn_s16(__rev1, __rev2));
60655  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
60656  return __ret;
60657}
60658#endif
60659
60660#ifdef __LITTLE_ENDIAN__
60661#define vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60662  poly64_t __s0 = __p0; \
60663  poly64x1_t __s1 = __p1; \
60664  poly64x1_t __ret; \
60665  __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
60666  __ret; \
60667})
60668#else
60669#define vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60670  poly64_t __s0 = __p0; \
60671  poly64x1_t __s1 = __p1; \
60672  poly64x1_t __ret; \
60673  __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
60674  __ret; \
60675})
60676#define __noswap_vset_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60677  poly64_t __s0 = __p0; \
60678  poly64x1_t __s1 = __p1; \
60679  poly64x1_t __ret; \
60680  __ret = (poly64x1_t) __builtin_neon_vset_lane_i64(__s0, (int8x8_t)__s1, __p2); \
60681  __ret; \
60682})
60683#endif
60684
60685#ifdef __LITTLE_ENDIAN__
60686#define vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60687  poly64_t __s0 = __p0; \
60688  poly64x2_t __s1 = __p1; \
60689  poly64x2_t __ret; \
60690  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
60691  __ret; \
60692})
60693#else
60694#define vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60695  poly64_t __s0 = __p0; \
60696  poly64x2_t __s1 = __p1; \
60697  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60698  poly64x2_t __ret; \
60699  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__rev1, __p2); \
60700  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
60701  __ret; \
60702})
60703#define __noswap_vsetq_lane_p64(__p0, __p1, __p2) __extension__ ({ \
60704  poly64_t __s0 = __p0; \
60705  poly64x2_t __s1 = __p1; \
60706  poly64x2_t __ret; \
60707  __ret = (poly64x2_t) __builtin_neon_vsetq_lane_i64(__s0, (int8x16_t)__s1, __p2); \
60708  __ret; \
60709})
60710#endif
60711
60712#ifdef __LITTLE_ENDIAN__
60713#define vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60714  float64_t __s0 = __p0; \
60715  float64x2_t __s1 = __p1; \
60716  float64x2_t __ret; \
60717  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__s1, __p2); \
60718  __ret; \
60719})
60720#else
60721#define vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60722  float64_t __s0 = __p0; \
60723  float64x2_t __s1 = __p1; \
60724  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
60725  float64x2_t __ret; \
60726  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__rev1, __p2); \
60727  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
60728  __ret; \
60729})
60730#define __noswap_vsetq_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60731  float64_t __s0 = __p0; \
60732  float64x2_t __s1 = __p1; \
60733  float64x2_t __ret; \
60734  __ret = (float64x2_t) __builtin_neon_vsetq_lane_f64(__s0, (int8x16_t)__s1, __p2); \
60735  __ret; \
60736})
60737#endif
60738
60739#ifdef __LITTLE_ENDIAN__
60740#define vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60741  float64_t __s0 = __p0; \
60742  float64x1_t __s1 = __p1; \
60743  float64x1_t __ret; \
60744  __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
60745  __ret; \
60746})
60747#else
60748#define vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60749  float64_t __s0 = __p0; \
60750  float64x1_t __s1 = __p1; \
60751  float64x1_t __ret; \
60752  __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
60753  __ret; \
60754})
60755#define __noswap_vset_lane_f64(__p0, __p1, __p2) __extension__ ({ \
60756  float64_t __s0 = __p0; \
60757  float64x1_t __s1 = __p1; \
60758  float64x1_t __ret; \
60759  __ret = (float64x1_t) __builtin_neon_vset_lane_f64(__s0, (int8x8_t)__s1, __p2); \
60760  __ret; \
60761})
60762#endif
60763
60764#ifdef __LITTLE_ENDIAN__
60765__ai uint64_t vshld_u64(uint64_t __p0, uint64_t __p1) {
60766  uint64_t __ret;
60767  __ret = (uint64_t) __builtin_neon_vshld_u64(__p0, __p1);
60768  return __ret;
60769}
60770#else
60771__ai uint64_t vshld_u64(uint64_t __p0, uint64_t __p1) {
60772  uint64_t __ret;
60773  __ret = (uint64_t) __builtin_neon_vshld_u64(__p0, __p1);
60774  return __ret;
60775}
60776#endif
60777
60778#ifdef __LITTLE_ENDIAN__
60779__ai int64_t vshld_s64(int64_t __p0, int64_t __p1) {
60780  int64_t __ret;
60781  __ret = (int64_t) __builtin_neon_vshld_s64(__p0, __p1);
60782  return __ret;
60783}
60784#else
60785__ai int64_t vshld_s64(int64_t __p0, int64_t __p1) {
60786  int64_t __ret;
60787  __ret = (int64_t) __builtin_neon_vshld_s64(__p0, __p1);
60788  return __ret;
60789}
60790#endif
60791
60792#ifdef __LITTLE_ENDIAN__
60793#define vshld_n_u64(__p0, __p1) __extension__ ({ \
60794  uint64_t __s0 = __p0; \
60795  uint64_t __ret; \
60796  __ret = (uint64_t) __builtin_neon_vshld_n_u64(__s0, __p1); \
60797  __ret; \
60798})
60799#else
60800#define vshld_n_u64(__p0, __p1) __extension__ ({ \
60801  uint64_t __s0 = __p0; \
60802  uint64_t __ret; \
60803  __ret = (uint64_t) __builtin_neon_vshld_n_u64(__s0, __p1); \
60804  __ret; \
60805})
60806#endif
60807
60808#ifdef __LITTLE_ENDIAN__
60809#define vshld_n_s64(__p0, __p1) __extension__ ({ \
60810  int64_t __s0 = __p0; \
60811  int64_t __ret; \
60812  __ret = (int64_t) __builtin_neon_vshld_n_s64(__s0, __p1); \
60813  __ret; \
60814})
60815#else
60816#define vshld_n_s64(__p0, __p1) __extension__ ({ \
60817  int64_t __s0 = __p0; \
60818  int64_t __ret; \
60819  __ret = (int64_t) __builtin_neon_vshld_n_s64(__s0, __p1); \
60820  __ret; \
60821})
60822#endif
60823
60824#ifdef __LITTLE_ENDIAN__
60825#define vshll_high_n_u8(__p0_218, __p1_218) __extension__ ({ \
60826  uint8x16_t __s0_218 = __p0_218; \
60827  uint16x8_t __ret_218; \
60828  __ret_218 = (uint16x8_t)(vshll_n_u8(vget_high_u8(__s0_218), __p1_218)); \
60829  __ret_218; \
60830})
60831#else
60832#define vshll_high_n_u8(__p0_219, __p1_219) __extension__ ({ \
60833  uint8x16_t __s0_219 = __p0_219; \
60834  uint8x16_t __rev0_219;  __rev0_219 = __builtin_shufflevector(__s0_219, __s0_219, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60835  uint16x8_t __ret_219; \
60836  __ret_219 = (uint16x8_t)(__noswap_vshll_n_u8(__noswap_vget_high_u8(__rev0_219), __p1_219)); \
60837  __ret_219 = __builtin_shufflevector(__ret_219, __ret_219, 7, 6, 5, 4, 3, 2, 1, 0); \
60838  __ret_219; \
60839})
60840#endif
60841
60842#ifdef __LITTLE_ENDIAN__
60843#define vshll_high_n_u32(__p0_220, __p1_220) __extension__ ({ \
60844  uint32x4_t __s0_220 = __p0_220; \
60845  uint64x2_t __ret_220; \
60846  __ret_220 = (uint64x2_t)(vshll_n_u32(vget_high_u32(__s0_220), __p1_220)); \
60847  __ret_220; \
60848})
60849#else
60850#define vshll_high_n_u32(__p0_221, __p1_221) __extension__ ({ \
60851  uint32x4_t __s0_221 = __p0_221; \
60852  uint32x4_t __rev0_221;  __rev0_221 = __builtin_shufflevector(__s0_221, __s0_221, 3, 2, 1, 0); \
60853  uint64x2_t __ret_221; \
60854  __ret_221 = (uint64x2_t)(__noswap_vshll_n_u32(__noswap_vget_high_u32(__rev0_221), __p1_221)); \
60855  __ret_221 = __builtin_shufflevector(__ret_221, __ret_221, 1, 0); \
60856  __ret_221; \
60857})
60858#endif
60859
60860#ifdef __LITTLE_ENDIAN__
60861#define vshll_high_n_u16(__p0_222, __p1_222) __extension__ ({ \
60862  uint16x8_t __s0_222 = __p0_222; \
60863  uint32x4_t __ret_222; \
60864  __ret_222 = (uint32x4_t)(vshll_n_u16(vget_high_u16(__s0_222), __p1_222)); \
60865  __ret_222; \
60866})
60867#else
60868#define vshll_high_n_u16(__p0_223, __p1_223) __extension__ ({ \
60869  uint16x8_t __s0_223 = __p0_223; \
60870  uint16x8_t __rev0_223;  __rev0_223 = __builtin_shufflevector(__s0_223, __s0_223, 7, 6, 5, 4, 3, 2, 1, 0); \
60871  uint32x4_t __ret_223; \
60872  __ret_223 = (uint32x4_t)(__noswap_vshll_n_u16(__noswap_vget_high_u16(__rev0_223), __p1_223)); \
60873  __ret_223 = __builtin_shufflevector(__ret_223, __ret_223, 3, 2, 1, 0); \
60874  __ret_223; \
60875})
60876#endif
60877
60878#ifdef __LITTLE_ENDIAN__
60879#define vshll_high_n_s8(__p0_224, __p1_224) __extension__ ({ \
60880  int8x16_t __s0_224 = __p0_224; \
60881  int16x8_t __ret_224; \
60882  __ret_224 = (int16x8_t)(vshll_n_s8(vget_high_s8(__s0_224), __p1_224)); \
60883  __ret_224; \
60884})
60885#else
60886#define vshll_high_n_s8(__p0_225, __p1_225) __extension__ ({ \
60887  int8x16_t __s0_225 = __p0_225; \
60888  int8x16_t __rev0_225;  __rev0_225 = __builtin_shufflevector(__s0_225, __s0_225, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
60889  int16x8_t __ret_225; \
60890  __ret_225 = (int16x8_t)(__noswap_vshll_n_s8(__noswap_vget_high_s8(__rev0_225), __p1_225)); \
60891  __ret_225 = __builtin_shufflevector(__ret_225, __ret_225, 7, 6, 5, 4, 3, 2, 1, 0); \
60892  __ret_225; \
60893})
60894#endif
60895
60896#ifdef __LITTLE_ENDIAN__
60897#define vshll_high_n_s32(__p0_226, __p1_226) __extension__ ({ \
60898  int32x4_t __s0_226 = __p0_226; \
60899  int64x2_t __ret_226; \
60900  __ret_226 = (int64x2_t)(vshll_n_s32(vget_high_s32(__s0_226), __p1_226)); \
60901  __ret_226; \
60902})
60903#else
60904#define vshll_high_n_s32(__p0_227, __p1_227) __extension__ ({ \
60905  int32x4_t __s0_227 = __p0_227; \
60906  int32x4_t __rev0_227;  __rev0_227 = __builtin_shufflevector(__s0_227, __s0_227, 3, 2, 1, 0); \
60907  int64x2_t __ret_227; \
60908  __ret_227 = (int64x2_t)(__noswap_vshll_n_s32(__noswap_vget_high_s32(__rev0_227), __p1_227)); \
60909  __ret_227 = __builtin_shufflevector(__ret_227, __ret_227, 1, 0); \
60910  __ret_227; \
60911})
60912#endif
60913
60914#ifdef __LITTLE_ENDIAN__
60915#define vshll_high_n_s16(__p0_228, __p1_228) __extension__ ({ \
60916  int16x8_t __s0_228 = __p0_228; \
60917  int32x4_t __ret_228; \
60918  __ret_228 = (int32x4_t)(vshll_n_s16(vget_high_s16(__s0_228), __p1_228)); \
60919  __ret_228; \
60920})
60921#else
60922#define vshll_high_n_s16(__p0_229, __p1_229) __extension__ ({ \
60923  int16x8_t __s0_229 = __p0_229; \
60924  int16x8_t __rev0_229;  __rev0_229 = __builtin_shufflevector(__s0_229, __s0_229, 7, 6, 5, 4, 3, 2, 1, 0); \
60925  int32x4_t __ret_229; \
60926  __ret_229 = (int32x4_t)(__noswap_vshll_n_s16(__noswap_vget_high_s16(__rev0_229), __p1_229)); \
60927  __ret_229 = __builtin_shufflevector(__ret_229, __ret_229, 3, 2, 1, 0); \
60928  __ret_229; \
60929})
60930#endif
60931
60932#ifdef __LITTLE_ENDIAN__
60933#define vshrd_n_u64(__p0, __p1) __extension__ ({ \
60934  uint64_t __s0 = __p0; \
60935  uint64_t __ret; \
60936  __ret = (uint64_t) __builtin_neon_vshrd_n_u64(__s0, __p1); \
60937  __ret; \
60938})
60939#else
60940#define vshrd_n_u64(__p0, __p1) __extension__ ({ \
60941  uint64_t __s0 = __p0; \
60942  uint64_t __ret; \
60943  __ret = (uint64_t) __builtin_neon_vshrd_n_u64(__s0, __p1); \
60944  __ret; \
60945})
60946#endif
60947
60948#ifdef __LITTLE_ENDIAN__
60949#define vshrd_n_s64(__p0, __p1) __extension__ ({ \
60950  int64_t __s0 = __p0; \
60951  int64_t __ret; \
60952  __ret = (int64_t) __builtin_neon_vshrd_n_s64(__s0, __p1); \
60953  __ret; \
60954})
60955#else
60956#define vshrd_n_s64(__p0, __p1) __extension__ ({ \
60957  int64_t __s0 = __p0; \
60958  int64_t __ret; \
60959  __ret = (int64_t) __builtin_neon_vshrd_n_s64(__s0, __p1); \
60960  __ret; \
60961})
60962#endif
60963
60964#ifdef __LITTLE_ENDIAN__
60965#define vshrn_high_n_u32(__p0_230, __p1_230, __p2_230) __extension__ ({ \
60966  uint16x4_t __s0_230 = __p0_230; \
60967  uint32x4_t __s1_230 = __p1_230; \
60968  uint16x8_t __ret_230; \
60969  __ret_230 = (uint16x8_t)(vcombine_u16((uint16x4_t)(__s0_230), (uint16x4_t)(vshrn_n_u32(__s1_230, __p2_230)))); \
60970  __ret_230; \
60971})
60972#else
60973#define vshrn_high_n_u32(__p0_231, __p1_231, __p2_231) __extension__ ({ \
60974  uint16x4_t __s0_231 = __p0_231; \
60975  uint32x4_t __s1_231 = __p1_231; \
60976  uint16x4_t __rev0_231;  __rev0_231 = __builtin_shufflevector(__s0_231, __s0_231, 3, 2, 1, 0); \
60977  uint32x4_t __rev1_231;  __rev1_231 = __builtin_shufflevector(__s1_231, __s1_231, 3, 2, 1, 0); \
60978  uint16x8_t __ret_231; \
60979  __ret_231 = (uint16x8_t)(__noswap_vcombine_u16((uint16x4_t)(__rev0_231), (uint16x4_t)(__noswap_vshrn_n_u32(__rev1_231, __p2_231)))); \
60980  __ret_231 = __builtin_shufflevector(__ret_231, __ret_231, 7, 6, 5, 4, 3, 2, 1, 0); \
60981  __ret_231; \
60982})
60983#endif
60984
60985#ifdef __LITTLE_ENDIAN__
60986#define vshrn_high_n_u64(__p0_232, __p1_232, __p2_232) __extension__ ({ \
60987  uint32x2_t __s0_232 = __p0_232; \
60988  uint64x2_t __s1_232 = __p1_232; \
60989  uint32x4_t __ret_232; \
60990  __ret_232 = (uint32x4_t)(vcombine_u32((uint32x2_t)(__s0_232), (uint32x2_t)(vshrn_n_u64(__s1_232, __p2_232)))); \
60991  __ret_232; \
60992})
60993#else
60994#define vshrn_high_n_u64(__p0_233, __p1_233, __p2_233) __extension__ ({ \
60995  uint32x2_t __s0_233 = __p0_233; \
60996  uint64x2_t __s1_233 = __p1_233; \
60997  uint32x2_t __rev0_233;  __rev0_233 = __builtin_shufflevector(__s0_233, __s0_233, 1, 0); \
60998  uint64x2_t __rev1_233;  __rev1_233 = __builtin_shufflevector(__s1_233, __s1_233, 1, 0); \
60999  uint32x4_t __ret_233; \
61000  __ret_233 = (uint32x4_t)(__noswap_vcombine_u32((uint32x2_t)(__rev0_233), (uint32x2_t)(__noswap_vshrn_n_u64(__rev1_233, __p2_233)))); \
61001  __ret_233 = __builtin_shufflevector(__ret_233, __ret_233, 3, 2, 1, 0); \
61002  __ret_233; \
61003})
61004#endif
61005
61006#ifdef __LITTLE_ENDIAN__
61007#define vshrn_high_n_u16(__p0_234, __p1_234, __p2_234) __extension__ ({ \
61008  uint8x8_t __s0_234 = __p0_234; \
61009  uint16x8_t __s1_234 = __p1_234; \
61010  uint8x16_t __ret_234; \
61011  __ret_234 = (uint8x16_t)(vcombine_u8((uint8x8_t)(__s0_234), (uint8x8_t)(vshrn_n_u16(__s1_234, __p2_234)))); \
61012  __ret_234; \
61013})
61014#else
61015#define vshrn_high_n_u16(__p0_235, __p1_235, __p2_235) __extension__ ({ \
61016  uint8x8_t __s0_235 = __p0_235; \
61017  uint16x8_t __s1_235 = __p1_235; \
61018  uint8x8_t __rev0_235;  __rev0_235 = __builtin_shufflevector(__s0_235, __s0_235, 7, 6, 5, 4, 3, 2, 1, 0); \
61019  uint16x8_t __rev1_235;  __rev1_235 = __builtin_shufflevector(__s1_235, __s1_235, 7, 6, 5, 4, 3, 2, 1, 0); \
61020  uint8x16_t __ret_235; \
61021  __ret_235 = (uint8x16_t)(__noswap_vcombine_u8((uint8x8_t)(__rev0_235), (uint8x8_t)(__noswap_vshrn_n_u16(__rev1_235, __p2_235)))); \
61022  __ret_235 = __builtin_shufflevector(__ret_235, __ret_235, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61023  __ret_235; \
61024})
61025#endif
61026
61027#ifdef __LITTLE_ENDIAN__
61028#define vshrn_high_n_s32(__p0_236, __p1_236, __p2_236) __extension__ ({ \
61029  int16x4_t __s0_236 = __p0_236; \
61030  int32x4_t __s1_236 = __p1_236; \
61031  int16x8_t __ret_236; \
61032  __ret_236 = (int16x8_t)(vcombine_s16((int16x4_t)(__s0_236), (int16x4_t)(vshrn_n_s32(__s1_236, __p2_236)))); \
61033  __ret_236; \
61034})
61035#else
61036#define vshrn_high_n_s32(__p0_237, __p1_237, __p2_237) __extension__ ({ \
61037  int16x4_t __s0_237 = __p0_237; \
61038  int32x4_t __s1_237 = __p1_237; \
61039  int16x4_t __rev0_237;  __rev0_237 = __builtin_shufflevector(__s0_237, __s0_237, 3, 2, 1, 0); \
61040  int32x4_t __rev1_237;  __rev1_237 = __builtin_shufflevector(__s1_237, __s1_237, 3, 2, 1, 0); \
61041  int16x8_t __ret_237; \
61042  __ret_237 = (int16x8_t)(__noswap_vcombine_s16((int16x4_t)(__rev0_237), (int16x4_t)(__noswap_vshrn_n_s32(__rev1_237, __p2_237)))); \
61043  __ret_237 = __builtin_shufflevector(__ret_237, __ret_237, 7, 6, 5, 4, 3, 2, 1, 0); \
61044  __ret_237; \
61045})
61046#endif
61047
61048#ifdef __LITTLE_ENDIAN__
61049#define vshrn_high_n_s64(__p0_238, __p1_238, __p2_238) __extension__ ({ \
61050  int32x2_t __s0_238 = __p0_238; \
61051  int64x2_t __s1_238 = __p1_238; \
61052  int32x4_t __ret_238; \
61053  __ret_238 = (int32x4_t)(vcombine_s32((int32x2_t)(__s0_238), (int32x2_t)(vshrn_n_s64(__s1_238, __p2_238)))); \
61054  __ret_238; \
61055})
61056#else
61057#define vshrn_high_n_s64(__p0_239, __p1_239, __p2_239) __extension__ ({ \
61058  int32x2_t __s0_239 = __p0_239; \
61059  int64x2_t __s1_239 = __p1_239; \
61060  int32x2_t __rev0_239;  __rev0_239 = __builtin_shufflevector(__s0_239, __s0_239, 1, 0); \
61061  int64x2_t __rev1_239;  __rev1_239 = __builtin_shufflevector(__s1_239, __s1_239, 1, 0); \
61062  int32x4_t __ret_239; \
61063  __ret_239 = (int32x4_t)(__noswap_vcombine_s32((int32x2_t)(__rev0_239), (int32x2_t)(__noswap_vshrn_n_s64(__rev1_239, __p2_239)))); \
61064  __ret_239 = __builtin_shufflevector(__ret_239, __ret_239, 3, 2, 1, 0); \
61065  __ret_239; \
61066})
61067#endif
61068
61069#ifdef __LITTLE_ENDIAN__
61070#define vshrn_high_n_s16(__p0_240, __p1_240, __p2_240) __extension__ ({ \
61071  int8x8_t __s0_240 = __p0_240; \
61072  int16x8_t __s1_240 = __p1_240; \
61073  int8x16_t __ret_240; \
61074  __ret_240 = (int8x16_t)(vcombine_s8((int8x8_t)(__s0_240), (int8x8_t)(vshrn_n_s16(__s1_240, __p2_240)))); \
61075  __ret_240; \
61076})
61077#else
61078#define vshrn_high_n_s16(__p0_241, __p1_241, __p2_241) __extension__ ({ \
61079  int8x8_t __s0_241 = __p0_241; \
61080  int16x8_t __s1_241 = __p1_241; \
61081  int8x8_t __rev0_241;  __rev0_241 = __builtin_shufflevector(__s0_241, __s0_241, 7, 6, 5, 4, 3, 2, 1, 0); \
61082  int16x8_t __rev1_241;  __rev1_241 = __builtin_shufflevector(__s1_241, __s1_241, 7, 6, 5, 4, 3, 2, 1, 0); \
61083  int8x16_t __ret_241; \
61084  __ret_241 = (int8x16_t)(__noswap_vcombine_s8((int8x8_t)(__rev0_241), (int8x8_t)(__noswap_vshrn_n_s16(__rev1_241, __p2_241)))); \
61085  __ret_241 = __builtin_shufflevector(__ret_241, __ret_241, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61086  __ret_241; \
61087})
61088#endif
61089
61090#ifdef __LITTLE_ENDIAN__
61091#define vslid_n_u64(__p0, __p1, __p2) __extension__ ({ \
61092  uint64_t __s0 = __p0; \
61093  uint64_t __s1 = __p1; \
61094  uint64_t __ret; \
61095  __ret = (uint64_t) __builtin_neon_vslid_n_u64(__s0, __s1, __p2); \
61096  __ret; \
61097})
61098#else
61099#define vslid_n_u64(__p0, __p1, __p2) __extension__ ({ \
61100  uint64_t __s0 = __p0; \
61101  uint64_t __s1 = __p1; \
61102  uint64_t __ret; \
61103  __ret = (uint64_t) __builtin_neon_vslid_n_u64(__s0, __s1, __p2); \
61104  __ret; \
61105})
61106#endif
61107
61108#ifdef __LITTLE_ENDIAN__
61109#define vslid_n_s64(__p0, __p1, __p2) __extension__ ({ \
61110  int64_t __s0 = __p0; \
61111  int64_t __s1 = __p1; \
61112  int64_t __ret; \
61113  __ret = (int64_t) __builtin_neon_vslid_n_s64(__s0, __s1, __p2); \
61114  __ret; \
61115})
61116#else
61117#define vslid_n_s64(__p0, __p1, __p2) __extension__ ({ \
61118  int64_t __s0 = __p0; \
61119  int64_t __s1 = __p1; \
61120  int64_t __ret; \
61121  __ret = (int64_t) __builtin_neon_vslid_n_s64(__s0, __s1, __p2); \
61122  __ret; \
61123})
61124#endif
61125
61126#ifdef __LITTLE_ENDIAN__
61127#define vsli_n_p64(__p0, __p1, __p2) __extension__ ({ \
61128  poly64x1_t __s0 = __p0; \
61129  poly64x1_t __s1 = __p1; \
61130  poly64x1_t __ret; \
61131  __ret = (poly64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
61132  __ret; \
61133})
61134#else
61135#define vsli_n_p64(__p0, __p1, __p2) __extension__ ({ \
61136  poly64x1_t __s0 = __p0; \
61137  poly64x1_t __s1 = __p1; \
61138  poly64x1_t __ret; \
61139  __ret = (poly64x1_t) __builtin_neon_vsli_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
61140  __ret; \
61141})
61142#endif
61143
61144#ifdef __LITTLE_ENDIAN__
61145#define vsliq_n_p64(__p0, __p1, __p2) __extension__ ({ \
61146  poly64x2_t __s0 = __p0; \
61147  poly64x2_t __s1 = __p1; \
61148  poly64x2_t __ret; \
61149  __ret = (poly64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
61150  __ret; \
61151})
61152#else
61153#define vsliq_n_p64(__p0, __p1, __p2) __extension__ ({ \
61154  poly64x2_t __s0 = __p0; \
61155  poly64x2_t __s1 = __p1; \
61156  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
61157  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61158  poly64x2_t __ret; \
61159  __ret = (poly64x2_t) __builtin_neon_vsliq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
61160  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
61161  __ret; \
61162})
61163#endif
61164
61165#ifdef __LITTLE_ENDIAN__
61166__ai uint8_t vsqaddb_u8(uint8_t __p0, uint8_t __p1) {
61167  uint8_t __ret;
61168  __ret = (uint8_t) __builtin_neon_vsqaddb_u8(__p0, __p1);
61169  return __ret;
61170}
61171#else
61172__ai uint8_t vsqaddb_u8(uint8_t __p0, uint8_t __p1) {
61173  uint8_t __ret;
61174  __ret = (uint8_t) __builtin_neon_vsqaddb_u8(__p0, __p1);
61175  return __ret;
61176}
61177#endif
61178
61179#ifdef __LITTLE_ENDIAN__
61180__ai uint32_t vsqadds_u32(uint32_t __p0, uint32_t __p1) {
61181  uint32_t __ret;
61182  __ret = (uint32_t) __builtin_neon_vsqadds_u32(__p0, __p1);
61183  return __ret;
61184}
61185#else
61186__ai uint32_t vsqadds_u32(uint32_t __p0, uint32_t __p1) {
61187  uint32_t __ret;
61188  __ret = (uint32_t) __builtin_neon_vsqadds_u32(__p0, __p1);
61189  return __ret;
61190}
61191#endif
61192
61193#ifdef __LITTLE_ENDIAN__
61194__ai uint64_t vsqaddd_u64(uint64_t __p0, uint64_t __p1) {
61195  uint64_t __ret;
61196  __ret = (uint64_t) __builtin_neon_vsqaddd_u64(__p0, __p1);
61197  return __ret;
61198}
61199#else
61200__ai uint64_t vsqaddd_u64(uint64_t __p0, uint64_t __p1) {
61201  uint64_t __ret;
61202  __ret = (uint64_t) __builtin_neon_vsqaddd_u64(__p0, __p1);
61203  return __ret;
61204}
61205#endif
61206
61207#ifdef __LITTLE_ENDIAN__
61208__ai uint16_t vsqaddh_u16(uint16_t __p0, uint16_t __p1) {
61209  uint16_t __ret;
61210  __ret = (uint16_t) __builtin_neon_vsqaddh_u16(__p0, __p1);
61211  return __ret;
61212}
61213#else
61214__ai uint16_t vsqaddh_u16(uint16_t __p0, uint16_t __p1) {
61215  uint16_t __ret;
61216  __ret = (uint16_t) __builtin_neon_vsqaddh_u16(__p0, __p1);
61217  return __ret;
61218}
61219#endif
61220
61221#ifdef __LITTLE_ENDIAN__
61222__ai uint8x16_t vsqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
61223  uint8x16_t __ret;
61224  __ret = (uint8x16_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 48);
61225  return __ret;
61226}
61227#else
61228__ai uint8x16_t vsqaddq_u8(uint8x16_t __p0, uint8x16_t __p1) {
61229  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61230  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61231  uint8x16_t __ret;
61232  __ret = (uint8x16_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 48);
61233  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
61234  return __ret;
61235}
61236#endif
61237
61238#ifdef __LITTLE_ENDIAN__
61239__ai uint32x4_t vsqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
61240  uint32x4_t __ret;
61241  __ret = (uint32x4_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 50);
61242  return __ret;
61243}
61244#else
61245__ai uint32x4_t vsqaddq_u32(uint32x4_t __p0, uint32x4_t __p1) {
61246  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61247  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61248  uint32x4_t __ret;
61249  __ret = (uint32x4_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50);
61250  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61251  return __ret;
61252}
61253#endif
61254
61255#ifdef __LITTLE_ENDIAN__
61256__ai uint64x2_t vsqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
61257  uint64x2_t __ret;
61258  __ret = (uint64x2_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
61259  return __ret;
61260}
61261#else
61262__ai uint64x2_t vsqaddq_u64(uint64x2_t __p0, uint64x2_t __p1) {
61263  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61264  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61265  uint64x2_t __ret;
61266  __ret = (uint64x2_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
61267  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61268  return __ret;
61269}
61270#endif
61271
61272#ifdef __LITTLE_ENDIAN__
61273__ai uint16x8_t vsqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
61274  uint16x8_t __ret;
61275  __ret = (uint16x8_t) __builtin_neon_vsqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 49);
61276  return __ret;
61277}
61278#else
61279__ai uint16x8_t vsqaddq_u16(uint16x8_t __p0, uint16x8_t __p1) {
61280  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61281  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61282  uint16x8_t __ret;
61283  __ret = (uint16x8_t) __builtin_neon_vsqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 49);
61284  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61285  return __ret;
61286}
61287#endif
61288
61289#ifdef __LITTLE_ENDIAN__
61290__ai uint8x8_t vsqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
61291  uint8x8_t __ret;
61292  __ret = (uint8x8_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 16);
61293  return __ret;
61294}
61295#else
61296__ai uint8x8_t vsqadd_u8(uint8x8_t __p0, uint8x8_t __p1) {
61297  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
61298  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
61299  uint8x8_t __ret;
61300  __ret = (uint8x8_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 16);
61301  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
61302  return __ret;
61303}
61304#endif
61305
61306#ifdef __LITTLE_ENDIAN__
61307__ai uint32x2_t vsqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
61308  uint32x2_t __ret;
61309  __ret = (uint32x2_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 18);
61310  return __ret;
61311}
61312#else
61313__ai uint32x2_t vsqadd_u32(uint32x2_t __p0, uint32x2_t __p1) {
61314  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61315  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
61316  uint32x2_t __ret;
61317  __ret = (uint32x2_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 18);
61318  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61319  return __ret;
61320}
61321#endif
61322
61323#ifdef __LITTLE_ENDIAN__
61324__ai uint64x1_t vsqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
61325  uint64x1_t __ret;
61326  __ret = (uint64x1_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
61327  return __ret;
61328}
61329#else
61330__ai uint64x1_t vsqadd_u64(uint64x1_t __p0, uint64x1_t __p1) {
61331  uint64x1_t __ret;
61332  __ret = (uint64x1_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
61333  return __ret;
61334}
61335#endif
61336
61337#ifdef __LITTLE_ENDIAN__
61338__ai uint16x4_t vsqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
61339  uint16x4_t __ret;
61340  __ret = (uint16x4_t) __builtin_neon_vsqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 17);
61341  return __ret;
61342}
61343#else
61344__ai uint16x4_t vsqadd_u16(uint16x4_t __p0, uint16x4_t __p1) {
61345  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61346  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
61347  uint16x4_t __ret;
61348  __ret = (uint16x4_t) __builtin_neon_vsqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 17);
61349  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61350  return __ret;
61351}
61352#endif
61353
61354#ifdef __LITTLE_ENDIAN__
61355__ai float64x2_t vsqrtq_f64(float64x2_t __p0) {
61356  float64x2_t __ret;
61357  __ret = (float64x2_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 42);
61358  return __ret;
61359}
61360#else
61361__ai float64x2_t vsqrtq_f64(float64x2_t __p0) {
61362  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61363  float64x2_t __ret;
61364  __ret = (float64x2_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 42);
61365  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61366  return __ret;
61367}
61368#endif
61369
61370#ifdef __LITTLE_ENDIAN__
61371__ai float32x4_t vsqrtq_f32(float32x4_t __p0) {
61372  float32x4_t __ret;
61373  __ret = (float32x4_t) __builtin_neon_vsqrtq_v((int8x16_t)__p0, 41);
61374  return __ret;
61375}
61376#else
61377__ai float32x4_t vsqrtq_f32(float32x4_t __p0) {
61378  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
61379  float32x4_t __ret;
61380  __ret = (float32x4_t) __builtin_neon_vsqrtq_v((int8x16_t)__rev0, 41);
61381  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
61382  return __ret;
61383}
61384#endif
61385
61386#ifdef __LITTLE_ENDIAN__
61387__ai float64x1_t vsqrt_f64(float64x1_t __p0) {
61388  float64x1_t __ret;
61389  __ret = (float64x1_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 10);
61390  return __ret;
61391}
61392#else
61393__ai float64x1_t vsqrt_f64(float64x1_t __p0) {
61394  float64x1_t __ret;
61395  __ret = (float64x1_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 10);
61396  return __ret;
61397}
61398#endif
61399
61400#ifdef __LITTLE_ENDIAN__
61401__ai float32x2_t vsqrt_f32(float32x2_t __p0) {
61402  float32x2_t __ret;
61403  __ret = (float32x2_t) __builtin_neon_vsqrt_v((int8x8_t)__p0, 9);
61404  return __ret;
61405}
61406#else
61407__ai float32x2_t vsqrt_f32(float32x2_t __p0) {
61408  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
61409  float32x2_t __ret;
61410  __ret = (float32x2_t) __builtin_neon_vsqrt_v((int8x8_t)__rev0, 9);
61411  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
61412  return __ret;
61413}
61414#endif
61415
61416#ifdef __LITTLE_ENDIAN__
61417#define vsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
61418  uint64_t __s0 = __p0; \
61419  uint64_t __s1 = __p1; \
61420  uint64_t __ret; \
61421  __ret = (uint64_t) __builtin_neon_vsrad_n_u64(__s0, __s1, __p2); \
61422  __ret; \
61423})
61424#else
61425#define vsrad_n_u64(__p0, __p1, __p2) __extension__ ({ \
61426  uint64_t __s0 = __p0; \
61427  uint64_t __s1 = __p1; \
61428  uint64_t __ret; \
61429  __ret = (uint64_t) __builtin_neon_vsrad_n_u64(__s0, __s1, __p2); \
61430  __ret; \
61431})
61432#endif
61433
61434#ifdef __LITTLE_ENDIAN__
61435#define vsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
61436  int64_t __s0 = __p0; \
61437  int64_t __s1 = __p1; \
61438  int64_t __ret; \
61439  __ret = (int64_t) __builtin_neon_vsrad_n_s64(__s0, __s1, __p2); \
61440  __ret; \
61441})
61442#else
61443#define vsrad_n_s64(__p0, __p1, __p2) __extension__ ({ \
61444  int64_t __s0 = __p0; \
61445  int64_t __s1 = __p1; \
61446  int64_t __ret; \
61447  __ret = (int64_t) __builtin_neon_vsrad_n_s64(__s0, __s1, __p2); \
61448  __ret; \
61449})
61450#endif
61451
61452#ifdef __LITTLE_ENDIAN__
61453#define vsrid_n_u64(__p0, __p1, __p2) __extension__ ({ \
61454  uint64_t __s0 = __p0; \
61455  uint64_t __s1 = __p1; \
61456  uint64_t __ret; \
61457  __ret = (uint64_t) __builtin_neon_vsrid_n_u64(__s0, __s1, __p2); \
61458  __ret; \
61459})
61460#else
61461#define vsrid_n_u64(__p0, __p1, __p2) __extension__ ({ \
61462  uint64_t __s0 = __p0; \
61463  uint64_t __s1 = __p1; \
61464  uint64_t __ret; \
61465  __ret = (uint64_t) __builtin_neon_vsrid_n_u64(__s0, __s1, __p2); \
61466  __ret; \
61467})
61468#endif
61469
61470#ifdef __LITTLE_ENDIAN__
61471#define vsrid_n_s64(__p0, __p1, __p2) __extension__ ({ \
61472  int64_t __s0 = __p0; \
61473  int64_t __s1 = __p1; \
61474  int64_t __ret; \
61475  __ret = (int64_t) __builtin_neon_vsrid_n_s64(__s0, __s1, __p2); \
61476  __ret; \
61477})
61478#else
61479#define vsrid_n_s64(__p0, __p1, __p2) __extension__ ({ \
61480  int64_t __s0 = __p0; \
61481  int64_t __s1 = __p1; \
61482  int64_t __ret; \
61483  __ret = (int64_t) __builtin_neon_vsrid_n_s64(__s0, __s1, __p2); \
61484  __ret; \
61485})
61486#endif
61487
61488#ifdef __LITTLE_ENDIAN__
61489#define vsri_n_p64(__p0, __p1, __p2) __extension__ ({ \
61490  poly64x1_t __s0 = __p0; \
61491  poly64x1_t __s1 = __p1; \
61492  poly64x1_t __ret; \
61493  __ret = (poly64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
61494  __ret; \
61495})
61496#else
61497#define vsri_n_p64(__p0, __p1, __p2) __extension__ ({ \
61498  poly64x1_t __s0 = __p0; \
61499  poly64x1_t __s1 = __p1; \
61500  poly64x1_t __ret; \
61501  __ret = (poly64x1_t) __builtin_neon_vsri_n_v((int8x8_t)__s0, (int8x8_t)__s1, __p2, 6); \
61502  __ret; \
61503})
61504#endif
61505
61506#ifdef __LITTLE_ENDIAN__
61507#define vsriq_n_p64(__p0, __p1, __p2) __extension__ ({ \
61508  poly64x2_t __s0 = __p0; \
61509  poly64x2_t __s1 = __p1; \
61510  poly64x2_t __ret; \
61511  __ret = (poly64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 38); \
61512  __ret; \
61513})
61514#else
61515#define vsriq_n_p64(__p0, __p1, __p2) __extension__ ({ \
61516  poly64x2_t __s0 = __p0; \
61517  poly64x2_t __s1 = __p1; \
61518  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
61519  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61520  poly64x2_t __ret; \
61521  __ret = (poly64x2_t) __builtin_neon_vsriq_n_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 38); \
61522  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
61523  __ret; \
61524})
61525#endif
61526
61527#ifdef __LITTLE_ENDIAN__
61528#define vst1_p64(__p0, __p1) __extension__ ({ \
61529  poly64x1_t __s1 = __p1; \
61530  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 6); \
61531})
61532#else
61533#define vst1_p64(__p0, __p1) __extension__ ({ \
61534  poly64x1_t __s1 = __p1; \
61535  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 6); \
61536})
61537#endif
61538
61539#ifdef __LITTLE_ENDIAN__
61540#define vst1q_p64(__p0, __p1) __extension__ ({ \
61541  poly64x2_t __s1 = __p1; \
61542  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 38); \
61543})
61544#else
61545#define vst1q_p64(__p0, __p1) __extension__ ({ \
61546  poly64x2_t __s1 = __p1; \
61547  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61548  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 38); \
61549})
61550#endif
61551
61552#ifdef __LITTLE_ENDIAN__
61553#define vst1q_f64(__p0, __p1) __extension__ ({ \
61554  float64x2_t __s1 = __p1; \
61555  __builtin_neon_vst1q_v(__p0, (int8x16_t)__s1, 42); \
61556})
61557#else
61558#define vst1q_f64(__p0, __p1) __extension__ ({ \
61559  float64x2_t __s1 = __p1; \
61560  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61561  __builtin_neon_vst1q_v(__p0, (int8x16_t)__rev1, 42); \
61562})
61563#endif
61564
61565#ifdef __LITTLE_ENDIAN__
61566#define vst1_f64(__p0, __p1) __extension__ ({ \
61567  float64x1_t __s1 = __p1; \
61568  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 10); \
61569})
61570#else
61571#define vst1_f64(__p0, __p1) __extension__ ({ \
61572  float64x1_t __s1 = __p1; \
61573  __builtin_neon_vst1_v(__p0, (int8x8_t)__s1, 10); \
61574})
61575#endif
61576
61577#ifdef __LITTLE_ENDIAN__
61578#define vst1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61579  poly64x1_t __s1 = __p1; \
61580  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
61581})
61582#else
61583#define vst1_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61584  poly64x1_t __s1 = __p1; \
61585  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 6); \
61586})
61587#endif
61588
61589#ifdef __LITTLE_ENDIAN__
61590#define vst1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61591  poly64x2_t __s1 = __p1; \
61592  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 38); \
61593})
61594#else
61595#define vst1q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
61596  poly64x2_t __s1 = __p1; \
61597  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61598  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 38); \
61599})
61600#endif
61601
61602#ifdef __LITTLE_ENDIAN__
61603#define vst1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61604  float64x2_t __s1 = __p1; \
61605  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__s1, __p2, 42); \
61606})
61607#else
61608#define vst1q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61609  float64x2_t __s1 = __p1; \
61610  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
61611  __builtin_neon_vst1q_lane_v(__p0, (int8x16_t)__rev1, __p2, 42); \
61612})
61613#endif
61614
61615#ifdef __LITTLE_ENDIAN__
61616#define vst1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61617  float64x1_t __s1 = __p1; \
61618  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
61619})
61620#else
61621#define vst1_lane_f64(__p0, __p1, __p2) __extension__ ({ \
61622  float64x1_t __s1 = __p1; \
61623  __builtin_neon_vst1_lane_v(__p0, (int8x8_t)__s1, __p2, 10); \
61624})
61625#endif
61626
61627#ifdef __LITTLE_ENDIAN__
61628#define vst1_p8_x2(__p0, __p1) __extension__ ({ \
61629  poly8x8x2_t __s1 = __p1; \
61630  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 4); \
61631})
61632#else
61633#define vst1_p8_x2(__p0, __p1) __extension__ ({ \
61634  poly8x8x2_t __s1 = __p1; \
61635  poly8x8x2_t __rev1; \
61636  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61637  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61638  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 4); \
61639})
61640#endif
61641
61642#ifdef __LITTLE_ENDIAN__
61643#define vst1_p64_x2(__p0, __p1) __extension__ ({ \
61644  poly64x1x2_t __s1 = __p1; \
61645  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
61646})
61647#else
61648#define vst1_p64_x2(__p0, __p1) __extension__ ({ \
61649  poly64x1x2_t __s1 = __p1; \
61650  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
61651})
61652#endif
61653
61654#ifdef __LITTLE_ENDIAN__
61655#define vst1_p16_x2(__p0, __p1) __extension__ ({ \
61656  poly16x4x2_t __s1 = __p1; \
61657  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 5); \
61658})
61659#else
61660#define vst1_p16_x2(__p0, __p1) __extension__ ({ \
61661  poly16x4x2_t __s1 = __p1; \
61662  poly16x4x2_t __rev1; \
61663  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61664  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61665  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 5); \
61666})
61667#endif
61668
61669#ifdef __LITTLE_ENDIAN__
61670#define vst1q_p8_x2(__p0, __p1) __extension__ ({ \
61671  poly8x16x2_t __s1 = __p1; \
61672  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 36); \
61673})
61674#else
61675#define vst1q_p8_x2(__p0, __p1) __extension__ ({ \
61676  poly8x16x2_t __s1 = __p1; \
61677  poly8x16x2_t __rev1; \
61678  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61679  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61680  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 36); \
61681})
61682#endif
61683
61684#ifdef __LITTLE_ENDIAN__
61685#define vst1q_p64_x2(__p0, __p1) __extension__ ({ \
61686  poly64x2x2_t __s1 = __p1; \
61687  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 38); \
61688})
61689#else
61690#define vst1q_p64_x2(__p0, __p1) __extension__ ({ \
61691  poly64x2x2_t __s1 = __p1; \
61692  poly64x2x2_t __rev1; \
61693  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61694  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61695  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 38); \
61696})
61697#endif
61698
61699#ifdef __LITTLE_ENDIAN__
61700#define vst1q_p16_x2(__p0, __p1) __extension__ ({ \
61701  poly16x8x2_t __s1 = __p1; \
61702  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 37); \
61703})
61704#else
61705#define vst1q_p16_x2(__p0, __p1) __extension__ ({ \
61706  poly16x8x2_t __s1 = __p1; \
61707  poly16x8x2_t __rev1; \
61708  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61709  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61710  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 37); \
61711})
61712#endif
61713
61714#ifdef __LITTLE_ENDIAN__
61715#define vst1q_u8_x2(__p0, __p1) __extension__ ({ \
61716  uint8x16x2_t __s1 = __p1; \
61717  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 48); \
61718})
61719#else
61720#define vst1q_u8_x2(__p0, __p1) __extension__ ({ \
61721  uint8x16x2_t __s1 = __p1; \
61722  uint8x16x2_t __rev1; \
61723  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61724  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61725  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 48); \
61726})
61727#endif
61728
61729#ifdef __LITTLE_ENDIAN__
61730#define vst1q_u32_x2(__p0, __p1) __extension__ ({ \
61731  uint32x4x2_t __s1 = __p1; \
61732  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 50); \
61733})
61734#else
61735#define vst1q_u32_x2(__p0, __p1) __extension__ ({ \
61736  uint32x4x2_t __s1 = __p1; \
61737  uint32x4x2_t __rev1; \
61738  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61739  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61740  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 50); \
61741})
61742#endif
61743
61744#ifdef __LITTLE_ENDIAN__
61745#define vst1q_u64_x2(__p0, __p1) __extension__ ({ \
61746  uint64x2x2_t __s1 = __p1; \
61747  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 51); \
61748})
61749#else
61750#define vst1q_u64_x2(__p0, __p1) __extension__ ({ \
61751  uint64x2x2_t __s1 = __p1; \
61752  uint64x2x2_t __rev1; \
61753  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61754  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61755  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 51); \
61756})
61757#endif
61758
61759#ifdef __LITTLE_ENDIAN__
61760#define vst1q_u16_x2(__p0, __p1) __extension__ ({ \
61761  uint16x8x2_t __s1 = __p1; \
61762  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 49); \
61763})
61764#else
61765#define vst1q_u16_x2(__p0, __p1) __extension__ ({ \
61766  uint16x8x2_t __s1 = __p1; \
61767  uint16x8x2_t __rev1; \
61768  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61769  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61770  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 49); \
61771})
61772#endif
61773
61774#ifdef __LITTLE_ENDIAN__
61775#define vst1q_s8_x2(__p0, __p1) __extension__ ({ \
61776  int8x16x2_t __s1 = __p1; \
61777  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 32); \
61778})
61779#else
61780#define vst1q_s8_x2(__p0, __p1) __extension__ ({ \
61781  int8x16x2_t __s1 = __p1; \
61782  int8x16x2_t __rev1; \
61783  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61784  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
61785  __builtin_neon_vst1q_x2_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 32); \
61786})
61787#endif
61788
61789#ifdef __LITTLE_ENDIAN__
61790#define vst1q_f64_x2(__p0, __p1) __extension__ ({ \
61791  float64x2x2_t __s1 = __p1; \
61792  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 42); \
61793})
61794#else
61795#define vst1q_f64_x2(__p0, __p1) __extension__ ({ \
61796  float64x2x2_t __s1 = __p1; \
61797  float64x2x2_t __rev1; \
61798  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61799  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61800  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 42); \
61801})
61802#endif
61803
61804#ifdef __LITTLE_ENDIAN__
61805#define vst1q_f32_x2(__p0, __p1) __extension__ ({ \
61806  float32x4x2_t __s1 = __p1; \
61807  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 41); \
61808})
61809#else
61810#define vst1q_f32_x2(__p0, __p1) __extension__ ({ \
61811  float32x4x2_t __s1 = __p1; \
61812  float32x4x2_t __rev1; \
61813  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61814  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61815  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 41); \
61816})
61817#endif
61818
61819#ifdef __LITTLE_ENDIAN__
61820#define vst1q_f16_x2(__p0, __p1) __extension__ ({ \
61821  float16x8x2_t __s1 = __p1; \
61822  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 40); \
61823})
61824#else
61825#define vst1q_f16_x2(__p0, __p1) __extension__ ({ \
61826  float16x8x2_t __s1 = __p1; \
61827  float16x8x2_t __rev1; \
61828  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61829  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61830  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 40); \
61831})
61832#endif
61833
61834#ifdef __LITTLE_ENDIAN__
61835#define vst1q_s32_x2(__p0, __p1) __extension__ ({ \
61836  int32x4x2_t __s1 = __p1; \
61837  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 34); \
61838})
61839#else
61840#define vst1q_s32_x2(__p0, __p1) __extension__ ({ \
61841  int32x4x2_t __s1 = __p1; \
61842  int32x4x2_t __rev1; \
61843  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61844  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61845  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 34); \
61846})
61847#endif
61848
61849#ifdef __LITTLE_ENDIAN__
61850#define vst1q_s64_x2(__p0, __p1) __extension__ ({ \
61851  int64x2x2_t __s1 = __p1; \
61852  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 35); \
61853})
61854#else
61855#define vst1q_s64_x2(__p0, __p1) __extension__ ({ \
61856  int64x2x2_t __s1 = __p1; \
61857  int64x2x2_t __rev1; \
61858  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61859  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61860  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 35); \
61861})
61862#endif
61863
61864#ifdef __LITTLE_ENDIAN__
61865#define vst1q_s16_x2(__p0, __p1) __extension__ ({ \
61866  int16x8x2_t __s1 = __p1; \
61867  __builtin_neon_vst1q_x2_v(__p0, __s1.val[0], __s1.val[1], 33); \
61868})
61869#else
61870#define vst1q_s16_x2(__p0, __p1) __extension__ ({ \
61871  int16x8x2_t __s1 = __p1; \
61872  int16x8x2_t __rev1; \
61873  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61874  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61875  __builtin_neon_vst1q_x2_v(__p0, __rev1.val[0], __rev1.val[1], 33); \
61876})
61877#endif
61878
61879#ifdef __LITTLE_ENDIAN__
61880#define vst1_u8_x2(__p0, __p1) __extension__ ({ \
61881  uint8x8x2_t __s1 = __p1; \
61882  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 16); \
61883})
61884#else
61885#define vst1_u8_x2(__p0, __p1) __extension__ ({ \
61886  uint8x8x2_t __s1 = __p1; \
61887  uint8x8x2_t __rev1; \
61888  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61889  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61890  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 16); \
61891})
61892#endif
61893
61894#ifdef __LITTLE_ENDIAN__
61895#define vst1_u32_x2(__p0, __p1) __extension__ ({ \
61896  uint32x2x2_t __s1 = __p1; \
61897  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 18); \
61898})
61899#else
61900#define vst1_u32_x2(__p0, __p1) __extension__ ({ \
61901  uint32x2x2_t __s1 = __p1; \
61902  uint32x2x2_t __rev1; \
61903  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61904  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61905  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 18); \
61906})
61907#endif
61908
61909#ifdef __LITTLE_ENDIAN__
61910#define vst1_u64_x2(__p0, __p1) __extension__ ({ \
61911  uint64x1x2_t __s1 = __p1; \
61912  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
61913})
61914#else
61915#define vst1_u64_x2(__p0, __p1) __extension__ ({ \
61916  uint64x1x2_t __s1 = __p1; \
61917  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 19); \
61918})
61919#endif
61920
61921#ifdef __LITTLE_ENDIAN__
61922#define vst1_u16_x2(__p0, __p1) __extension__ ({ \
61923  uint16x4x2_t __s1 = __p1; \
61924  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 17); \
61925})
61926#else
61927#define vst1_u16_x2(__p0, __p1) __extension__ ({ \
61928  uint16x4x2_t __s1 = __p1; \
61929  uint16x4x2_t __rev1; \
61930  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61931  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61932  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 17); \
61933})
61934#endif
61935
61936#ifdef __LITTLE_ENDIAN__
61937#define vst1_s8_x2(__p0, __p1) __extension__ ({ \
61938  int8x8x2_t __s1 = __p1; \
61939  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 0); \
61940})
61941#else
61942#define vst1_s8_x2(__p0, __p1) __extension__ ({ \
61943  int8x8x2_t __s1 = __p1; \
61944  int8x8x2_t __rev1; \
61945  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
61946  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
61947  __builtin_neon_vst1_x2_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], 0); \
61948})
61949#endif
61950
61951#ifdef __LITTLE_ENDIAN__
61952#define vst1_f64_x2(__p0, __p1) __extension__ ({ \
61953  float64x1x2_t __s1 = __p1; \
61954  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 10); \
61955})
61956#else
61957#define vst1_f64_x2(__p0, __p1) __extension__ ({ \
61958  float64x1x2_t __s1 = __p1; \
61959  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 10); \
61960})
61961#endif
61962
61963#ifdef __LITTLE_ENDIAN__
61964#define vst1_f32_x2(__p0, __p1) __extension__ ({ \
61965  float32x2x2_t __s1 = __p1; \
61966  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 9); \
61967})
61968#else
61969#define vst1_f32_x2(__p0, __p1) __extension__ ({ \
61970  float32x2x2_t __s1 = __p1; \
61971  float32x2x2_t __rev1; \
61972  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
61973  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
61974  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 9); \
61975})
61976#endif
61977
61978#ifdef __LITTLE_ENDIAN__
61979#define vst1_f16_x2(__p0, __p1) __extension__ ({ \
61980  float16x4x2_t __s1 = __p1; \
61981  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 8); \
61982})
61983#else
61984#define vst1_f16_x2(__p0, __p1) __extension__ ({ \
61985  float16x4x2_t __s1 = __p1; \
61986  float16x4x2_t __rev1; \
61987  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
61988  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
61989  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 8); \
61990})
61991#endif
61992
61993#ifdef __LITTLE_ENDIAN__
61994#define vst1_s32_x2(__p0, __p1) __extension__ ({ \
61995  int32x2x2_t __s1 = __p1; \
61996  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 2); \
61997})
61998#else
61999#define vst1_s32_x2(__p0, __p1) __extension__ ({ \
62000  int32x2x2_t __s1 = __p1; \
62001  int32x2x2_t __rev1; \
62002  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62003  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62004  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 2); \
62005})
62006#endif
62007
62008#ifdef __LITTLE_ENDIAN__
62009#define vst1_s64_x2(__p0, __p1) __extension__ ({ \
62010  int64x1x2_t __s1 = __p1; \
62011  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 3); \
62012})
62013#else
62014#define vst1_s64_x2(__p0, __p1) __extension__ ({ \
62015  int64x1x2_t __s1 = __p1; \
62016  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 3); \
62017})
62018#endif
62019
62020#ifdef __LITTLE_ENDIAN__
62021#define vst1_s16_x2(__p0, __p1) __extension__ ({ \
62022  int16x4x2_t __s1 = __p1; \
62023  __builtin_neon_vst1_x2_v(__p0, __s1.val[0], __s1.val[1], 1); \
62024})
62025#else
62026#define vst1_s16_x2(__p0, __p1) __extension__ ({ \
62027  int16x4x2_t __s1 = __p1; \
62028  int16x4x2_t __rev1; \
62029  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62030  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62031  __builtin_neon_vst1_x2_v(__p0, __rev1.val[0], __rev1.val[1], 1); \
62032})
62033#endif
62034
62035#ifdef __LITTLE_ENDIAN__
62036#define vst1_p8_x3(__p0, __p1) __extension__ ({ \
62037  poly8x8x3_t __s1 = __p1; \
62038  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 4); \
62039})
62040#else
62041#define vst1_p8_x3(__p0, __p1) __extension__ ({ \
62042  poly8x8x3_t __s1 = __p1; \
62043  poly8x8x3_t __rev1; \
62044  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62045  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62046  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62047  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 4); \
62048})
62049#endif
62050
62051#ifdef __LITTLE_ENDIAN__
62052#define vst1_p64_x3(__p0, __p1) __extension__ ({ \
62053  poly64x1x3_t __s1 = __p1; \
62054  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
62055})
62056#else
62057#define vst1_p64_x3(__p0, __p1) __extension__ ({ \
62058  poly64x1x3_t __s1 = __p1; \
62059  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
62060})
62061#endif
62062
62063#ifdef __LITTLE_ENDIAN__
62064#define vst1_p16_x3(__p0, __p1) __extension__ ({ \
62065  poly16x4x3_t __s1 = __p1; \
62066  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 5); \
62067})
62068#else
62069#define vst1_p16_x3(__p0, __p1) __extension__ ({ \
62070  poly16x4x3_t __s1 = __p1; \
62071  poly16x4x3_t __rev1; \
62072  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62073  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62074  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62075  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 5); \
62076})
62077#endif
62078
62079#ifdef __LITTLE_ENDIAN__
62080#define vst1q_p8_x3(__p0, __p1) __extension__ ({ \
62081  poly8x16x3_t __s1 = __p1; \
62082  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 36); \
62083})
62084#else
62085#define vst1q_p8_x3(__p0, __p1) __extension__ ({ \
62086  poly8x16x3_t __s1 = __p1; \
62087  poly8x16x3_t __rev1; \
62088  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62089  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62090  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62091  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 36); \
62092})
62093#endif
62094
62095#ifdef __LITTLE_ENDIAN__
62096#define vst1q_p64_x3(__p0, __p1) __extension__ ({ \
62097  poly64x2x3_t __s1 = __p1; \
62098  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 38); \
62099})
62100#else
62101#define vst1q_p64_x3(__p0, __p1) __extension__ ({ \
62102  poly64x2x3_t __s1 = __p1; \
62103  poly64x2x3_t __rev1; \
62104  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62105  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62106  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62107  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 38); \
62108})
62109#endif
62110
62111#ifdef __LITTLE_ENDIAN__
62112#define vst1q_p16_x3(__p0, __p1) __extension__ ({ \
62113  poly16x8x3_t __s1 = __p1; \
62114  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 37); \
62115})
62116#else
62117#define vst1q_p16_x3(__p0, __p1) __extension__ ({ \
62118  poly16x8x3_t __s1 = __p1; \
62119  poly16x8x3_t __rev1; \
62120  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62121  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62122  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62123  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 37); \
62124})
62125#endif
62126
62127#ifdef __LITTLE_ENDIAN__
62128#define vst1q_u8_x3(__p0, __p1) __extension__ ({ \
62129  uint8x16x3_t __s1 = __p1; \
62130  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 48); \
62131})
62132#else
62133#define vst1q_u8_x3(__p0, __p1) __extension__ ({ \
62134  uint8x16x3_t __s1 = __p1; \
62135  uint8x16x3_t __rev1; \
62136  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62137  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62138  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62139  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 48); \
62140})
62141#endif
62142
62143#ifdef __LITTLE_ENDIAN__
62144#define vst1q_u32_x3(__p0, __p1) __extension__ ({ \
62145  uint32x4x3_t __s1 = __p1; \
62146  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 50); \
62147})
62148#else
62149#define vst1q_u32_x3(__p0, __p1) __extension__ ({ \
62150  uint32x4x3_t __s1 = __p1; \
62151  uint32x4x3_t __rev1; \
62152  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62153  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62154  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62155  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 50); \
62156})
62157#endif
62158
62159#ifdef __LITTLE_ENDIAN__
62160#define vst1q_u64_x3(__p0, __p1) __extension__ ({ \
62161  uint64x2x3_t __s1 = __p1; \
62162  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 51); \
62163})
62164#else
62165#define vst1q_u64_x3(__p0, __p1) __extension__ ({ \
62166  uint64x2x3_t __s1 = __p1; \
62167  uint64x2x3_t __rev1; \
62168  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62169  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62170  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62171  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 51); \
62172})
62173#endif
62174
62175#ifdef __LITTLE_ENDIAN__
62176#define vst1q_u16_x3(__p0, __p1) __extension__ ({ \
62177  uint16x8x3_t __s1 = __p1; \
62178  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 49); \
62179})
62180#else
62181#define vst1q_u16_x3(__p0, __p1) __extension__ ({ \
62182  uint16x8x3_t __s1 = __p1; \
62183  uint16x8x3_t __rev1; \
62184  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62185  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62186  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62187  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 49); \
62188})
62189#endif
62190
62191#ifdef __LITTLE_ENDIAN__
62192#define vst1q_s8_x3(__p0, __p1) __extension__ ({ \
62193  int8x16x3_t __s1 = __p1; \
62194  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 32); \
62195})
62196#else
62197#define vst1q_s8_x3(__p0, __p1) __extension__ ({ \
62198  int8x16x3_t __s1 = __p1; \
62199  int8x16x3_t __rev1; \
62200  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62201  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62202  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62203  __builtin_neon_vst1q_x3_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 32); \
62204})
62205#endif
62206
62207#ifdef __LITTLE_ENDIAN__
62208#define vst1q_f64_x3(__p0, __p1) __extension__ ({ \
62209  float64x2x3_t __s1 = __p1; \
62210  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 42); \
62211})
62212#else
62213#define vst1q_f64_x3(__p0, __p1) __extension__ ({ \
62214  float64x2x3_t __s1 = __p1; \
62215  float64x2x3_t __rev1; \
62216  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62217  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62218  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62219  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 42); \
62220})
62221#endif
62222
62223#ifdef __LITTLE_ENDIAN__
62224#define vst1q_f32_x3(__p0, __p1) __extension__ ({ \
62225  float32x4x3_t __s1 = __p1; \
62226  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 41); \
62227})
62228#else
62229#define vst1q_f32_x3(__p0, __p1) __extension__ ({ \
62230  float32x4x3_t __s1 = __p1; \
62231  float32x4x3_t __rev1; \
62232  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62233  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62234  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62235  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 41); \
62236})
62237#endif
62238
62239#ifdef __LITTLE_ENDIAN__
62240#define vst1q_f16_x3(__p0, __p1) __extension__ ({ \
62241  float16x8x3_t __s1 = __p1; \
62242  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 40); \
62243})
62244#else
62245#define vst1q_f16_x3(__p0, __p1) __extension__ ({ \
62246  float16x8x3_t __s1 = __p1; \
62247  float16x8x3_t __rev1; \
62248  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62249  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62250  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62251  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 40); \
62252})
62253#endif
62254
62255#ifdef __LITTLE_ENDIAN__
62256#define vst1q_s32_x3(__p0, __p1) __extension__ ({ \
62257  int32x4x3_t __s1 = __p1; \
62258  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 34); \
62259})
62260#else
62261#define vst1q_s32_x3(__p0, __p1) __extension__ ({ \
62262  int32x4x3_t __s1 = __p1; \
62263  int32x4x3_t __rev1; \
62264  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62265  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62266  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62267  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 34); \
62268})
62269#endif
62270
62271#ifdef __LITTLE_ENDIAN__
62272#define vst1q_s64_x3(__p0, __p1) __extension__ ({ \
62273  int64x2x3_t __s1 = __p1; \
62274  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 35); \
62275})
62276#else
62277#define vst1q_s64_x3(__p0, __p1) __extension__ ({ \
62278  int64x2x3_t __s1 = __p1; \
62279  int64x2x3_t __rev1; \
62280  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62281  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62282  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62283  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 35); \
62284})
62285#endif
62286
62287#ifdef __LITTLE_ENDIAN__
62288#define vst1q_s16_x3(__p0, __p1) __extension__ ({ \
62289  int16x8x3_t __s1 = __p1; \
62290  __builtin_neon_vst1q_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 33); \
62291})
62292#else
62293#define vst1q_s16_x3(__p0, __p1) __extension__ ({ \
62294  int16x8x3_t __s1 = __p1; \
62295  int16x8x3_t __rev1; \
62296  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62297  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62298  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62299  __builtin_neon_vst1q_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 33); \
62300})
62301#endif
62302
62303#ifdef __LITTLE_ENDIAN__
62304#define vst1_u8_x3(__p0, __p1) __extension__ ({ \
62305  uint8x8x3_t __s1 = __p1; \
62306  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 16); \
62307})
62308#else
62309#define vst1_u8_x3(__p0, __p1) __extension__ ({ \
62310  uint8x8x3_t __s1 = __p1; \
62311  uint8x8x3_t __rev1; \
62312  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62313  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62314  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62315  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 16); \
62316})
62317#endif
62318
62319#ifdef __LITTLE_ENDIAN__
62320#define vst1_u32_x3(__p0, __p1) __extension__ ({ \
62321  uint32x2x3_t __s1 = __p1; \
62322  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 18); \
62323})
62324#else
62325#define vst1_u32_x3(__p0, __p1) __extension__ ({ \
62326  uint32x2x3_t __s1 = __p1; \
62327  uint32x2x3_t __rev1; \
62328  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62329  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62330  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62331  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 18); \
62332})
62333#endif
62334
62335#ifdef __LITTLE_ENDIAN__
62336#define vst1_u64_x3(__p0, __p1) __extension__ ({ \
62337  uint64x1x3_t __s1 = __p1; \
62338  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
62339})
62340#else
62341#define vst1_u64_x3(__p0, __p1) __extension__ ({ \
62342  uint64x1x3_t __s1 = __p1; \
62343  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 19); \
62344})
62345#endif
62346
62347#ifdef __LITTLE_ENDIAN__
62348#define vst1_u16_x3(__p0, __p1) __extension__ ({ \
62349  uint16x4x3_t __s1 = __p1; \
62350  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 17); \
62351})
62352#else
62353#define vst1_u16_x3(__p0, __p1) __extension__ ({ \
62354  uint16x4x3_t __s1 = __p1; \
62355  uint16x4x3_t __rev1; \
62356  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62357  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62358  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62359  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 17); \
62360})
62361#endif
62362
62363#ifdef __LITTLE_ENDIAN__
62364#define vst1_s8_x3(__p0, __p1) __extension__ ({ \
62365  int8x8x3_t __s1 = __p1; \
62366  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 0); \
62367})
62368#else
62369#define vst1_s8_x3(__p0, __p1) __extension__ ({ \
62370  int8x8x3_t __s1 = __p1; \
62371  int8x8x3_t __rev1; \
62372  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62373  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62374  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62375  __builtin_neon_vst1_x3_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], 0); \
62376})
62377#endif
62378
62379#ifdef __LITTLE_ENDIAN__
62380#define vst1_f64_x3(__p0, __p1) __extension__ ({ \
62381  float64x1x3_t __s1 = __p1; \
62382  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
62383})
62384#else
62385#define vst1_f64_x3(__p0, __p1) __extension__ ({ \
62386  float64x1x3_t __s1 = __p1; \
62387  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
62388})
62389#endif
62390
62391#ifdef __LITTLE_ENDIAN__
62392#define vst1_f32_x3(__p0, __p1) __extension__ ({ \
62393  float32x2x3_t __s1 = __p1; \
62394  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 9); \
62395})
62396#else
62397#define vst1_f32_x3(__p0, __p1) __extension__ ({ \
62398  float32x2x3_t __s1 = __p1; \
62399  float32x2x3_t __rev1; \
62400  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62401  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62402  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62403  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 9); \
62404})
62405#endif
62406
62407#ifdef __LITTLE_ENDIAN__
62408#define vst1_f16_x3(__p0, __p1) __extension__ ({ \
62409  float16x4x3_t __s1 = __p1; \
62410  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 8); \
62411})
62412#else
62413#define vst1_f16_x3(__p0, __p1) __extension__ ({ \
62414  float16x4x3_t __s1 = __p1; \
62415  float16x4x3_t __rev1; \
62416  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62417  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62418  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62419  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 8); \
62420})
62421#endif
62422
62423#ifdef __LITTLE_ENDIAN__
62424#define vst1_s32_x3(__p0, __p1) __extension__ ({ \
62425  int32x2x3_t __s1 = __p1; \
62426  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 2); \
62427})
62428#else
62429#define vst1_s32_x3(__p0, __p1) __extension__ ({ \
62430  int32x2x3_t __s1 = __p1; \
62431  int32x2x3_t __rev1; \
62432  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62433  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62434  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62435  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 2); \
62436})
62437#endif
62438
62439#ifdef __LITTLE_ENDIAN__
62440#define vst1_s64_x3(__p0, __p1) __extension__ ({ \
62441  int64x1x3_t __s1 = __p1; \
62442  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
62443})
62444#else
62445#define vst1_s64_x3(__p0, __p1) __extension__ ({ \
62446  int64x1x3_t __s1 = __p1; \
62447  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 3); \
62448})
62449#endif
62450
62451#ifdef __LITTLE_ENDIAN__
62452#define vst1_s16_x3(__p0, __p1) __extension__ ({ \
62453  int16x4x3_t __s1 = __p1; \
62454  __builtin_neon_vst1_x3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 1); \
62455})
62456#else
62457#define vst1_s16_x3(__p0, __p1) __extension__ ({ \
62458  int16x4x3_t __s1 = __p1; \
62459  int16x4x3_t __rev1; \
62460  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62461  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62462  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62463  __builtin_neon_vst1_x3_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 1); \
62464})
62465#endif
62466
62467#ifdef __LITTLE_ENDIAN__
62468#define vst1_p8_x4(__p0, __p1) __extension__ ({ \
62469  poly8x8x4_t __s1 = __p1; \
62470  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 4); \
62471})
62472#else
62473#define vst1_p8_x4(__p0, __p1) __extension__ ({ \
62474  poly8x8x4_t __s1 = __p1; \
62475  poly8x8x4_t __rev1; \
62476  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62477  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62478  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62479  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62480  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 4); \
62481})
62482#endif
62483
62484#ifdef __LITTLE_ENDIAN__
62485#define vst1_p64_x4(__p0, __p1) __extension__ ({ \
62486  poly64x1x4_t __s1 = __p1; \
62487  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
62488})
62489#else
62490#define vst1_p64_x4(__p0, __p1) __extension__ ({ \
62491  poly64x1x4_t __s1 = __p1; \
62492  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
62493})
62494#endif
62495
62496#ifdef __LITTLE_ENDIAN__
62497#define vst1_p16_x4(__p0, __p1) __extension__ ({ \
62498  poly16x4x4_t __s1 = __p1; \
62499  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 5); \
62500})
62501#else
62502#define vst1_p16_x4(__p0, __p1) __extension__ ({ \
62503  poly16x4x4_t __s1 = __p1; \
62504  poly16x4x4_t __rev1; \
62505  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62506  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62507  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62508  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62509  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 5); \
62510})
62511#endif
62512
62513#ifdef __LITTLE_ENDIAN__
62514#define vst1q_p8_x4(__p0, __p1) __extension__ ({ \
62515  poly8x16x4_t __s1 = __p1; \
62516  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 36); \
62517})
62518#else
62519#define vst1q_p8_x4(__p0, __p1) __extension__ ({ \
62520  poly8x16x4_t __s1 = __p1; \
62521  poly8x16x4_t __rev1; \
62522  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62523  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62524  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62525  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62526  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 36); \
62527})
62528#endif
62529
62530#ifdef __LITTLE_ENDIAN__
62531#define vst1q_p64_x4(__p0, __p1) __extension__ ({ \
62532  poly64x2x4_t __s1 = __p1; \
62533  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 38); \
62534})
62535#else
62536#define vst1q_p64_x4(__p0, __p1) __extension__ ({ \
62537  poly64x2x4_t __s1 = __p1; \
62538  poly64x2x4_t __rev1; \
62539  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62540  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62541  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62542  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62543  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 38); \
62544})
62545#endif
62546
62547#ifdef __LITTLE_ENDIAN__
62548#define vst1q_p16_x4(__p0, __p1) __extension__ ({ \
62549  poly16x8x4_t __s1 = __p1; \
62550  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 37); \
62551})
62552#else
62553#define vst1q_p16_x4(__p0, __p1) __extension__ ({ \
62554  poly16x8x4_t __s1 = __p1; \
62555  poly16x8x4_t __rev1; \
62556  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62557  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62558  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62559  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62560  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 37); \
62561})
62562#endif
62563
62564#ifdef __LITTLE_ENDIAN__
62565#define vst1q_u8_x4(__p0, __p1) __extension__ ({ \
62566  uint8x16x4_t __s1 = __p1; \
62567  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 48); \
62568})
62569#else
62570#define vst1q_u8_x4(__p0, __p1) __extension__ ({ \
62571  uint8x16x4_t __s1 = __p1; \
62572  uint8x16x4_t __rev1; \
62573  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62574  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62575  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62576  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62577  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 48); \
62578})
62579#endif
62580
62581#ifdef __LITTLE_ENDIAN__
62582#define vst1q_u32_x4(__p0, __p1) __extension__ ({ \
62583  uint32x4x4_t __s1 = __p1; \
62584  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 50); \
62585})
62586#else
62587#define vst1q_u32_x4(__p0, __p1) __extension__ ({ \
62588  uint32x4x4_t __s1 = __p1; \
62589  uint32x4x4_t __rev1; \
62590  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62591  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62592  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62593  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62594  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 50); \
62595})
62596#endif
62597
62598#ifdef __LITTLE_ENDIAN__
62599#define vst1q_u64_x4(__p0, __p1) __extension__ ({ \
62600  uint64x2x4_t __s1 = __p1; \
62601  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 51); \
62602})
62603#else
62604#define vst1q_u64_x4(__p0, __p1) __extension__ ({ \
62605  uint64x2x4_t __s1 = __p1; \
62606  uint64x2x4_t __rev1; \
62607  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62608  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62609  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62610  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62611  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 51); \
62612})
62613#endif
62614
62615#ifdef __LITTLE_ENDIAN__
62616#define vst1q_u16_x4(__p0, __p1) __extension__ ({ \
62617  uint16x8x4_t __s1 = __p1; \
62618  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 49); \
62619})
62620#else
62621#define vst1q_u16_x4(__p0, __p1) __extension__ ({ \
62622  uint16x8x4_t __s1 = __p1; \
62623  uint16x8x4_t __rev1; \
62624  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62625  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62626  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62627  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62628  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 49); \
62629})
62630#endif
62631
62632#ifdef __LITTLE_ENDIAN__
62633#define vst1q_s8_x4(__p0, __p1) __extension__ ({ \
62634  int8x16x4_t __s1 = __p1; \
62635  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 32); \
62636})
62637#else
62638#define vst1q_s8_x4(__p0, __p1) __extension__ ({ \
62639  int8x16x4_t __s1 = __p1; \
62640  int8x16x4_t __rev1; \
62641  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62642  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62643  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62644  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
62645  __builtin_neon_vst1q_x4_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 32); \
62646})
62647#endif
62648
62649#ifdef __LITTLE_ENDIAN__
62650#define vst1q_f64_x4(__p0, __p1) __extension__ ({ \
62651  float64x2x4_t __s1 = __p1; \
62652  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 42); \
62653})
62654#else
62655#define vst1q_f64_x4(__p0, __p1) __extension__ ({ \
62656  float64x2x4_t __s1 = __p1; \
62657  float64x2x4_t __rev1; \
62658  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62659  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62660  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62661  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62662  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 42); \
62663})
62664#endif
62665
62666#ifdef __LITTLE_ENDIAN__
62667#define vst1q_f32_x4(__p0, __p1) __extension__ ({ \
62668  float32x4x4_t __s1 = __p1; \
62669  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 41); \
62670})
62671#else
62672#define vst1q_f32_x4(__p0, __p1) __extension__ ({ \
62673  float32x4x4_t __s1 = __p1; \
62674  float32x4x4_t __rev1; \
62675  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62676  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62677  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62678  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62679  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 41); \
62680})
62681#endif
62682
62683#ifdef __LITTLE_ENDIAN__
62684#define vst1q_f16_x4(__p0, __p1) __extension__ ({ \
62685  float16x8x4_t __s1 = __p1; \
62686  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 40); \
62687})
62688#else
62689#define vst1q_f16_x4(__p0, __p1) __extension__ ({ \
62690  float16x8x4_t __s1 = __p1; \
62691  float16x8x4_t __rev1; \
62692  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62693  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62694  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62695  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62696  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 40); \
62697})
62698#endif
62699
62700#ifdef __LITTLE_ENDIAN__
62701#define vst1q_s32_x4(__p0, __p1) __extension__ ({ \
62702  int32x4x4_t __s1 = __p1; \
62703  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 34); \
62704})
62705#else
62706#define vst1q_s32_x4(__p0, __p1) __extension__ ({ \
62707  int32x4x4_t __s1 = __p1; \
62708  int32x4x4_t __rev1; \
62709  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62710  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62711  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62712  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62713  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 34); \
62714})
62715#endif
62716
62717#ifdef __LITTLE_ENDIAN__
62718#define vst1q_s64_x4(__p0, __p1) __extension__ ({ \
62719  int64x2x4_t __s1 = __p1; \
62720  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 35); \
62721})
62722#else
62723#define vst1q_s64_x4(__p0, __p1) __extension__ ({ \
62724  int64x2x4_t __s1 = __p1; \
62725  int64x2x4_t __rev1; \
62726  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62727  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62728  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62729  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62730  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 35); \
62731})
62732#endif
62733
62734#ifdef __LITTLE_ENDIAN__
62735#define vst1q_s16_x4(__p0, __p1) __extension__ ({ \
62736  int16x8x4_t __s1 = __p1; \
62737  __builtin_neon_vst1q_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 33); \
62738})
62739#else
62740#define vst1q_s16_x4(__p0, __p1) __extension__ ({ \
62741  int16x8x4_t __s1 = __p1; \
62742  int16x8x4_t __rev1; \
62743  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62744  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62745  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62746  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62747  __builtin_neon_vst1q_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 33); \
62748})
62749#endif
62750
62751#ifdef __LITTLE_ENDIAN__
62752#define vst1_u8_x4(__p0, __p1) __extension__ ({ \
62753  uint8x8x4_t __s1 = __p1; \
62754  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 16); \
62755})
62756#else
62757#define vst1_u8_x4(__p0, __p1) __extension__ ({ \
62758  uint8x8x4_t __s1 = __p1; \
62759  uint8x8x4_t __rev1; \
62760  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62761  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62762  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62763  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62764  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 16); \
62765})
62766#endif
62767
62768#ifdef __LITTLE_ENDIAN__
62769#define vst1_u32_x4(__p0, __p1) __extension__ ({ \
62770  uint32x2x4_t __s1 = __p1; \
62771  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 18); \
62772})
62773#else
62774#define vst1_u32_x4(__p0, __p1) __extension__ ({ \
62775  uint32x2x4_t __s1 = __p1; \
62776  uint32x2x4_t __rev1; \
62777  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62778  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62779  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62780  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62781  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 18); \
62782})
62783#endif
62784
62785#ifdef __LITTLE_ENDIAN__
62786#define vst1_u64_x4(__p0, __p1) __extension__ ({ \
62787  uint64x1x4_t __s1 = __p1; \
62788  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
62789})
62790#else
62791#define vst1_u64_x4(__p0, __p1) __extension__ ({ \
62792  uint64x1x4_t __s1 = __p1; \
62793  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 19); \
62794})
62795#endif
62796
62797#ifdef __LITTLE_ENDIAN__
62798#define vst1_u16_x4(__p0, __p1) __extension__ ({ \
62799  uint16x4x4_t __s1 = __p1; \
62800  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 17); \
62801})
62802#else
62803#define vst1_u16_x4(__p0, __p1) __extension__ ({ \
62804  uint16x4x4_t __s1 = __p1; \
62805  uint16x4x4_t __rev1; \
62806  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62807  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62808  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62809  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62810  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 17); \
62811})
62812#endif
62813
62814#ifdef __LITTLE_ENDIAN__
62815#define vst1_s8_x4(__p0, __p1) __extension__ ({ \
62816  int8x8x4_t __s1 = __p1; \
62817  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 0); \
62818})
62819#else
62820#define vst1_s8_x4(__p0, __p1) __extension__ ({ \
62821  int8x8x4_t __s1 = __p1; \
62822  int8x8x4_t __rev1; \
62823  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 7, 6, 5, 4, 3, 2, 1, 0); \
62824  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 7, 6, 5, 4, 3, 2, 1, 0); \
62825  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 7, 6, 5, 4, 3, 2, 1, 0); \
62826  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 7, 6, 5, 4, 3, 2, 1, 0); \
62827  __builtin_neon_vst1_x4_v(__p0, (int8x8_t)__rev1.val[0], (int8x8_t)__rev1.val[1], (int8x8_t)__rev1.val[2], (int8x8_t)__rev1.val[3], 0); \
62828})
62829#endif
62830
62831#ifdef __LITTLE_ENDIAN__
62832#define vst1_f64_x4(__p0, __p1) __extension__ ({ \
62833  float64x1x4_t __s1 = __p1; \
62834  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
62835})
62836#else
62837#define vst1_f64_x4(__p0, __p1) __extension__ ({ \
62838  float64x1x4_t __s1 = __p1; \
62839  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
62840})
62841#endif
62842
62843#ifdef __LITTLE_ENDIAN__
62844#define vst1_f32_x4(__p0, __p1) __extension__ ({ \
62845  float32x2x4_t __s1 = __p1; \
62846  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 9); \
62847})
62848#else
62849#define vst1_f32_x4(__p0, __p1) __extension__ ({ \
62850  float32x2x4_t __s1 = __p1; \
62851  float32x2x4_t __rev1; \
62852  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62853  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62854  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62855  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62856  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 9); \
62857})
62858#endif
62859
62860#ifdef __LITTLE_ENDIAN__
62861#define vst1_f16_x4(__p0, __p1) __extension__ ({ \
62862  float16x4x4_t __s1 = __p1; \
62863  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 8); \
62864})
62865#else
62866#define vst1_f16_x4(__p0, __p1) __extension__ ({ \
62867  float16x4x4_t __s1 = __p1; \
62868  float16x4x4_t __rev1; \
62869  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62870  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62871  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62872  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62873  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 8); \
62874})
62875#endif
62876
62877#ifdef __LITTLE_ENDIAN__
62878#define vst1_s32_x4(__p0, __p1) __extension__ ({ \
62879  int32x2x4_t __s1 = __p1; \
62880  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 2); \
62881})
62882#else
62883#define vst1_s32_x4(__p0, __p1) __extension__ ({ \
62884  int32x2x4_t __s1 = __p1; \
62885  int32x2x4_t __rev1; \
62886  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62887  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62888  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
62889  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
62890  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 2); \
62891})
62892#endif
62893
62894#ifdef __LITTLE_ENDIAN__
62895#define vst1_s64_x4(__p0, __p1) __extension__ ({ \
62896  int64x1x4_t __s1 = __p1; \
62897  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
62898})
62899#else
62900#define vst1_s64_x4(__p0, __p1) __extension__ ({ \
62901  int64x1x4_t __s1 = __p1; \
62902  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 3); \
62903})
62904#endif
62905
62906#ifdef __LITTLE_ENDIAN__
62907#define vst1_s16_x4(__p0, __p1) __extension__ ({ \
62908  int16x4x4_t __s1 = __p1; \
62909  __builtin_neon_vst1_x4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 1); \
62910})
62911#else
62912#define vst1_s16_x4(__p0, __p1) __extension__ ({ \
62913  int16x4x4_t __s1 = __p1; \
62914  int16x4x4_t __rev1; \
62915  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 3, 2, 1, 0); \
62916  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 3, 2, 1, 0); \
62917  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 3, 2, 1, 0); \
62918  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 3, 2, 1, 0); \
62919  __builtin_neon_vst1_x4_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 1); \
62920})
62921#endif
62922
62923#ifdef __LITTLE_ENDIAN__
62924#define vst2_p64(__p0, __p1) __extension__ ({ \
62925  poly64x1x2_t __s1 = __p1; \
62926  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
62927})
62928#else
62929#define vst2_p64(__p0, __p1) __extension__ ({ \
62930  poly64x1x2_t __s1 = __p1; \
62931  __builtin_neon_vst2_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], 6); \
62932})
62933#endif
62934
62935#ifdef __LITTLE_ENDIAN__
62936#define vst2q_p64(__p0, __p1) __extension__ ({ \
62937  poly64x2x2_t __s1 = __p1; \
62938  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 38); \
62939})
62940#else
62941#define vst2q_p64(__p0, __p1) __extension__ ({ \
62942  poly64x2x2_t __s1 = __p1; \
62943  poly64x2x2_t __rev1; \
62944  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62945  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62946  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 38); \
62947})
62948#endif
62949
62950#ifdef __LITTLE_ENDIAN__
62951#define vst2q_u64(__p0, __p1) __extension__ ({ \
62952  uint64x2x2_t __s1 = __p1; \
62953  __builtin_neon_vst2q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], 51); \
62954})
62955#else
62956#define vst2q_u64(__p0, __p1) __extension__ ({ \
62957  uint64x2x2_t __s1 = __p1; \
62958  uint64x2x2_t __rev1; \
62959  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62960  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62961  __builtin_neon_vst2q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], 51); \
62962})
62963#endif
62964
62965#ifdef __LITTLE_ENDIAN__
62966#define vst2q_f64(__p0, __p1) __extension__ ({ \
62967  float64x2x2_t __s1 = __p1; \
62968  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 42); \
62969})
62970#else
62971#define vst2q_f64(__p0, __p1) __extension__ ({ \
62972  float64x2x2_t __s1 = __p1; \
62973  float64x2x2_t __rev1; \
62974  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62975  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62976  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 42); \
62977})
62978#endif
62979
62980#ifdef __LITTLE_ENDIAN__
62981#define vst2q_s64(__p0, __p1) __extension__ ({ \
62982  int64x2x2_t __s1 = __p1; \
62983  __builtin_neon_vst2q_v(__p0, __s1.val[0], __s1.val[1], 35); \
62984})
62985#else
62986#define vst2q_s64(__p0, __p1) __extension__ ({ \
62987  int64x2x2_t __s1 = __p1; \
62988  int64x2x2_t __rev1; \
62989  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
62990  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
62991  __builtin_neon_vst2q_v(__p0, __rev1.val[0], __rev1.val[1], 35); \
62992})
62993#endif
62994
62995#ifdef __LITTLE_ENDIAN__
62996#define vst2_f64(__p0, __p1) __extension__ ({ \
62997  float64x1x2_t __s1 = __p1; \
62998  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 10); \
62999})
63000#else
63001#define vst2_f64(__p0, __p1) __extension__ ({ \
63002  float64x1x2_t __s1 = __p1; \
63003  __builtin_neon_vst2_v(__p0, __s1.val[0], __s1.val[1], 10); \
63004})
63005#endif
63006
63007#ifdef __LITTLE_ENDIAN__
63008#define vst2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63009  poly64x1x2_t __s1 = __p1; \
63010  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
63011})
63012#else
63013#define vst2_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63014  poly64x1x2_t __s1 = __p1; \
63015  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 6); \
63016})
63017#endif
63018
63019#ifdef __LITTLE_ENDIAN__
63020#define vst2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63021  poly8x16x2_t __s1 = __p1; \
63022  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 36); \
63023})
63024#else
63025#define vst2q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63026  poly8x16x2_t __s1 = __p1; \
63027  poly8x16x2_t __rev1; \
63028  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63029  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63030  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 36); \
63031})
63032#endif
63033
63034#ifdef __LITTLE_ENDIAN__
63035#define vst2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63036  poly64x2x2_t __s1 = __p1; \
63037  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 38); \
63038})
63039#else
63040#define vst2q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63041  poly64x2x2_t __s1 = __p1; \
63042  poly64x2x2_t __rev1; \
63043  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63044  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63045  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 38); \
63046})
63047#endif
63048
63049#ifdef __LITTLE_ENDIAN__
63050#define vst2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63051  uint8x16x2_t __s1 = __p1; \
63052  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 48); \
63053})
63054#else
63055#define vst2q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63056  uint8x16x2_t __s1 = __p1; \
63057  uint8x16x2_t __rev1; \
63058  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63059  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63060  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 48); \
63061})
63062#endif
63063
63064#ifdef __LITTLE_ENDIAN__
63065#define vst2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63066  uint64x2x2_t __s1 = __p1; \
63067  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 51); \
63068})
63069#else
63070#define vst2q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63071  uint64x2x2_t __s1 = __p1; \
63072  uint64x2x2_t __rev1; \
63073  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63074  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63075  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 51); \
63076})
63077#endif
63078
63079#ifdef __LITTLE_ENDIAN__
63080#define vst2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63081  int8x16x2_t __s1 = __p1; \
63082  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], __p2, 32); \
63083})
63084#else
63085#define vst2q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63086  int8x16x2_t __s1 = __p1; \
63087  int8x16x2_t __rev1; \
63088  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63089  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63090  __builtin_neon_vst2q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], __p2, 32); \
63091})
63092#endif
63093
63094#ifdef __LITTLE_ENDIAN__
63095#define vst2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63096  float64x2x2_t __s1 = __p1; \
63097  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 42); \
63098})
63099#else
63100#define vst2q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63101  float64x2x2_t __s1 = __p1; \
63102  float64x2x2_t __rev1; \
63103  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63104  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63105  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 42); \
63106})
63107#endif
63108
63109#ifdef __LITTLE_ENDIAN__
63110#define vst2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63111  int64x2x2_t __s1 = __p1; \
63112  __builtin_neon_vst2q_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 35); \
63113})
63114#else
63115#define vst2q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63116  int64x2x2_t __s1 = __p1; \
63117  int64x2x2_t __rev1; \
63118  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63119  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63120  __builtin_neon_vst2q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __p2, 35); \
63121})
63122#endif
63123
63124#ifdef __LITTLE_ENDIAN__
63125#define vst2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63126  uint64x1x2_t __s1 = __p1; \
63127  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
63128})
63129#else
63130#define vst2_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63131  uint64x1x2_t __s1 = __p1; \
63132  __builtin_neon_vst2_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], __p2, 19); \
63133})
63134#endif
63135
63136#ifdef __LITTLE_ENDIAN__
63137#define vst2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63138  float64x1x2_t __s1 = __p1; \
63139  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 10); \
63140})
63141#else
63142#define vst2_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63143  float64x1x2_t __s1 = __p1; \
63144  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 10); \
63145})
63146#endif
63147
63148#ifdef __LITTLE_ENDIAN__
63149#define vst2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63150  int64x1x2_t __s1 = __p1; \
63151  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 3); \
63152})
63153#else
63154#define vst2_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63155  int64x1x2_t __s1 = __p1; \
63156  __builtin_neon_vst2_lane_v(__p0, __s1.val[0], __s1.val[1], __p2, 3); \
63157})
63158#endif
63159
63160#ifdef __LITTLE_ENDIAN__
63161#define vst3_p64(__p0, __p1) __extension__ ({ \
63162  poly64x1x3_t __s1 = __p1; \
63163  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
63164})
63165#else
63166#define vst3_p64(__p0, __p1) __extension__ ({ \
63167  poly64x1x3_t __s1 = __p1; \
63168  __builtin_neon_vst3_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], 6); \
63169})
63170#endif
63171
63172#ifdef __LITTLE_ENDIAN__
63173#define vst3q_p64(__p0, __p1) __extension__ ({ \
63174  poly64x2x3_t __s1 = __p1; \
63175  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 38); \
63176})
63177#else
63178#define vst3q_p64(__p0, __p1) __extension__ ({ \
63179  poly64x2x3_t __s1 = __p1; \
63180  poly64x2x3_t __rev1; \
63181  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63182  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63183  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63184  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 38); \
63185})
63186#endif
63187
63188#ifdef __LITTLE_ENDIAN__
63189#define vst3q_u64(__p0, __p1) __extension__ ({ \
63190  uint64x2x3_t __s1 = __p1; \
63191  __builtin_neon_vst3q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], 51); \
63192})
63193#else
63194#define vst3q_u64(__p0, __p1) __extension__ ({ \
63195  uint64x2x3_t __s1 = __p1; \
63196  uint64x2x3_t __rev1; \
63197  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63198  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63199  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63200  __builtin_neon_vst3q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], 51); \
63201})
63202#endif
63203
63204#ifdef __LITTLE_ENDIAN__
63205#define vst3q_f64(__p0, __p1) __extension__ ({ \
63206  float64x2x3_t __s1 = __p1; \
63207  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 42); \
63208})
63209#else
63210#define vst3q_f64(__p0, __p1) __extension__ ({ \
63211  float64x2x3_t __s1 = __p1; \
63212  float64x2x3_t __rev1; \
63213  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63214  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63215  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63216  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 42); \
63217})
63218#endif
63219
63220#ifdef __LITTLE_ENDIAN__
63221#define vst3q_s64(__p0, __p1) __extension__ ({ \
63222  int64x2x3_t __s1 = __p1; \
63223  __builtin_neon_vst3q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 35); \
63224})
63225#else
63226#define vst3q_s64(__p0, __p1) __extension__ ({ \
63227  int64x2x3_t __s1 = __p1; \
63228  int64x2x3_t __rev1; \
63229  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63230  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63231  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63232  __builtin_neon_vst3q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], 35); \
63233})
63234#endif
63235
63236#ifdef __LITTLE_ENDIAN__
63237#define vst3_f64(__p0, __p1) __extension__ ({ \
63238  float64x1x3_t __s1 = __p1; \
63239  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
63240})
63241#else
63242#define vst3_f64(__p0, __p1) __extension__ ({ \
63243  float64x1x3_t __s1 = __p1; \
63244  __builtin_neon_vst3_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], 10); \
63245})
63246#endif
63247
63248#ifdef __LITTLE_ENDIAN__
63249#define vst3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63250  poly64x1x3_t __s1 = __p1; \
63251  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
63252})
63253#else
63254#define vst3_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63255  poly64x1x3_t __s1 = __p1; \
63256  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 6); \
63257})
63258#endif
63259
63260#ifdef __LITTLE_ENDIAN__
63261#define vst3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63262  poly8x16x3_t __s1 = __p1; \
63263  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 36); \
63264})
63265#else
63266#define vst3q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63267  poly8x16x3_t __s1 = __p1; \
63268  poly8x16x3_t __rev1; \
63269  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63270  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63271  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63272  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 36); \
63273})
63274#endif
63275
63276#ifdef __LITTLE_ENDIAN__
63277#define vst3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63278  poly64x2x3_t __s1 = __p1; \
63279  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 38); \
63280})
63281#else
63282#define vst3q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63283  poly64x2x3_t __s1 = __p1; \
63284  poly64x2x3_t __rev1; \
63285  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63286  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63287  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63288  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 38); \
63289})
63290#endif
63291
63292#ifdef __LITTLE_ENDIAN__
63293#define vst3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63294  uint8x16x3_t __s1 = __p1; \
63295  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 48); \
63296})
63297#else
63298#define vst3q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63299  uint8x16x3_t __s1 = __p1; \
63300  uint8x16x3_t __rev1; \
63301  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63302  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63303  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63304  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 48); \
63305})
63306#endif
63307
63308#ifdef __LITTLE_ENDIAN__
63309#define vst3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63310  uint64x2x3_t __s1 = __p1; \
63311  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 51); \
63312})
63313#else
63314#define vst3q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63315  uint64x2x3_t __s1 = __p1; \
63316  uint64x2x3_t __rev1; \
63317  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63318  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63319  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63320  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 51); \
63321})
63322#endif
63323
63324#ifdef __LITTLE_ENDIAN__
63325#define vst3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63326  int8x16x3_t __s1 = __p1; \
63327  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], __p2, 32); \
63328})
63329#else
63330#define vst3q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63331  int8x16x3_t __s1 = __p1; \
63332  int8x16x3_t __rev1; \
63333  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63334  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63335  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63336  __builtin_neon_vst3q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], __p2, 32); \
63337})
63338#endif
63339
63340#ifdef __LITTLE_ENDIAN__
63341#define vst3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63342  float64x2x3_t __s1 = __p1; \
63343  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 42); \
63344})
63345#else
63346#define vst3q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63347  float64x2x3_t __s1 = __p1; \
63348  float64x2x3_t __rev1; \
63349  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63350  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63351  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63352  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 42); \
63353})
63354#endif
63355
63356#ifdef __LITTLE_ENDIAN__
63357#define vst3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63358  int64x2x3_t __s1 = __p1; \
63359  __builtin_neon_vst3q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 35); \
63360})
63361#else
63362#define vst3q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63363  int64x2x3_t __s1 = __p1; \
63364  int64x2x3_t __rev1; \
63365  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63366  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63367  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63368  __builtin_neon_vst3q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __p2, 35); \
63369})
63370#endif
63371
63372#ifdef __LITTLE_ENDIAN__
63373#define vst3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63374  uint64x1x3_t __s1 = __p1; \
63375  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
63376})
63377#else
63378#define vst3_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63379  uint64x1x3_t __s1 = __p1; \
63380  __builtin_neon_vst3_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], __p2, 19); \
63381})
63382#endif
63383
63384#ifdef __LITTLE_ENDIAN__
63385#define vst3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63386  float64x1x3_t __s1 = __p1; \
63387  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
63388})
63389#else
63390#define vst3_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63391  float64x1x3_t __s1 = __p1; \
63392  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 10); \
63393})
63394#endif
63395
63396#ifdef __LITTLE_ENDIAN__
63397#define vst3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63398  int64x1x3_t __s1 = __p1; \
63399  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
63400})
63401#else
63402#define vst3_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63403  int64x1x3_t __s1 = __p1; \
63404  __builtin_neon_vst3_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __p2, 3); \
63405})
63406#endif
63407
63408#ifdef __LITTLE_ENDIAN__
63409#define vst4_p64(__p0, __p1) __extension__ ({ \
63410  poly64x1x4_t __s1 = __p1; \
63411  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
63412})
63413#else
63414#define vst4_p64(__p0, __p1) __extension__ ({ \
63415  poly64x1x4_t __s1 = __p1; \
63416  __builtin_neon_vst4_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], 6); \
63417})
63418#endif
63419
63420#ifdef __LITTLE_ENDIAN__
63421#define vst4q_p64(__p0, __p1) __extension__ ({ \
63422  poly64x2x4_t __s1 = __p1; \
63423  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 38); \
63424})
63425#else
63426#define vst4q_p64(__p0, __p1) __extension__ ({ \
63427  poly64x2x4_t __s1 = __p1; \
63428  poly64x2x4_t __rev1; \
63429  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63430  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63431  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63432  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63433  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 38); \
63434})
63435#endif
63436
63437#ifdef __LITTLE_ENDIAN__
63438#define vst4q_u64(__p0, __p1) __extension__ ({ \
63439  uint64x2x4_t __s1 = __p1; \
63440  __builtin_neon_vst4q_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], 51); \
63441})
63442#else
63443#define vst4q_u64(__p0, __p1) __extension__ ({ \
63444  uint64x2x4_t __s1 = __p1; \
63445  uint64x2x4_t __rev1; \
63446  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63447  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63448  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63449  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63450  __builtin_neon_vst4q_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], 51); \
63451})
63452#endif
63453
63454#ifdef __LITTLE_ENDIAN__
63455#define vst4q_f64(__p0, __p1) __extension__ ({ \
63456  float64x2x4_t __s1 = __p1; \
63457  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 42); \
63458})
63459#else
63460#define vst4q_f64(__p0, __p1) __extension__ ({ \
63461  float64x2x4_t __s1 = __p1; \
63462  float64x2x4_t __rev1; \
63463  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63464  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63465  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63466  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63467  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 42); \
63468})
63469#endif
63470
63471#ifdef __LITTLE_ENDIAN__
63472#define vst4q_s64(__p0, __p1) __extension__ ({ \
63473  int64x2x4_t __s1 = __p1; \
63474  __builtin_neon_vst4q_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 35); \
63475})
63476#else
63477#define vst4q_s64(__p0, __p1) __extension__ ({ \
63478  int64x2x4_t __s1 = __p1; \
63479  int64x2x4_t __rev1; \
63480  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63481  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63482  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63483  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63484  __builtin_neon_vst4q_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], 35); \
63485})
63486#endif
63487
63488#ifdef __LITTLE_ENDIAN__
63489#define vst4_f64(__p0, __p1) __extension__ ({ \
63490  float64x1x4_t __s1 = __p1; \
63491  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
63492})
63493#else
63494#define vst4_f64(__p0, __p1) __extension__ ({ \
63495  float64x1x4_t __s1 = __p1; \
63496  __builtin_neon_vst4_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], 10); \
63497})
63498#endif
63499
63500#ifdef __LITTLE_ENDIAN__
63501#define vst4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63502  poly64x1x4_t __s1 = __p1; \
63503  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
63504})
63505#else
63506#define vst4_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63507  poly64x1x4_t __s1 = __p1; \
63508  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 6); \
63509})
63510#endif
63511
63512#ifdef __LITTLE_ENDIAN__
63513#define vst4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63514  poly8x16x4_t __s1 = __p1; \
63515  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 36); \
63516})
63517#else
63518#define vst4q_lane_p8(__p0, __p1, __p2) __extension__ ({ \
63519  poly8x16x4_t __s1 = __p1; \
63520  poly8x16x4_t __rev1; \
63521  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63522  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63523  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63524  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63525  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 36); \
63526})
63527#endif
63528
63529#ifdef __LITTLE_ENDIAN__
63530#define vst4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63531  poly64x2x4_t __s1 = __p1; \
63532  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 38); \
63533})
63534#else
63535#define vst4q_lane_p64(__p0, __p1, __p2) __extension__ ({ \
63536  poly64x2x4_t __s1 = __p1; \
63537  poly64x2x4_t __rev1; \
63538  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63539  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63540  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63541  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63542  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 38); \
63543})
63544#endif
63545
63546#ifdef __LITTLE_ENDIAN__
63547#define vst4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63548  uint8x16x4_t __s1 = __p1; \
63549  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 48); \
63550})
63551#else
63552#define vst4q_lane_u8(__p0, __p1, __p2) __extension__ ({ \
63553  uint8x16x4_t __s1 = __p1; \
63554  uint8x16x4_t __rev1; \
63555  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63556  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63557  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63558  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63559  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 48); \
63560})
63561#endif
63562
63563#ifdef __LITTLE_ENDIAN__
63564#define vst4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63565  uint64x2x4_t __s1 = __p1; \
63566  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 51); \
63567})
63568#else
63569#define vst4q_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63570  uint64x2x4_t __s1 = __p1; \
63571  uint64x2x4_t __rev1; \
63572  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63573  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63574  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63575  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63576  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 51); \
63577})
63578#endif
63579
63580#ifdef __LITTLE_ENDIAN__
63581#define vst4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63582  int8x16x4_t __s1 = __p1; \
63583  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__s1.val[0], (int8x16_t)__s1.val[1], (int8x16_t)__s1.val[2], (int8x16_t)__s1.val[3], __p2, 32); \
63584})
63585#else
63586#define vst4q_lane_s8(__p0, __p1, __p2) __extension__ ({ \
63587  int8x16x4_t __s1 = __p1; \
63588  int8x16x4_t __rev1; \
63589  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63590  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63591  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63592  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); \
63593  __builtin_neon_vst4q_lane_v(__p0, (int8x16_t)__rev1.val[0], (int8x16_t)__rev1.val[1], (int8x16_t)__rev1.val[2], (int8x16_t)__rev1.val[3], __p2, 32); \
63594})
63595#endif
63596
63597#ifdef __LITTLE_ENDIAN__
63598#define vst4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63599  float64x2x4_t __s1 = __p1; \
63600  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 42); \
63601})
63602#else
63603#define vst4q_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63604  float64x2x4_t __s1 = __p1; \
63605  float64x2x4_t __rev1; \
63606  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63607  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63608  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63609  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63610  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 42); \
63611})
63612#endif
63613
63614#ifdef __LITTLE_ENDIAN__
63615#define vst4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63616  int64x2x4_t __s1 = __p1; \
63617  __builtin_neon_vst4q_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 35); \
63618})
63619#else
63620#define vst4q_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63621  int64x2x4_t __s1 = __p1; \
63622  int64x2x4_t __rev1; \
63623  __rev1.val[0] = __builtin_shufflevector(__s1.val[0], __s1.val[0], 1, 0); \
63624  __rev1.val[1] = __builtin_shufflevector(__s1.val[1], __s1.val[1], 1, 0); \
63625  __rev1.val[2] = __builtin_shufflevector(__s1.val[2], __s1.val[2], 1, 0); \
63626  __rev1.val[3] = __builtin_shufflevector(__s1.val[3], __s1.val[3], 1, 0); \
63627  __builtin_neon_vst4q_lane_v(__p0, __rev1.val[0], __rev1.val[1], __rev1.val[2], __rev1.val[3], __p2, 35); \
63628})
63629#endif
63630
63631#ifdef __LITTLE_ENDIAN__
63632#define vst4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63633  uint64x1x4_t __s1 = __p1; \
63634  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
63635})
63636#else
63637#define vst4_lane_u64(__p0, __p1, __p2) __extension__ ({ \
63638  uint64x1x4_t __s1 = __p1; \
63639  __builtin_neon_vst4_lane_v(__p0, (int8x8_t)__s1.val[0], (int8x8_t)__s1.val[1], (int8x8_t)__s1.val[2], (int8x8_t)__s1.val[3], __p2, 19); \
63640})
63641#endif
63642
63643#ifdef __LITTLE_ENDIAN__
63644#define vst4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63645  float64x1x4_t __s1 = __p1; \
63646  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
63647})
63648#else
63649#define vst4_lane_f64(__p0, __p1, __p2) __extension__ ({ \
63650  float64x1x4_t __s1 = __p1; \
63651  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 10); \
63652})
63653#endif
63654
63655#ifdef __LITTLE_ENDIAN__
63656#define vst4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63657  int64x1x4_t __s1 = __p1; \
63658  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
63659})
63660#else
63661#define vst4_lane_s64(__p0, __p1, __p2) __extension__ ({ \
63662  int64x1x4_t __s1 = __p1; \
63663  __builtin_neon_vst4_lane_v(__p0, __s1.val[0], __s1.val[1], __s1.val[2], __s1.val[3], __p2, 3); \
63664})
63665#endif
63666
63667#ifdef __LITTLE_ENDIAN__
63668#define vstrq_p128(__p0, __p1) __extension__ ({ \
63669  poly128_t __s1 = __p1; \
63670  __builtin_neon_vstrq_p128(__p0, __s1); \
63671})
63672#else
63673#define vstrq_p128(__p0, __p1) __extension__ ({ \
63674  poly128_t __s1 = __p1; \
63675  __builtin_neon_vstrq_p128(__p0, __s1); \
63676})
63677#endif
63678
63679#ifdef __LITTLE_ENDIAN__
63680__ai uint64_t vsubd_u64(uint64_t __p0, uint64_t __p1) {
63681  uint64_t __ret;
63682  __ret = (uint64_t) __builtin_neon_vsubd_u64(__p0, __p1);
63683  return __ret;
63684}
63685#else
63686__ai uint64_t vsubd_u64(uint64_t __p0, uint64_t __p1) {
63687  uint64_t __ret;
63688  __ret = (uint64_t) __builtin_neon_vsubd_u64(__p0, __p1);
63689  return __ret;
63690}
63691#endif
63692
63693#ifdef __LITTLE_ENDIAN__
63694__ai int64_t vsubd_s64(int64_t __p0, int64_t __p1) {
63695  int64_t __ret;
63696  __ret = (int64_t) __builtin_neon_vsubd_s64(__p0, __p1);
63697  return __ret;
63698}
63699#else
63700__ai int64_t vsubd_s64(int64_t __p0, int64_t __p1) {
63701  int64_t __ret;
63702  __ret = (int64_t) __builtin_neon_vsubd_s64(__p0, __p1);
63703  return __ret;
63704}
63705#endif
63706
63707#ifdef __LITTLE_ENDIAN__
63708__ai float64x2_t vsubq_f64(float64x2_t __p0, float64x2_t __p1) {
63709  float64x2_t __ret;
63710  __ret = __p0 - __p1;
63711  return __ret;
63712}
63713#else
63714__ai float64x2_t vsubq_f64(float64x2_t __p0, float64x2_t __p1) {
63715  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63716  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63717  float64x2_t __ret;
63718  __ret = __rev0 - __rev1;
63719  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63720  return __ret;
63721}
63722#endif
63723
63724#ifdef __LITTLE_ENDIAN__
63725__ai float64x1_t vsub_f64(float64x1_t __p0, float64x1_t __p1) {
63726  float64x1_t __ret;
63727  __ret = __p0 - __p1;
63728  return __ret;
63729}
63730#else
63731__ai float64x1_t vsub_f64(float64x1_t __p0, float64x1_t __p1) {
63732  float64x1_t __ret;
63733  __ret = __p0 - __p1;
63734  return __ret;
63735}
63736#endif
63737
63738#ifdef __LITTLE_ENDIAN__
63739__ai uint16x8_t vsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
63740  uint16x8_t __ret;
63741  __ret = vcombine_u16(__p0, vsubhn_u32(__p1, __p2));
63742  return __ret;
63743}
63744#else
63745__ai uint16x8_t vsubhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
63746  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63747  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63748  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
63749  uint16x8_t __ret;
63750  __ret = __noswap_vcombine_u16(__rev0, __noswap_vsubhn_u32(__rev1, __rev2));
63751  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63752  return __ret;
63753}
63754#endif
63755
63756#ifdef __LITTLE_ENDIAN__
63757__ai uint32x4_t vsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
63758  uint32x4_t __ret;
63759  __ret = vcombine_u32(__p0, vsubhn_u64(__p1, __p2));
63760  return __ret;
63761}
63762#else
63763__ai uint32x4_t vsubhn_high_u64(uint32x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) {
63764  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63765  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63766  uint64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
63767  uint32x4_t __ret;
63768  __ret = __noswap_vcombine_u32(__rev0, __noswap_vsubhn_u64(__rev1, __rev2));
63769  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63770  return __ret;
63771}
63772#endif
63773
63774#ifdef __LITTLE_ENDIAN__
63775__ai uint8x16_t vsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
63776  uint8x16_t __ret;
63777  __ret = vcombine_u8(__p0, vsubhn_u16(__p1, __p2));
63778  return __ret;
63779}
63780#else
63781__ai uint8x16_t vsubhn_high_u16(uint8x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
63782  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63783  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63784  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
63785  uint8x16_t __ret;
63786  __ret = __noswap_vcombine_u8(__rev0, __noswap_vsubhn_u16(__rev1, __rev2));
63787  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63788  return __ret;
63789}
63790#endif
63791
63792#ifdef __LITTLE_ENDIAN__
63793__ai int16x8_t vsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
63794  int16x8_t __ret;
63795  __ret = vcombine_s16(__p0, vsubhn_s32(__p1, __p2));
63796  return __ret;
63797}
63798#else
63799__ai int16x8_t vsubhn_high_s32(int16x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
63800  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63801  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63802  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
63803  int16x8_t __ret;
63804  __ret = __noswap_vcombine_s16(__rev0, __noswap_vsubhn_s32(__rev1, __rev2));
63805  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63806  return __ret;
63807}
63808#endif
63809
63810#ifdef __LITTLE_ENDIAN__
63811__ai int32x4_t vsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
63812  int32x4_t __ret;
63813  __ret = vcombine_s32(__p0, vsubhn_s64(__p1, __p2));
63814  return __ret;
63815}
63816#else
63817__ai int32x4_t vsubhn_high_s64(int32x2_t __p0, int64x2_t __p1, int64x2_t __p2) {
63818  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63819  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
63820  int64x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
63821  int32x4_t __ret;
63822  __ret = __noswap_vcombine_s32(__rev0, __noswap_vsubhn_s64(__rev1, __rev2));
63823  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63824  return __ret;
63825}
63826#endif
63827
63828#ifdef __LITTLE_ENDIAN__
63829__ai int8x16_t vsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
63830  int8x16_t __ret;
63831  __ret = vcombine_s8(__p0, vsubhn_s16(__p1, __p2));
63832  return __ret;
63833}
63834#else
63835__ai int8x16_t vsubhn_high_s16(int8x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
63836  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63837  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63838  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
63839  int8x16_t __ret;
63840  __ret = __noswap_vcombine_s8(__rev0, __noswap_vsubhn_s16(__rev1, __rev2));
63841  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63842  return __ret;
63843}
63844#endif
63845
63846#ifdef __LITTLE_ENDIAN__
63847__ai uint16x8_t vsubl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
63848  uint16x8_t __ret;
63849  __ret = vmovl_high_u8(__p0) - vmovl_high_u8(__p1);
63850  return __ret;
63851}
63852#else
63853__ai uint16x8_t vsubl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
63854  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63855  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63856  uint16x8_t __ret;
63857  __ret = __noswap_vmovl_high_u8(__rev0) - __noswap_vmovl_high_u8(__rev1);
63858  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63859  return __ret;
63860}
63861#endif
63862
63863#ifdef __LITTLE_ENDIAN__
63864__ai uint64x2_t vsubl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
63865  uint64x2_t __ret;
63866  __ret = vmovl_high_u32(__p0) - vmovl_high_u32(__p1);
63867  return __ret;
63868}
63869#else
63870__ai uint64x2_t vsubl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
63871  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63872  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63873  uint64x2_t __ret;
63874  __ret = __noswap_vmovl_high_u32(__rev0) - __noswap_vmovl_high_u32(__rev1);
63875  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63876  return __ret;
63877}
63878#endif
63879
63880#ifdef __LITTLE_ENDIAN__
63881__ai uint32x4_t vsubl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
63882  uint32x4_t __ret;
63883  __ret = vmovl_high_u16(__p0) - vmovl_high_u16(__p1);
63884  return __ret;
63885}
63886#else
63887__ai uint32x4_t vsubl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
63888  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63889  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63890  uint32x4_t __ret;
63891  __ret = __noswap_vmovl_high_u16(__rev0) - __noswap_vmovl_high_u16(__rev1);
63892  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63893  return __ret;
63894}
63895#endif
63896
63897#ifdef __LITTLE_ENDIAN__
63898__ai int16x8_t vsubl_high_s8(int8x16_t __p0, int8x16_t __p1) {
63899  int16x8_t __ret;
63900  __ret = vmovl_high_s8(__p0) - vmovl_high_s8(__p1);
63901  return __ret;
63902}
63903#else
63904__ai int16x8_t vsubl_high_s8(int8x16_t __p0, int8x16_t __p1) {
63905  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63906  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63907  int16x8_t __ret;
63908  __ret = __noswap_vmovl_high_s8(__rev0) - __noswap_vmovl_high_s8(__rev1);
63909  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63910  return __ret;
63911}
63912#endif
63913
63914#ifdef __LITTLE_ENDIAN__
63915__ai int64x2_t vsubl_high_s32(int32x4_t __p0, int32x4_t __p1) {
63916  int64x2_t __ret;
63917  __ret = vmovl_high_s32(__p0) - vmovl_high_s32(__p1);
63918  return __ret;
63919}
63920#else
63921__ai int64x2_t vsubl_high_s32(int32x4_t __p0, int32x4_t __p1) {
63922  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63923  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63924  int64x2_t __ret;
63925  __ret = __noswap_vmovl_high_s32(__rev0) - __noswap_vmovl_high_s32(__rev1);
63926  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63927  return __ret;
63928}
63929#endif
63930
63931#ifdef __LITTLE_ENDIAN__
63932__ai int32x4_t vsubl_high_s16(int16x8_t __p0, int16x8_t __p1) {
63933  int32x4_t __ret;
63934  __ret = vmovl_high_s16(__p0) - vmovl_high_s16(__p1);
63935  return __ret;
63936}
63937#else
63938__ai int32x4_t vsubl_high_s16(int16x8_t __p0, int16x8_t __p1) {
63939  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63940  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63941  int32x4_t __ret;
63942  __ret = __noswap_vmovl_high_s16(__rev0) - __noswap_vmovl_high_s16(__rev1);
63943  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63944  return __ret;
63945}
63946#endif
63947
63948#ifdef __LITTLE_ENDIAN__
63949__ai uint16x8_t vsubw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
63950  uint16x8_t __ret;
63951  __ret = __p0 - vmovl_high_u8(__p1);
63952  return __ret;
63953}
63954#else
63955__ai uint16x8_t vsubw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
63956  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
63957  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
63958  uint16x8_t __ret;
63959  __ret = __rev0 - __noswap_vmovl_high_u8(__rev1);
63960  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
63961  return __ret;
63962}
63963#endif
63964
63965#ifdef __LITTLE_ENDIAN__
63966__ai uint64x2_t vsubw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
63967  uint64x2_t __ret;
63968  __ret = __p0 - vmovl_high_u32(__p1);
63969  return __ret;
63970}
63971#else
63972__ai uint64x2_t vsubw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
63973  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
63974  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
63975  uint64x2_t __ret;
63976  __ret = __rev0 - __noswap_vmovl_high_u32(__rev1);
63977  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
63978  return __ret;
63979}
63980#endif
63981
63982#ifdef __LITTLE_ENDIAN__
63983__ai uint32x4_t vsubw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
63984  uint32x4_t __ret;
63985  __ret = __p0 - vmovl_high_u16(__p1);
63986  return __ret;
63987}
63988#else
63989__ai uint32x4_t vsubw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
63990  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
63991  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
63992  uint32x4_t __ret;
63993  __ret = __rev0 - __noswap_vmovl_high_u16(__rev1);
63994  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
63995  return __ret;
63996}
63997#endif
63998
63999#ifdef __LITTLE_ENDIAN__
64000__ai int16x8_t vsubw_high_s8(int16x8_t __p0, int8x16_t __p1) {
64001  int16x8_t __ret;
64002  __ret = __p0 - vmovl_high_s8(__p1);
64003  return __ret;
64004}
64005#else
64006__ai int16x8_t vsubw_high_s8(int16x8_t __p0, int8x16_t __p1) {
64007  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64008  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64009  int16x8_t __ret;
64010  __ret = __rev0 - __noswap_vmovl_high_s8(__rev1);
64011  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64012  return __ret;
64013}
64014#endif
64015
64016#ifdef __LITTLE_ENDIAN__
64017__ai int64x2_t vsubw_high_s32(int64x2_t __p0, int32x4_t __p1) {
64018  int64x2_t __ret;
64019  __ret = __p0 - vmovl_high_s32(__p1);
64020  return __ret;
64021}
64022#else
64023__ai int64x2_t vsubw_high_s32(int64x2_t __p0, int32x4_t __p1) {
64024  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64025  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64026  int64x2_t __ret;
64027  __ret = __rev0 - __noswap_vmovl_high_s32(__rev1);
64028  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64029  return __ret;
64030}
64031#endif
64032
64033#ifdef __LITTLE_ENDIAN__
64034__ai int32x4_t vsubw_high_s16(int32x4_t __p0, int16x8_t __p1) {
64035  int32x4_t __ret;
64036  __ret = __p0 - vmovl_high_s16(__p1);
64037  return __ret;
64038}
64039#else
64040__ai int32x4_t vsubw_high_s16(int32x4_t __p0, int16x8_t __p1) {
64041  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64042  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64043  int32x4_t __ret;
64044  __ret = __rev0 - __noswap_vmovl_high_s16(__rev1);
64045  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64046  return __ret;
64047}
64048#endif
64049
64050#ifdef __LITTLE_ENDIAN__
64051__ai poly8x8_t vtrn1_p8(poly8x8_t __p0, poly8x8_t __p1) {
64052  poly8x8_t __ret;
64053  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64054  return __ret;
64055}
64056#else
64057__ai poly8x8_t vtrn1_p8(poly8x8_t __p0, poly8x8_t __p1) {
64058  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64059  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64060  poly8x8_t __ret;
64061  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64062  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64063  return __ret;
64064}
64065#endif
64066
64067#ifdef __LITTLE_ENDIAN__
64068__ai poly16x4_t vtrn1_p16(poly16x4_t __p0, poly16x4_t __p1) {
64069  poly16x4_t __ret;
64070  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64071  return __ret;
64072}
64073#else
64074__ai poly16x4_t vtrn1_p16(poly16x4_t __p0, poly16x4_t __p1) {
64075  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64076  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64077  poly16x4_t __ret;
64078  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64079  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64080  return __ret;
64081}
64082#endif
64083
64084#ifdef __LITTLE_ENDIAN__
64085__ai poly8x16_t vtrn1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64086  poly8x16_t __ret;
64087  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64088  return __ret;
64089}
64090#else
64091__ai poly8x16_t vtrn1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64092  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64093  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64094  poly8x16_t __ret;
64095  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64096  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64097  return __ret;
64098}
64099#endif
64100
64101#ifdef __LITTLE_ENDIAN__
64102__ai poly64x2_t vtrn1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
64103  poly64x2_t __ret;
64104  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64105  return __ret;
64106}
64107#else
64108__ai poly64x2_t vtrn1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
64109  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64110  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64111  poly64x2_t __ret;
64112  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64113  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64114  return __ret;
64115}
64116#endif
64117
64118#ifdef __LITTLE_ENDIAN__
64119__ai poly16x8_t vtrn1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
64120  poly16x8_t __ret;
64121  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64122  return __ret;
64123}
64124#else
64125__ai poly16x8_t vtrn1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
64126  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64127  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64128  poly16x8_t __ret;
64129  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64130  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64131  return __ret;
64132}
64133#endif
64134
64135#ifdef __LITTLE_ENDIAN__
64136__ai uint8x16_t vtrn1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
64137  uint8x16_t __ret;
64138  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64139  return __ret;
64140}
64141#else
64142__ai uint8x16_t vtrn1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
64143  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64144  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64145  uint8x16_t __ret;
64146  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64147  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64148  return __ret;
64149}
64150#endif
64151
64152#ifdef __LITTLE_ENDIAN__
64153__ai uint32x4_t vtrn1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
64154  uint32x4_t __ret;
64155  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64156  return __ret;
64157}
64158#else
64159__ai uint32x4_t vtrn1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
64160  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64161  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64162  uint32x4_t __ret;
64163  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64164  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64165  return __ret;
64166}
64167#endif
64168
64169#ifdef __LITTLE_ENDIAN__
64170__ai uint64x2_t vtrn1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64171  uint64x2_t __ret;
64172  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64173  return __ret;
64174}
64175#else
64176__ai uint64x2_t vtrn1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64177  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64178  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64179  uint64x2_t __ret;
64180  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64181  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64182  return __ret;
64183}
64184#endif
64185
64186#ifdef __LITTLE_ENDIAN__
64187__ai uint16x8_t vtrn1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64188  uint16x8_t __ret;
64189  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64190  return __ret;
64191}
64192#else
64193__ai uint16x8_t vtrn1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64194  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64195  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64196  uint16x8_t __ret;
64197  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64198  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64199  return __ret;
64200}
64201#endif
64202
64203#ifdef __LITTLE_ENDIAN__
64204__ai int8x16_t vtrn1q_s8(int8x16_t __p0, int8x16_t __p1) {
64205  int8x16_t __ret;
64206  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64207  return __ret;
64208}
64209#else
64210__ai int8x16_t vtrn1q_s8(int8x16_t __p0, int8x16_t __p1) {
64211  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64212  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64213  int8x16_t __ret;
64214  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 2, 18, 4, 20, 6, 22, 8, 24, 10, 26, 12, 28, 14, 30);
64215  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64216  return __ret;
64217}
64218#endif
64219
64220#ifdef __LITTLE_ENDIAN__
64221__ai float64x2_t vtrn1q_f64(float64x2_t __p0, float64x2_t __p1) {
64222  float64x2_t __ret;
64223  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64224  return __ret;
64225}
64226#else
64227__ai float64x2_t vtrn1q_f64(float64x2_t __p0, float64x2_t __p1) {
64228  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64229  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64230  float64x2_t __ret;
64231  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64232  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64233  return __ret;
64234}
64235#endif
64236
64237#ifdef __LITTLE_ENDIAN__
64238__ai float32x4_t vtrn1q_f32(float32x4_t __p0, float32x4_t __p1) {
64239  float32x4_t __ret;
64240  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64241  return __ret;
64242}
64243#else
64244__ai float32x4_t vtrn1q_f32(float32x4_t __p0, float32x4_t __p1) {
64245  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64246  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64247  float32x4_t __ret;
64248  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64249  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64250  return __ret;
64251}
64252#endif
64253
64254#ifdef __LITTLE_ENDIAN__
64255__ai int32x4_t vtrn1q_s32(int32x4_t __p0, int32x4_t __p1) {
64256  int32x4_t __ret;
64257  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64258  return __ret;
64259}
64260#else
64261__ai int32x4_t vtrn1q_s32(int32x4_t __p0, int32x4_t __p1) {
64262  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64263  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64264  int32x4_t __ret;
64265  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64266  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64267  return __ret;
64268}
64269#endif
64270
64271#ifdef __LITTLE_ENDIAN__
64272__ai int64x2_t vtrn1q_s64(int64x2_t __p0, int64x2_t __p1) {
64273  int64x2_t __ret;
64274  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64275  return __ret;
64276}
64277#else
64278__ai int64x2_t vtrn1q_s64(int64x2_t __p0, int64x2_t __p1) {
64279  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64280  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64281  int64x2_t __ret;
64282  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64283  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64284  return __ret;
64285}
64286#endif
64287
64288#ifdef __LITTLE_ENDIAN__
64289__ai int16x8_t vtrn1q_s16(int16x8_t __p0, int16x8_t __p1) {
64290  int16x8_t __ret;
64291  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64292  return __ret;
64293}
64294#else
64295__ai int16x8_t vtrn1q_s16(int16x8_t __p0, int16x8_t __p1) {
64296  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64297  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64298  int16x8_t __ret;
64299  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64300  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64301  return __ret;
64302}
64303#endif
64304
64305#ifdef __LITTLE_ENDIAN__
64306__ai uint8x8_t vtrn1_u8(uint8x8_t __p0, uint8x8_t __p1) {
64307  uint8x8_t __ret;
64308  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64309  return __ret;
64310}
64311#else
64312__ai uint8x8_t vtrn1_u8(uint8x8_t __p0, uint8x8_t __p1) {
64313  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64314  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64315  uint8x8_t __ret;
64316  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64317  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64318  return __ret;
64319}
64320#endif
64321
64322#ifdef __LITTLE_ENDIAN__
64323__ai uint32x2_t vtrn1_u32(uint32x2_t __p0, uint32x2_t __p1) {
64324  uint32x2_t __ret;
64325  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64326  return __ret;
64327}
64328#else
64329__ai uint32x2_t vtrn1_u32(uint32x2_t __p0, uint32x2_t __p1) {
64330  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64331  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64332  uint32x2_t __ret;
64333  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64334  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64335  return __ret;
64336}
64337#endif
64338
64339#ifdef __LITTLE_ENDIAN__
64340__ai uint16x4_t vtrn1_u16(uint16x4_t __p0, uint16x4_t __p1) {
64341  uint16x4_t __ret;
64342  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64343  return __ret;
64344}
64345#else
64346__ai uint16x4_t vtrn1_u16(uint16x4_t __p0, uint16x4_t __p1) {
64347  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64348  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64349  uint16x4_t __ret;
64350  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64351  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64352  return __ret;
64353}
64354#endif
64355
64356#ifdef __LITTLE_ENDIAN__
64357__ai int8x8_t vtrn1_s8(int8x8_t __p0, int8x8_t __p1) {
64358  int8x8_t __ret;
64359  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 2, 10, 4, 12, 6, 14);
64360  return __ret;
64361}
64362#else
64363__ai int8x8_t vtrn1_s8(int8x8_t __p0, int8x8_t __p1) {
64364  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64365  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64366  int8x8_t __ret;
64367  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 2, 10, 4, 12, 6, 14);
64368  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64369  return __ret;
64370}
64371#endif
64372
64373#ifdef __LITTLE_ENDIAN__
64374__ai float32x2_t vtrn1_f32(float32x2_t __p0, float32x2_t __p1) {
64375  float32x2_t __ret;
64376  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64377  return __ret;
64378}
64379#else
64380__ai float32x2_t vtrn1_f32(float32x2_t __p0, float32x2_t __p1) {
64381  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64382  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64383  float32x2_t __ret;
64384  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64385  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64386  return __ret;
64387}
64388#endif
64389
64390#ifdef __LITTLE_ENDIAN__
64391__ai int32x2_t vtrn1_s32(int32x2_t __p0, int32x2_t __p1) {
64392  int32x2_t __ret;
64393  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
64394  return __ret;
64395}
64396#else
64397__ai int32x2_t vtrn1_s32(int32x2_t __p0, int32x2_t __p1) {
64398  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64399  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64400  int32x2_t __ret;
64401  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
64402  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64403  return __ret;
64404}
64405#endif
64406
64407#ifdef __LITTLE_ENDIAN__
64408__ai int16x4_t vtrn1_s16(int16x4_t __p0, int16x4_t __p1) {
64409  int16x4_t __ret;
64410  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 2, 6);
64411  return __ret;
64412}
64413#else
64414__ai int16x4_t vtrn1_s16(int16x4_t __p0, int16x4_t __p1) {
64415  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64416  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64417  int16x4_t __ret;
64418  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 2, 6);
64419  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64420  return __ret;
64421}
64422#endif
64423
64424#ifdef __LITTLE_ENDIAN__
64425__ai poly8x8_t vtrn2_p8(poly8x8_t __p0, poly8x8_t __p1) {
64426  poly8x8_t __ret;
64427  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64428  return __ret;
64429}
64430#else
64431__ai poly8x8_t vtrn2_p8(poly8x8_t __p0, poly8x8_t __p1) {
64432  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64433  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64434  poly8x8_t __ret;
64435  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64436  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64437  return __ret;
64438}
64439#endif
64440
64441#ifdef __LITTLE_ENDIAN__
64442__ai poly16x4_t vtrn2_p16(poly16x4_t __p0, poly16x4_t __p1) {
64443  poly16x4_t __ret;
64444  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64445  return __ret;
64446}
64447#else
64448__ai poly16x4_t vtrn2_p16(poly16x4_t __p0, poly16x4_t __p1) {
64449  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64450  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64451  poly16x4_t __ret;
64452  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64453  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64454  return __ret;
64455}
64456#endif
64457
64458#ifdef __LITTLE_ENDIAN__
64459__ai poly8x16_t vtrn2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64460  poly8x16_t __ret;
64461  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64462  return __ret;
64463}
64464#else
64465__ai poly8x16_t vtrn2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
64466  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64467  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64468  poly8x16_t __ret;
64469  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64470  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64471  return __ret;
64472}
64473#endif
64474
64475#ifdef __LITTLE_ENDIAN__
64476__ai poly64x2_t vtrn2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
64477  poly64x2_t __ret;
64478  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64479  return __ret;
64480}
64481#else
64482__ai poly64x2_t vtrn2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
64483  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64484  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64485  poly64x2_t __ret;
64486  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64487  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64488  return __ret;
64489}
64490#endif
64491
64492#ifdef __LITTLE_ENDIAN__
64493__ai poly16x8_t vtrn2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
64494  poly16x8_t __ret;
64495  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64496  return __ret;
64497}
64498#else
64499__ai poly16x8_t vtrn2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
64500  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64501  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64502  poly16x8_t __ret;
64503  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64504  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64505  return __ret;
64506}
64507#endif
64508
64509#ifdef __LITTLE_ENDIAN__
64510__ai uint8x16_t vtrn2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
64511  uint8x16_t __ret;
64512  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64513  return __ret;
64514}
64515#else
64516__ai uint8x16_t vtrn2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
64517  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64518  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64519  uint8x16_t __ret;
64520  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64521  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64522  return __ret;
64523}
64524#endif
64525
64526#ifdef __LITTLE_ENDIAN__
64527__ai uint32x4_t vtrn2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
64528  uint32x4_t __ret;
64529  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64530  return __ret;
64531}
64532#else
64533__ai uint32x4_t vtrn2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
64534  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64535  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64536  uint32x4_t __ret;
64537  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64538  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64539  return __ret;
64540}
64541#endif
64542
64543#ifdef __LITTLE_ENDIAN__
64544__ai uint64x2_t vtrn2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64545  uint64x2_t __ret;
64546  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64547  return __ret;
64548}
64549#else
64550__ai uint64x2_t vtrn2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
64551  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64552  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64553  uint64x2_t __ret;
64554  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64555  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64556  return __ret;
64557}
64558#endif
64559
64560#ifdef __LITTLE_ENDIAN__
64561__ai uint16x8_t vtrn2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64562  uint16x8_t __ret;
64563  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64564  return __ret;
64565}
64566#else
64567__ai uint16x8_t vtrn2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
64568  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64569  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64570  uint16x8_t __ret;
64571  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64572  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64573  return __ret;
64574}
64575#endif
64576
64577#ifdef __LITTLE_ENDIAN__
64578__ai int8x16_t vtrn2q_s8(int8x16_t __p0, int8x16_t __p1) {
64579  int8x16_t __ret;
64580  __ret = __builtin_shufflevector(__p0, __p1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64581  return __ret;
64582}
64583#else
64584__ai int8x16_t vtrn2q_s8(int8x16_t __p0, int8x16_t __p1) {
64585  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64586  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64587  int8x16_t __ret;
64588  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 17, 3, 19, 5, 21, 7, 23, 9, 25, 11, 27, 13, 29, 15, 31);
64589  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64590  return __ret;
64591}
64592#endif
64593
64594#ifdef __LITTLE_ENDIAN__
64595__ai float64x2_t vtrn2q_f64(float64x2_t __p0, float64x2_t __p1) {
64596  float64x2_t __ret;
64597  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64598  return __ret;
64599}
64600#else
64601__ai float64x2_t vtrn2q_f64(float64x2_t __p0, float64x2_t __p1) {
64602  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64603  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64604  float64x2_t __ret;
64605  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64606  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64607  return __ret;
64608}
64609#endif
64610
64611#ifdef __LITTLE_ENDIAN__
64612__ai float32x4_t vtrn2q_f32(float32x4_t __p0, float32x4_t __p1) {
64613  float32x4_t __ret;
64614  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64615  return __ret;
64616}
64617#else
64618__ai float32x4_t vtrn2q_f32(float32x4_t __p0, float32x4_t __p1) {
64619  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64620  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64621  float32x4_t __ret;
64622  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64623  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64624  return __ret;
64625}
64626#endif
64627
64628#ifdef __LITTLE_ENDIAN__
64629__ai int32x4_t vtrn2q_s32(int32x4_t __p0, int32x4_t __p1) {
64630  int32x4_t __ret;
64631  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64632  return __ret;
64633}
64634#else
64635__ai int32x4_t vtrn2q_s32(int32x4_t __p0, int32x4_t __p1) {
64636  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64637  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64638  int32x4_t __ret;
64639  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64640  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64641  return __ret;
64642}
64643#endif
64644
64645#ifdef __LITTLE_ENDIAN__
64646__ai int64x2_t vtrn2q_s64(int64x2_t __p0, int64x2_t __p1) {
64647  int64x2_t __ret;
64648  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64649  return __ret;
64650}
64651#else
64652__ai int64x2_t vtrn2q_s64(int64x2_t __p0, int64x2_t __p1) {
64653  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64654  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64655  int64x2_t __ret;
64656  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64657  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64658  return __ret;
64659}
64660#endif
64661
64662#ifdef __LITTLE_ENDIAN__
64663__ai int16x8_t vtrn2q_s16(int16x8_t __p0, int16x8_t __p1) {
64664  int16x8_t __ret;
64665  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64666  return __ret;
64667}
64668#else
64669__ai int16x8_t vtrn2q_s16(int16x8_t __p0, int16x8_t __p1) {
64670  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64671  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64672  int16x8_t __ret;
64673  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64674  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64675  return __ret;
64676}
64677#endif
64678
64679#ifdef __LITTLE_ENDIAN__
64680__ai uint8x8_t vtrn2_u8(uint8x8_t __p0, uint8x8_t __p1) {
64681  uint8x8_t __ret;
64682  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64683  return __ret;
64684}
64685#else
64686__ai uint8x8_t vtrn2_u8(uint8x8_t __p0, uint8x8_t __p1) {
64687  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64688  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64689  uint8x8_t __ret;
64690  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64691  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64692  return __ret;
64693}
64694#endif
64695
64696#ifdef __LITTLE_ENDIAN__
64697__ai uint32x2_t vtrn2_u32(uint32x2_t __p0, uint32x2_t __p1) {
64698  uint32x2_t __ret;
64699  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64700  return __ret;
64701}
64702#else
64703__ai uint32x2_t vtrn2_u32(uint32x2_t __p0, uint32x2_t __p1) {
64704  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64705  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64706  uint32x2_t __ret;
64707  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64708  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64709  return __ret;
64710}
64711#endif
64712
64713#ifdef __LITTLE_ENDIAN__
64714__ai uint16x4_t vtrn2_u16(uint16x4_t __p0, uint16x4_t __p1) {
64715  uint16x4_t __ret;
64716  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64717  return __ret;
64718}
64719#else
64720__ai uint16x4_t vtrn2_u16(uint16x4_t __p0, uint16x4_t __p1) {
64721  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64722  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64723  uint16x4_t __ret;
64724  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64725  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64726  return __ret;
64727}
64728#endif
64729
64730#ifdef __LITTLE_ENDIAN__
64731__ai int8x8_t vtrn2_s8(int8x8_t __p0, int8x8_t __p1) {
64732  int8x8_t __ret;
64733  __ret = __builtin_shufflevector(__p0, __p1, 1, 9, 3, 11, 5, 13, 7, 15);
64734  return __ret;
64735}
64736#else
64737__ai int8x8_t vtrn2_s8(int8x8_t __p0, int8x8_t __p1) {
64738  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
64739  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
64740  int8x8_t __ret;
64741  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 9, 3, 11, 5, 13, 7, 15);
64742  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
64743  return __ret;
64744}
64745#endif
64746
64747#ifdef __LITTLE_ENDIAN__
64748__ai float32x2_t vtrn2_f32(float32x2_t __p0, float32x2_t __p1) {
64749  float32x2_t __ret;
64750  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64751  return __ret;
64752}
64753#else
64754__ai float32x2_t vtrn2_f32(float32x2_t __p0, float32x2_t __p1) {
64755  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64756  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64757  float32x2_t __ret;
64758  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64759  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64760  return __ret;
64761}
64762#endif
64763
64764#ifdef __LITTLE_ENDIAN__
64765__ai int32x2_t vtrn2_s32(int32x2_t __p0, int32x2_t __p1) {
64766  int32x2_t __ret;
64767  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
64768  return __ret;
64769}
64770#else
64771__ai int32x2_t vtrn2_s32(int32x2_t __p0, int32x2_t __p1) {
64772  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64773  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64774  int32x2_t __ret;
64775  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
64776  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64777  return __ret;
64778}
64779#endif
64780
64781#ifdef __LITTLE_ENDIAN__
64782__ai int16x4_t vtrn2_s16(int16x4_t __p0, int16x4_t __p1) {
64783  int16x4_t __ret;
64784  __ret = __builtin_shufflevector(__p0, __p1, 1, 5, 3, 7);
64785  return __ret;
64786}
64787#else
64788__ai int16x4_t vtrn2_s16(int16x4_t __p0, int16x4_t __p1) {
64789  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
64790  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
64791  int16x4_t __ret;
64792  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 5, 3, 7);
64793  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
64794  return __ret;
64795}
64796#endif
64797
64798#ifdef __LITTLE_ENDIAN__
64799__ai uint64x1_t vtst_p64(poly64x1_t __p0, poly64x1_t __p1) {
64800  uint64x1_t __ret;
64801  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64802  return __ret;
64803}
64804#else
64805__ai uint64x1_t vtst_p64(poly64x1_t __p0, poly64x1_t __p1) {
64806  uint64x1_t __ret;
64807  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64808  return __ret;
64809}
64810#endif
64811
64812#ifdef __LITTLE_ENDIAN__
64813__ai uint64x2_t vtstq_p64(poly64x2_t __p0, poly64x2_t __p1) {
64814  uint64x2_t __ret;
64815  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
64816  return __ret;
64817}
64818#else
64819__ai uint64x2_t vtstq_p64(poly64x2_t __p0, poly64x2_t __p1) {
64820  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64821  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64822  uint64x2_t __ret;
64823  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
64824  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64825  return __ret;
64826}
64827#endif
64828
64829#ifdef __LITTLE_ENDIAN__
64830__ai uint64x2_t vtstq_u64(uint64x2_t __p0, uint64x2_t __p1) {
64831  uint64x2_t __ret;
64832  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
64833  return __ret;
64834}
64835#else
64836__ai uint64x2_t vtstq_u64(uint64x2_t __p0, uint64x2_t __p1) {
64837  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64838  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64839  uint64x2_t __ret;
64840  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
64841  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64842  return __ret;
64843}
64844#endif
64845
64846#ifdef __LITTLE_ENDIAN__
64847__ai uint64x2_t vtstq_s64(int64x2_t __p0, int64x2_t __p1) {
64848  uint64x2_t __ret;
64849  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__p0, (int8x16_t)__p1, 51);
64850  return __ret;
64851}
64852#else
64853__ai uint64x2_t vtstq_s64(int64x2_t __p0, int64x2_t __p1) {
64854  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
64855  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
64856  uint64x2_t __ret;
64857  __ret = (uint64x2_t) __builtin_neon_vtstq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51);
64858  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
64859  return __ret;
64860}
64861#endif
64862
64863#ifdef __LITTLE_ENDIAN__
64864__ai uint64x1_t vtst_u64(uint64x1_t __p0, uint64x1_t __p1) {
64865  uint64x1_t __ret;
64866  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64867  return __ret;
64868}
64869#else
64870__ai uint64x1_t vtst_u64(uint64x1_t __p0, uint64x1_t __p1) {
64871  uint64x1_t __ret;
64872  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64873  return __ret;
64874}
64875#endif
64876
64877#ifdef __LITTLE_ENDIAN__
64878__ai uint64x1_t vtst_s64(int64x1_t __p0, int64x1_t __p1) {
64879  uint64x1_t __ret;
64880  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64881  return __ret;
64882}
64883#else
64884__ai uint64x1_t vtst_s64(int64x1_t __p0, int64x1_t __p1) {
64885  uint64x1_t __ret;
64886  __ret = (uint64x1_t) __builtin_neon_vtst_v((int8x8_t)__p0, (int8x8_t)__p1, 19);
64887  return __ret;
64888}
64889#endif
64890
64891#ifdef __LITTLE_ENDIAN__
64892__ai uint64_t vtstd_u64(uint64_t __p0, uint64_t __p1) {
64893  uint64_t __ret;
64894  __ret = (uint64_t) __builtin_neon_vtstd_u64(__p0, __p1);
64895  return __ret;
64896}
64897#else
64898__ai uint64_t vtstd_u64(uint64_t __p0, uint64_t __p1) {
64899  uint64_t __ret;
64900  __ret = (uint64_t) __builtin_neon_vtstd_u64(__p0, __p1);
64901  return __ret;
64902}
64903#endif
64904
64905#ifdef __LITTLE_ENDIAN__
64906__ai int64_t vtstd_s64(int64_t __p0, int64_t __p1) {
64907  int64_t __ret;
64908  __ret = (int64_t) __builtin_neon_vtstd_s64(__p0, __p1);
64909  return __ret;
64910}
64911#else
64912__ai int64_t vtstd_s64(int64_t __p0, int64_t __p1) {
64913  int64_t __ret;
64914  __ret = (int64_t) __builtin_neon_vtstd_s64(__p0, __p1);
64915  return __ret;
64916}
64917#endif
64918
64919#ifdef __LITTLE_ENDIAN__
64920__ai int8_t vuqaddb_s8(int8_t __p0, int8_t __p1) {
64921  int8_t __ret;
64922  __ret = (int8_t) __builtin_neon_vuqaddb_s8(__p0, __p1);
64923  return __ret;
64924}
64925#else
64926__ai int8_t vuqaddb_s8(int8_t __p0, int8_t __p1) {
64927  int8_t __ret;
64928  __ret = (int8_t) __builtin_neon_vuqaddb_s8(__p0, __p1);
64929  return __ret;
64930}
64931#endif
64932
64933#ifdef __LITTLE_ENDIAN__
64934__ai int32_t vuqadds_s32(int32_t __p0, int32_t __p1) {
64935  int32_t __ret;
64936  __ret = (int32_t) __builtin_neon_vuqadds_s32(__p0, __p1);
64937  return __ret;
64938}
64939#else
64940__ai int32_t vuqadds_s32(int32_t __p0, int32_t __p1) {
64941  int32_t __ret;
64942  __ret = (int32_t) __builtin_neon_vuqadds_s32(__p0, __p1);
64943  return __ret;
64944}
64945#endif
64946
64947#ifdef __LITTLE_ENDIAN__
64948__ai int64_t vuqaddd_s64(int64_t __p0, int64_t __p1) {
64949  int64_t __ret;
64950  __ret = (int64_t) __builtin_neon_vuqaddd_s64(__p0, __p1);
64951  return __ret;
64952}
64953#else
64954__ai int64_t vuqaddd_s64(int64_t __p0, int64_t __p1) {
64955  int64_t __ret;
64956  __ret = (int64_t) __builtin_neon_vuqaddd_s64(__p0, __p1);
64957  return __ret;
64958}
64959#endif
64960
64961#ifdef __LITTLE_ENDIAN__
64962__ai int16_t vuqaddh_s16(int16_t __p0, int16_t __p1) {
64963  int16_t __ret;
64964  __ret = (int16_t) __builtin_neon_vuqaddh_s16(__p0, __p1);
64965  return __ret;
64966}
64967#else
64968__ai int16_t vuqaddh_s16(int16_t __p0, int16_t __p1) {
64969  int16_t __ret;
64970  __ret = (int16_t) __builtin_neon_vuqaddh_s16(__p0, __p1);
64971  return __ret;
64972}
64973#endif
64974
64975#ifdef __LITTLE_ENDIAN__
64976__ai int8x16_t vuqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
64977  int8x16_t __ret;
64978  __ret = (int8x16_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 32);
64979  return __ret;
64980}
64981#else
64982__ai int8x16_t vuqaddq_s8(int8x16_t __p0, int8x16_t __p1) {
64983  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64984  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64985  int8x16_t __ret;
64986  __ret = (int8x16_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 32);
64987  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
64988  return __ret;
64989}
64990#endif
64991
64992#ifdef __LITTLE_ENDIAN__
64993__ai int32x4_t vuqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
64994  int32x4_t __ret;
64995  __ret = (int32x4_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 34);
64996  return __ret;
64997}
64998#else
64999__ai int32x4_t vuqaddq_s32(int32x4_t __p0, int32x4_t __p1) {
65000  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65001  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65002  int32x4_t __ret;
65003  __ret = (int32x4_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 34);
65004  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65005  return __ret;
65006}
65007#endif
65008
65009#ifdef __LITTLE_ENDIAN__
65010__ai int64x2_t vuqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
65011  int64x2_t __ret;
65012  __ret = (int64x2_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 35);
65013  return __ret;
65014}
65015#else
65016__ai int64x2_t vuqaddq_s64(int64x2_t __p0, int64x2_t __p1) {
65017  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65018  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65019  int64x2_t __ret;
65020  __ret = (int64x2_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 35);
65021  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65022  return __ret;
65023}
65024#endif
65025
65026#ifdef __LITTLE_ENDIAN__
65027__ai int16x8_t vuqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
65028  int16x8_t __ret;
65029  __ret = (int16x8_t) __builtin_neon_vuqaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 33);
65030  return __ret;
65031}
65032#else
65033__ai int16x8_t vuqaddq_s16(int16x8_t __p0, int16x8_t __p1) {
65034  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65035  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65036  int16x8_t __ret;
65037  __ret = (int16x8_t) __builtin_neon_vuqaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 33);
65038  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65039  return __ret;
65040}
65041#endif
65042
65043#ifdef __LITTLE_ENDIAN__
65044__ai int8x8_t vuqadd_s8(int8x8_t __p0, int8x8_t __p1) {
65045  int8x8_t __ret;
65046  __ret = (int8x8_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 0);
65047  return __ret;
65048}
65049#else
65050__ai int8x8_t vuqadd_s8(int8x8_t __p0, int8x8_t __p1) {
65051  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65052  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65053  int8x8_t __ret;
65054  __ret = (int8x8_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 0);
65055  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65056  return __ret;
65057}
65058#endif
65059
65060#ifdef __LITTLE_ENDIAN__
65061__ai int32x2_t vuqadd_s32(int32x2_t __p0, int32x2_t __p1) {
65062  int32x2_t __ret;
65063  __ret = (int32x2_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 2);
65064  return __ret;
65065}
65066#else
65067__ai int32x2_t vuqadd_s32(int32x2_t __p0, int32x2_t __p1) {
65068  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65069  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65070  int32x2_t __ret;
65071  __ret = (int32x2_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 2);
65072  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65073  return __ret;
65074}
65075#endif
65076
65077#ifdef __LITTLE_ENDIAN__
65078__ai int64x1_t vuqadd_s64(int64x1_t __p0, int64x1_t __p1) {
65079  int64x1_t __ret;
65080  __ret = (int64x1_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
65081  return __ret;
65082}
65083#else
65084__ai int64x1_t vuqadd_s64(int64x1_t __p0, int64x1_t __p1) {
65085  int64x1_t __ret;
65086  __ret = (int64x1_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 3);
65087  return __ret;
65088}
65089#endif
65090
65091#ifdef __LITTLE_ENDIAN__
65092__ai int16x4_t vuqadd_s16(int16x4_t __p0, int16x4_t __p1) {
65093  int16x4_t __ret;
65094  __ret = (int16x4_t) __builtin_neon_vuqadd_v((int8x8_t)__p0, (int8x8_t)__p1, 1);
65095  return __ret;
65096}
65097#else
65098__ai int16x4_t vuqadd_s16(int16x4_t __p0, int16x4_t __p1) {
65099  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65100  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65101  int16x4_t __ret;
65102  __ret = (int16x4_t) __builtin_neon_vuqadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 1);
65103  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65104  return __ret;
65105}
65106#endif
65107
65108#ifdef __LITTLE_ENDIAN__
65109__ai poly8x8_t vuzp1_p8(poly8x8_t __p0, poly8x8_t __p1) {
65110  poly8x8_t __ret;
65111  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65112  return __ret;
65113}
65114#else
65115__ai poly8x8_t vuzp1_p8(poly8x8_t __p0, poly8x8_t __p1) {
65116  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65117  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65118  poly8x8_t __ret;
65119  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65120  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65121  return __ret;
65122}
65123#endif
65124
65125#ifdef __LITTLE_ENDIAN__
65126__ai poly16x4_t vuzp1_p16(poly16x4_t __p0, poly16x4_t __p1) {
65127  poly16x4_t __ret;
65128  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65129  return __ret;
65130}
65131#else
65132__ai poly16x4_t vuzp1_p16(poly16x4_t __p0, poly16x4_t __p1) {
65133  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65134  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65135  poly16x4_t __ret;
65136  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65137  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65138  return __ret;
65139}
65140#endif
65141
65142#ifdef __LITTLE_ENDIAN__
65143__ai poly8x16_t vuzp1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65144  poly8x16_t __ret;
65145  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65146  return __ret;
65147}
65148#else
65149__ai poly8x16_t vuzp1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65150  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65151  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65152  poly8x16_t __ret;
65153  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65154  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65155  return __ret;
65156}
65157#endif
65158
65159#ifdef __LITTLE_ENDIAN__
65160__ai poly64x2_t vuzp1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65161  poly64x2_t __ret;
65162  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65163  return __ret;
65164}
65165#else
65166__ai poly64x2_t vuzp1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65167  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65168  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65169  poly64x2_t __ret;
65170  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65171  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65172  return __ret;
65173}
65174#endif
65175
65176#ifdef __LITTLE_ENDIAN__
65177__ai poly16x8_t vuzp1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65178  poly16x8_t __ret;
65179  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65180  return __ret;
65181}
65182#else
65183__ai poly16x8_t vuzp1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65184  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65185  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65186  poly16x8_t __ret;
65187  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65188  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65189  return __ret;
65190}
65191#endif
65192
65193#ifdef __LITTLE_ENDIAN__
65194__ai uint8x16_t vuzp1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65195  uint8x16_t __ret;
65196  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65197  return __ret;
65198}
65199#else
65200__ai uint8x16_t vuzp1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65201  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65202  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65203  uint8x16_t __ret;
65204  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65205  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65206  return __ret;
65207}
65208#endif
65209
65210#ifdef __LITTLE_ENDIAN__
65211__ai uint32x4_t vuzp1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65212  uint32x4_t __ret;
65213  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65214  return __ret;
65215}
65216#else
65217__ai uint32x4_t vuzp1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65218  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65219  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65220  uint32x4_t __ret;
65221  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65222  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65223  return __ret;
65224}
65225#endif
65226
65227#ifdef __LITTLE_ENDIAN__
65228__ai uint64x2_t vuzp1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65229  uint64x2_t __ret;
65230  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65231  return __ret;
65232}
65233#else
65234__ai uint64x2_t vuzp1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65235  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65236  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65237  uint64x2_t __ret;
65238  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65239  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65240  return __ret;
65241}
65242#endif
65243
65244#ifdef __LITTLE_ENDIAN__
65245__ai uint16x8_t vuzp1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65246  uint16x8_t __ret;
65247  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65248  return __ret;
65249}
65250#else
65251__ai uint16x8_t vuzp1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65252  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65253  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65254  uint16x8_t __ret;
65255  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65256  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65257  return __ret;
65258}
65259#endif
65260
65261#ifdef __LITTLE_ENDIAN__
65262__ai int8x16_t vuzp1q_s8(int8x16_t __p0, int8x16_t __p1) {
65263  int8x16_t __ret;
65264  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65265  return __ret;
65266}
65267#else
65268__ai int8x16_t vuzp1q_s8(int8x16_t __p0, int8x16_t __p1) {
65269  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65270  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65271  int8x16_t __ret;
65272  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28, 30);
65273  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65274  return __ret;
65275}
65276#endif
65277
65278#ifdef __LITTLE_ENDIAN__
65279__ai float64x2_t vuzp1q_f64(float64x2_t __p0, float64x2_t __p1) {
65280  float64x2_t __ret;
65281  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65282  return __ret;
65283}
65284#else
65285__ai float64x2_t vuzp1q_f64(float64x2_t __p0, float64x2_t __p1) {
65286  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65287  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65288  float64x2_t __ret;
65289  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65290  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65291  return __ret;
65292}
65293#endif
65294
65295#ifdef __LITTLE_ENDIAN__
65296__ai float32x4_t vuzp1q_f32(float32x4_t __p0, float32x4_t __p1) {
65297  float32x4_t __ret;
65298  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65299  return __ret;
65300}
65301#else
65302__ai float32x4_t vuzp1q_f32(float32x4_t __p0, float32x4_t __p1) {
65303  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65304  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65305  float32x4_t __ret;
65306  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65307  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65308  return __ret;
65309}
65310#endif
65311
65312#ifdef __LITTLE_ENDIAN__
65313__ai int32x4_t vuzp1q_s32(int32x4_t __p0, int32x4_t __p1) {
65314  int32x4_t __ret;
65315  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65316  return __ret;
65317}
65318#else
65319__ai int32x4_t vuzp1q_s32(int32x4_t __p0, int32x4_t __p1) {
65320  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65321  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65322  int32x4_t __ret;
65323  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65324  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65325  return __ret;
65326}
65327#endif
65328
65329#ifdef __LITTLE_ENDIAN__
65330__ai int64x2_t vuzp1q_s64(int64x2_t __p0, int64x2_t __p1) {
65331  int64x2_t __ret;
65332  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65333  return __ret;
65334}
65335#else
65336__ai int64x2_t vuzp1q_s64(int64x2_t __p0, int64x2_t __p1) {
65337  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65338  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65339  int64x2_t __ret;
65340  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65341  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65342  return __ret;
65343}
65344#endif
65345
65346#ifdef __LITTLE_ENDIAN__
65347__ai int16x8_t vuzp1q_s16(int16x8_t __p0, int16x8_t __p1) {
65348  int16x8_t __ret;
65349  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65350  return __ret;
65351}
65352#else
65353__ai int16x8_t vuzp1q_s16(int16x8_t __p0, int16x8_t __p1) {
65354  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65355  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65356  int16x8_t __ret;
65357  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65358  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65359  return __ret;
65360}
65361#endif
65362
65363#ifdef __LITTLE_ENDIAN__
65364__ai uint8x8_t vuzp1_u8(uint8x8_t __p0, uint8x8_t __p1) {
65365  uint8x8_t __ret;
65366  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65367  return __ret;
65368}
65369#else
65370__ai uint8x8_t vuzp1_u8(uint8x8_t __p0, uint8x8_t __p1) {
65371  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65372  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65373  uint8x8_t __ret;
65374  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65375  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65376  return __ret;
65377}
65378#endif
65379
65380#ifdef __LITTLE_ENDIAN__
65381__ai uint32x2_t vuzp1_u32(uint32x2_t __p0, uint32x2_t __p1) {
65382  uint32x2_t __ret;
65383  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65384  return __ret;
65385}
65386#else
65387__ai uint32x2_t vuzp1_u32(uint32x2_t __p0, uint32x2_t __p1) {
65388  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65389  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65390  uint32x2_t __ret;
65391  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65392  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65393  return __ret;
65394}
65395#endif
65396
65397#ifdef __LITTLE_ENDIAN__
65398__ai uint16x4_t vuzp1_u16(uint16x4_t __p0, uint16x4_t __p1) {
65399  uint16x4_t __ret;
65400  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65401  return __ret;
65402}
65403#else
65404__ai uint16x4_t vuzp1_u16(uint16x4_t __p0, uint16x4_t __p1) {
65405  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65406  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65407  uint16x4_t __ret;
65408  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65409  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65410  return __ret;
65411}
65412#endif
65413
65414#ifdef __LITTLE_ENDIAN__
65415__ai int8x8_t vuzp1_s8(int8x8_t __p0, int8x8_t __p1) {
65416  int8x8_t __ret;
65417  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6, 8, 10, 12, 14);
65418  return __ret;
65419}
65420#else
65421__ai int8x8_t vuzp1_s8(int8x8_t __p0, int8x8_t __p1) {
65422  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65423  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65424  int8x8_t __ret;
65425  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6, 8, 10, 12, 14);
65426  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65427  return __ret;
65428}
65429#endif
65430
65431#ifdef __LITTLE_ENDIAN__
65432__ai float32x2_t vuzp1_f32(float32x2_t __p0, float32x2_t __p1) {
65433  float32x2_t __ret;
65434  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65435  return __ret;
65436}
65437#else
65438__ai float32x2_t vuzp1_f32(float32x2_t __p0, float32x2_t __p1) {
65439  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65440  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65441  float32x2_t __ret;
65442  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65443  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65444  return __ret;
65445}
65446#endif
65447
65448#ifdef __LITTLE_ENDIAN__
65449__ai int32x2_t vuzp1_s32(int32x2_t __p0, int32x2_t __p1) {
65450  int32x2_t __ret;
65451  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65452  return __ret;
65453}
65454#else
65455__ai int32x2_t vuzp1_s32(int32x2_t __p0, int32x2_t __p1) {
65456  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65457  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65458  int32x2_t __ret;
65459  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65460  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65461  return __ret;
65462}
65463#endif
65464
65465#ifdef __LITTLE_ENDIAN__
65466__ai int16x4_t vuzp1_s16(int16x4_t __p0, int16x4_t __p1) {
65467  int16x4_t __ret;
65468  __ret = __builtin_shufflevector(__p0, __p1, 0, 2, 4, 6);
65469  return __ret;
65470}
65471#else
65472__ai int16x4_t vuzp1_s16(int16x4_t __p0, int16x4_t __p1) {
65473  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65474  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65475  int16x4_t __ret;
65476  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2, 4, 6);
65477  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65478  return __ret;
65479}
65480#endif
65481
65482#ifdef __LITTLE_ENDIAN__
65483__ai poly8x8_t vuzp2_p8(poly8x8_t __p0, poly8x8_t __p1) {
65484  poly8x8_t __ret;
65485  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65486  return __ret;
65487}
65488#else
65489__ai poly8x8_t vuzp2_p8(poly8x8_t __p0, poly8x8_t __p1) {
65490  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65491  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65492  poly8x8_t __ret;
65493  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65494  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65495  return __ret;
65496}
65497#endif
65498
65499#ifdef __LITTLE_ENDIAN__
65500__ai poly16x4_t vuzp2_p16(poly16x4_t __p0, poly16x4_t __p1) {
65501  poly16x4_t __ret;
65502  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65503  return __ret;
65504}
65505#else
65506__ai poly16x4_t vuzp2_p16(poly16x4_t __p0, poly16x4_t __p1) {
65507  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65508  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65509  poly16x4_t __ret;
65510  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65511  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65512  return __ret;
65513}
65514#endif
65515
65516#ifdef __LITTLE_ENDIAN__
65517__ai poly8x16_t vuzp2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65518  poly8x16_t __ret;
65519  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65520  return __ret;
65521}
65522#else
65523__ai poly8x16_t vuzp2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65524  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65525  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65526  poly8x16_t __ret;
65527  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65528  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65529  return __ret;
65530}
65531#endif
65532
65533#ifdef __LITTLE_ENDIAN__
65534__ai poly64x2_t vuzp2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65535  poly64x2_t __ret;
65536  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65537  return __ret;
65538}
65539#else
65540__ai poly64x2_t vuzp2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65541  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65542  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65543  poly64x2_t __ret;
65544  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65545  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65546  return __ret;
65547}
65548#endif
65549
65550#ifdef __LITTLE_ENDIAN__
65551__ai poly16x8_t vuzp2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65552  poly16x8_t __ret;
65553  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65554  return __ret;
65555}
65556#else
65557__ai poly16x8_t vuzp2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65558  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65559  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65560  poly16x8_t __ret;
65561  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65562  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65563  return __ret;
65564}
65565#endif
65566
65567#ifdef __LITTLE_ENDIAN__
65568__ai uint8x16_t vuzp2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65569  uint8x16_t __ret;
65570  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65571  return __ret;
65572}
65573#else
65574__ai uint8x16_t vuzp2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65575  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65576  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65577  uint8x16_t __ret;
65578  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65579  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65580  return __ret;
65581}
65582#endif
65583
65584#ifdef __LITTLE_ENDIAN__
65585__ai uint32x4_t vuzp2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65586  uint32x4_t __ret;
65587  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65588  return __ret;
65589}
65590#else
65591__ai uint32x4_t vuzp2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65592  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65593  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65594  uint32x4_t __ret;
65595  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65596  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65597  return __ret;
65598}
65599#endif
65600
65601#ifdef __LITTLE_ENDIAN__
65602__ai uint64x2_t vuzp2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65603  uint64x2_t __ret;
65604  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65605  return __ret;
65606}
65607#else
65608__ai uint64x2_t vuzp2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65609  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65610  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65611  uint64x2_t __ret;
65612  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65613  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65614  return __ret;
65615}
65616#endif
65617
65618#ifdef __LITTLE_ENDIAN__
65619__ai uint16x8_t vuzp2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65620  uint16x8_t __ret;
65621  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65622  return __ret;
65623}
65624#else
65625__ai uint16x8_t vuzp2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65626  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65627  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65628  uint16x8_t __ret;
65629  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65630  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65631  return __ret;
65632}
65633#endif
65634
65635#ifdef __LITTLE_ENDIAN__
65636__ai int8x16_t vuzp2q_s8(int8x16_t __p0, int8x16_t __p1) {
65637  int8x16_t __ret;
65638  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65639  return __ret;
65640}
65641#else
65642__ai int8x16_t vuzp2q_s8(int8x16_t __p0, int8x16_t __p1) {
65643  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65644  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65645  int8x16_t __ret;
65646  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31);
65647  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65648  return __ret;
65649}
65650#endif
65651
65652#ifdef __LITTLE_ENDIAN__
65653__ai float64x2_t vuzp2q_f64(float64x2_t __p0, float64x2_t __p1) {
65654  float64x2_t __ret;
65655  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65656  return __ret;
65657}
65658#else
65659__ai float64x2_t vuzp2q_f64(float64x2_t __p0, float64x2_t __p1) {
65660  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65661  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65662  float64x2_t __ret;
65663  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65664  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65665  return __ret;
65666}
65667#endif
65668
65669#ifdef __LITTLE_ENDIAN__
65670__ai float32x4_t vuzp2q_f32(float32x4_t __p0, float32x4_t __p1) {
65671  float32x4_t __ret;
65672  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65673  return __ret;
65674}
65675#else
65676__ai float32x4_t vuzp2q_f32(float32x4_t __p0, float32x4_t __p1) {
65677  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65678  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65679  float32x4_t __ret;
65680  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65681  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65682  return __ret;
65683}
65684#endif
65685
65686#ifdef __LITTLE_ENDIAN__
65687__ai int32x4_t vuzp2q_s32(int32x4_t __p0, int32x4_t __p1) {
65688  int32x4_t __ret;
65689  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65690  return __ret;
65691}
65692#else
65693__ai int32x4_t vuzp2q_s32(int32x4_t __p0, int32x4_t __p1) {
65694  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65695  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65696  int32x4_t __ret;
65697  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65698  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65699  return __ret;
65700}
65701#endif
65702
65703#ifdef __LITTLE_ENDIAN__
65704__ai int64x2_t vuzp2q_s64(int64x2_t __p0, int64x2_t __p1) {
65705  int64x2_t __ret;
65706  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65707  return __ret;
65708}
65709#else
65710__ai int64x2_t vuzp2q_s64(int64x2_t __p0, int64x2_t __p1) {
65711  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65712  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65713  int64x2_t __ret;
65714  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65715  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65716  return __ret;
65717}
65718#endif
65719
65720#ifdef __LITTLE_ENDIAN__
65721__ai int16x8_t vuzp2q_s16(int16x8_t __p0, int16x8_t __p1) {
65722  int16x8_t __ret;
65723  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65724  return __ret;
65725}
65726#else
65727__ai int16x8_t vuzp2q_s16(int16x8_t __p0, int16x8_t __p1) {
65728  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65729  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65730  int16x8_t __ret;
65731  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65732  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65733  return __ret;
65734}
65735#endif
65736
65737#ifdef __LITTLE_ENDIAN__
65738__ai uint8x8_t vuzp2_u8(uint8x8_t __p0, uint8x8_t __p1) {
65739  uint8x8_t __ret;
65740  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65741  return __ret;
65742}
65743#else
65744__ai uint8x8_t vuzp2_u8(uint8x8_t __p0, uint8x8_t __p1) {
65745  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65746  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65747  uint8x8_t __ret;
65748  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65749  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65750  return __ret;
65751}
65752#endif
65753
65754#ifdef __LITTLE_ENDIAN__
65755__ai uint32x2_t vuzp2_u32(uint32x2_t __p0, uint32x2_t __p1) {
65756  uint32x2_t __ret;
65757  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65758  return __ret;
65759}
65760#else
65761__ai uint32x2_t vuzp2_u32(uint32x2_t __p0, uint32x2_t __p1) {
65762  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65763  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65764  uint32x2_t __ret;
65765  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65766  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65767  return __ret;
65768}
65769#endif
65770
65771#ifdef __LITTLE_ENDIAN__
65772__ai uint16x4_t vuzp2_u16(uint16x4_t __p0, uint16x4_t __p1) {
65773  uint16x4_t __ret;
65774  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65775  return __ret;
65776}
65777#else
65778__ai uint16x4_t vuzp2_u16(uint16x4_t __p0, uint16x4_t __p1) {
65779  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65780  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65781  uint16x4_t __ret;
65782  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65783  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65784  return __ret;
65785}
65786#endif
65787
65788#ifdef __LITTLE_ENDIAN__
65789__ai int8x8_t vuzp2_s8(int8x8_t __p0, int8x8_t __p1) {
65790  int8x8_t __ret;
65791  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7, 9, 11, 13, 15);
65792  return __ret;
65793}
65794#else
65795__ai int8x8_t vuzp2_s8(int8x8_t __p0, int8x8_t __p1) {
65796  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65797  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65798  int8x8_t __ret;
65799  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7, 9, 11, 13, 15);
65800  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65801  return __ret;
65802}
65803#endif
65804
65805#ifdef __LITTLE_ENDIAN__
65806__ai float32x2_t vuzp2_f32(float32x2_t __p0, float32x2_t __p1) {
65807  float32x2_t __ret;
65808  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65809  return __ret;
65810}
65811#else
65812__ai float32x2_t vuzp2_f32(float32x2_t __p0, float32x2_t __p1) {
65813  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65814  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65815  float32x2_t __ret;
65816  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65817  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65818  return __ret;
65819}
65820#endif
65821
65822#ifdef __LITTLE_ENDIAN__
65823__ai int32x2_t vuzp2_s32(int32x2_t __p0, int32x2_t __p1) {
65824  int32x2_t __ret;
65825  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
65826  return __ret;
65827}
65828#else
65829__ai int32x2_t vuzp2_s32(int32x2_t __p0, int32x2_t __p1) {
65830  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65831  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65832  int32x2_t __ret;
65833  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
65834  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65835  return __ret;
65836}
65837#endif
65838
65839#ifdef __LITTLE_ENDIAN__
65840__ai int16x4_t vuzp2_s16(int16x4_t __p0, int16x4_t __p1) {
65841  int16x4_t __ret;
65842  __ret = __builtin_shufflevector(__p0, __p1, 1, 3, 5, 7);
65843  return __ret;
65844}
65845#else
65846__ai int16x4_t vuzp2_s16(int16x4_t __p0, int16x4_t __p1) {
65847  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65848  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65849  int16x4_t __ret;
65850  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3, 5, 7);
65851  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65852  return __ret;
65853}
65854#endif
65855
65856#ifdef __LITTLE_ENDIAN__
65857__ai poly8x8_t vzip1_p8(poly8x8_t __p0, poly8x8_t __p1) {
65858  poly8x8_t __ret;
65859  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65860  return __ret;
65861}
65862#else
65863__ai poly8x8_t vzip1_p8(poly8x8_t __p0, poly8x8_t __p1) {
65864  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65865  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65866  poly8x8_t __ret;
65867  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
65868  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65869  return __ret;
65870}
65871#endif
65872
65873#ifdef __LITTLE_ENDIAN__
65874__ai poly16x4_t vzip1_p16(poly16x4_t __p0, poly16x4_t __p1) {
65875  poly16x4_t __ret;
65876  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
65877  return __ret;
65878}
65879#else
65880__ai poly16x4_t vzip1_p16(poly16x4_t __p0, poly16x4_t __p1) {
65881  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65882  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65883  poly16x4_t __ret;
65884  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
65885  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65886  return __ret;
65887}
65888#endif
65889
65890#ifdef __LITTLE_ENDIAN__
65891__ai poly8x16_t vzip1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65892  poly8x16_t __ret;
65893  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65894  return __ret;
65895}
65896#else
65897__ai poly8x16_t vzip1q_p8(poly8x16_t __p0, poly8x16_t __p1) {
65898  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65899  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65900  poly8x16_t __ret;
65901  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65902  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65903  return __ret;
65904}
65905#endif
65906
65907#ifdef __LITTLE_ENDIAN__
65908__ai poly64x2_t vzip1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65909  poly64x2_t __ret;
65910  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65911  return __ret;
65912}
65913#else
65914__ai poly64x2_t vzip1q_p64(poly64x2_t __p0, poly64x2_t __p1) {
65915  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65916  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65917  poly64x2_t __ret;
65918  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65919  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65920  return __ret;
65921}
65922#endif
65923
65924#ifdef __LITTLE_ENDIAN__
65925__ai poly16x8_t vzip1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65926  poly16x8_t __ret;
65927  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65928  return __ret;
65929}
65930#else
65931__ai poly16x8_t vzip1q_p16(poly16x8_t __p0, poly16x8_t __p1) {
65932  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
65933  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
65934  poly16x8_t __ret;
65935  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
65936  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
65937  return __ret;
65938}
65939#endif
65940
65941#ifdef __LITTLE_ENDIAN__
65942__ai uint8x16_t vzip1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65943  uint8x16_t __ret;
65944  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65945  return __ret;
65946}
65947#else
65948__ai uint8x16_t vzip1q_u8(uint8x16_t __p0, uint8x16_t __p1) {
65949  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65950  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65951  uint8x16_t __ret;
65952  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
65953  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
65954  return __ret;
65955}
65956#endif
65957
65958#ifdef __LITTLE_ENDIAN__
65959__ai uint32x4_t vzip1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65960  uint32x4_t __ret;
65961  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
65962  return __ret;
65963}
65964#else
65965__ai uint32x4_t vzip1q_u32(uint32x4_t __p0, uint32x4_t __p1) {
65966  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
65967  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
65968  uint32x4_t __ret;
65969  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
65970  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
65971  return __ret;
65972}
65973#endif
65974
65975#ifdef __LITTLE_ENDIAN__
65976__ai uint64x2_t vzip1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65977  uint64x2_t __ret;
65978  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
65979  return __ret;
65980}
65981#else
65982__ai uint64x2_t vzip1q_u64(uint64x2_t __p0, uint64x2_t __p1) {
65983  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
65984  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
65985  uint64x2_t __ret;
65986  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
65987  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
65988  return __ret;
65989}
65990#endif
65991
65992#ifdef __LITTLE_ENDIAN__
65993__ai uint16x8_t vzip1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
65994  uint16x8_t __ret;
65995  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
65996  return __ret;
65997}
65998#else
65999__ai uint16x8_t vzip1q_u16(uint16x8_t __p0, uint16x8_t __p1) {
66000  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66001  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66002  uint16x8_t __ret;
66003  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
66004  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66005  return __ret;
66006}
66007#endif
66008
66009#ifdef __LITTLE_ENDIAN__
66010__ai int8x16_t vzip1q_s8(int8x16_t __p0, int8x16_t __p1) {
66011  int8x16_t __ret;
66012  __ret = __builtin_shufflevector(__p0, __p1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
66013  return __ret;
66014}
66015#else
66016__ai int8x16_t vzip1q_s8(int8x16_t __p0, int8x16_t __p1) {
66017  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66018  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66019  int8x16_t __ret;
66020  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23);
66021  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66022  return __ret;
66023}
66024#endif
66025
66026#ifdef __LITTLE_ENDIAN__
66027__ai float64x2_t vzip1q_f64(float64x2_t __p0, float64x2_t __p1) {
66028  float64x2_t __ret;
66029  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66030  return __ret;
66031}
66032#else
66033__ai float64x2_t vzip1q_f64(float64x2_t __p0, float64x2_t __p1) {
66034  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66035  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66036  float64x2_t __ret;
66037  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66038  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66039  return __ret;
66040}
66041#endif
66042
66043#ifdef __LITTLE_ENDIAN__
66044__ai float32x4_t vzip1q_f32(float32x4_t __p0, float32x4_t __p1) {
66045  float32x4_t __ret;
66046  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
66047  return __ret;
66048}
66049#else
66050__ai float32x4_t vzip1q_f32(float32x4_t __p0, float32x4_t __p1) {
66051  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66052  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66053  float32x4_t __ret;
66054  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
66055  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66056  return __ret;
66057}
66058#endif
66059
66060#ifdef __LITTLE_ENDIAN__
66061__ai int32x4_t vzip1q_s32(int32x4_t __p0, int32x4_t __p1) {
66062  int32x4_t __ret;
66063  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
66064  return __ret;
66065}
66066#else
66067__ai int32x4_t vzip1q_s32(int32x4_t __p0, int32x4_t __p1) {
66068  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66069  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66070  int32x4_t __ret;
66071  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
66072  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66073  return __ret;
66074}
66075#endif
66076
66077#ifdef __LITTLE_ENDIAN__
66078__ai int64x2_t vzip1q_s64(int64x2_t __p0, int64x2_t __p1) {
66079  int64x2_t __ret;
66080  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66081  return __ret;
66082}
66083#else
66084__ai int64x2_t vzip1q_s64(int64x2_t __p0, int64x2_t __p1) {
66085  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66086  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66087  int64x2_t __ret;
66088  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66089  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66090  return __ret;
66091}
66092#endif
66093
66094#ifdef __LITTLE_ENDIAN__
66095__ai int16x8_t vzip1q_s16(int16x8_t __p0, int16x8_t __p1) {
66096  int16x8_t __ret;
66097  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
66098  return __ret;
66099}
66100#else
66101__ai int16x8_t vzip1q_s16(int16x8_t __p0, int16x8_t __p1) {
66102  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66103  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66104  int16x8_t __ret;
66105  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
66106  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66107  return __ret;
66108}
66109#endif
66110
66111#ifdef __LITTLE_ENDIAN__
66112__ai uint8x8_t vzip1_u8(uint8x8_t __p0, uint8x8_t __p1) {
66113  uint8x8_t __ret;
66114  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
66115  return __ret;
66116}
66117#else
66118__ai uint8x8_t vzip1_u8(uint8x8_t __p0, uint8x8_t __p1) {
66119  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66120  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66121  uint8x8_t __ret;
66122  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
66123  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66124  return __ret;
66125}
66126#endif
66127
66128#ifdef __LITTLE_ENDIAN__
66129__ai uint32x2_t vzip1_u32(uint32x2_t __p0, uint32x2_t __p1) {
66130  uint32x2_t __ret;
66131  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66132  return __ret;
66133}
66134#else
66135__ai uint32x2_t vzip1_u32(uint32x2_t __p0, uint32x2_t __p1) {
66136  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66137  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66138  uint32x2_t __ret;
66139  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66140  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66141  return __ret;
66142}
66143#endif
66144
66145#ifdef __LITTLE_ENDIAN__
66146__ai uint16x4_t vzip1_u16(uint16x4_t __p0, uint16x4_t __p1) {
66147  uint16x4_t __ret;
66148  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
66149  return __ret;
66150}
66151#else
66152__ai uint16x4_t vzip1_u16(uint16x4_t __p0, uint16x4_t __p1) {
66153  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66154  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66155  uint16x4_t __ret;
66156  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
66157  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66158  return __ret;
66159}
66160#endif
66161
66162#ifdef __LITTLE_ENDIAN__
66163__ai int8x8_t vzip1_s8(int8x8_t __p0, int8x8_t __p1) {
66164  int8x8_t __ret;
66165  __ret = __builtin_shufflevector(__p0, __p1, 0, 8, 1, 9, 2, 10, 3, 11);
66166  return __ret;
66167}
66168#else
66169__ai int8x8_t vzip1_s8(int8x8_t __p0, int8x8_t __p1) {
66170  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66171  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66172  int8x8_t __ret;
66173  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 8, 1, 9, 2, 10, 3, 11);
66174  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66175  return __ret;
66176}
66177#endif
66178
66179#ifdef __LITTLE_ENDIAN__
66180__ai float32x2_t vzip1_f32(float32x2_t __p0, float32x2_t __p1) {
66181  float32x2_t __ret;
66182  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66183  return __ret;
66184}
66185#else
66186__ai float32x2_t vzip1_f32(float32x2_t __p0, float32x2_t __p1) {
66187  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66188  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66189  float32x2_t __ret;
66190  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66191  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66192  return __ret;
66193}
66194#endif
66195
66196#ifdef __LITTLE_ENDIAN__
66197__ai int32x2_t vzip1_s32(int32x2_t __p0, int32x2_t __p1) {
66198  int32x2_t __ret;
66199  __ret = __builtin_shufflevector(__p0, __p1, 0, 2);
66200  return __ret;
66201}
66202#else
66203__ai int32x2_t vzip1_s32(int32x2_t __p0, int32x2_t __p1) {
66204  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66205  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66206  int32x2_t __ret;
66207  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 2);
66208  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66209  return __ret;
66210}
66211#endif
66212
66213#ifdef __LITTLE_ENDIAN__
66214__ai int16x4_t vzip1_s16(int16x4_t __p0, int16x4_t __p1) {
66215  int16x4_t __ret;
66216  __ret = __builtin_shufflevector(__p0, __p1, 0, 4, 1, 5);
66217  return __ret;
66218}
66219#else
66220__ai int16x4_t vzip1_s16(int16x4_t __p0, int16x4_t __p1) {
66221  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66222  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66223  int16x4_t __ret;
66224  __ret = __builtin_shufflevector(__rev0, __rev1, 0, 4, 1, 5);
66225  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66226  return __ret;
66227}
66228#endif
66229
66230#ifdef __LITTLE_ENDIAN__
66231__ai poly8x8_t vzip2_p8(poly8x8_t __p0, poly8x8_t __p1) {
66232  poly8x8_t __ret;
66233  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66234  return __ret;
66235}
66236#else
66237__ai poly8x8_t vzip2_p8(poly8x8_t __p0, poly8x8_t __p1) {
66238  poly8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66239  poly8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66240  poly8x8_t __ret;
66241  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66242  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66243  return __ret;
66244}
66245#endif
66246
66247#ifdef __LITTLE_ENDIAN__
66248__ai poly16x4_t vzip2_p16(poly16x4_t __p0, poly16x4_t __p1) {
66249  poly16x4_t __ret;
66250  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66251  return __ret;
66252}
66253#else
66254__ai poly16x4_t vzip2_p16(poly16x4_t __p0, poly16x4_t __p1) {
66255  poly16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66256  poly16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66257  poly16x4_t __ret;
66258  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66259  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66260  return __ret;
66261}
66262#endif
66263
66264#ifdef __LITTLE_ENDIAN__
66265__ai poly8x16_t vzip2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
66266  poly8x16_t __ret;
66267  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66268  return __ret;
66269}
66270#else
66271__ai poly8x16_t vzip2q_p8(poly8x16_t __p0, poly8x16_t __p1) {
66272  poly8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66273  poly8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66274  poly8x16_t __ret;
66275  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66276  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66277  return __ret;
66278}
66279#endif
66280
66281#ifdef __LITTLE_ENDIAN__
66282__ai poly64x2_t vzip2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
66283  poly64x2_t __ret;
66284  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66285  return __ret;
66286}
66287#else
66288__ai poly64x2_t vzip2q_p64(poly64x2_t __p0, poly64x2_t __p1) {
66289  poly64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66290  poly64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66291  poly64x2_t __ret;
66292  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66293  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66294  return __ret;
66295}
66296#endif
66297
66298#ifdef __LITTLE_ENDIAN__
66299__ai poly16x8_t vzip2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
66300  poly16x8_t __ret;
66301  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66302  return __ret;
66303}
66304#else
66305__ai poly16x8_t vzip2q_p16(poly16x8_t __p0, poly16x8_t __p1) {
66306  poly16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66307  poly16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66308  poly16x8_t __ret;
66309  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66310  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66311  return __ret;
66312}
66313#endif
66314
66315#ifdef __LITTLE_ENDIAN__
66316__ai uint8x16_t vzip2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
66317  uint8x16_t __ret;
66318  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66319  return __ret;
66320}
66321#else
66322__ai uint8x16_t vzip2q_u8(uint8x16_t __p0, uint8x16_t __p1) {
66323  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66324  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66325  uint8x16_t __ret;
66326  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66327  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66328  return __ret;
66329}
66330#endif
66331
66332#ifdef __LITTLE_ENDIAN__
66333__ai uint32x4_t vzip2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
66334  uint32x4_t __ret;
66335  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66336  return __ret;
66337}
66338#else
66339__ai uint32x4_t vzip2q_u32(uint32x4_t __p0, uint32x4_t __p1) {
66340  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66341  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66342  uint32x4_t __ret;
66343  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66344  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66345  return __ret;
66346}
66347#endif
66348
66349#ifdef __LITTLE_ENDIAN__
66350__ai uint64x2_t vzip2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
66351  uint64x2_t __ret;
66352  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66353  return __ret;
66354}
66355#else
66356__ai uint64x2_t vzip2q_u64(uint64x2_t __p0, uint64x2_t __p1) {
66357  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66358  uint64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66359  uint64x2_t __ret;
66360  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66361  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66362  return __ret;
66363}
66364#endif
66365
66366#ifdef __LITTLE_ENDIAN__
66367__ai uint16x8_t vzip2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
66368  uint16x8_t __ret;
66369  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66370  return __ret;
66371}
66372#else
66373__ai uint16x8_t vzip2q_u16(uint16x8_t __p0, uint16x8_t __p1) {
66374  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66375  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66376  uint16x8_t __ret;
66377  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66378  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66379  return __ret;
66380}
66381#endif
66382
66383#ifdef __LITTLE_ENDIAN__
66384__ai int8x16_t vzip2q_s8(int8x16_t __p0, int8x16_t __p1) {
66385  int8x16_t __ret;
66386  __ret = __builtin_shufflevector(__p0, __p1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66387  return __ret;
66388}
66389#else
66390__ai int8x16_t vzip2q_s8(int8x16_t __p0, int8x16_t __p1) {
66391  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66392  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66393  int8x16_t __ret;
66394  __ret = __builtin_shufflevector(__rev0, __rev1, 8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31);
66395  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66396  return __ret;
66397}
66398#endif
66399
66400#ifdef __LITTLE_ENDIAN__
66401__ai float64x2_t vzip2q_f64(float64x2_t __p0, float64x2_t __p1) {
66402  float64x2_t __ret;
66403  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66404  return __ret;
66405}
66406#else
66407__ai float64x2_t vzip2q_f64(float64x2_t __p0, float64x2_t __p1) {
66408  float64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66409  float64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66410  float64x2_t __ret;
66411  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66412  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66413  return __ret;
66414}
66415#endif
66416
66417#ifdef __LITTLE_ENDIAN__
66418__ai float32x4_t vzip2q_f32(float32x4_t __p0, float32x4_t __p1) {
66419  float32x4_t __ret;
66420  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66421  return __ret;
66422}
66423#else
66424__ai float32x4_t vzip2q_f32(float32x4_t __p0, float32x4_t __p1) {
66425  float32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66426  float32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66427  float32x4_t __ret;
66428  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66429  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66430  return __ret;
66431}
66432#endif
66433
66434#ifdef __LITTLE_ENDIAN__
66435__ai int32x4_t vzip2q_s32(int32x4_t __p0, int32x4_t __p1) {
66436  int32x4_t __ret;
66437  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66438  return __ret;
66439}
66440#else
66441__ai int32x4_t vzip2q_s32(int32x4_t __p0, int32x4_t __p1) {
66442  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66443  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66444  int32x4_t __ret;
66445  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66446  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66447  return __ret;
66448}
66449#endif
66450
66451#ifdef __LITTLE_ENDIAN__
66452__ai int64x2_t vzip2q_s64(int64x2_t __p0, int64x2_t __p1) {
66453  int64x2_t __ret;
66454  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66455  return __ret;
66456}
66457#else
66458__ai int64x2_t vzip2q_s64(int64x2_t __p0, int64x2_t __p1) {
66459  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66460  int64x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66461  int64x2_t __ret;
66462  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66463  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66464  return __ret;
66465}
66466#endif
66467
66468#ifdef __LITTLE_ENDIAN__
66469__ai int16x8_t vzip2q_s16(int16x8_t __p0, int16x8_t __p1) {
66470  int16x8_t __ret;
66471  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66472  return __ret;
66473}
66474#else
66475__ai int16x8_t vzip2q_s16(int16x8_t __p0, int16x8_t __p1) {
66476  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66477  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66478  int16x8_t __ret;
66479  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66480  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66481  return __ret;
66482}
66483#endif
66484
66485#ifdef __LITTLE_ENDIAN__
66486__ai uint8x8_t vzip2_u8(uint8x8_t __p0, uint8x8_t __p1) {
66487  uint8x8_t __ret;
66488  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66489  return __ret;
66490}
66491#else
66492__ai uint8x8_t vzip2_u8(uint8x8_t __p0, uint8x8_t __p1) {
66493  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66494  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66495  uint8x8_t __ret;
66496  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66497  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66498  return __ret;
66499}
66500#endif
66501
66502#ifdef __LITTLE_ENDIAN__
66503__ai uint32x2_t vzip2_u32(uint32x2_t __p0, uint32x2_t __p1) {
66504  uint32x2_t __ret;
66505  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66506  return __ret;
66507}
66508#else
66509__ai uint32x2_t vzip2_u32(uint32x2_t __p0, uint32x2_t __p1) {
66510  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66511  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66512  uint32x2_t __ret;
66513  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66514  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66515  return __ret;
66516}
66517#endif
66518
66519#ifdef __LITTLE_ENDIAN__
66520__ai uint16x4_t vzip2_u16(uint16x4_t __p0, uint16x4_t __p1) {
66521  uint16x4_t __ret;
66522  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66523  return __ret;
66524}
66525#else
66526__ai uint16x4_t vzip2_u16(uint16x4_t __p0, uint16x4_t __p1) {
66527  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66528  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66529  uint16x4_t __ret;
66530  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66531  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66532  return __ret;
66533}
66534#endif
66535
66536#ifdef __LITTLE_ENDIAN__
66537__ai int8x8_t vzip2_s8(int8x8_t __p0, int8x8_t __p1) {
66538  int8x8_t __ret;
66539  __ret = __builtin_shufflevector(__p0, __p1, 4, 12, 5, 13, 6, 14, 7, 15);
66540  return __ret;
66541}
66542#else
66543__ai int8x8_t vzip2_s8(int8x8_t __p0, int8x8_t __p1) {
66544  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66545  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66546  int8x8_t __ret;
66547  __ret = __builtin_shufflevector(__rev0, __rev1, 4, 12, 5, 13, 6, 14, 7, 15);
66548  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66549  return __ret;
66550}
66551#endif
66552
66553#ifdef __LITTLE_ENDIAN__
66554__ai float32x2_t vzip2_f32(float32x2_t __p0, float32x2_t __p1) {
66555  float32x2_t __ret;
66556  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66557  return __ret;
66558}
66559#else
66560__ai float32x2_t vzip2_f32(float32x2_t __p0, float32x2_t __p1) {
66561  float32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66562  float32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66563  float32x2_t __ret;
66564  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66565  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66566  return __ret;
66567}
66568#endif
66569
66570#ifdef __LITTLE_ENDIAN__
66571__ai int32x2_t vzip2_s32(int32x2_t __p0, int32x2_t __p1) {
66572  int32x2_t __ret;
66573  __ret = __builtin_shufflevector(__p0, __p1, 1, 3);
66574  return __ret;
66575}
66576#else
66577__ai int32x2_t vzip2_s32(int32x2_t __p0, int32x2_t __p1) {
66578  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66579  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66580  int32x2_t __ret;
66581  __ret = __builtin_shufflevector(__rev0, __rev1, 1, 3);
66582  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66583  return __ret;
66584}
66585#endif
66586
66587#ifdef __LITTLE_ENDIAN__
66588__ai int16x4_t vzip2_s16(int16x4_t __p0, int16x4_t __p1) {
66589  int16x4_t __ret;
66590  __ret = __builtin_shufflevector(__p0, __p1, 2, 6, 3, 7);
66591  return __ret;
66592}
66593#else
66594__ai int16x4_t vzip2_s16(int16x4_t __p0, int16x4_t __p1) {
66595  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66596  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66597  int16x4_t __ret;
66598  __ret = __builtin_shufflevector(__rev0, __rev1, 2, 6, 3, 7);
66599  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66600  return __ret;
66601}
66602#endif
66603
66604#endif
66605#ifdef __LITTLE_ENDIAN__
66606__ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
66607  uint8x16_t __ret;
66608  __ret = __p0 + vabdq_u8(__p1, __p2);
66609  return __ret;
66610}
66611#else
66612__ai uint8x16_t vabaq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
66613  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66614  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66615  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66616  uint8x16_t __ret;
66617  __ret = __rev0 + __noswap_vabdq_u8(__rev1, __rev2);
66618  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66619  return __ret;
66620}
66621#endif
66622
66623#ifdef __LITTLE_ENDIAN__
66624__ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
66625  uint32x4_t __ret;
66626  __ret = __p0 + vabdq_u32(__p1, __p2);
66627  return __ret;
66628}
66629#else
66630__ai uint32x4_t vabaq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
66631  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66632  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66633  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66634  uint32x4_t __ret;
66635  __ret = __rev0 + __noswap_vabdq_u32(__rev1, __rev2);
66636  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66637  return __ret;
66638}
66639#endif
66640
66641#ifdef __LITTLE_ENDIAN__
66642__ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
66643  uint16x8_t __ret;
66644  __ret = __p0 + vabdq_u16(__p1, __p2);
66645  return __ret;
66646}
66647#else
66648__ai uint16x8_t vabaq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
66649  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66650  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66651  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66652  uint16x8_t __ret;
66653  __ret = __rev0 + __noswap_vabdq_u16(__rev1, __rev2);
66654  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66655  return __ret;
66656}
66657#endif
66658
66659#ifdef __LITTLE_ENDIAN__
66660__ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
66661  int8x16_t __ret;
66662  __ret = __p0 + vabdq_s8(__p1, __p2);
66663  return __ret;
66664}
66665#else
66666__ai int8x16_t vabaq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) {
66667  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66668  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66669  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66670  int8x16_t __ret;
66671  __ret = __rev0 + __noswap_vabdq_s8(__rev1, __rev2);
66672  __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
66673  return __ret;
66674}
66675#endif
66676
66677#ifdef __LITTLE_ENDIAN__
66678__ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
66679  int32x4_t __ret;
66680  __ret = __p0 + vabdq_s32(__p1, __p2);
66681  return __ret;
66682}
66683#else
66684__ai int32x4_t vabaq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) {
66685  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66686  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66687  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66688  int32x4_t __ret;
66689  __ret = __rev0 + __noswap_vabdq_s32(__rev1, __rev2);
66690  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66691  return __ret;
66692}
66693#endif
66694
66695#ifdef __LITTLE_ENDIAN__
66696__ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
66697  int16x8_t __ret;
66698  __ret = __p0 + vabdq_s16(__p1, __p2);
66699  return __ret;
66700}
66701#else
66702__ai int16x8_t vabaq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) {
66703  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66704  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66705  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66706  int16x8_t __ret;
66707  __ret = __rev0 + __noswap_vabdq_s16(__rev1, __rev2);
66708  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66709  return __ret;
66710}
66711#endif
66712
66713#ifdef __LITTLE_ENDIAN__
66714__ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66715  uint8x8_t __ret;
66716  __ret = __p0 + vabd_u8(__p1, __p2);
66717  return __ret;
66718}
66719#else
66720__ai uint8x8_t vaba_u8(uint8x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
66721  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66722  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66723  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66724  uint8x8_t __ret;
66725  __ret = __rev0 + __noswap_vabd_u8(__rev1, __rev2);
66726  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66727  return __ret;
66728}
66729#endif
66730
66731#ifdef __LITTLE_ENDIAN__
66732__ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66733  uint32x2_t __ret;
66734  __ret = __p0 + vabd_u32(__p1, __p2);
66735  return __ret;
66736}
66737#else
66738__ai uint32x2_t vaba_u32(uint32x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
66739  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66740  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66741  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66742  uint32x2_t __ret;
66743  __ret = __rev0 + __noswap_vabd_u32(__rev1, __rev2);
66744  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66745  return __ret;
66746}
66747#endif
66748
66749#ifdef __LITTLE_ENDIAN__
66750__ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66751  uint16x4_t __ret;
66752  __ret = __p0 + vabd_u16(__p1, __p2);
66753  return __ret;
66754}
66755#else
66756__ai uint16x4_t vaba_u16(uint16x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
66757  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66758  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66759  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66760  uint16x4_t __ret;
66761  __ret = __rev0 + __noswap_vabd_u16(__rev1, __rev2);
66762  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66763  return __ret;
66764}
66765#endif
66766
66767#ifdef __LITTLE_ENDIAN__
66768__ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
66769  int8x8_t __ret;
66770  __ret = __p0 + vabd_s8(__p1, __p2);
66771  return __ret;
66772}
66773#else
66774__ai int8x8_t vaba_s8(int8x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
66775  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66776  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66777  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
66778  int8x8_t __ret;
66779  __ret = __rev0 + __noswap_vabd_s8(__rev1, __rev2);
66780  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66781  return __ret;
66782}
66783#endif
66784
66785#ifdef __LITTLE_ENDIAN__
66786__ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
66787  int32x2_t __ret;
66788  __ret = __p0 + vabd_s32(__p1, __p2);
66789  return __ret;
66790}
66791#else
66792__ai int32x2_t vaba_s32(int32x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
66793  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66794  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66795  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
66796  int32x2_t __ret;
66797  __ret = __rev0 + __noswap_vabd_s32(__rev1, __rev2);
66798  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66799  return __ret;
66800}
66801#endif
66802
66803#ifdef __LITTLE_ENDIAN__
66804__ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
66805  int16x4_t __ret;
66806  __ret = __p0 + vabd_s16(__p1, __p2);
66807  return __ret;
66808}
66809#else
66810__ai int16x4_t vaba_s16(int16x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
66811  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66812  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66813  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
66814  int16x4_t __ret;
66815  __ret = __rev0 + __noswap_vabd_s16(__rev1, __rev2);
66816  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66817  return __ret;
66818}
66819#endif
66820
66821#ifdef __LITTLE_ENDIAN__
66822__ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66823  uint16x8_t __ret;
66824  __ret = (uint16x8_t)(vmovl_u8((uint8x8_t)(vabd_u8(__p0, __p1))));
66825  return __ret;
66826}
66827#else
66828__ai uint16x8_t vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66829  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66830  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66831  uint16x8_t __ret;
66832  __ret = (uint16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_u8(__rev0, __rev1))));
66833  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66834  return __ret;
66835}
66836__ai uint16x8_t __noswap_vabdl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66837  uint16x8_t __ret;
66838  __ret = (uint16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_u8(__p0, __p1))));
66839  return __ret;
66840}
66841#endif
66842
66843#ifdef __LITTLE_ENDIAN__
66844__ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66845  uint64x2_t __ret;
66846  __ret = (uint64x2_t)(vmovl_u32((uint32x2_t)(vabd_u32(__p0, __p1))));
66847  return __ret;
66848}
66849#else
66850__ai uint64x2_t vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66851  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66852  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66853  uint64x2_t __ret;
66854  __ret = (uint64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_u32(__rev0, __rev1))));
66855  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66856  return __ret;
66857}
66858__ai uint64x2_t __noswap_vabdl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66859  uint64x2_t __ret;
66860  __ret = (uint64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_u32(__p0, __p1))));
66861  return __ret;
66862}
66863#endif
66864
66865#ifdef __LITTLE_ENDIAN__
66866__ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66867  uint32x4_t __ret;
66868  __ret = (uint32x4_t)(vmovl_u16((uint16x4_t)(vabd_u16(__p0, __p1))));
66869  return __ret;
66870}
66871#else
66872__ai uint32x4_t vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66873  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66874  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66875  uint32x4_t __ret;
66876  __ret = (uint32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_u16(__rev0, __rev1))));
66877  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66878  return __ret;
66879}
66880__ai uint32x4_t __noswap_vabdl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66881  uint32x4_t __ret;
66882  __ret = (uint32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_u16(__p0, __p1))));
66883  return __ret;
66884}
66885#endif
66886
66887#ifdef __LITTLE_ENDIAN__
66888__ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66889  int16x8_t __ret;
66890  __ret = (int16x8_t)(vmovl_u8((uint8x8_t)(vabd_s8(__p0, __p1))));
66891  return __ret;
66892}
66893#else
66894__ai int16x8_t vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66895  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66896  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66897  int16x8_t __ret;
66898  __ret = (int16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_s8(__rev0, __rev1))));
66899  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66900  return __ret;
66901}
66902__ai int16x8_t __noswap_vabdl_s8(int8x8_t __p0, int8x8_t __p1) {
66903  int16x8_t __ret;
66904  __ret = (int16x8_t)(__noswap_vmovl_u8((uint8x8_t)(__noswap_vabd_s8(__p0, __p1))));
66905  return __ret;
66906}
66907#endif
66908
66909#ifdef __LITTLE_ENDIAN__
66910__ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66911  int64x2_t __ret;
66912  __ret = (int64x2_t)(vmovl_u32((uint32x2_t)(vabd_s32(__p0, __p1))));
66913  return __ret;
66914}
66915#else
66916__ai int64x2_t vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66917  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66918  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66919  int64x2_t __ret;
66920  __ret = (int64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_s32(__rev0, __rev1))));
66921  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66922  return __ret;
66923}
66924__ai int64x2_t __noswap_vabdl_s32(int32x2_t __p0, int32x2_t __p1) {
66925  int64x2_t __ret;
66926  __ret = (int64x2_t)(__noswap_vmovl_u32((uint32x2_t)(__noswap_vabd_s32(__p0, __p1))));
66927  return __ret;
66928}
66929#endif
66930
66931#ifdef __LITTLE_ENDIAN__
66932__ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66933  int32x4_t __ret;
66934  __ret = (int32x4_t)(vmovl_u16((uint16x4_t)(vabd_s16(__p0, __p1))));
66935  return __ret;
66936}
66937#else
66938__ai int32x4_t vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66939  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66940  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66941  int32x4_t __ret;
66942  __ret = (int32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_s16(__rev0, __rev1))));
66943  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
66944  return __ret;
66945}
66946__ai int32x4_t __noswap_vabdl_s16(int16x4_t __p0, int16x4_t __p1) {
66947  int32x4_t __ret;
66948  __ret = (int32x4_t)(__noswap_vmovl_u16((uint16x4_t)(__noswap_vabd_s16(__p0, __p1))));
66949  return __ret;
66950}
66951#endif
66952
66953#ifdef __LITTLE_ENDIAN__
66954__ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66955  uint16x8_t __ret;
66956  __ret = vmovl_u8(__p0) + vmovl_u8(__p1);
66957  return __ret;
66958}
66959#else
66960__ai uint16x8_t vaddl_u8(uint8x8_t __p0, uint8x8_t __p1) {
66961  uint8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
66962  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
66963  uint16x8_t __ret;
66964  __ret = __noswap_vmovl_u8(__rev0) + __noswap_vmovl_u8(__rev1);
66965  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
66966  return __ret;
66967}
66968#endif
66969
66970#ifdef __LITTLE_ENDIAN__
66971__ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66972  uint64x2_t __ret;
66973  __ret = vmovl_u32(__p0) + vmovl_u32(__p1);
66974  return __ret;
66975}
66976#else
66977__ai uint64x2_t vaddl_u32(uint32x2_t __p0, uint32x2_t __p1) {
66978  uint32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
66979  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
66980  uint64x2_t __ret;
66981  __ret = __noswap_vmovl_u32(__rev0) + __noswap_vmovl_u32(__rev1);
66982  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
66983  return __ret;
66984}
66985#endif
66986
66987#ifdef __LITTLE_ENDIAN__
66988__ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66989  uint32x4_t __ret;
66990  __ret = vmovl_u16(__p0) + vmovl_u16(__p1);
66991  return __ret;
66992}
66993#else
66994__ai uint32x4_t vaddl_u16(uint16x4_t __p0, uint16x4_t __p1) {
66995  uint16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
66996  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
66997  uint32x4_t __ret;
66998  __ret = __noswap_vmovl_u16(__rev0) + __noswap_vmovl_u16(__rev1);
66999  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67000  return __ret;
67001}
67002#endif
67003
67004#ifdef __LITTLE_ENDIAN__
67005__ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1) {
67006  int16x8_t __ret;
67007  __ret = vmovl_s8(__p0) + vmovl_s8(__p1);
67008  return __ret;
67009}
67010#else
67011__ai int16x8_t vaddl_s8(int8x8_t __p0, int8x8_t __p1) {
67012  int8x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67013  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67014  int16x8_t __ret;
67015  __ret = __noswap_vmovl_s8(__rev0) + __noswap_vmovl_s8(__rev1);
67016  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67017  return __ret;
67018}
67019#endif
67020
67021#ifdef __LITTLE_ENDIAN__
67022__ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1) {
67023  int64x2_t __ret;
67024  __ret = vmovl_s32(__p0) + vmovl_s32(__p1);
67025  return __ret;
67026}
67027#else
67028__ai int64x2_t vaddl_s32(int32x2_t __p0, int32x2_t __p1) {
67029  int32x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67030  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67031  int64x2_t __ret;
67032  __ret = __noswap_vmovl_s32(__rev0) + __noswap_vmovl_s32(__rev1);
67033  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67034  return __ret;
67035}
67036#endif
67037
67038#ifdef __LITTLE_ENDIAN__
67039__ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1) {
67040  int32x4_t __ret;
67041  __ret = vmovl_s16(__p0) + vmovl_s16(__p1);
67042  return __ret;
67043}
67044#else
67045__ai int32x4_t vaddl_s16(int16x4_t __p0, int16x4_t __p1) {
67046  int16x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67047  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67048  int32x4_t __ret;
67049  __ret = __noswap_vmovl_s16(__rev0) + __noswap_vmovl_s16(__rev1);
67050  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67051  return __ret;
67052}
67053#endif
67054
67055#ifdef __LITTLE_ENDIAN__
67056__ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1) {
67057  uint16x8_t __ret;
67058  __ret = __p0 + vmovl_u8(__p1);
67059  return __ret;
67060}
67061#else
67062__ai uint16x8_t vaddw_u8(uint16x8_t __p0, uint8x8_t __p1) {
67063  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67064  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67065  uint16x8_t __ret;
67066  __ret = __rev0 + __noswap_vmovl_u8(__rev1);
67067  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67068  return __ret;
67069}
67070#endif
67071
67072#ifdef __LITTLE_ENDIAN__
67073__ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1) {
67074  uint64x2_t __ret;
67075  __ret = __p0 + vmovl_u32(__p1);
67076  return __ret;
67077}
67078#else
67079__ai uint64x2_t vaddw_u32(uint64x2_t __p0, uint32x2_t __p1) {
67080  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67081  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67082  uint64x2_t __ret;
67083  __ret = __rev0 + __noswap_vmovl_u32(__rev1);
67084  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67085  return __ret;
67086}
67087#endif
67088
67089#ifdef __LITTLE_ENDIAN__
67090__ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1) {
67091  uint32x4_t __ret;
67092  __ret = __p0 + vmovl_u16(__p1);
67093  return __ret;
67094}
67095#else
67096__ai uint32x4_t vaddw_u16(uint32x4_t __p0, uint16x4_t __p1) {
67097  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67098  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67099  uint32x4_t __ret;
67100  __ret = __rev0 + __noswap_vmovl_u16(__rev1);
67101  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67102  return __ret;
67103}
67104#endif
67105
67106#ifdef __LITTLE_ENDIAN__
67107__ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1) {
67108  int16x8_t __ret;
67109  __ret = __p0 + vmovl_s8(__p1);
67110  return __ret;
67111}
67112#else
67113__ai int16x8_t vaddw_s8(int16x8_t __p0, int8x8_t __p1) {
67114  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67115  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67116  int16x8_t __ret;
67117  __ret = __rev0 + __noswap_vmovl_s8(__rev1);
67118  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67119  return __ret;
67120}
67121#endif
67122
67123#ifdef __LITTLE_ENDIAN__
67124__ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1) {
67125  int64x2_t __ret;
67126  __ret = __p0 + vmovl_s32(__p1);
67127  return __ret;
67128}
67129#else
67130__ai int64x2_t vaddw_s32(int64x2_t __p0, int32x2_t __p1) {
67131  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67132  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67133  int64x2_t __ret;
67134  __ret = __rev0 + __noswap_vmovl_s32(__rev1);
67135  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67136  return __ret;
67137}
67138#endif
67139
67140#ifdef __LITTLE_ENDIAN__
67141__ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1) {
67142  int32x4_t __ret;
67143  __ret = __p0 + vmovl_s16(__p1);
67144  return __ret;
67145}
67146#else
67147__ai int32x4_t vaddw_s16(int32x4_t __p0, int16x4_t __p1) {
67148  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67149  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67150  int32x4_t __ret;
67151  __ret = __rev0 + __noswap_vmovl_s16(__rev1);
67152  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67153  return __ret;
67154}
67155#endif
67156
67157#ifdef __LITTLE_ENDIAN__
67158#define vget_lane_f16(__p0_242, __p1_242) __extension__ ({ \
67159  float16x4_t __s0_242 = __p0_242; \
67160  float16_t __ret_242; \
67161float16x4_t __reint_242 = __s0_242; \
67162int16_t __reint1_242 = vget_lane_s16(*(int16x4_t *) &__reint_242, __p1_242); \
67163  __ret_242 = *(float16_t *) &__reint1_242; \
67164  __ret_242; \
67165})
67166#else
67167#define vget_lane_f16(__p0_243, __p1_243) __extension__ ({ \
67168  float16x4_t __s0_243 = __p0_243; \
67169  float16x4_t __rev0_243;  __rev0_243 = __builtin_shufflevector(__s0_243, __s0_243, 3, 2, 1, 0); \
67170  float16_t __ret_243; \
67171float16x4_t __reint_243 = __rev0_243; \
67172int16_t __reint1_243 = __noswap_vget_lane_s16(*(int16x4_t *) &__reint_243, __p1_243); \
67173  __ret_243 = *(float16_t *) &__reint1_243; \
67174  __ret_243; \
67175})
67176#endif
67177
67178#ifdef __LITTLE_ENDIAN__
67179#define vgetq_lane_f16(__p0_244, __p1_244) __extension__ ({ \
67180  float16x8_t __s0_244 = __p0_244; \
67181  float16_t __ret_244; \
67182float16x8_t __reint_244 = __s0_244; \
67183int16_t __reint1_244 = vgetq_lane_s16(*(int16x8_t *) &__reint_244, __p1_244); \
67184  __ret_244 = *(float16_t *) &__reint1_244; \
67185  __ret_244; \
67186})
67187#else
67188#define vgetq_lane_f16(__p0_245, __p1_245) __extension__ ({ \
67189  float16x8_t __s0_245 = __p0_245; \
67190  float16x8_t __rev0_245;  __rev0_245 = __builtin_shufflevector(__s0_245, __s0_245, 7, 6, 5, 4, 3, 2, 1, 0); \
67191  float16_t __ret_245; \
67192float16x8_t __reint_245 = __rev0_245; \
67193int16_t __reint1_245 = __noswap_vgetq_lane_s16(*(int16x8_t *) &__reint_245, __p1_245); \
67194  __ret_245 = *(float16_t *) &__reint1_245; \
67195  __ret_245; \
67196})
67197#endif
67198
67199#ifdef __LITTLE_ENDIAN__
67200__ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67201  uint16x8_t __ret;
67202  __ret = __p0 + vmull_u8(__p1, __p2);
67203  return __ret;
67204}
67205#else
67206__ai uint16x8_t vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67207  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67208  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67209  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67210  uint16x8_t __ret;
67211  __ret = __rev0 + __noswap_vmull_u8(__rev1, __rev2);
67212  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67213  return __ret;
67214}
67215__ai uint16x8_t __noswap_vmlal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67216  uint16x8_t __ret;
67217  __ret = __p0 + __noswap_vmull_u8(__p1, __p2);
67218  return __ret;
67219}
67220#endif
67221
67222#ifdef __LITTLE_ENDIAN__
67223__ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67224  uint64x2_t __ret;
67225  __ret = __p0 + vmull_u32(__p1, __p2);
67226  return __ret;
67227}
67228#else
67229__ai uint64x2_t vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67230  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67231  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67232  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67233  uint64x2_t __ret;
67234  __ret = __rev0 + __noswap_vmull_u32(__rev1, __rev2);
67235  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67236  return __ret;
67237}
67238__ai uint64x2_t __noswap_vmlal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67239  uint64x2_t __ret;
67240  __ret = __p0 + __noswap_vmull_u32(__p1, __p2);
67241  return __ret;
67242}
67243#endif
67244
67245#ifdef __LITTLE_ENDIAN__
67246__ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67247  uint32x4_t __ret;
67248  __ret = __p0 + vmull_u16(__p1, __p2);
67249  return __ret;
67250}
67251#else
67252__ai uint32x4_t vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67253  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67254  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67255  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67256  uint32x4_t __ret;
67257  __ret = __rev0 + __noswap_vmull_u16(__rev1, __rev2);
67258  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67259  return __ret;
67260}
67261__ai uint32x4_t __noswap_vmlal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67262  uint32x4_t __ret;
67263  __ret = __p0 + __noswap_vmull_u16(__p1, __p2);
67264  return __ret;
67265}
67266#endif
67267
67268#ifdef __LITTLE_ENDIAN__
67269__ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67270  int16x8_t __ret;
67271  __ret = __p0 + vmull_s8(__p1, __p2);
67272  return __ret;
67273}
67274#else
67275__ai int16x8_t vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67276  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67277  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67278  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67279  int16x8_t __ret;
67280  __ret = __rev0 + __noswap_vmull_s8(__rev1, __rev2);
67281  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67282  return __ret;
67283}
67284__ai int16x8_t __noswap_vmlal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67285  int16x8_t __ret;
67286  __ret = __p0 + __noswap_vmull_s8(__p1, __p2);
67287  return __ret;
67288}
67289#endif
67290
67291#ifdef __LITTLE_ENDIAN__
67292__ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67293  int64x2_t __ret;
67294  __ret = __p0 + vmull_s32(__p1, __p2);
67295  return __ret;
67296}
67297#else
67298__ai int64x2_t vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67299  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67300  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67301  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67302  int64x2_t __ret;
67303  __ret = __rev0 + __noswap_vmull_s32(__rev1, __rev2);
67304  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67305  return __ret;
67306}
67307__ai int64x2_t __noswap_vmlal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67308  int64x2_t __ret;
67309  __ret = __p0 + __noswap_vmull_s32(__p1, __p2);
67310  return __ret;
67311}
67312#endif
67313
67314#ifdef __LITTLE_ENDIAN__
67315__ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67316  int32x4_t __ret;
67317  __ret = __p0 + vmull_s16(__p1, __p2);
67318  return __ret;
67319}
67320#else
67321__ai int32x4_t vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67322  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67323  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67324  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67325  int32x4_t __ret;
67326  __ret = __rev0 + __noswap_vmull_s16(__rev1, __rev2);
67327  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67328  return __ret;
67329}
67330__ai int32x4_t __noswap_vmlal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67331  int32x4_t __ret;
67332  __ret = __p0 + __noswap_vmull_s16(__p1, __p2);
67333  return __ret;
67334}
67335#endif
67336
67337#ifdef __LITTLE_ENDIAN__
67338#define vmlal_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
67339  uint64x2_t __s0 = __p0; \
67340  uint32x2_t __s1 = __p1; \
67341  uint32x2_t __s2 = __p2; \
67342  uint64x2_t __ret; \
67343  __ret = __s0 + vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
67344  __ret; \
67345})
67346#else
67347#define vmlal_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
67348  uint64x2_t __s0 = __p0; \
67349  uint32x2_t __s1 = __p1; \
67350  uint32x2_t __s2 = __p2; \
67351  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
67352  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
67353  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
67354  uint64x2_t __ret; \
67355  __ret = __rev0 + __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
67356  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
67357  __ret; \
67358})
67359#endif
67360
67361#ifdef __LITTLE_ENDIAN__
67362#define vmlal_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
67363  uint32x4_t __s0 = __p0; \
67364  uint16x4_t __s1 = __p1; \
67365  uint16x4_t __s2 = __p2; \
67366  uint32x4_t __ret; \
67367  __ret = __s0 + vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
67368  __ret; \
67369})
67370#else
67371#define vmlal_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
67372  uint32x4_t __s0 = __p0; \
67373  uint16x4_t __s1 = __p1; \
67374  uint16x4_t __s2 = __p2; \
67375  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
67376  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
67377  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
67378  uint32x4_t __ret; \
67379  __ret = __rev0 + __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
67380  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
67381  __ret; \
67382})
67383#endif
67384
67385#ifdef __LITTLE_ENDIAN__
67386#define vmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
67387  int64x2_t __s0 = __p0; \
67388  int32x2_t __s1 = __p1; \
67389  int32x2_t __s2 = __p2; \
67390  int64x2_t __ret; \
67391  __ret = __s0 + vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
67392  __ret; \
67393})
67394#else
67395#define vmlal_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
67396  int64x2_t __s0 = __p0; \
67397  int32x2_t __s1 = __p1; \
67398  int32x2_t __s2 = __p2; \
67399  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
67400  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
67401  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
67402  int64x2_t __ret; \
67403  __ret = __rev0 + __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
67404  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
67405  __ret; \
67406})
67407#endif
67408
67409#ifdef __LITTLE_ENDIAN__
67410#define vmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
67411  int32x4_t __s0 = __p0; \
67412  int16x4_t __s1 = __p1; \
67413  int16x4_t __s2 = __p2; \
67414  int32x4_t __ret; \
67415  __ret = __s0 + vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
67416  __ret; \
67417})
67418#else
67419#define vmlal_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
67420  int32x4_t __s0 = __p0; \
67421  int16x4_t __s1 = __p1; \
67422  int16x4_t __s2 = __p2; \
67423  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
67424  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
67425  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
67426  int32x4_t __ret; \
67427  __ret = __rev0 + __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
67428  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
67429  __ret; \
67430})
67431#endif
67432
67433#ifdef __LITTLE_ENDIAN__
67434__ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67435  uint64x2_t __ret;
67436  __ret = __p0 + vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67437  return __ret;
67438}
67439#else
67440__ai uint64x2_t vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67441  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67442  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67443  uint64x2_t __ret;
67444  __ret = __rev0 + __noswap_vmull_u32(__rev1, (uint32x2_t) {__p2, __p2});
67445  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67446  return __ret;
67447}
67448__ai uint64x2_t __noswap_vmlal_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67449  uint64x2_t __ret;
67450  __ret = __p0 + __noswap_vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67451  return __ret;
67452}
67453#endif
67454
67455#ifdef __LITTLE_ENDIAN__
67456__ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67457  uint32x4_t __ret;
67458  __ret = __p0 + vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67459  return __ret;
67460}
67461#else
67462__ai uint32x4_t vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67463  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67464  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67465  uint32x4_t __ret;
67466  __ret = __rev0 + __noswap_vmull_u16(__rev1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67467  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67468  return __ret;
67469}
67470__ai uint32x4_t __noswap_vmlal_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67471  uint32x4_t __ret;
67472  __ret = __p0 + __noswap_vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67473  return __ret;
67474}
67475#endif
67476
67477#ifdef __LITTLE_ENDIAN__
67478__ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67479  int64x2_t __ret;
67480  __ret = __p0 + vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67481  return __ret;
67482}
67483#else
67484__ai int64x2_t vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67485  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67486  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67487  int64x2_t __ret;
67488  __ret = __rev0 + __noswap_vmull_s32(__rev1, (int32x2_t) {__p2, __p2});
67489  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67490  return __ret;
67491}
67492__ai int64x2_t __noswap_vmlal_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67493  int64x2_t __ret;
67494  __ret = __p0 + __noswap_vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67495  return __ret;
67496}
67497#endif
67498
67499#ifdef __LITTLE_ENDIAN__
67500__ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67501  int32x4_t __ret;
67502  __ret = __p0 + vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67503  return __ret;
67504}
67505#else
67506__ai int32x4_t vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67507  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67508  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67509  int32x4_t __ret;
67510  __ret = __rev0 + __noswap_vmull_s16(__rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
67511  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67512  return __ret;
67513}
67514__ai int32x4_t __noswap_vmlal_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67515  int32x4_t __ret;
67516  __ret = __p0 + __noswap_vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67517  return __ret;
67518}
67519#endif
67520
67521#ifdef __LITTLE_ENDIAN__
67522__ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67523  uint16x8_t __ret;
67524  __ret = __p0 - vmull_u8(__p1, __p2);
67525  return __ret;
67526}
67527#else
67528__ai uint16x8_t vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67529  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67530  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67531  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67532  uint16x8_t __ret;
67533  __ret = __rev0 - __noswap_vmull_u8(__rev1, __rev2);
67534  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67535  return __ret;
67536}
67537__ai uint16x8_t __noswap_vmlsl_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
67538  uint16x8_t __ret;
67539  __ret = __p0 - __noswap_vmull_u8(__p1, __p2);
67540  return __ret;
67541}
67542#endif
67543
67544#ifdef __LITTLE_ENDIAN__
67545__ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67546  uint64x2_t __ret;
67547  __ret = __p0 - vmull_u32(__p1, __p2);
67548  return __ret;
67549}
67550#else
67551__ai uint64x2_t vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67552  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67553  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67554  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67555  uint64x2_t __ret;
67556  __ret = __rev0 - __noswap_vmull_u32(__rev1, __rev2);
67557  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67558  return __ret;
67559}
67560__ai uint64x2_t __noswap_vmlsl_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
67561  uint64x2_t __ret;
67562  __ret = __p0 - __noswap_vmull_u32(__p1, __p2);
67563  return __ret;
67564}
67565#endif
67566
67567#ifdef __LITTLE_ENDIAN__
67568__ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67569  uint32x4_t __ret;
67570  __ret = __p0 - vmull_u16(__p1, __p2);
67571  return __ret;
67572}
67573#else
67574__ai uint32x4_t vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67575  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67576  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67577  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67578  uint32x4_t __ret;
67579  __ret = __rev0 - __noswap_vmull_u16(__rev1, __rev2);
67580  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67581  return __ret;
67582}
67583__ai uint32x4_t __noswap_vmlsl_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
67584  uint32x4_t __ret;
67585  __ret = __p0 - __noswap_vmull_u16(__p1, __p2);
67586  return __ret;
67587}
67588#endif
67589
67590#ifdef __LITTLE_ENDIAN__
67591__ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67592  int16x8_t __ret;
67593  __ret = __p0 - vmull_s8(__p1, __p2);
67594  return __ret;
67595}
67596#else
67597__ai int16x8_t vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67598  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
67599  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
67600  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
67601  int16x8_t __ret;
67602  __ret = __rev0 - __noswap_vmull_s8(__rev1, __rev2);
67603  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
67604  return __ret;
67605}
67606__ai int16x8_t __noswap_vmlsl_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
67607  int16x8_t __ret;
67608  __ret = __p0 - __noswap_vmull_s8(__p1, __p2);
67609  return __ret;
67610}
67611#endif
67612
67613#ifdef __LITTLE_ENDIAN__
67614__ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67615  int64x2_t __ret;
67616  __ret = __p0 - vmull_s32(__p1, __p2);
67617  return __ret;
67618}
67619#else
67620__ai int64x2_t vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67621  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67622  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67623  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
67624  int64x2_t __ret;
67625  __ret = __rev0 - __noswap_vmull_s32(__rev1, __rev2);
67626  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67627  return __ret;
67628}
67629__ai int64x2_t __noswap_vmlsl_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
67630  int64x2_t __ret;
67631  __ret = __p0 - __noswap_vmull_s32(__p1, __p2);
67632  return __ret;
67633}
67634#endif
67635
67636#ifdef __LITTLE_ENDIAN__
67637__ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67638  int32x4_t __ret;
67639  __ret = __p0 - vmull_s16(__p1, __p2);
67640  return __ret;
67641}
67642#else
67643__ai int32x4_t vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67644  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67645  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67646  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
67647  int32x4_t __ret;
67648  __ret = __rev0 - __noswap_vmull_s16(__rev1, __rev2);
67649  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67650  return __ret;
67651}
67652__ai int32x4_t __noswap_vmlsl_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
67653  int32x4_t __ret;
67654  __ret = __p0 - __noswap_vmull_s16(__p1, __p2);
67655  return __ret;
67656}
67657#endif
67658
67659#ifdef __LITTLE_ENDIAN__
67660#define vmlsl_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
67661  uint64x2_t __s0 = __p0; \
67662  uint32x2_t __s1 = __p1; \
67663  uint32x2_t __s2 = __p2; \
67664  uint64x2_t __ret; \
67665  __ret = __s0 - vmull_u32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
67666  __ret; \
67667})
67668#else
67669#define vmlsl_lane_u32(__p0, __p1, __p2, __p3) __extension__ ({ \
67670  uint64x2_t __s0 = __p0; \
67671  uint32x2_t __s1 = __p1; \
67672  uint32x2_t __s2 = __p2; \
67673  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
67674  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
67675  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
67676  uint64x2_t __ret; \
67677  __ret = __rev0 - __noswap_vmull_u32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
67678  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
67679  __ret; \
67680})
67681#endif
67682
67683#ifdef __LITTLE_ENDIAN__
67684#define vmlsl_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
67685  uint32x4_t __s0 = __p0; \
67686  uint16x4_t __s1 = __p1; \
67687  uint16x4_t __s2 = __p2; \
67688  uint32x4_t __ret; \
67689  __ret = __s0 - vmull_u16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
67690  __ret; \
67691})
67692#else
67693#define vmlsl_lane_u16(__p0, __p1, __p2, __p3) __extension__ ({ \
67694  uint32x4_t __s0 = __p0; \
67695  uint16x4_t __s1 = __p1; \
67696  uint16x4_t __s2 = __p2; \
67697  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
67698  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
67699  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
67700  uint32x4_t __ret; \
67701  __ret = __rev0 - __noswap_vmull_u16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
67702  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
67703  __ret; \
67704})
67705#endif
67706
67707#ifdef __LITTLE_ENDIAN__
67708#define vmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
67709  int64x2_t __s0 = __p0; \
67710  int32x2_t __s1 = __p1; \
67711  int32x2_t __s2 = __p2; \
67712  int64x2_t __ret; \
67713  __ret = __s0 - vmull_s32(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3)); \
67714  __ret; \
67715})
67716#else
67717#define vmlsl_lane_s32(__p0, __p1, __p2, __p3) __extension__ ({ \
67718  int64x2_t __s0 = __p0; \
67719  int32x2_t __s1 = __p1; \
67720  int32x2_t __s2 = __p2; \
67721  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \
67722  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \
67723  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 1, 0); \
67724  int64x2_t __ret; \
67725  __ret = __rev0 - __noswap_vmull_s32(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3)); \
67726  __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \
67727  __ret; \
67728})
67729#endif
67730
67731#ifdef __LITTLE_ENDIAN__
67732#define vmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
67733  int32x4_t __s0 = __p0; \
67734  int16x4_t __s1 = __p1; \
67735  int16x4_t __s2 = __p2; \
67736  int32x4_t __ret; \
67737  __ret = __s0 - vmull_s16(__s1, __builtin_shufflevector(__s2, __s2, __p3, __p3, __p3, __p3)); \
67738  __ret; \
67739})
67740#else
67741#define vmlsl_lane_s16(__p0, __p1, __p2, __p3) __extension__ ({ \
67742  int32x4_t __s0 = __p0; \
67743  int16x4_t __s1 = __p1; \
67744  int16x4_t __s2 = __p2; \
67745  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \
67746  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \
67747  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \
67748  int32x4_t __ret; \
67749  __ret = __rev0 - __noswap_vmull_s16(__rev1, __builtin_shufflevector(__rev2, __rev2, __p3, __p3, __p3, __p3)); \
67750  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \
67751  __ret; \
67752})
67753#endif
67754
67755#ifdef __LITTLE_ENDIAN__
67756__ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67757  uint64x2_t __ret;
67758  __ret = __p0 - vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67759  return __ret;
67760}
67761#else
67762__ai uint64x2_t vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67763  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67764  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67765  uint64x2_t __ret;
67766  __ret = __rev0 - __noswap_vmull_u32(__rev1, (uint32x2_t) {__p2, __p2});
67767  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67768  return __ret;
67769}
67770__ai uint64x2_t __noswap_vmlsl_n_u32(uint64x2_t __p0, uint32x2_t __p1, uint32_t __p2) {
67771  uint64x2_t __ret;
67772  __ret = __p0 - __noswap_vmull_u32(__p1, (uint32x2_t) {__p2, __p2});
67773  return __ret;
67774}
67775#endif
67776
67777#ifdef __LITTLE_ENDIAN__
67778__ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67779  uint32x4_t __ret;
67780  __ret = __p0 - vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67781  return __ret;
67782}
67783#else
67784__ai uint32x4_t vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67785  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67786  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67787  uint32x4_t __ret;
67788  __ret = __rev0 - __noswap_vmull_u16(__rev1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67789  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67790  return __ret;
67791}
67792__ai uint32x4_t __noswap_vmlsl_n_u16(uint32x4_t __p0, uint16x4_t __p1, uint16_t __p2) {
67793  uint32x4_t __ret;
67794  __ret = __p0 - __noswap_vmull_u16(__p1, (uint16x4_t) {__p2, __p2, __p2, __p2});
67795  return __ret;
67796}
67797#endif
67798
67799#ifdef __LITTLE_ENDIAN__
67800__ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67801  int64x2_t __ret;
67802  __ret = __p0 - vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67803  return __ret;
67804}
67805#else
67806__ai int64x2_t vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67807  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
67808  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
67809  int64x2_t __ret;
67810  __ret = __rev0 - __noswap_vmull_s32(__rev1, (int32x2_t) {__p2, __p2});
67811  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
67812  return __ret;
67813}
67814__ai int64x2_t __noswap_vmlsl_n_s32(int64x2_t __p0, int32x2_t __p1, int32_t __p2) {
67815  int64x2_t __ret;
67816  __ret = __p0 - __noswap_vmull_s32(__p1, (int32x2_t) {__p2, __p2});
67817  return __ret;
67818}
67819#endif
67820
67821#ifdef __LITTLE_ENDIAN__
67822__ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67823  int32x4_t __ret;
67824  __ret = __p0 - vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67825  return __ret;
67826}
67827#else
67828__ai int32x4_t vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67829  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
67830  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
67831  int32x4_t __ret;
67832  __ret = __rev0 - __noswap_vmull_s16(__rev1, (int16x4_t) {__p2, __p2, __p2, __p2});
67833  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
67834  return __ret;
67835}
67836__ai int32x4_t __noswap_vmlsl_n_s16(int32x4_t __p0, int16x4_t __p1, int16_t __p2) {
67837  int32x4_t __ret;
67838  __ret = __p0 - __noswap_vmull_s16(__p1, (int16x4_t) {__p2, __p2, __p2, __p2});
67839  return __ret;
67840}
67841#endif
67842
67843#ifdef __LITTLE_ENDIAN__
67844#define vset_lane_f16(__p0_246, __p1_246, __p2_246) __extension__ ({ \
67845  float16_t __s0_246 = __p0_246; \
67846  float16x4_t __s1_246 = __p1_246; \
67847  float16x4_t __ret_246; \
67848float16_t __reint_246 = __s0_246; \
67849float16x4_t __reint1_246 = __s1_246; \
67850int16x4_t __reint2_246 = vset_lane_s16(*(int16_t *) &__reint_246, *(int16x4_t *) &__reint1_246, __p2_246); \
67851  __ret_246 = *(float16x4_t *) &__reint2_246; \
67852  __ret_246; \
67853})
67854#else
67855#define vset_lane_f16(__p0_247, __p1_247, __p2_247) __extension__ ({ \
67856  float16_t __s0_247 = __p0_247; \
67857  float16x4_t __s1_247 = __p1_247; \
67858  float16x4_t __rev1_247;  __rev1_247 = __builtin_shufflevector(__s1_247, __s1_247, 3, 2, 1, 0); \
67859  float16x4_t __ret_247; \
67860float16_t __reint_247 = __s0_247; \
67861float16x4_t __reint1_247 = __rev1_247; \
67862int16x4_t __reint2_247 = __noswap_vset_lane_s16(*(int16_t *) &__reint_247, *(int16x4_t *) &__reint1_247, __p2_247); \
67863  __ret_247 = *(float16x4_t *) &__reint2_247; \
67864  __ret_247 = __builtin_shufflevector(__ret_247, __ret_247, 3, 2, 1, 0); \
67865  __ret_247; \
67866})
67867#endif
67868
67869#ifdef __LITTLE_ENDIAN__
67870#define vsetq_lane_f16(__p0_248, __p1_248, __p2_248) __extension__ ({ \
67871  float16_t __s0_248 = __p0_248; \
67872  float16x8_t __s1_248 = __p1_248; \
67873  float16x8_t __ret_248; \
67874float16_t __reint_248 = __s0_248; \
67875float16x8_t __reint1_248 = __s1_248; \
67876int16x8_t __reint2_248 = vsetq_lane_s16(*(int16_t *) &__reint_248, *(int16x8_t *) &__reint1_248, __p2_248); \
67877  __ret_248 = *(float16x8_t *) &__reint2_248; \
67878  __ret_248; \
67879})
67880#else
67881#define vsetq_lane_f16(__p0_249, __p1_249, __p2_249) __extension__ ({ \
67882  float16_t __s0_249 = __p0_249; \
67883  float16x8_t __s1_249 = __p1_249; \
67884  float16x8_t __rev1_249;  __rev1_249 = __builtin_shufflevector(__s1_249, __s1_249, 7, 6, 5, 4, 3, 2, 1, 0); \
67885  float16x8_t __ret_249; \
67886float16_t __reint_249 = __s0_249; \
67887float16x8_t __reint1_249 = __rev1_249; \
67888int16x8_t __reint2_249 = __noswap_vsetq_lane_s16(*(int16_t *) &__reint_249, *(int16x8_t *) &__reint1_249, __p2_249); \
67889  __ret_249 = *(float16x8_t *) &__reint2_249; \
67890  __ret_249 = __builtin_shufflevector(__ret_249, __ret_249, 7, 6, 5, 4, 3, 2, 1, 0); \
67891  __ret_249; \
67892})
67893#endif
67894
67895#if defined(__ARM_FEATURE_QRDMX) && defined(__aarch64__)
67896#ifdef __LITTLE_ENDIAN__
67897__ai int32_t vqrdmlahs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
67898  int32_t __ret;
67899  __ret = vqadds_s32(__p0, vqrdmulhs_s32(__p1, __p2));
67900  return __ret;
67901}
67902#else
67903__ai int32_t vqrdmlahs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
67904  int32_t __ret;
67905  __ret = __noswap_vqadds_s32(__p0, __noswap_vqrdmulhs_s32(__p1, __p2));
67906  return __ret;
67907}
67908#endif
67909
67910#ifdef __LITTLE_ENDIAN__
67911__ai int16_t vqrdmlahh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
67912  int16_t __ret;
67913  __ret = vqaddh_s16(__p0, vqrdmulhh_s16(__p1, __p2));
67914  return __ret;
67915}
67916#else
67917__ai int16_t vqrdmlahh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
67918  int16_t __ret;
67919  __ret = __noswap_vqaddh_s16(__p0, __noswap_vqrdmulhh_s16(__p1, __p2));
67920  return __ret;
67921}
67922#endif
67923
67924#ifdef __LITTLE_ENDIAN__
67925#define vqrdmlahs_lane_s32(__p0_250, __p1_250, __p2_250, __p3_250) __extension__ ({ \
67926  int32_t __s0_250 = __p0_250; \
67927  int32_t __s1_250 = __p1_250; \
67928  int32x2_t __s2_250 = __p2_250; \
67929  int32_t __ret_250; \
67930  __ret_250 = vqadds_s32(__s0_250, vqrdmulhs_s32(__s1_250, vget_lane_s32(__s2_250, __p3_250))); \
67931  __ret_250; \
67932})
67933#else
67934#define vqrdmlahs_lane_s32(__p0_251, __p1_251, __p2_251, __p3_251) __extension__ ({ \
67935  int32_t __s0_251 = __p0_251; \
67936  int32_t __s1_251 = __p1_251; \
67937  int32x2_t __s2_251 = __p2_251; \
67938  int32x2_t __rev2_251;  __rev2_251 = __builtin_shufflevector(__s2_251, __s2_251, 1, 0); \
67939  int32_t __ret_251; \
67940  __ret_251 = __noswap_vqadds_s32(__s0_251, __noswap_vqrdmulhs_s32(__s1_251, __noswap_vget_lane_s32(__rev2_251, __p3_251))); \
67941  __ret_251; \
67942})
67943#endif
67944
67945#ifdef __LITTLE_ENDIAN__
67946#define vqrdmlahh_lane_s16(__p0_252, __p1_252, __p2_252, __p3_252) __extension__ ({ \
67947  int16_t __s0_252 = __p0_252; \
67948  int16_t __s1_252 = __p1_252; \
67949  int16x4_t __s2_252 = __p2_252; \
67950  int16_t __ret_252; \
67951  __ret_252 = vqaddh_s16(__s0_252, vqrdmulhh_s16(__s1_252, vget_lane_s16(__s2_252, __p3_252))); \
67952  __ret_252; \
67953})
67954#else
67955#define vqrdmlahh_lane_s16(__p0_253, __p1_253, __p2_253, __p3_253) __extension__ ({ \
67956  int16_t __s0_253 = __p0_253; \
67957  int16_t __s1_253 = __p1_253; \
67958  int16x4_t __s2_253 = __p2_253; \
67959  int16x4_t __rev2_253;  __rev2_253 = __builtin_shufflevector(__s2_253, __s2_253, 3, 2, 1, 0); \
67960  int16_t __ret_253; \
67961  __ret_253 = __noswap_vqaddh_s16(__s0_253, __noswap_vqrdmulhh_s16(__s1_253, __noswap_vget_lane_s16(__rev2_253, __p3_253))); \
67962  __ret_253; \
67963})
67964#endif
67965
67966#ifdef __LITTLE_ENDIAN__
67967#define vqrdmlahs_laneq_s32(__p0_254, __p1_254, __p2_254, __p3_254) __extension__ ({ \
67968  int32_t __s0_254 = __p0_254; \
67969  int32_t __s1_254 = __p1_254; \
67970  int32x4_t __s2_254 = __p2_254; \
67971  int32_t __ret_254; \
67972  __ret_254 = vqadds_s32(__s0_254, vqrdmulhs_s32(__s1_254, vgetq_lane_s32(__s2_254, __p3_254))); \
67973  __ret_254; \
67974})
67975#else
67976#define vqrdmlahs_laneq_s32(__p0_255, __p1_255, __p2_255, __p3_255) __extension__ ({ \
67977  int32_t __s0_255 = __p0_255; \
67978  int32_t __s1_255 = __p1_255; \
67979  int32x4_t __s2_255 = __p2_255; \
67980  int32x4_t __rev2_255;  __rev2_255 = __builtin_shufflevector(__s2_255, __s2_255, 3, 2, 1, 0); \
67981  int32_t __ret_255; \
67982  __ret_255 = __noswap_vqadds_s32(__s0_255, __noswap_vqrdmulhs_s32(__s1_255, __noswap_vgetq_lane_s32(__rev2_255, __p3_255))); \
67983  __ret_255; \
67984})
67985#endif
67986
67987#ifdef __LITTLE_ENDIAN__
67988#define vqrdmlahh_laneq_s16(__p0_256, __p1_256, __p2_256, __p3_256) __extension__ ({ \
67989  int16_t __s0_256 = __p0_256; \
67990  int16_t __s1_256 = __p1_256; \
67991  int16x8_t __s2_256 = __p2_256; \
67992  int16_t __ret_256; \
67993  __ret_256 = vqaddh_s16(__s0_256, vqrdmulhh_s16(__s1_256, vgetq_lane_s16(__s2_256, __p3_256))); \
67994  __ret_256; \
67995})
67996#else
67997#define vqrdmlahh_laneq_s16(__p0_257, __p1_257, __p2_257, __p3_257) __extension__ ({ \
67998  int16_t __s0_257 = __p0_257; \
67999  int16_t __s1_257 = __p1_257; \
68000  int16x8_t __s2_257 = __p2_257; \
68001  int16x8_t __rev2_257;  __rev2_257 = __builtin_shufflevector(__s2_257, __s2_257, 7, 6, 5, 4, 3, 2, 1, 0); \
68002  int16_t __ret_257; \
68003  __ret_257 = __noswap_vqaddh_s16(__s0_257, __noswap_vqrdmulhh_s16(__s1_257, __noswap_vgetq_lane_s16(__rev2_257, __p3_257))); \
68004  __ret_257; \
68005})
68006#endif
68007
68008#ifdef __LITTLE_ENDIAN__
68009__ai int32_t vqrdmlshs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
68010  int32_t __ret;
68011  __ret = vqsubs_s32(__p0, vqrdmulhs_s32(__p1, __p2));
68012  return __ret;
68013}
68014#else
68015__ai int32_t vqrdmlshs_s32(int32_t __p0, int32_t __p1, int32_t __p2) {
68016  int32_t __ret;
68017  __ret = __noswap_vqsubs_s32(__p0, __noswap_vqrdmulhs_s32(__p1, __p2));
68018  return __ret;
68019}
68020#endif
68021
68022#ifdef __LITTLE_ENDIAN__
68023__ai int16_t vqrdmlshh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
68024  int16_t __ret;
68025  __ret = vqsubh_s16(__p0, vqrdmulhh_s16(__p1, __p2));
68026  return __ret;
68027}
68028#else
68029__ai int16_t vqrdmlshh_s16(int16_t __p0, int16_t __p1, int16_t __p2) {
68030  int16_t __ret;
68031  __ret = __noswap_vqsubh_s16(__p0, __noswap_vqrdmulhh_s16(__p1, __p2));
68032  return __ret;
68033}
68034#endif
68035
68036#ifdef __LITTLE_ENDIAN__
68037#define vqrdmlshs_lane_s32(__p0_258, __p1_258, __p2_258, __p3_258) __extension__ ({ \
68038  int32_t __s0_258 = __p0_258; \
68039  int32_t __s1_258 = __p1_258; \
68040  int32x2_t __s2_258 = __p2_258; \
68041  int32_t __ret_258; \
68042  __ret_258 = vqsubs_s32(__s0_258, vqrdmulhs_s32(__s1_258, vget_lane_s32(__s2_258, __p3_258))); \
68043  __ret_258; \
68044})
68045#else
68046#define vqrdmlshs_lane_s32(__p0_259, __p1_259, __p2_259, __p3_259) __extension__ ({ \
68047  int32_t __s0_259 = __p0_259; \
68048  int32_t __s1_259 = __p1_259; \
68049  int32x2_t __s2_259 = __p2_259; \
68050  int32x2_t __rev2_259;  __rev2_259 = __builtin_shufflevector(__s2_259, __s2_259, 1, 0); \
68051  int32_t __ret_259; \
68052  __ret_259 = __noswap_vqsubs_s32(__s0_259, __noswap_vqrdmulhs_s32(__s1_259, __noswap_vget_lane_s32(__rev2_259, __p3_259))); \
68053  __ret_259; \
68054})
68055#endif
68056
68057#ifdef __LITTLE_ENDIAN__
68058#define vqrdmlshh_lane_s16(__p0_260, __p1_260, __p2_260, __p3_260) __extension__ ({ \
68059  int16_t __s0_260 = __p0_260; \
68060  int16_t __s1_260 = __p1_260; \
68061  int16x4_t __s2_260 = __p2_260; \
68062  int16_t __ret_260; \
68063  __ret_260 = vqsubh_s16(__s0_260, vqrdmulhh_s16(__s1_260, vget_lane_s16(__s2_260, __p3_260))); \
68064  __ret_260; \
68065})
68066#else
68067#define vqrdmlshh_lane_s16(__p0_261, __p1_261, __p2_261, __p3_261) __extension__ ({ \
68068  int16_t __s0_261 = __p0_261; \
68069  int16_t __s1_261 = __p1_261; \
68070  int16x4_t __s2_261 = __p2_261; \
68071  int16x4_t __rev2_261;  __rev2_261 = __builtin_shufflevector(__s2_261, __s2_261, 3, 2, 1, 0); \
68072  int16_t __ret_261; \
68073  __ret_261 = __noswap_vqsubh_s16(__s0_261, __noswap_vqrdmulhh_s16(__s1_261, __noswap_vget_lane_s16(__rev2_261, __p3_261))); \
68074  __ret_261; \
68075})
68076#endif
68077
68078#ifdef __LITTLE_ENDIAN__
68079#define vqrdmlshs_laneq_s32(__p0_262, __p1_262, __p2_262, __p3_262) __extension__ ({ \
68080  int32_t __s0_262 = __p0_262; \
68081  int32_t __s1_262 = __p1_262; \
68082  int32x4_t __s2_262 = __p2_262; \
68083  int32_t __ret_262; \
68084  __ret_262 = vqsubs_s32(__s0_262, vqrdmulhs_s32(__s1_262, vgetq_lane_s32(__s2_262, __p3_262))); \
68085  __ret_262; \
68086})
68087#else
68088#define vqrdmlshs_laneq_s32(__p0_263, __p1_263, __p2_263, __p3_263) __extension__ ({ \
68089  int32_t __s0_263 = __p0_263; \
68090  int32_t __s1_263 = __p1_263; \
68091  int32x4_t __s2_263 = __p2_263; \
68092  int32x4_t __rev2_263;  __rev2_263 = __builtin_shufflevector(__s2_263, __s2_263, 3, 2, 1, 0); \
68093  int32_t __ret_263; \
68094  __ret_263 = __noswap_vqsubs_s32(__s0_263, __noswap_vqrdmulhs_s32(__s1_263, __noswap_vgetq_lane_s32(__rev2_263, __p3_263))); \
68095  __ret_263; \
68096})
68097#endif
68098
68099#ifdef __LITTLE_ENDIAN__
68100#define vqrdmlshh_laneq_s16(__p0_264, __p1_264, __p2_264, __p3_264) __extension__ ({ \
68101  int16_t __s0_264 = __p0_264; \
68102  int16_t __s1_264 = __p1_264; \
68103  int16x8_t __s2_264 = __p2_264; \
68104  int16_t __ret_264; \
68105  __ret_264 = vqsubh_s16(__s0_264, vqrdmulhh_s16(__s1_264, vgetq_lane_s16(__s2_264, __p3_264))); \
68106  __ret_264; \
68107})
68108#else
68109#define vqrdmlshh_laneq_s16(__p0_265, __p1_265, __p2_265, __p3_265) __extension__ ({ \
68110  int16_t __s0_265 = __p0_265; \
68111  int16_t __s1_265 = __p1_265; \
68112  int16x8_t __s2_265 = __p2_265; \
68113  int16x8_t __rev2_265;  __rev2_265 = __builtin_shufflevector(__s2_265, __s2_265, 7, 6, 5, 4, 3, 2, 1, 0); \
68114  int16_t __ret_265; \
68115  __ret_265 = __noswap_vqsubh_s16(__s0_265, __noswap_vqrdmulhh_s16(__s1_265, __noswap_vgetq_lane_s16(__rev2_265, __p3_265))); \
68116  __ret_265; \
68117})
68118#endif
68119
68120#endif
68121#if defined(__aarch64__)
68122#ifdef __LITTLE_ENDIAN__
68123__ai uint16x8_t vabdl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68124  uint16x8_t __ret;
68125  __ret = vabdl_u8(vget_high_u8(__p0), vget_high_u8(__p1));
68126  return __ret;
68127}
68128#else
68129__ai uint16x8_t vabdl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68130  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68131  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68132  uint16x8_t __ret;
68133  __ret = __noswap_vabdl_u8(__noswap_vget_high_u8(__rev0), __noswap_vget_high_u8(__rev1));
68134  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68135  return __ret;
68136}
68137#endif
68138
68139#ifdef __LITTLE_ENDIAN__
68140__ai uint64x2_t vabdl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68141  uint64x2_t __ret;
68142  __ret = vabdl_u32(vget_high_u32(__p0), vget_high_u32(__p1));
68143  return __ret;
68144}
68145#else
68146__ai uint64x2_t vabdl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68147  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68148  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68149  uint64x2_t __ret;
68150  __ret = __noswap_vabdl_u32(__noswap_vget_high_u32(__rev0), __noswap_vget_high_u32(__rev1));
68151  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68152  return __ret;
68153}
68154#endif
68155
68156#ifdef __LITTLE_ENDIAN__
68157__ai uint32x4_t vabdl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68158  uint32x4_t __ret;
68159  __ret = vabdl_u16(vget_high_u16(__p0), vget_high_u16(__p1));
68160  return __ret;
68161}
68162#else
68163__ai uint32x4_t vabdl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68164  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68165  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68166  uint32x4_t __ret;
68167  __ret = __noswap_vabdl_u16(__noswap_vget_high_u16(__rev0), __noswap_vget_high_u16(__rev1));
68168  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68169  return __ret;
68170}
68171#endif
68172
68173#ifdef __LITTLE_ENDIAN__
68174__ai int16x8_t vabdl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68175  int16x8_t __ret;
68176  __ret = vabdl_s8(vget_high_s8(__p0), vget_high_s8(__p1));
68177  return __ret;
68178}
68179#else
68180__ai int16x8_t vabdl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68181  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68182  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68183  int16x8_t __ret;
68184  __ret = __noswap_vabdl_s8(__noswap_vget_high_s8(__rev0), __noswap_vget_high_s8(__rev1));
68185  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68186  return __ret;
68187}
68188#endif
68189
68190#ifdef __LITTLE_ENDIAN__
68191__ai int64x2_t vabdl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68192  int64x2_t __ret;
68193  __ret = vabdl_s32(vget_high_s32(__p0), vget_high_s32(__p1));
68194  return __ret;
68195}
68196#else
68197__ai int64x2_t vabdl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68198  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68199  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68200  int64x2_t __ret;
68201  __ret = __noswap_vabdl_s32(__noswap_vget_high_s32(__rev0), __noswap_vget_high_s32(__rev1));
68202  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68203  return __ret;
68204}
68205#endif
68206
68207#ifdef __LITTLE_ENDIAN__
68208__ai int32x4_t vabdl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68209  int32x4_t __ret;
68210  __ret = vabdl_s16(vget_high_s16(__p0), vget_high_s16(__p1));
68211  return __ret;
68212}
68213#else
68214__ai int32x4_t vabdl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68215  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68216  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68217  int32x4_t __ret;
68218  __ret = __noswap_vabdl_s16(__noswap_vget_high_s16(__rev0), __noswap_vget_high_s16(__rev1));
68219  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68220  return __ret;
68221}
68222#endif
68223
68224#ifdef __LITTLE_ENDIAN__
68225__ai uint16x8_t vaddl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68226  uint16x8_t __ret;
68227  __ret = vmovl_high_u8(__p0) + vmovl_high_u8(__p1);
68228  return __ret;
68229}
68230#else
68231__ai uint16x8_t vaddl_high_u8(uint8x16_t __p0, uint8x16_t __p1) {
68232  uint8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68233  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68234  uint16x8_t __ret;
68235  __ret = __noswap_vmovl_high_u8(__rev0) + __noswap_vmovl_high_u8(__rev1);
68236  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68237  return __ret;
68238}
68239#endif
68240
68241#ifdef __LITTLE_ENDIAN__
68242__ai uint64x2_t vaddl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68243  uint64x2_t __ret;
68244  __ret = vmovl_high_u32(__p0) + vmovl_high_u32(__p1);
68245  return __ret;
68246}
68247#else
68248__ai uint64x2_t vaddl_high_u32(uint32x4_t __p0, uint32x4_t __p1) {
68249  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68250  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68251  uint64x2_t __ret;
68252  __ret = __noswap_vmovl_high_u32(__rev0) + __noswap_vmovl_high_u32(__rev1);
68253  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68254  return __ret;
68255}
68256#endif
68257
68258#ifdef __LITTLE_ENDIAN__
68259__ai uint32x4_t vaddl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68260  uint32x4_t __ret;
68261  __ret = vmovl_high_u16(__p0) + vmovl_high_u16(__p1);
68262  return __ret;
68263}
68264#else
68265__ai uint32x4_t vaddl_high_u16(uint16x8_t __p0, uint16x8_t __p1) {
68266  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68267  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68268  uint32x4_t __ret;
68269  __ret = __noswap_vmovl_high_u16(__rev0) + __noswap_vmovl_high_u16(__rev1);
68270  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68271  return __ret;
68272}
68273#endif
68274
68275#ifdef __LITTLE_ENDIAN__
68276__ai int16x8_t vaddl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68277  int16x8_t __ret;
68278  __ret = vmovl_high_s8(__p0) + vmovl_high_s8(__p1);
68279  return __ret;
68280}
68281#else
68282__ai int16x8_t vaddl_high_s8(int8x16_t __p0, int8x16_t __p1) {
68283  int8x16_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68284  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68285  int16x8_t __ret;
68286  __ret = __noswap_vmovl_high_s8(__rev0) + __noswap_vmovl_high_s8(__rev1);
68287  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68288  return __ret;
68289}
68290#endif
68291
68292#ifdef __LITTLE_ENDIAN__
68293__ai int64x2_t vaddl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68294  int64x2_t __ret;
68295  __ret = vmovl_high_s32(__p0) + vmovl_high_s32(__p1);
68296  return __ret;
68297}
68298#else
68299__ai int64x2_t vaddl_high_s32(int32x4_t __p0, int32x4_t __p1) {
68300  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68301  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68302  int64x2_t __ret;
68303  __ret = __noswap_vmovl_high_s32(__rev0) + __noswap_vmovl_high_s32(__rev1);
68304  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68305  return __ret;
68306}
68307#endif
68308
68309#ifdef __LITTLE_ENDIAN__
68310__ai int32x4_t vaddl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68311  int32x4_t __ret;
68312  __ret = vmovl_high_s16(__p0) + vmovl_high_s16(__p1);
68313  return __ret;
68314}
68315#else
68316__ai int32x4_t vaddl_high_s16(int16x8_t __p0, int16x8_t __p1) {
68317  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68318  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68319  int32x4_t __ret;
68320  __ret = __noswap_vmovl_high_s16(__rev0) + __noswap_vmovl_high_s16(__rev1);
68321  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68322  return __ret;
68323}
68324#endif
68325
68326#ifdef __LITTLE_ENDIAN__
68327__ai uint16x8_t vaddw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
68328  uint16x8_t __ret;
68329  __ret = __p0 + vmovl_high_u8(__p1);
68330  return __ret;
68331}
68332#else
68333__ai uint16x8_t vaddw_high_u8(uint16x8_t __p0, uint8x16_t __p1) {
68334  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68335  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68336  uint16x8_t __ret;
68337  __ret = __rev0 + __noswap_vmovl_high_u8(__rev1);
68338  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68339  return __ret;
68340}
68341#endif
68342
68343#ifdef __LITTLE_ENDIAN__
68344__ai uint64x2_t vaddw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
68345  uint64x2_t __ret;
68346  __ret = __p0 + vmovl_high_u32(__p1);
68347  return __ret;
68348}
68349#else
68350__ai uint64x2_t vaddw_high_u32(uint64x2_t __p0, uint32x4_t __p1) {
68351  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68352  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68353  uint64x2_t __ret;
68354  __ret = __rev0 + __noswap_vmovl_high_u32(__rev1);
68355  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68356  return __ret;
68357}
68358#endif
68359
68360#ifdef __LITTLE_ENDIAN__
68361__ai uint32x4_t vaddw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
68362  uint32x4_t __ret;
68363  __ret = __p0 + vmovl_high_u16(__p1);
68364  return __ret;
68365}
68366#else
68367__ai uint32x4_t vaddw_high_u16(uint32x4_t __p0, uint16x8_t __p1) {
68368  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68369  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68370  uint32x4_t __ret;
68371  __ret = __rev0 + __noswap_vmovl_high_u16(__rev1);
68372  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68373  return __ret;
68374}
68375#endif
68376
68377#ifdef __LITTLE_ENDIAN__
68378__ai int16x8_t vaddw_high_s8(int16x8_t __p0, int8x16_t __p1) {
68379  int16x8_t __ret;
68380  __ret = __p0 + vmovl_high_s8(__p1);
68381  return __ret;
68382}
68383#else
68384__ai int16x8_t vaddw_high_s8(int16x8_t __p0, int8x16_t __p1) {
68385  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68386  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68387  int16x8_t __ret;
68388  __ret = __rev0 + __noswap_vmovl_high_s8(__rev1);
68389  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68390  return __ret;
68391}
68392#endif
68393
68394#ifdef __LITTLE_ENDIAN__
68395__ai int64x2_t vaddw_high_s32(int64x2_t __p0, int32x4_t __p1) {
68396  int64x2_t __ret;
68397  __ret = __p0 + vmovl_high_s32(__p1);
68398  return __ret;
68399}
68400#else
68401__ai int64x2_t vaddw_high_s32(int64x2_t __p0, int32x4_t __p1) {
68402  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68403  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68404  int64x2_t __ret;
68405  __ret = __rev0 + __noswap_vmovl_high_s32(__rev1);
68406  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68407  return __ret;
68408}
68409#endif
68410
68411#ifdef __LITTLE_ENDIAN__
68412__ai int32x4_t vaddw_high_s16(int32x4_t __p0, int16x8_t __p1) {
68413  int32x4_t __ret;
68414  __ret = __p0 + vmovl_high_s16(__p1);
68415  return __ret;
68416}
68417#else
68418__ai int32x4_t vaddw_high_s16(int32x4_t __p0, int16x8_t __p1) {
68419  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68420  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68421  int32x4_t __ret;
68422  __ret = __rev0 + __noswap_vmovl_high_s16(__rev1);
68423  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68424  return __ret;
68425}
68426#endif
68427
68428#ifdef __LITTLE_ENDIAN__
68429#define vcopyq_lane_p64(__p0_266, __p1_266, __p2_266, __p3_266) __extension__ ({ \
68430  poly64x2_t __s0_266 = __p0_266; \
68431  poly64x1_t __s2_266 = __p2_266; \
68432  poly64x2_t __ret_266; \
68433  __ret_266 = vsetq_lane_p64(vget_lane_p64(__s2_266, __p3_266), __s0_266, __p1_266); \
68434  __ret_266; \
68435})
68436#else
68437#define vcopyq_lane_p64(__p0_267, __p1_267, __p2_267, __p3_267) __extension__ ({ \
68438  poly64x2_t __s0_267 = __p0_267; \
68439  poly64x1_t __s2_267 = __p2_267; \
68440  poly64x2_t __rev0_267;  __rev0_267 = __builtin_shufflevector(__s0_267, __s0_267, 1, 0); \
68441  poly64x2_t __ret_267; \
68442  __ret_267 = __noswap_vsetq_lane_p64(__noswap_vget_lane_p64(__s2_267, __p3_267), __rev0_267, __p1_267); \
68443  __ret_267 = __builtin_shufflevector(__ret_267, __ret_267, 1, 0); \
68444  __ret_267; \
68445})
68446#endif
68447
68448#ifdef __LITTLE_ENDIAN__
68449#define vcopyq_lane_f64(__p0_268, __p1_268, __p2_268, __p3_268) __extension__ ({ \
68450  float64x2_t __s0_268 = __p0_268; \
68451  float64x1_t __s2_268 = __p2_268; \
68452  float64x2_t __ret_268; \
68453  __ret_268 = vsetq_lane_f64(vget_lane_f64(__s2_268, __p3_268), __s0_268, __p1_268); \
68454  __ret_268; \
68455})
68456#else
68457#define vcopyq_lane_f64(__p0_269, __p1_269, __p2_269, __p3_269) __extension__ ({ \
68458  float64x2_t __s0_269 = __p0_269; \
68459  float64x1_t __s2_269 = __p2_269; \
68460  float64x2_t __rev0_269;  __rev0_269 = __builtin_shufflevector(__s0_269, __s0_269, 1, 0); \
68461  float64x2_t __ret_269; \
68462  __ret_269 = __noswap_vsetq_lane_f64(__noswap_vget_lane_f64(__s2_269, __p3_269), __rev0_269, __p1_269); \
68463  __ret_269 = __builtin_shufflevector(__ret_269, __ret_269, 1, 0); \
68464  __ret_269; \
68465})
68466#endif
68467
68468#ifdef __LITTLE_ENDIAN__
68469#define vcopy_lane_p64(__p0_270, __p1_270, __p2_270, __p3_270) __extension__ ({ \
68470  poly64x1_t __s0_270 = __p0_270; \
68471  poly64x1_t __s2_270 = __p2_270; \
68472  poly64x1_t __ret_270; \
68473  __ret_270 = vset_lane_p64(vget_lane_p64(__s2_270, __p3_270), __s0_270, __p1_270); \
68474  __ret_270; \
68475})
68476#else
68477#define vcopy_lane_p64(__p0_271, __p1_271, __p2_271, __p3_271) __extension__ ({ \
68478  poly64x1_t __s0_271 = __p0_271; \
68479  poly64x1_t __s2_271 = __p2_271; \
68480  poly64x1_t __ret_271; \
68481  __ret_271 = __noswap_vset_lane_p64(__noswap_vget_lane_p64(__s2_271, __p3_271), __s0_271, __p1_271); \
68482  __ret_271; \
68483})
68484#endif
68485
68486#ifdef __LITTLE_ENDIAN__
68487#define vcopy_lane_f64(__p0_272, __p1_272, __p2_272, __p3_272) __extension__ ({ \
68488  float64x1_t __s0_272 = __p0_272; \
68489  float64x1_t __s2_272 = __p2_272; \
68490  float64x1_t __ret_272; \
68491  __ret_272 = vset_lane_f64(vget_lane_f64(__s2_272, __p3_272), __s0_272, __p1_272); \
68492  __ret_272; \
68493})
68494#else
68495#define vcopy_lane_f64(__p0_273, __p1_273, __p2_273, __p3_273) __extension__ ({ \
68496  float64x1_t __s0_273 = __p0_273; \
68497  float64x1_t __s2_273 = __p2_273; \
68498  float64x1_t __ret_273; \
68499  __ret_273 = __noswap_vset_lane_f64(__noswap_vget_lane_f64(__s2_273, __p3_273), __s0_273, __p1_273); \
68500  __ret_273; \
68501})
68502#endif
68503
68504#ifdef __LITTLE_ENDIAN__
68505#define vcopyq_laneq_p64(__p0_274, __p1_274, __p2_274, __p3_274) __extension__ ({ \
68506  poly64x2_t __s0_274 = __p0_274; \
68507  poly64x2_t __s2_274 = __p2_274; \
68508  poly64x2_t __ret_274; \
68509  __ret_274 = vsetq_lane_p64(vgetq_lane_p64(__s2_274, __p3_274), __s0_274, __p1_274); \
68510  __ret_274; \
68511})
68512#else
68513#define vcopyq_laneq_p64(__p0_275, __p1_275, __p2_275, __p3_275) __extension__ ({ \
68514  poly64x2_t __s0_275 = __p0_275; \
68515  poly64x2_t __s2_275 = __p2_275; \
68516  poly64x2_t __rev0_275;  __rev0_275 = __builtin_shufflevector(__s0_275, __s0_275, 1, 0); \
68517  poly64x2_t __rev2_275;  __rev2_275 = __builtin_shufflevector(__s2_275, __s2_275, 1, 0); \
68518  poly64x2_t __ret_275; \
68519  __ret_275 = __noswap_vsetq_lane_p64(__noswap_vgetq_lane_p64(__rev2_275, __p3_275), __rev0_275, __p1_275); \
68520  __ret_275 = __builtin_shufflevector(__ret_275, __ret_275, 1, 0); \
68521  __ret_275; \
68522})
68523#endif
68524
68525#ifdef __LITTLE_ENDIAN__
68526#define vcopyq_laneq_f64(__p0_276, __p1_276, __p2_276, __p3_276) __extension__ ({ \
68527  float64x2_t __s0_276 = __p0_276; \
68528  float64x2_t __s2_276 = __p2_276; \
68529  float64x2_t __ret_276; \
68530  __ret_276 = vsetq_lane_f64(vgetq_lane_f64(__s2_276, __p3_276), __s0_276, __p1_276); \
68531  __ret_276; \
68532})
68533#else
68534#define vcopyq_laneq_f64(__p0_277, __p1_277, __p2_277, __p3_277) __extension__ ({ \
68535  float64x2_t __s0_277 = __p0_277; \
68536  float64x2_t __s2_277 = __p2_277; \
68537  float64x2_t __rev0_277;  __rev0_277 = __builtin_shufflevector(__s0_277, __s0_277, 1, 0); \
68538  float64x2_t __rev2_277;  __rev2_277 = __builtin_shufflevector(__s2_277, __s2_277, 1, 0); \
68539  float64x2_t __ret_277; \
68540  __ret_277 = __noswap_vsetq_lane_f64(__noswap_vgetq_lane_f64(__rev2_277, __p3_277), __rev0_277, __p1_277); \
68541  __ret_277 = __builtin_shufflevector(__ret_277, __ret_277, 1, 0); \
68542  __ret_277; \
68543})
68544#endif
68545
68546#ifdef __LITTLE_ENDIAN__
68547#define vcopy_laneq_p64(__p0_278, __p1_278, __p2_278, __p3_278) __extension__ ({ \
68548  poly64x1_t __s0_278 = __p0_278; \
68549  poly64x2_t __s2_278 = __p2_278; \
68550  poly64x1_t __ret_278; \
68551  __ret_278 = vset_lane_p64(vgetq_lane_p64(__s2_278, __p3_278), __s0_278, __p1_278); \
68552  __ret_278; \
68553})
68554#else
68555#define vcopy_laneq_p64(__p0_279, __p1_279, __p2_279, __p3_279) __extension__ ({ \
68556  poly64x1_t __s0_279 = __p0_279; \
68557  poly64x2_t __s2_279 = __p2_279; \
68558  poly64x2_t __rev2_279;  __rev2_279 = __builtin_shufflevector(__s2_279, __s2_279, 1, 0); \
68559  poly64x1_t __ret_279; \
68560  __ret_279 = __noswap_vset_lane_p64(__noswap_vgetq_lane_p64(__rev2_279, __p3_279), __s0_279, __p1_279); \
68561  __ret_279; \
68562})
68563#endif
68564
68565#ifdef __LITTLE_ENDIAN__
68566#define vcopy_laneq_f64(__p0_280, __p1_280, __p2_280, __p3_280) __extension__ ({ \
68567  float64x1_t __s0_280 = __p0_280; \
68568  float64x2_t __s2_280 = __p2_280; \
68569  float64x1_t __ret_280; \
68570  __ret_280 = vset_lane_f64(vgetq_lane_f64(__s2_280, __p3_280), __s0_280, __p1_280); \
68571  __ret_280; \
68572})
68573#else
68574#define vcopy_laneq_f64(__p0_281, __p1_281, __p2_281, __p3_281) __extension__ ({ \
68575  float64x1_t __s0_281 = __p0_281; \
68576  float64x2_t __s2_281 = __p2_281; \
68577  float64x2_t __rev2_281;  __rev2_281 = __builtin_shufflevector(__s2_281, __s2_281, 1, 0); \
68578  float64x1_t __ret_281; \
68579  __ret_281 = __noswap_vset_lane_f64(__noswap_vgetq_lane_f64(__rev2_281, __p3_281), __s0_281, __p1_281); \
68580  __ret_281; \
68581})
68582#endif
68583
68584#ifdef __LITTLE_ENDIAN__
68585__ai uint16x8_t vmlal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68586  uint16x8_t __ret;
68587  __ret = vmlal_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
68588  return __ret;
68589}
68590#else
68591__ai uint16x8_t vmlal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68592  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68593  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68594  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68595  uint16x8_t __ret;
68596  __ret = __noswap_vmlal_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
68597  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68598  return __ret;
68599}
68600#endif
68601
68602#ifdef __LITTLE_ENDIAN__
68603__ai uint64x2_t vmlal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68604  uint64x2_t __ret;
68605  __ret = vmlal_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
68606  return __ret;
68607}
68608#else
68609__ai uint64x2_t vmlal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68610  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68611  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68612  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68613  uint64x2_t __ret;
68614  __ret = __noswap_vmlal_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
68615  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68616  return __ret;
68617}
68618#endif
68619
68620#ifdef __LITTLE_ENDIAN__
68621__ai uint32x4_t vmlal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68622  uint32x4_t __ret;
68623  __ret = vmlal_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
68624  return __ret;
68625}
68626#else
68627__ai uint32x4_t vmlal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68628  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68629  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68630  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68631  uint32x4_t __ret;
68632  __ret = __noswap_vmlal_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
68633  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68634  return __ret;
68635}
68636#endif
68637
68638#ifdef __LITTLE_ENDIAN__
68639__ai int16x8_t vmlal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68640  int16x8_t __ret;
68641  __ret = vmlal_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
68642  return __ret;
68643}
68644#else
68645__ai int16x8_t vmlal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68646  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68647  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68648  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68649  int16x8_t __ret;
68650  __ret = __noswap_vmlal_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
68651  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68652  return __ret;
68653}
68654#endif
68655
68656#ifdef __LITTLE_ENDIAN__
68657__ai int64x2_t vmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68658  int64x2_t __ret;
68659  __ret = vmlal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
68660  return __ret;
68661}
68662#else
68663__ai int64x2_t vmlal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68664  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68665  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68666  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68667  int64x2_t __ret;
68668  __ret = __noswap_vmlal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
68669  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68670  return __ret;
68671}
68672#endif
68673
68674#ifdef __LITTLE_ENDIAN__
68675__ai int32x4_t vmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68676  int32x4_t __ret;
68677  __ret = vmlal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
68678  return __ret;
68679}
68680#else
68681__ai int32x4_t vmlal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68682  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68683  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68684  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68685  int32x4_t __ret;
68686  __ret = __noswap_vmlal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
68687  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68688  return __ret;
68689}
68690#endif
68691
68692#ifdef __LITTLE_ENDIAN__
68693__ai uint64x2_t vmlal_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68694  uint64x2_t __ret;
68695  __ret = vmlal_n_u32(__p0, vget_high_u32(__p1), __p2);
68696  return __ret;
68697}
68698#else
68699__ai uint64x2_t vmlal_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68700  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68701  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68702  uint64x2_t __ret;
68703  __ret = __noswap_vmlal_n_u32(__rev0, __noswap_vget_high_u32(__rev1), __p2);
68704  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68705  return __ret;
68706}
68707#endif
68708
68709#ifdef __LITTLE_ENDIAN__
68710__ai uint32x4_t vmlal_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68711  uint32x4_t __ret;
68712  __ret = vmlal_n_u16(__p0, vget_high_u16(__p1), __p2);
68713  return __ret;
68714}
68715#else
68716__ai uint32x4_t vmlal_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68717  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68718  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68719  uint32x4_t __ret;
68720  __ret = __noswap_vmlal_n_u16(__rev0, __noswap_vget_high_u16(__rev1), __p2);
68721  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68722  return __ret;
68723}
68724#endif
68725
68726#ifdef __LITTLE_ENDIAN__
68727__ai int64x2_t vmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68728  int64x2_t __ret;
68729  __ret = vmlal_n_s32(__p0, vget_high_s32(__p1), __p2);
68730  return __ret;
68731}
68732#else
68733__ai int64x2_t vmlal_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68734  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68735  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68736  int64x2_t __ret;
68737  __ret = __noswap_vmlal_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
68738  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68739  return __ret;
68740}
68741#endif
68742
68743#ifdef __LITTLE_ENDIAN__
68744__ai int32x4_t vmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68745  int32x4_t __ret;
68746  __ret = vmlal_n_s16(__p0, vget_high_s16(__p1), __p2);
68747  return __ret;
68748}
68749#else
68750__ai int32x4_t vmlal_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68751  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68752  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68753  int32x4_t __ret;
68754  __ret = __noswap_vmlal_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
68755  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68756  return __ret;
68757}
68758#endif
68759
68760#ifdef __LITTLE_ENDIAN__
68761__ai uint16x8_t vmlsl_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68762  uint16x8_t __ret;
68763  __ret = vmlsl_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
68764  return __ret;
68765}
68766#else
68767__ai uint16x8_t vmlsl_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
68768  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68769  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68770  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68771  uint16x8_t __ret;
68772  __ret = __noswap_vmlsl_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
68773  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68774  return __ret;
68775}
68776#endif
68777
68778#ifdef __LITTLE_ENDIAN__
68779__ai uint64x2_t vmlsl_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68780  uint64x2_t __ret;
68781  __ret = vmlsl_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
68782  return __ret;
68783}
68784#else
68785__ai uint64x2_t vmlsl_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
68786  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68787  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68788  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68789  uint64x2_t __ret;
68790  __ret = __noswap_vmlsl_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
68791  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68792  return __ret;
68793}
68794#endif
68795
68796#ifdef __LITTLE_ENDIAN__
68797__ai uint32x4_t vmlsl_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68798  uint32x4_t __ret;
68799  __ret = vmlsl_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
68800  return __ret;
68801}
68802#else
68803__ai uint32x4_t vmlsl_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
68804  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68805  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68806  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68807  uint32x4_t __ret;
68808  __ret = __noswap_vmlsl_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
68809  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68810  return __ret;
68811}
68812#endif
68813
68814#ifdef __LITTLE_ENDIAN__
68815__ai int16x8_t vmlsl_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68816  int16x8_t __ret;
68817  __ret = vmlsl_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
68818  return __ret;
68819}
68820#else
68821__ai int16x8_t vmlsl_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
68822  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68823  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68824  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
68825  int16x8_t __ret;
68826  __ret = __noswap_vmlsl_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
68827  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
68828  return __ret;
68829}
68830#endif
68831
68832#ifdef __LITTLE_ENDIAN__
68833__ai int64x2_t vmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68834  int64x2_t __ret;
68835  __ret = vmlsl_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
68836  return __ret;
68837}
68838#else
68839__ai int64x2_t vmlsl_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
68840  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68841  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68842  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
68843  int64x2_t __ret;
68844  __ret = __noswap_vmlsl_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
68845  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68846  return __ret;
68847}
68848#endif
68849
68850#ifdef __LITTLE_ENDIAN__
68851__ai int32x4_t vmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68852  int32x4_t __ret;
68853  __ret = vmlsl_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
68854  return __ret;
68855}
68856#else
68857__ai int32x4_t vmlsl_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
68858  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68859  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68860  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68861  int32x4_t __ret;
68862  __ret = __noswap_vmlsl_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
68863  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68864  return __ret;
68865}
68866#endif
68867
68868#ifdef __LITTLE_ENDIAN__
68869__ai uint64x2_t vmlsl_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68870  uint64x2_t __ret;
68871  __ret = vmlsl_n_u32(__p0, vget_high_u32(__p1), __p2);
68872  return __ret;
68873}
68874#else
68875__ai uint64x2_t vmlsl_high_n_u32(uint64x2_t __p0, uint32x4_t __p1, uint32_t __p2) {
68876  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68877  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68878  uint64x2_t __ret;
68879  __ret = __noswap_vmlsl_n_u32(__rev0, __noswap_vget_high_u32(__rev1), __p2);
68880  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68881  return __ret;
68882}
68883#endif
68884
68885#ifdef __LITTLE_ENDIAN__
68886__ai uint32x4_t vmlsl_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68887  uint32x4_t __ret;
68888  __ret = vmlsl_n_u16(__p0, vget_high_u16(__p1), __p2);
68889  return __ret;
68890}
68891#else
68892__ai uint32x4_t vmlsl_high_n_u16(uint32x4_t __p0, uint16x8_t __p1, uint16_t __p2) {
68893  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68894  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68895  uint32x4_t __ret;
68896  __ret = __noswap_vmlsl_n_u16(__rev0, __noswap_vget_high_u16(__rev1), __p2);
68897  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68898  return __ret;
68899}
68900#endif
68901
68902#ifdef __LITTLE_ENDIAN__
68903__ai int64x2_t vmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68904  int64x2_t __ret;
68905  __ret = vmlsl_n_s32(__p0, vget_high_s32(__p1), __p2);
68906  return __ret;
68907}
68908#else
68909__ai int64x2_t vmlsl_high_n_s32(int64x2_t __p0, int32x4_t __p1, int32_t __p2) {
68910  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
68911  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
68912  int64x2_t __ret;
68913  __ret = __noswap_vmlsl_n_s32(__rev0, __noswap_vget_high_s32(__rev1), __p2);
68914  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
68915  return __ret;
68916}
68917#endif
68918
68919#ifdef __LITTLE_ENDIAN__
68920__ai int32x4_t vmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68921  int32x4_t __ret;
68922  __ret = vmlsl_n_s16(__p0, vget_high_s16(__p1), __p2);
68923  return __ret;
68924}
68925#else
68926__ai int32x4_t vmlsl_high_n_s16(int32x4_t __p0, int16x8_t __p1, int16_t __p2) {
68927  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
68928  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68929  int32x4_t __ret;
68930  __ret = __noswap_vmlsl_n_s16(__rev0, __noswap_vget_high_s16(__rev1), __p2);
68931  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
68932  return __ret;
68933}
68934#endif
68935
68936#ifdef __LITTLE_ENDIAN__
68937#define vmulx_lane_f64(__p0_282, __p1_282, __p2_282) __extension__ ({ \
68938  float64x1_t __s0_282 = __p0_282; \
68939  float64x1_t __s1_282 = __p1_282; \
68940  float64x1_t __ret_282; \
68941  float64_t __x_282 = vget_lane_f64(__s0_282, 0); \
68942  float64_t __y_282 = vget_lane_f64(__s1_282, __p2_282); \
68943  float64_t __z_282 = vmulxd_f64(__x_282, __y_282); \
68944  __ret_282 = vset_lane_f64(__z_282, __s0_282, __p2_282); \
68945  __ret_282; \
68946})
68947#else
68948#define vmulx_lane_f64(__p0_283, __p1_283, __p2_283) __extension__ ({ \
68949  float64x1_t __s0_283 = __p0_283; \
68950  float64x1_t __s1_283 = __p1_283; \
68951  float64x1_t __ret_283; \
68952  float64_t __x_283 = __noswap_vget_lane_f64(__s0_283, 0); \
68953  float64_t __y_283 = __noswap_vget_lane_f64(__s1_283, __p2_283); \
68954  float64_t __z_283 = __noswap_vmulxd_f64(__x_283, __y_283); \
68955  __ret_283 = __noswap_vset_lane_f64(__z_283, __s0_283, __p2_283); \
68956  __ret_283; \
68957})
68958#endif
68959
68960#ifdef __LITTLE_ENDIAN__
68961#define vmulx_laneq_f64(__p0_284, __p1_284, __p2_284) __extension__ ({ \
68962  float64x1_t __s0_284 = __p0_284; \
68963  float64x2_t __s1_284 = __p1_284; \
68964  float64x1_t __ret_284; \
68965  float64_t __x_284 = vget_lane_f64(__s0_284, 0); \
68966  float64_t __y_284 = vgetq_lane_f64(__s1_284, __p2_284); \
68967  float64_t __z_284 = vmulxd_f64(__x_284, __y_284); \
68968  __ret_284 = vset_lane_f64(__z_284, __s0_284, 0); \
68969  __ret_284; \
68970})
68971#else
68972#define vmulx_laneq_f64(__p0_285, __p1_285, __p2_285) __extension__ ({ \
68973  float64x1_t __s0_285 = __p0_285; \
68974  float64x2_t __s1_285 = __p1_285; \
68975  float64x2_t __rev1_285;  __rev1_285 = __builtin_shufflevector(__s1_285, __s1_285, 1, 0); \
68976  float64x1_t __ret_285; \
68977  float64_t __x_285 = __noswap_vget_lane_f64(__s0_285, 0); \
68978  float64_t __y_285 = __noswap_vgetq_lane_f64(__rev1_285, __p2_285); \
68979  float64_t __z_285 = __noswap_vmulxd_f64(__x_285, __y_285); \
68980  __ret_285 = __noswap_vset_lane_f64(__z_285, __s0_285, 0); \
68981  __ret_285; \
68982})
68983#endif
68984
68985#endif
68986#ifdef __LITTLE_ENDIAN__
68987__ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
68988  uint16x8_t __ret;
68989  __ret = __p0 + vabdl_u8(__p1, __p2);
68990  return __ret;
68991}
68992#else
68993__ai uint16x8_t vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
68994  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
68995  uint8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
68996  uint8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
68997  uint16x8_t __ret;
68998  __ret = __rev0 + __noswap_vabdl_u8(__rev1, __rev2);
68999  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69000  return __ret;
69001}
69002__ai uint16x8_t __noswap_vabal_u8(uint16x8_t __p0, uint8x8_t __p1, uint8x8_t __p2) {
69003  uint16x8_t __ret;
69004  __ret = __p0 + __noswap_vabdl_u8(__p1, __p2);
69005  return __ret;
69006}
69007#endif
69008
69009#ifdef __LITTLE_ENDIAN__
69010__ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
69011  uint64x2_t __ret;
69012  __ret = __p0 + vabdl_u32(__p1, __p2);
69013  return __ret;
69014}
69015#else
69016__ai uint64x2_t vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
69017  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69018  uint32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
69019  uint32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
69020  uint64x2_t __ret;
69021  __ret = __rev0 + __noswap_vabdl_u32(__rev1, __rev2);
69022  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69023  return __ret;
69024}
69025__ai uint64x2_t __noswap_vabal_u32(uint64x2_t __p0, uint32x2_t __p1, uint32x2_t __p2) {
69026  uint64x2_t __ret;
69027  __ret = __p0 + __noswap_vabdl_u32(__p1, __p2);
69028  return __ret;
69029}
69030#endif
69031
69032#ifdef __LITTLE_ENDIAN__
69033__ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
69034  uint32x4_t __ret;
69035  __ret = __p0 + vabdl_u16(__p1, __p2);
69036  return __ret;
69037}
69038#else
69039__ai uint32x4_t vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
69040  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69041  uint16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69042  uint16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69043  uint32x4_t __ret;
69044  __ret = __rev0 + __noswap_vabdl_u16(__rev1, __rev2);
69045  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69046  return __ret;
69047}
69048__ai uint32x4_t __noswap_vabal_u16(uint32x4_t __p0, uint16x4_t __p1, uint16x4_t __p2) {
69049  uint32x4_t __ret;
69050  __ret = __p0 + __noswap_vabdl_u16(__p1, __p2);
69051  return __ret;
69052}
69053#endif
69054
69055#ifdef __LITTLE_ENDIAN__
69056__ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
69057  int16x8_t __ret;
69058  __ret = __p0 + vabdl_s8(__p1, __p2);
69059  return __ret;
69060}
69061#else
69062__ai int16x8_t vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
69063  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
69064  int8x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
69065  int8x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
69066  int16x8_t __ret;
69067  __ret = __rev0 + __noswap_vabdl_s8(__rev1, __rev2);
69068  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69069  return __ret;
69070}
69071__ai int16x8_t __noswap_vabal_s8(int16x8_t __p0, int8x8_t __p1, int8x8_t __p2) {
69072  int16x8_t __ret;
69073  __ret = __p0 + __noswap_vabdl_s8(__p1, __p2);
69074  return __ret;
69075}
69076#endif
69077
69078#ifdef __LITTLE_ENDIAN__
69079__ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
69080  int64x2_t __ret;
69081  __ret = __p0 + vabdl_s32(__p1, __p2);
69082  return __ret;
69083}
69084#else
69085__ai int64x2_t vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
69086  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69087  int32x2_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0);
69088  int32x2_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0);
69089  int64x2_t __ret;
69090  __ret = __rev0 + __noswap_vabdl_s32(__rev1, __rev2);
69091  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69092  return __ret;
69093}
69094__ai int64x2_t __noswap_vabal_s32(int64x2_t __p0, int32x2_t __p1, int32x2_t __p2) {
69095  int64x2_t __ret;
69096  __ret = __p0 + __noswap_vabdl_s32(__p1, __p2);
69097  return __ret;
69098}
69099#endif
69100
69101#ifdef __LITTLE_ENDIAN__
69102__ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
69103  int32x4_t __ret;
69104  __ret = __p0 + vabdl_s16(__p1, __p2);
69105  return __ret;
69106}
69107#else
69108__ai int32x4_t vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
69109  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69110  int16x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69111  int16x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69112  int32x4_t __ret;
69113  __ret = __rev0 + __noswap_vabdl_s16(__rev1, __rev2);
69114  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69115  return __ret;
69116}
69117__ai int32x4_t __noswap_vabal_s16(int32x4_t __p0, int16x4_t __p1, int16x4_t __p2) {
69118  int32x4_t __ret;
69119  __ret = __p0 + __noswap_vabdl_s16(__p1, __p2);
69120  return __ret;
69121}
69122#endif
69123
69124#if defined(__aarch64__)
69125#ifdef __LITTLE_ENDIAN__
69126__ai uint16x8_t vabal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
69127  uint16x8_t __ret;
69128  __ret = vabal_u8(__p0, vget_high_u8(__p1), vget_high_u8(__p2));
69129  return __ret;
69130}
69131#else
69132__ai uint16x8_t vabal_high_u8(uint16x8_t __p0, uint8x16_t __p1, uint8x16_t __p2) {
69133  uint16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
69134  uint8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69135  uint8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69136  uint16x8_t __ret;
69137  __ret = __noswap_vabal_u8(__rev0, __noswap_vget_high_u8(__rev1), __noswap_vget_high_u8(__rev2));
69138  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69139  return __ret;
69140}
69141#endif
69142
69143#ifdef __LITTLE_ENDIAN__
69144__ai uint64x2_t vabal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
69145  uint64x2_t __ret;
69146  __ret = vabal_u32(__p0, vget_high_u32(__p1), vget_high_u32(__p2));
69147  return __ret;
69148}
69149#else
69150__ai uint64x2_t vabal_high_u32(uint64x2_t __p0, uint32x4_t __p1, uint32x4_t __p2) {
69151  uint64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69152  uint32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69153  uint32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69154  uint64x2_t __ret;
69155  __ret = __noswap_vabal_u32(__rev0, __noswap_vget_high_u32(__rev1), __noswap_vget_high_u32(__rev2));
69156  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69157  return __ret;
69158}
69159#endif
69160
69161#ifdef __LITTLE_ENDIAN__
69162__ai uint32x4_t vabal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
69163  uint32x4_t __ret;
69164  __ret = vabal_u16(__p0, vget_high_u16(__p1), vget_high_u16(__p2));
69165  return __ret;
69166}
69167#else
69168__ai uint32x4_t vabal_high_u16(uint32x4_t __p0, uint16x8_t __p1, uint16x8_t __p2) {
69169  uint32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69170  uint16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
69171  uint16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
69172  uint32x4_t __ret;
69173  __ret = __noswap_vabal_u16(__rev0, __noswap_vget_high_u16(__rev1), __noswap_vget_high_u16(__rev2));
69174  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69175  return __ret;
69176}
69177#endif
69178
69179#ifdef __LITTLE_ENDIAN__
69180__ai int16x8_t vabal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
69181  int16x8_t __ret;
69182  __ret = vabal_s8(__p0, vget_high_s8(__p1), vget_high_s8(__p2));
69183  return __ret;
69184}
69185#else
69186__ai int16x8_t vabal_high_s8(int16x8_t __p0, int8x16_t __p1, int8x16_t __p2) {
69187  int16x8_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0);
69188  int8x16_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69189  int8x16_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0);
69190  int16x8_t __ret;
69191  __ret = __noswap_vabal_s8(__rev0, __noswap_vget_high_s8(__rev1), __noswap_vget_high_s8(__rev2));
69192  __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0);
69193  return __ret;
69194}
69195#endif
69196
69197#ifdef __LITTLE_ENDIAN__
69198__ai int64x2_t vabal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
69199  int64x2_t __ret;
69200  __ret = vabal_s32(__p0, vget_high_s32(__p1), vget_high_s32(__p2));
69201  return __ret;
69202}
69203#else
69204__ai int64x2_t vabal_high_s32(int64x2_t __p0, int32x4_t __p1, int32x4_t __p2) {
69205  int64x2_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0);
69206  int32x4_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0);
69207  int32x4_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0);
69208  int64x2_t __ret;
69209  __ret = __noswap_vabal_s32(__rev0, __noswap_vget_high_s32(__rev1), __noswap_vget_high_s32(__rev2));
69210  __ret = __builtin_shufflevector(__ret, __ret, 1, 0);
69211  return __ret;
69212}
69213#endif
69214
69215#ifdef __LITTLE_ENDIAN__
69216__ai int32x4_t vabal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
69217  int32x4_t __ret;
69218  __ret = vabal_s16(__p0, vget_high_s16(__p1), vget_high_s16(__p2));
69219  return __ret;
69220}
69221#else
69222__ai int32x4_t vabal_high_s16(int32x4_t __p0, int16x8_t __p1, int16x8_t __p2) {
69223  int32x4_t __rev0;  __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0);
69224  int16x8_t __rev1;  __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0);
69225  int16x8_t __rev2;  __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0);
69226  int32x4_t __ret;
69227  __ret = __noswap_vabal_s16(__rev0, __noswap_vget_high_s16(__rev1), __noswap_vget_high_s16(__rev2));
69228  __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0);
69229  return __ret;
69230}
69231#endif
69232
69233#endif
69234
69235#undef __ai
69236
69237#endif /* __ARM_NEON_H */
69238